var/home/core/zuul-output/0000755000175000017500000000000015114041702014520 5ustar corecorevar/home/core/zuul-output/logs/0000755000175000017500000000000015114054374015475 5ustar corecorevar/home/core/zuul-output/logs/kubelet.log0000644000000000000000005264607315114054366017717 0ustar rootrootDec 03 14:09:29 crc systemd[1]: Starting Kubernetes Kubelet... Dec 03 14:09:29 crc restorecon[4691]: Relabeled /var/lib/kubelet/config.json from system_u:object_r:unlabeled_t:s0 to system_u:object_r:container_var_lib_t:s0 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/device-plugins not reset as customized by admin to system_u:object_r:container_file_t:s0 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/device-plugins/kubelet.sock not reset as customized by admin to system_u:object_r:container_file_t:s0 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/volumes/kubernetes.io~configmap/nginx-conf/..2025_02_23_05_40_35.4114275528/nginx.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/containers/networking-console-plugin/22e96971 not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/containers/networking-console-plugin/21c98286 not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/containers/networking-console-plugin/0f1869e1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c215,c682 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/setup/46889d52 not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/setup/5b6a5969 not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c963 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/setup/6c7921f5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c215,c682 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/4804f443 not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/2a46b283 not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/a6b5573e not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/4f88ee5b not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/5a4eee4b not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c963 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/cd87c521 not reset as customized by admin to system_u:object_r:container_file_t:s0:c215,c682 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_33_42.2574241751 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_33_42.2574241751/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/38602af4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/1483b002 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/0346718b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/d3ed4ada not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/3bb473a5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/8cd075a9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/00ab4760 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/54a21c09 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c589,c726 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/70478888 not reset as customized by admin to system_u:object_r:container_file_t:s0:c176,c499 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/43802770 not reset as customized by admin to system_u:object_r:container_file_t:s0:c176,c499 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/955a0edc not reset as customized by admin to system_u:object_r:container_file_t:s0:c176,c499 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/bca2d009 not reset as customized by admin to system_u:object_r:container_file_t:s0:c140,c1009 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/b295f9bd not reset as customized by admin to system_u:object_r:container_file_t:s0:c589,c726 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/..2025_02_23_05_21_22.3617465230 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/..2025_02_23_05_21_22.3617465230/cnibincopy.sh not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/cnibincopy.sh not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/..2025_02_23_05_21_22.2050650026 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/..2025_02_23_05_21_22.2050650026/allowlist.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/allowlist.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/egress-router-binary-copy/bc46ea27 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/egress-router-binary-copy/5731fc1b not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/egress-router-binary-copy/5e1b2a3c not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/cni-plugins/943f0936 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/cni-plugins/3f764ee4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/cni-plugins/8695e3f9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/bond-cni-plugin/aed7aa86 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/bond-cni-plugin/c64d7448 not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/bond-cni-plugin/0ba16bd2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/routeoverride-cni/207a939f not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/routeoverride-cni/54aa8cdb not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/routeoverride-cni/1f5fa595 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni-bincopy/bf9c8153 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni-bincopy/47fba4ea not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni-bincopy/7ae55ce9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni/7906a268 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni/ce43fa69 not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni/7fc7ea3a not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/kube-multus-additional-cni-plugins/d8c38b7d not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/kube-multus-additional-cni-plugins/9ef015fb not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/kube-multus-additional-cni-plugins/b9db6a41 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c432,c991 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/network-metrics-daemon/b1733d79 not reset as customized by admin to system_u:object_r:container_file_t:s0:c476,c820 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/network-metrics-daemon/afccd338 not reset as customized by admin to system_u:object_r:container_file_t:s0:c272,c818 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/network-metrics-daemon/9df0a185 not reset as customized by admin to system_u:object_r:container_file_t:s0:c432,c991 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/kube-rbac-proxy/18938cf8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c476,c820 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/kube-rbac-proxy/7ab4eb23 not reset as customized by admin to system_u:object_r:container_file_t:s0:c272,c818 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/kube-rbac-proxy/56930be6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c432,c991 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/env-overrides not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/env-overrides/..2025_02_23_05_21_35.630010865 not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/env-overrides/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/..2025_02_23_05_21_35.1088506337 not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/..2025_02_23_05_21_35.1088506337/ovnkube.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/ovnkube.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/kube-rbac-proxy/0d8e3722 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/kube-rbac-proxy/d22b2e76 not reset as customized by admin to system_u:object_r:container_file_t:s0:c382,c850 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/kube-rbac-proxy/e036759f not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/2734c483 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/57878fe7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/3f3c2e58 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/375bec3e not reset as customized by admin to system_u:object_r:container_file_t:s0:c382,c850 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/7bc41e08 not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/containers/download-server/48c7a72d not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/containers/download-server/4b66701f not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/containers/download-server/a5a1c202 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..2025_02_23_05_21_40.3350632666 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..2025_02_23_05_21_40.3350632666/additional-cert-acceptance-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..2025_02_23_05_21_40.3350632666/additional-pod-admission-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/additional-cert-acceptance-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/additional-pod-admission-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/env-overrides not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/env-overrides/..2025_02_23_05_21_40.1388695756 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/env-overrides/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/webhook/26f3df5b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/webhook/6d8fb21d not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/webhook/50e94777 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/208473b3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/ec9e08ba not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/3b787c39 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/208eaed5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/93aa3a2b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/3c697968 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/containers/network-check-target-container/ba950ec9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/containers/network-check-target-container/cb5cdb37 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/containers/network-check-target-container/f2df9827 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/..2025_02_23_05_22_30.473230615 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/..2025_02_23_05_22_30.473230615/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_24_06_22_02.1904938450 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_24_06_22_02.1904938450/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/machine-config-operator/fedaa673 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/machine-config-operator/9ca2df95 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/machine-config-operator/b2d7460e not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/kube-rbac-proxy/2207853c not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/kube-rbac-proxy/241c1c29 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/kube-rbac-proxy/2d910eaf not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/..2025_02_23_05_23_49.3726007728 not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/..2025_02_23_05_23_49.3726007728/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/..2025_02_23_05_23_49.841175008 not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/..2025_02_23_05_23_49.841175008/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.843437178 not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.843437178/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/c6c0f2e7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/399edc97 not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/8049f7cc not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/0cec5484 not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/312446d0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c406,c828 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/8e56a35d not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.133159589 not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.133159589/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/2d30ddb9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c380,c909 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/eca8053d not reset as customized by admin to system_u:object_r:container_file_t:s0:c380,c909 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/c3a25c9a not reset as customized by admin to system_u:object_r:container_file_t:s0:c168,c522 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/b9609c22 not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c968,c969 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/dns-operator/e8b0eca9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c106,c418 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/dns-operator/b36a9c3f not reset as customized by admin to system_u:object_r:container_file_t:s0:c529,c711 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/dns-operator/38af7b07 not reset as customized by admin to system_u:object_r:container_file_t:s0:c968,c969 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/kube-rbac-proxy/ae821620 not reset as customized by admin to system_u:object_r:container_file_t:s0:c106,c418 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/kube-rbac-proxy/baa23338 not reset as customized by admin to system_u:object_r:container_file_t:s0:c529,c711 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/kube-rbac-proxy/2c534809 not reset as customized by admin to system_u:object_r:container_file_t:s0:c968,c969 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3532625537 not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3532625537/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/59b29eae not reset as customized by admin to system_u:object_r:container_file_t:s0:c338,c381 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/c91a8e4f not reset as customized by admin to system_u:object_r:container_file_t:s0:c338,c381 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/4d87494a not reset as customized by admin to system_u:object_r:container_file_t:s0:c442,c857 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/1e33ca63 not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/kube-rbac-proxy/8dea7be2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/kube-rbac-proxy/d0b04a99 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/kube-rbac-proxy/d84f01e7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/package-server-manager/4109059b not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/package-server-manager/a7258a3e not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/package-server-manager/05bdf2b6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/f3261b51 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/315d045e not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/5fdcf278 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/d053f757 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/c2850dc7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/..2025_02_23_05_22_30.2390596521 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/..2025_02_23_05_22_30.2390596521/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/fcfb0b2b not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/c7ac9b7d not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/fa0c0d52 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/c609b6ba not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/2be6c296 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/89a32653 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/4eb9afeb not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/13af6efa not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/containers/olm-operator/b03f9724 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/containers/olm-operator/e3d105cc not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/containers/olm-operator/3aed4d83 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1906041176 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1906041176/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/0765fa6e not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/2cefc627 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/3dcc6345 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/365af391 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-SelfManagedHA-Default.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-SelfManagedHA-TechPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-SelfManagedHA-DevPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-Hypershift-TechPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-Hypershift-DevPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-Hypershift-Default.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-api/b1130c0f not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-api/236a5913 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-api/b9432e26 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/5ddb0e3f not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/986dc4fd not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/8a23ff9a not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/9728ae68 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/665f31d0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1255385357 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1255385357/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_23_57.573792656 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_23_57.573792656/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_23_05_22_30.3254245399 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_23_05_22_30.3254245399/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/136c9b42 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/98a1575b not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/cac69136 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/5deb77a7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/2ae53400 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3608339744 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3608339744/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/e46f2326 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/dc688d3c not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/3497c3cd not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/177eb008 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3819292994 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3819292994/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/af5a2afa not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/d780cb1f not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/49b0f374 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/26fbb125 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.3244779536 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.3244779536/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/cf14125a not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/b7f86972 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/e51d739c not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/88ba6a69 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/669a9acf not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/5cd51231 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/75349ec7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/15c26839 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/45023dcd not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/2bb66a50 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/kube-rbac-proxy/64d03bdd not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/kube-rbac-proxy/ab8e7ca0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/kube-rbac-proxy/bb9be25f not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.2034221258 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.2034221258/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/containers/cluster-image-registry-operator/9a0b61d3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/containers/cluster-image-registry-operator/d471b9d2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/containers/cluster-image-registry-operator/8cb76b8e not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/containers/catalog-operator/11a00840 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/containers/catalog-operator/ec355a92 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/containers/catalog-operator/992f735e not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1782968797 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1782968797/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/d59cdbbc not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/72133ff0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/c56c834c not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/d13724c7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/0a498258 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/containers/machine-config-server/fa471982 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/containers/machine-config-server/fc900d92 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/containers/machine-config-server/fa7d68da not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/migrator/4bacf9b4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/migrator/424021b1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/migrator/fc2e31a3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/graceful-termination/f51eefac not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/graceful-termination/c8997f2f not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/graceful-termination/7481f599 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/..2025_02_23_05_22_49.2255460704 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/..2025_02_23_05_22_49.2255460704/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/fdafea19 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/d0e1c571 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/ee398915 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/682bb6b8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/setup/a3e67855 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/setup/a989f289 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/setup/915431bd not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-ensure-env-vars/7796fdab not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-ensure-env-vars/dcdb5f19 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-ensure-env-vars/a3aaa88c not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-resources-copy/5508e3e6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-resources-copy/160585de not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-resources-copy/e99f8da3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcdctl/8bc85570 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcdctl/a5861c91 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcdctl/84db1135 not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd/9e1a6043 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd/c1aba1c2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd/d55ccd6d not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-metrics/971cc9f6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-metrics/8f2e3dcf not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-metrics/ceb35e9c not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-readyz/1c192745 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-readyz/5209e501 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-readyz/f83de4df not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-rev/e7b978ac not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-rev/c64304a1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-rev/5384386b not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c268,c620 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/multus-admission-controller/cce3e3ff not reset as customized by admin to system_u:object_r:container_file_t:s0:c435,c756 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/multus-admission-controller/8fb75465 not reset as customized by admin to system_u:object_r:container_file_t:s0:c268,c620 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/kube-rbac-proxy/740f573e not reset as customized by admin to system_u:object_r:container_file_t:s0:c435,c756 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/kube-rbac-proxy/32fd1134 not reset as customized by admin to system_u:object_r:container_file_t:s0:c268,c620 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/containers/serve-healthcheck-canary/0a861bd3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/containers/serve-healthcheck-canary/80363026 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/containers/serve-healthcheck-canary/bfa952a8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_23_05_33_31.2122464563 not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_23_05_33_31.2122464563/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/config/..2025_02_23_05_33_31.333075221 not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/kube-rbac-proxy/793bf43d not reset as customized by admin to system_u:object_r:container_file_t:s0:c381,c387 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/kube-rbac-proxy/7db1bb6e not reset as customized by admin to system_u:object_r:container_file_t:s0:c142,c438 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/kube-rbac-proxy/4f6a0368 not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/c12c7d86 not reset as customized by admin to system_u:object_r:container_file_t:s0:c381,c387 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/36c4a773 not reset as customized by admin to system_u:object_r:container_file_t:s0:c142,c438 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/4c1e98ae not reset as customized by admin to system_u:object_r:container_file_t:s0:c142,c438 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/a4c8115c not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/setup/7db1802e not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver/a008a7ab not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-cert-syncer/2c836bac not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-cert-regeneration-controller/0ce62299 not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-insecure-readyz/945d2457 not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-check-endpoints/7d5c1dd8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/advanced-cluster-management not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/advanced-cluster-management/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-broker-rhel8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-broker-rhel8/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-online not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-online/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams-console not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams-console/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq7-interconnect-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq7-interconnect-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-automation-platform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-automation-platform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-cloud-addons-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-cloud-addons-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry-3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry-3/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-load-balancer-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-load-balancer-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-businessautomation-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-businessautomation-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-kogito-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-kogito-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator/index.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/businessautomation-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/businessautomation-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cephcsi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cephcsi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cincinnati-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cincinnati-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-kube-descheduler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-kube-descheduler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-logging not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-logging/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-observability-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-observability-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/compliance-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/compliance-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/container-security-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/container-security-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/costmanagement-metrics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/costmanagement-metrics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cryostat-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cryostat-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datagrid not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datagrid/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devspaces not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devspaces/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devworkspace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devworkspace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dpu-network-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dpu-network-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eap not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eap/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-dns-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-dns-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/file-integrity-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/file-integrity-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-apicurito not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-apicurito/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-console not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-console/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-online not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-online/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gatekeeper-operator-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gatekeeper-operator-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jws-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jws-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management-hub not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management-hub/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kiali-ossm not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kiali-ossm/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubevirt-hyperconverged not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubevirt-hyperconverged/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logic-operator-rhel8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logic-operator-rhel8/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lvms-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lvms-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mcg-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mcg-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mta-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mta-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtr-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtr-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-engine not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-engine/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-observability-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-observability-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-client-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-client-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-csi-addons-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-csi-addons-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-multicluster-orchestrator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-multicluster-orchestrator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-prometheus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-prometheus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-cluster-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-cluster-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-hub-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-hub-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator/bundle-v1.15.0.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator/channel.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator/package.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-custom-metrics-autoscaler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-custom-metrics-autoscaler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-gitops-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-gitops-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-pipelines-operator-rh not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-pipelines-operator-rh/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-secondary-scheduler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-secondary-scheduler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-bridge-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-bridge-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/recipe not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/recipe/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-camel-k not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-camel-k/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-hawtio-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-hawtio-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redhat-oadp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redhat-oadp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rh-service-binding-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rh-service-binding-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhacs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhacs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhbk-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhbk-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhdh not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhdh/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-prometheus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-prometheus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhpam-kogito-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhpam-kogito-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhsso-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhsso-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rook-ceph-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rook-ceph-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/run-once-duration-override-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/run-once-duration-override-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sandboxed-containers-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sandboxed-containers-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/security-profiles-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/security-profiles-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/serverless-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/serverless-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-registry-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-registry-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator3/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/submariner not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/submariner/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tang-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tang-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustee-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustee-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volsync-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volsync-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/web-terminal not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/web-terminal/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-utilities/bc8d0691 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-utilities/6b76097a not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-utilities/34d1af30 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-content/312ba61c not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-content/645d5dd1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-content/16e825f0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/registry-server/4cf51fc9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/registry-server/2a23d348 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/registry-server/075dbd49 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/containers/node-ca/dd585ddd not reset as customized by admin to system_u:object_r:container_file_t:s0:c377,c642 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/containers/node-ca/17ebd0ab not reset as customized by admin to system_u:object_r:container_file_t:s0:c338,c343 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/containers/node-ca/005579f4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_23_05_23_11.449897510 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_23_05_23_11.449897510/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_23_05_23_11.1287037894 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/..2025_02_23_05_23_11.1301053334 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/..2025_02_23_05_23_11.1301053334/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/fix-audit-permissions/bf5f3b9c not reset as customized by admin to system_u:object_r:container_file_t:s0:c49,c263 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/fix-audit-permissions/af276eb7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c701 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/fix-audit-permissions/ea28e322 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/oauth-apiserver/692e6683 not reset as customized by admin to system_u:object_r:container_file_t:s0:c49,c263 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/oauth-apiserver/871746a7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c701 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/oauth-apiserver/4eb2e958 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/..2025_02_24_06_09_06.2875086261 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/..2025_02_24_06_09_06.2875086261/console-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/console-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_09_06.286118152 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_09_06.286118152/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/..2025_02_24_06_09_06.3865795478 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/..2025_02_24_06_09_06.3865795478/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/..2025_02_24_06_09_06.584414814 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/..2025_02_24_06_09_06.584414814/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/containers/console/ca9b62da not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/containers/console/0edd6fce not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837 not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/openshift-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/openshift-controller-manager.openshift-global-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/openshift-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/openshift-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/openshift-controller-manager.openshift-global-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/openshift-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.1071801880 not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.1071801880/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/..2025_02_24_06_20_07.2494444877 not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/..2025_02_24_06_20_07.2494444877/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/containers/controller-manager/89b4555f not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/..2025_02_23_05_23_22.4071100442 not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/..2025_02_23_05_23_22.4071100442/Corefile not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/Corefile not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/dns/655fcd71 not reset as customized by admin to system_u:object_r:container_file_t:s0:c457,c841 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/dns/0d43c002 not reset as customized by admin to system_u:object_r:container_file_t:s0:c55,c1022 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/dns/e68efd17 not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/kube-rbac-proxy/9acf9b65 not reset as customized by admin to system_u:object_r:container_file_t:s0:c457,c841 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/kube-rbac-proxy/5ae3ff11 not reset as customized by admin to system_u:object_r:container_file_t:s0:c55,c1022 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/kube-rbac-proxy/1e59206a not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/containers/dns-node-resolver/27af16d1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c304,c1017 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/containers/dns-node-resolver/7918e729 not reset as customized by admin to system_u:object_r:container_file_t:s0:c853,c893 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/containers/dns-node-resolver/5d976d0e not reset as customized by admin to system_u:object_r:container_file_t:s0:c585,c981 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/..2025_02_23_05_38_56.1112187283 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/..2025_02_23_05_38_56.1112187283/controller-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/controller-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_38_56.2839772658 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_38_56.2839772658/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/d7f55cbb not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/f0812073 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/1a56cbeb not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/7fdd437e not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/cdfb5652 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_24_06_17_29.3844392896 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_24_06_17_29.3844392896/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/..2025_02_24_06_17_29.848549803 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/..2025_02_24_06_17_29.848549803/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/..2025_02_24_06_17_29.780046231 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/..2025_02_24_06_17_29.780046231/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_17_29.2729721485 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_17_29.2729721485/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/containers/fix-audit-permissions/fb93119e not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/containers/openshift-apiserver/f1e8fc0e not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/containers/openshift-apiserver-check-endpoints/218511f3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes/kubernetes.io~empty-dir/tmpfs not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes/kubernetes.io~empty-dir/tmpfs/k8s-webhook-server not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes/kubernetes.io~empty-dir/tmpfs/k8s-webhook-server/serving-certs not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/containers/packageserver/ca8af7b3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/containers/packageserver/72cc8a75 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/containers/packageserver/6e8a3760 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/..2025_02_23_05_27_30.557428972 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/..2025_02_23_05_27_30.557428972/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/4c3455c0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/2278acb0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/4b453e4f not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/3ec09bda not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_24_06_25_03.422633132 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_24_06_25_03.422633132/anchors not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_24_06_25_03.422633132/anchors/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/anchors not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/edk2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/edk2/cacerts.bin not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/java not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/java/cacerts not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/openssl not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/openssl/ca-bundle.trust.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/email-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/objsign-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2ae6433e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fde84897.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/75680d2e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/openshift-service-serving-signer_1740288168.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/facfc4fa.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8f5a969c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CFCA_EV_ROOT.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9ef4a08a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ingress-operator_1740288202.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2f332aed.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/248c8271.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d10a21f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ACCVRAIZ1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a94d09e5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3c9a4d3b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/40193066.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AC_RAIZ_FNMT-RCM.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cd8c0d63.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b936d1c6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CA_Disig_Root_R2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4fd49c6c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AC_RAIZ_FNMT-RCM_SERVIDORES_SEGUROS.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b81b93f0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f9a69fa.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certigna.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b30d5fda.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ANF_Secure_Server_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b433981b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/93851c9e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9282e51c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e7dd1bc4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Actalis_Authentication_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/930ac5d2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f47b495.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e113c810.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5931b5bc.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Commercial.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2b349938.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e48193cf.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/302904dd.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a716d4ed.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Networking.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/93bc0acc.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/86212b19.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certigna_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Premium.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b727005e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dbc54cab.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f51bb24c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c28a8a30.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Premium_ECC.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9c8dfbd4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ccc52f49.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cb1c3204.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ce5e74ef.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fd08c599.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_Trusted_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6d41d539.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fb5fa911.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e35234b1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8cb5ee0f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a7c655d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f8fc53da.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/de6d66f3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d41b5e2a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/41a3f684.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1df5a75f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Atos_TrustedRoot_2011.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e36a6752.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b872f2b4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9576d26b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/228f89db.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Atos_TrustedRoot_Root_CA_ECC_TLS_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fb717492.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2d21b73c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0b1b94ef.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/595e996b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Atos_TrustedRoot_Root_CA_RSA_TLS_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9b46e03d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/128f4b91.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Buypass_Class_3_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/81f2d2b1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Autoridad_de_Certificacion_Firmaprofesional_CIF_A62634068.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3bde41ac.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d16a5865.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_EC-384_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/BJCA_Global_Root_CA1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0179095f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ffa7f1eb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9482e63a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d4dae3dd.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/BJCA_Global_Root_CA2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3e359ba6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7e067d03.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/95aff9e3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d7746a63.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Baltimore_CyberTrust_Root.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/653b494a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3ad48a91.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_Trusted_Network_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Buypass_Class_2_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/54657681.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/82223c44.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e8de2f56.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2d9dafe4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d96b65e2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ee64a828.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/COMODO_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/40547a79.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5a3f0ff8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a780d93.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/34d996fb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/COMODO_ECC_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/eed8c118.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/89c02a45.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certainly_Root_R1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b1159c4c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/COMODO_RSA_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d6325660.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d4c339cb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8312c4c1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certainly_Root_E1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8508e720.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5fdd185d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/48bec511.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/69105f4f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0b9bc432.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_Trusted_Network_CA_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/32888f65.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_ECC_Root-01.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6b03dec0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/219d9499.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_ECC_Root-02.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5acf816d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cbf06781.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_RSA_Root-01.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dc99f41e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_RSA_Root-02.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AAA_Certificate_Services.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/985c1f52.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8794b4e3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_BR_Root_CA_1_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e7c037b4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ef954a4e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_EV_Root_CA_1_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2add47b6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/90c5a3c8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_Root_Class_3_CA_2_2009.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b0f3e76e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/53a1b57a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_Root_Class_3_CA_2_EV_2009.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Assured_ID_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5ad8a5d6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/68dd7389.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Assured_ID_Root_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9d04f354.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d6437c3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/062cdee6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bd43e1dd.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Assured_ID_Root_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7f3d5d1d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c491639e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign_Root_E46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Global_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3513523f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/399e7759.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/feffd413.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d18e9066.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Global_Root_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/607986c7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c90bc37d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1b0f7e5c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e08bfd1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Global_Root_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dd8e9d41.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ed39abd0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a3418fda.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bc3f2570.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_High_Assurance_EV_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/244b5494.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/81b9768f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4be590e0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_TLS_ECC_P384_Root_G5.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9846683b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/252252d2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e8e7201.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ISRG_Root_X1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_TLS_RSA4096_Root_G5.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d52c538d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c44cc0c0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign_Root_R46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Trusted_Root_G4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/75d1b2ed.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a2c66da8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ecccd8db.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust.net_Certification_Authority__2048_.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/aee5f10d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3e7271e8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b0e59380.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4c3982f2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6b99d060.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bf64f35b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0a775a30.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/002c0b4f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cc450945.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority_-_EC1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/106f3e4d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b3fb433b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4042bcee.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/02265526.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/455f1b52.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0d69c7e1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9f727ac7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority_-_G4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5e98733a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f0cd152c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dc4d6a89.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6187b673.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/FIRMAPROFESIONAL_CA_ROOT-A_WEB.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ba8887ce.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/068570d1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f081611a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/48a195d8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GDCA_TrustAUTH_R5_ROOT.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0f6fa695.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ab59055e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b92fd57f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GLOBALTRUST_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fa5da96b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1ec40989.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7719f463.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1001acf7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f013ecaf.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/626dceaf.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c559d742.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1d3472b9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9479c8c3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a81e292b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4bfab552.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Go_Daddy_Class_2_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Sectigo_Public_Server_Authentication_Root_E46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Go_Daddy_Root_Certificate_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e071171e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/57bcb2da.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/HARICA_TLS_ECC_Root_CA_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ab5346f4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5046c355.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/HARICA_TLS_RSA_Root_CA_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/865fbdf9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/da0cfd1d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/85cde254.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Hellenic_Academic_and_Research_Institutions_ECC_RootCA_2015.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cbb3f32b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SecureSign_RootCA11.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Hellenic_Academic_and_Research_Institutions_RootCA_2015.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5860aaa6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/31188b5e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/HiPKI_Root_CA_-_G1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c7f1359b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f15c80c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Hongkong_Post_Root_CA_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/09789157.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ISRG_Root_X2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/18856ac4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e09d511.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/IdenTrust_Commercial_Root_CA_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cf701eeb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d06393bb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/IdenTrust_Public_Sector_Root_CA_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/10531352.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Izenpe.com.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SecureTrust_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b0ed035a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Microsec_e-Szigno_Root_CA_2009.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8160b96c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e8651083.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2c63f966.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Security_Communication_RootCA2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Microsoft_ECC_Root_Certificate_Authority_2017.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d89cda1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/01419da9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_TLS_RSA_Root_CA_2022.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b7a5b843.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Microsoft_RSA_Root_Certificate_Authority_2017.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bf53fb88.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9591a472.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3afde786.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SwissSign_Gold_CA_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/NAVER_Global_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3fb36b73.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d39b0a2c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a89d74c2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cd58d51e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b7db1890.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/NetLock_Arany__Class_Gold__F__tan__s__tv__ny.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/988a38cb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/60afe812.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f39fc864.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5443e9e3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/OISTE_WISeKey_Global_Root_GB_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e73d606e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dfc0fe80.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b66938e9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e1eab7c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/OISTE_WISeKey_Global_Root_GC_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/773e07ad.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3c899c73.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d59297b8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ddcda989.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_1_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/749e9e03.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/52b525c7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Security_Communication_RootCA3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d7e8dc79.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a819ef2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/08063a00.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6b483515.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_2_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/064e0aa9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1f58a078.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6f7454b3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7fa05551.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/76faf6c0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9339512a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f387163d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ee37c333.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_3_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e18bfb83.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e442e424.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fe8a2cd8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/23f4c490.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5cd81ad7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_EV_Root_Certification_Authority_ECC.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f0c70a8d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7892ad52.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SZAFIR_ROOT_CA2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4f316efb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_EV_Root_Certification_Authority_RSA_R2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/06dc52d5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/583d0756.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Sectigo_Public_Server_Authentication_Root_R46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_Root_Certification_Authority_ECC.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0bf05006.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/88950faa.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9046744a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3c860d51.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_Root_Certification_Authority_RSA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6fa5da56.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/33ee480d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Secure_Global_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/63a2c897.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_TLS_ECC_Root_CA_2022.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bdacca6f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ff34af3f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dbff3a01.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Security_Communication_ECC_RootCA1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_Root_CA_-_C1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Starfield_Class_2_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/406c9bb1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Starfield_Root_Certificate_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_ECC_Root_CA_-_C3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Starfield_Services_Root_Certificate_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SwissSign_Silver_CA_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/99e1b953.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/T-TeleSec_GlobalRoot_Class_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/vTrus_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/T-TeleSec_GlobalRoot_Class_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/14bc7599.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TUBITAK_Kamu_SM_SSL_Kok_Sertifikasi_-_Surum_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TWCA_Global_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a3adc42.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TWCA_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f459871d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Telekom_Security_TLS_ECC_Root_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_Root_CA_-_G1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Telekom_Security_TLS_RSA_Root_2023.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TeliaSonera_Root_CA_v1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Telia_Root_CA_v2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8f103249.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f058632f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ca-certificates.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TrustAsia_Global_Root_CA_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9bf03295.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/98aaf404.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TrustAsia_Global_Root_CA_G4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1cef98f5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/073bfcc5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2923b3f9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Trustwave_Global_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f249de83.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/edcbddb5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_ECC_Root_CA_-_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Trustwave_Global_ECC_P256_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9b5697b0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1ae85e5e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b74d2bd5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Trustwave_Global_ECC_P384_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d887a5bb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9aef356c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TunTrust_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fd64f3fc.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e13665f9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/UCA_Extended_Validation_Root.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0f5dc4f3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/da7377f6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/UCA_Global_G2_Root.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c01eb047.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/304d27c3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ed858448.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/USERTrust_ECC_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f30dd6ad.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/04f60c28.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/vTrus_ECC_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/USERTrust_RSA_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fc5a8f99.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/35105088.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ee532fd5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/XRamp_Global_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/706f604c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/76579174.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/certSIGN_ROOT_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d86cdd1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/882de061.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/certSIGN_ROOT_CA_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f618aec.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a9d40e02.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e-Szigno_Root_CA_2017.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e868b802.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/83e9984f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ePKI_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ca6e4ad9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9d6523ce.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4b718d9b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/869fbf79.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/containers/registry/f8d22bdb not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator/6e8bbfac not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator/54dd7996 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator/a4f1bb05 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator-watch/207129da not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator-watch/c1df39e1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator-watch/15b8f1cd not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3523263858 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3523263858/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/..2025_02_23_05_27_49.3256605594 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/..2025_02_23_05_27_49.3256605594/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/kube-rbac-proxy/77bd6913 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/kube-rbac-proxy/2382c1b1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/kube-rbac-proxy/704ce128 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/machine-api-operator/70d16fe0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/machine-api-operator/bfb95535 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/machine-api-operator/57a8e8e2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3413793711 not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3413793711/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/containers/kube-apiserver-operator/1b9d3e5e not reset as customized by admin to system_u:object_r:container_file_t:s0:c107,c917 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/containers/kube-apiserver-operator/fddb173c not reset as customized by admin to system_u:object_r:container_file_t:s0:c202,c983 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/containers/kube-apiserver-operator/95d3c6c4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/containers/check-endpoints/bfb5fff5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/containers/check-endpoints/2aef40aa not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/containers/check-endpoints/c0391cad not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager/1119e69d not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager/660608b4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager/8220bd53 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/cluster-policy-controller/85f99d5c not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/cluster-policy-controller/4b0225f6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-cert-syncer/9c2a3394 not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-cert-syncer/e820b243 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-recovery-controller/1ca52ea0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-recovery-controller/e6988e45 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/..2025_02_24_06_09_21.2517297950 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/..2025_02_24_06_09_21.2517297950/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/machine-config-controller/6655f00b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/machine-config-controller/98bc3986 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/machine-config-controller/08e3458a not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/kube-rbac-proxy/2a191cb0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/kube-rbac-proxy/6c4eeefb not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/kube-rbac-proxy/f61a549c not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/hostpath-provisioner/24891863 not reset as customized by admin to system_u:object_r:container_file_t:s0:c37,c572 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/hostpath-provisioner/fbdfd89c not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/liveness-probe/9b63b3bc not reset as customized by admin to system_u:object_r:container_file_t:s0:c37,c572 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/liveness-probe/8acde6d6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/node-driver-registrar/59ecbba3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/csi-provisioner/685d4be3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 03 14:09:29 crc restorecon[4691]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300/openshift-route-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300/openshift-route-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/openshift-route-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/openshift-route-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.2950937851 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.2950937851/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/containers/route-controller-manager/feaea55e not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abinitio-runtime-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abinitio-runtime-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/accuknox-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/accuknox-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aci-containers-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aci-containers-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airlock-microgateway not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airlock-microgateway/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ako-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ako-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloy not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloy/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anchore-engine not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anchore-engine/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-cloud-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-cloud-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-dcap-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-dcap-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cfm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cfm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium-enterprise not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium-enterprise/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloud-native-postgresql not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloud-native-postgresql/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudera-streams-messaging-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudera-streams-messaging-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudnative-pg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudnative-pg/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cnfv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cnfv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/conjur-follower-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/conjur-follower-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/coroot-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/coroot-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cte-k8s-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cte-k8s-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-deploy-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-deploy-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-release-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-release-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edb-hcp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edb-hcp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-eck-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-eck-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/federatorai-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/federatorai-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fujitsu-enterprise-postgres-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fujitsu-enterprise-postgres-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/function-mesh not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/function-mesh/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/harness-gitops-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/harness-gitops-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hcp-terraform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hcp-terraform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hpe-ezmeral-csi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hpe-ezmeral-csi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-application-gateway-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-application-gateway-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-directory-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-directory-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-dr-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-dr-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-licensing-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-licensing-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-sds-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-sds-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infrastructure-asset-orchestrator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infrastructure-asset-orchestrator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-device-plugins-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-device-plugins-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-kubernetes-power-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-kubernetes-power-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-openshift-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-openshift-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8s-triliovault not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8s-triliovault/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-ati-updates not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-ati-updates/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-framework not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-framework/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-ingress not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-ingress/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-licensing not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-licensing/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-sso not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-sso/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-keycloak-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-keycloak-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-load-core not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-load-core/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-loadcore-agents not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-loadcore-agents/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nats-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nats-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nimbusmosaic-dusim not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nimbusmosaic-dusim/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-rest-api-browser-v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-rest-api-browser-v1/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-appsec not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-appsec/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-core not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-core/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-db/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-diagnostics not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-diagnostics/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-logging not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-logging/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-migration not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-migration/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-msg-broker not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-msg-broker/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-notifications not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-notifications/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-stats-dashboards not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-stats-dashboards/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-storage not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-storage/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-test-core not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-test-core/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-ui not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-ui/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-websocket-service not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-websocket-service/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kong-gateway-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kong-gateway-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubearmor-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubearmor-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lenovo-locd-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lenovo-locd-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memcached-operator-ogaye not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memcached-operator-ogaye/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memory-machine-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memory-machine-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-enterprise not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-enterprise/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netapp-spark-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netapp-spark-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-adm-agent-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-adm-agent-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-repository-ha-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-repository-ha-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nginx-ingress-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nginx-ingress-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nim-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nim-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxiq-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxiq-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxrm-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxrm-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odigos-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odigos-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/open-liberty-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/open-liberty-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftartifactoryha-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftartifactoryha-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftxray-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftxray-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/operator-certification-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/operator-certification-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pmem-csi-operator-os not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pmem-csi-operator-os/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-component-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-component-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-fabric-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-fabric-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sanstoragecsi-operator-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sanstoragecsi-operator-bundle/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/smilecdr-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/smilecdr-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sriov-fec not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sriov-fec/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-commons-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-commons-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-zookeeper-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-zookeeper-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-tsc-client-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-tsc-client-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tawon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tawon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tigera-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tigera-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-secrets-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-secrets-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vcp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vcp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/webotx-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/webotx-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-utilities/63709497 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-utilities/d966b7fd not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-utilities/f5773757 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-content/81c9edb9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-content/57bf57ee not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-content/86f5e6aa not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/registry-server/0aabe31d not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/registry-server/d2af85c2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/registry-server/09d157d9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acm-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acm-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acmpca-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acmpca-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigateway-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigateway-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigatewayv2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigatewayv2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-applicationautoscaling-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-applicationautoscaling-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-athena-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-athena-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudfront-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudfront-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudtrail-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudtrail-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatch-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatch-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatchlogs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatchlogs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-documentdb-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-documentdb-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-dynamodb-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-dynamodb-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ec2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ec2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecr-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecr-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-efs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-efs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eks-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eks-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elasticache-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elasticache-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elbv2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elbv2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-emrcontainers-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-emrcontainers-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eventbridge-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eventbridge-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-iam-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-iam-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kafka-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kafka-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-keyspaces-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-keyspaces-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kinesis-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kinesis-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kms-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kms-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-lambda-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-lambda-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-memorydb-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-memorydb-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-mq-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-mq-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-networkfirewall-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-networkfirewall-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-opensearchservice-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-opensearchservice-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-organizations-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-organizations-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-pipes-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-pipes-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-prometheusservice-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-prometheusservice-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-rds-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-rds-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-recyclebin-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-recyclebin-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53resolver-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53resolver-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-s3-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-s3-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sagemaker-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sagemaker-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-secretsmanager-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-secretsmanager-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ses-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ses-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sfn-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sfn-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sns-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sns-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sqs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sqs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ssm-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ssm-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-wafv2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-wafv2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airflow-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airflow-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloydb-omni-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloydb-omni-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alvearie-imaging-ingestion not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alvearie-imaging-ingestion/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amd-gpu-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amd-gpu-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/analytics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/analytics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/annotationlab not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/annotationlab/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-api-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-api-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurito not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurito/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apimatic-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apimatic-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/application-services-metering-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/application-services-metering-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/argocd-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/argocd-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/assisted-service-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/assisted-service-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/automotive-infra not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/automotive-infra/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-efs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-efs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/awss3-operator-registry not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/awss3-operator-registry/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/azure-service-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/azure-service-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/beegfs-csi-driver-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/beegfs-csi-driver-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-k not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-k/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-karavan-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-karavan-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator-community not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator-community/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-utils-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-utils-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-aas-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-aas-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-impairment-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-impairment-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/codeflare-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/codeflare-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-kubevirt-hyperconverged not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-kubevirt-hyperconverged/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-trivy-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-trivy-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-windows-machine-config-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-windows-machine-config-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/customized-user-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/customized-user-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cxl-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cxl-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dapr-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dapr-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datatrucker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datatrucker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dbaas-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dbaas-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/debezium-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/debezium-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/deployment-validation-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/deployment-validation-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devopsinabox not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devopsinabox/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dns-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dns-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-amlen-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-amlen-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-che not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-che/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ecr-secret-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ecr-secret-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edp-keycloak-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edp-keycloak-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/egressip-ipam-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/egressip-ipam-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ember-csi-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ember-csi-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/etcd not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/etcd/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eventing-kogito not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eventing-kogito/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-secrets-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-secrets-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flink-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flink-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8gb not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8gb/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fossul-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fossul-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/github-arc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/github-arc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitops-primer not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitops-primer/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitwebhook-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitwebhook-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/global-load-balancer-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/global-load-balancer-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/grafana-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/grafana-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/group-sync-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/group-sync-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hawtio-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hawtio-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hedvig-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hedvig-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hive-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hive-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/horreum-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/horreum-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hyperfoil-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hyperfoil-bundle/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator-community not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator-community/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-spectrum-scale-csi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-spectrum-scale-csi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibmcloud-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibmcloud-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infinispan not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infinispan/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/integrity-shield-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/integrity-shield-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ipfs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ipfs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/istio-workspace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/istio-workspace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kaoto-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kaoto-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keda not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keda/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keepalived-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keepalived-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-permissions-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-permissions-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/klusterlet not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/klusterlet/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kogito-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kogito-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/koku-metrics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/koku-metrics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/konveyor-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/konveyor-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/korrel8r not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/korrel8r/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kuadrant-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kuadrant-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kube-green not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kube-green/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubernetes-imagepuller-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubernetes-imagepuller-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/l5-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/l5-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/layer7-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/layer7-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lbconfig-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lbconfig-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lib-bucket-provisioner not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lib-bucket-provisioner/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/limitador-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/limitador-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logging-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logging-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mariadb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mariadb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marin3r not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marin3r/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mercury-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mercury-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/microcks not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/microcks/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/move2kube-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/move2kube-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multi-nic-cni-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multi-nic-cni-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-global-hub-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-global-hub-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-operators-subscription not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-operators-subscription/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/must-gather-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/must-gather-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/namespace-configuration-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/namespace-configuration-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ncn-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ncn-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ndmspc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ndmspc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator-m88i not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator-m88i/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nfs-provisioner-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nfs-provisioner-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nlp-server not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nlp-server/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-discovery-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-discovery-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nsm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nsm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oadp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oadp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/observability-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/observability-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oci-ccm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oci-ccm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odoo-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odoo-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opendatahub-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opendatahub-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openebs not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openebs/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-nfd-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-nfd-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-node-upgrade-mutex-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-node-upgrade-mutex-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-qiskit-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-qiskit-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patch-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patch-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patterns-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patterns-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pelorus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pelorus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/percona-xtradb-cluster-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/percona-xtradb-cluster-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-essentials not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-essentials/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/postgresql not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/postgresql/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/proactive-node-scaling-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/proactive-node-scaling-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/project-quay not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/project-quay/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus-exporter-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus-exporter-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pulp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pulp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-cluster-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-cluster-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-messaging-topology-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-messaging-topology-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/reportportal-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/reportportal-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/resource-locker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/resource-locker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhoas-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhoas-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ripsaw not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ripsaw/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sailoperator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sailoperator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-commerce-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-commerce-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-data-intelligence-observer-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-data-intelligence-observer-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-hana-express-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-hana-express-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-binding-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-binding-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/shipwright-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/shipwright-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sigstore-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sigstore-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snapscheduler not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snapscheduler/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snyk-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snyk-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/socmmd not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/socmmd/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonar-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonar-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosivio not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosivio/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonataflow-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonataflow-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosreport-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosreport-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/spark-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/spark-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/special-resource-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/special-resource-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron-engine not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron-engine/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/strimzi-kafka-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/strimzi-kafka-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/syndesis not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/syndesis/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tagger not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tagger/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tf-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tf-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tidb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tidb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trident-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trident-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustify-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustify-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ucs-ci-solutions-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ucs-ci-solutions-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/universal-crossplane not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/universal-crossplane/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/varnish-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/varnish-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-config-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-config-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/verticadb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/verticadb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volume-expander-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volume-expander-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/wandb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/wandb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/windup-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/windup-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yaks not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yaks/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-utilities/c0fe7256 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-utilities/c30319e4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-utilities/e6b1dd45 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-content/2bb643f0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-content/920de426 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-content/70fa1e87 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/registry-server/a1c12a2f not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/registry-server/9442e6c7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/registry-server/5b45ec72 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abot-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abot-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/entando-k8s-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/entando-k8s-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-paygo-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-paygo-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-term-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-term-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/linstor-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/linstor-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-deploy-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-deploy-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-paygo-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-paygo-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vfunction-server-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vfunction-server-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yugabyte-platform-operator-bundle-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yugabyte-platform-operator-bundle-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-utilities/3c9f3a59 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-utilities/1091c11b not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-utilities/9a6821c6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-content/ec0c35e2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-content/517f37e7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-content/6214fe78 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/registry-server/ba189c8b not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/registry-server/351e4f31 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/registry-server/c0f219ff not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/wait-for-host-port/8069f607 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/wait-for-host-port/559c3d82 not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/wait-for-host-port/605ad488 not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler/148df488 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler/3bf6dcb4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler/022a2feb not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-cert-syncer/938c3924 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-cert-syncer/729fe23e not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-cert-syncer/1fd5cbd4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-recovery-controller/a96697e1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-recovery-controller/e155ddca not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-recovery-controller/10dd0e0f not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/..2025_02_24_06_09_35.3018472960 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/..2025_02_24_06_09_35.3018472960/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/..2025_02_24_06_09_35.4262376737 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/..2025_02_24_06_09_35.4262376737/audit.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/audit.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/..2025_02_24_06_09_35.2630275752 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/..2025_02_24_06_09_35.2630275752/v4-0-config-system-cliconfig not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/v4-0-config-system-cliconfig not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/..2025_02_24_06_09_35.2376963788 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/..2025_02_24_06_09_35.2376963788/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/containers/oauth-openshift/6f2c8392 not reset as customized by admin to system_u:object_r:container_file_t:s0:c267,c588 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/containers/oauth-openshift/bd241ad9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/plugins not reset as customized by admin to system_u:object_r:container_file_t:s0 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/plugins/csi-hostpath not reset as customized by admin to system_u:object_r:container_file_t:s0 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/plugins/csi-hostpath/csi.sock not reset as customized by admin to system_u:object_r:container_file_t:s0 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/plugins/kubernetes.io not reset as customized by admin to system_u:object_r:container_file_t:s0 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/plugins/kubernetes.io/csi not reset as customized by admin to system_u:object_r:container_file_t:s0 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner not reset as customized by admin to system_u:object_r:container_file_t:s0 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983 not reset as customized by admin to system_u:object_r:container_file_t:s0 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/globalmount not reset as customized by admin to system_u:object_r:container_file_t:s0 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/vol_data.json not reset as customized by admin to system_u:object_r:container_file_t:s0 Dec 03 14:09:30 crc restorecon[4691]: /var/lib/kubelet/plugins_registry not reset as customized by admin to system_u:object_r:container_file_t:s0 Dec 03 14:09:30 crc restorecon[4691]: Relabeled /var/usrlocal/bin/kubenswrapper from system_u:object_r:bin_t:s0 to system_u:object_r:kubelet_exec_t:s0 Dec 03 14:09:30 crc kubenswrapper[4805]: Flag --container-runtime-endpoint has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Dec 03 14:09:30 crc kubenswrapper[4805]: Flag --minimum-container-ttl-duration has been deprecated, Use --eviction-hard or --eviction-soft instead. Will be removed in a future version. Dec 03 14:09:30 crc kubenswrapper[4805]: Flag --volume-plugin-dir has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Dec 03 14:09:30 crc kubenswrapper[4805]: Flag --register-with-taints has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Dec 03 14:09:30 crc kubenswrapper[4805]: Flag --pod-infra-container-image has been deprecated, will be removed in a future release. Image garbage collector will get sandbox image information from CRI. Dec 03 14:09:30 crc kubenswrapper[4805]: Flag --system-reserved has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Dec 03 14:09:30 crc kubenswrapper[4805]: I1203 14:09:30.525135 4805 server.go:211] "--pod-infra-container-image will not be pruned by the image garbage collector in kubelet and should also be set in the remote runtime" Dec 03 14:09:30 crc kubenswrapper[4805]: W1203 14:09:30.528961 4805 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Dec 03 14:09:30 crc kubenswrapper[4805]: W1203 14:09:30.528989 4805 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Dec 03 14:09:30 crc kubenswrapper[4805]: W1203 14:09:30.528996 4805 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Dec 03 14:09:30 crc kubenswrapper[4805]: W1203 14:09:30.529001 4805 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Dec 03 14:09:30 crc kubenswrapper[4805]: W1203 14:09:30.529007 4805 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Dec 03 14:09:30 crc kubenswrapper[4805]: W1203 14:09:30.529013 4805 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Dec 03 14:09:30 crc kubenswrapper[4805]: W1203 14:09:30.529019 4805 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Dec 03 14:09:30 crc kubenswrapper[4805]: W1203 14:09:30.529024 4805 feature_gate.go:330] unrecognized feature gate: Example Dec 03 14:09:30 crc kubenswrapper[4805]: W1203 14:09:30.529030 4805 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Dec 03 14:09:30 crc kubenswrapper[4805]: W1203 14:09:30.529035 4805 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Dec 03 14:09:30 crc kubenswrapper[4805]: W1203 14:09:30.529040 4805 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Dec 03 14:09:30 crc kubenswrapper[4805]: W1203 14:09:30.529045 4805 feature_gate.go:330] unrecognized feature gate: InsightsConfig Dec 03 14:09:30 crc kubenswrapper[4805]: W1203 14:09:30.529050 4805 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Dec 03 14:09:30 crc kubenswrapper[4805]: W1203 14:09:30.529055 4805 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Dec 03 14:09:30 crc kubenswrapper[4805]: W1203 14:09:30.529060 4805 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Dec 03 14:09:30 crc kubenswrapper[4805]: W1203 14:09:30.529065 4805 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Dec 03 14:09:30 crc kubenswrapper[4805]: W1203 14:09:30.529071 4805 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Dec 03 14:09:30 crc kubenswrapper[4805]: W1203 14:09:30.529076 4805 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Dec 03 14:09:30 crc kubenswrapper[4805]: W1203 14:09:30.529081 4805 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Dec 03 14:09:30 crc kubenswrapper[4805]: W1203 14:09:30.529086 4805 feature_gate.go:330] unrecognized feature gate: OVNObservability Dec 03 14:09:30 crc kubenswrapper[4805]: W1203 14:09:30.529091 4805 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Dec 03 14:09:30 crc kubenswrapper[4805]: W1203 14:09:30.529096 4805 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Dec 03 14:09:30 crc kubenswrapper[4805]: W1203 14:09:30.529102 4805 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Dec 03 14:09:30 crc kubenswrapper[4805]: W1203 14:09:30.529107 4805 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Dec 03 14:09:30 crc kubenswrapper[4805]: W1203 14:09:30.529113 4805 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Dec 03 14:09:30 crc kubenswrapper[4805]: W1203 14:09:30.529118 4805 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Dec 03 14:09:30 crc kubenswrapper[4805]: W1203 14:09:30.529125 4805 feature_gate.go:330] unrecognized feature gate: SignatureStores Dec 03 14:09:30 crc kubenswrapper[4805]: W1203 14:09:30.529131 4805 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Dec 03 14:09:30 crc kubenswrapper[4805]: W1203 14:09:30.529136 4805 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Dec 03 14:09:30 crc kubenswrapper[4805]: W1203 14:09:30.529143 4805 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Dec 03 14:09:30 crc kubenswrapper[4805]: W1203 14:09:30.529148 4805 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Dec 03 14:09:30 crc kubenswrapper[4805]: W1203 14:09:30.529153 4805 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Dec 03 14:09:30 crc kubenswrapper[4805]: W1203 14:09:30.529159 4805 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Dec 03 14:09:30 crc kubenswrapper[4805]: W1203 14:09:30.529164 4805 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Dec 03 14:09:30 crc kubenswrapper[4805]: W1203 14:09:30.529169 4805 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Dec 03 14:09:30 crc kubenswrapper[4805]: W1203 14:09:30.529176 4805 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Dec 03 14:09:30 crc kubenswrapper[4805]: W1203 14:09:30.529183 4805 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Dec 03 14:09:30 crc kubenswrapper[4805]: W1203 14:09:30.529190 4805 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Dec 03 14:09:30 crc kubenswrapper[4805]: W1203 14:09:30.529195 4805 feature_gate.go:330] unrecognized feature gate: PinnedImages Dec 03 14:09:30 crc kubenswrapper[4805]: W1203 14:09:30.529200 4805 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Dec 03 14:09:30 crc kubenswrapper[4805]: W1203 14:09:30.529205 4805 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Dec 03 14:09:30 crc kubenswrapper[4805]: W1203 14:09:30.529211 4805 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Dec 03 14:09:30 crc kubenswrapper[4805]: W1203 14:09:30.529216 4805 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Dec 03 14:09:30 crc kubenswrapper[4805]: W1203 14:09:30.529222 4805 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Dec 03 14:09:30 crc kubenswrapper[4805]: W1203 14:09:30.529227 4805 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Dec 03 14:09:30 crc kubenswrapper[4805]: W1203 14:09:30.529232 4805 feature_gate.go:330] unrecognized feature gate: NewOLM Dec 03 14:09:30 crc kubenswrapper[4805]: W1203 14:09:30.529237 4805 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Dec 03 14:09:30 crc kubenswrapper[4805]: W1203 14:09:30.529243 4805 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Dec 03 14:09:30 crc kubenswrapper[4805]: W1203 14:09:30.529248 4805 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Dec 03 14:09:30 crc kubenswrapper[4805]: W1203 14:09:30.529253 4805 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Dec 03 14:09:30 crc kubenswrapper[4805]: W1203 14:09:30.529259 4805 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Dec 03 14:09:30 crc kubenswrapper[4805]: W1203 14:09:30.529264 4805 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Dec 03 14:09:30 crc kubenswrapper[4805]: W1203 14:09:30.529269 4805 feature_gate.go:330] unrecognized feature gate: GatewayAPI Dec 03 14:09:30 crc kubenswrapper[4805]: W1203 14:09:30.529274 4805 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Dec 03 14:09:30 crc kubenswrapper[4805]: W1203 14:09:30.529280 4805 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Dec 03 14:09:30 crc kubenswrapper[4805]: W1203 14:09:30.529286 4805 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Dec 03 14:09:30 crc kubenswrapper[4805]: W1203 14:09:30.529292 4805 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Dec 03 14:09:30 crc kubenswrapper[4805]: W1203 14:09:30.529297 4805 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Dec 03 14:09:30 crc kubenswrapper[4805]: W1203 14:09:30.529302 4805 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Dec 03 14:09:30 crc kubenswrapper[4805]: W1203 14:09:30.529308 4805 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Dec 03 14:09:30 crc kubenswrapper[4805]: W1203 14:09:30.529315 4805 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Dec 03 14:09:30 crc kubenswrapper[4805]: W1203 14:09:30.529322 4805 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Dec 03 14:09:30 crc kubenswrapper[4805]: W1203 14:09:30.529327 4805 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Dec 03 14:09:30 crc kubenswrapper[4805]: W1203 14:09:30.529336 4805 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Dec 03 14:09:30 crc kubenswrapper[4805]: W1203 14:09:30.529342 4805 feature_gate.go:330] unrecognized feature gate: PlatformOperators Dec 03 14:09:30 crc kubenswrapper[4805]: W1203 14:09:30.529348 4805 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Dec 03 14:09:30 crc kubenswrapper[4805]: W1203 14:09:30.529353 4805 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Dec 03 14:09:30 crc kubenswrapper[4805]: W1203 14:09:30.529358 4805 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Dec 03 14:09:30 crc kubenswrapper[4805]: W1203 14:09:30.529363 4805 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Dec 03 14:09:30 crc kubenswrapper[4805]: W1203 14:09:30.529370 4805 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Dec 03 14:09:30 crc kubenswrapper[4805]: W1203 14:09:30.529376 4805 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Dec 03 14:09:30 crc kubenswrapper[4805]: I1203 14:09:30.529481 4805 flags.go:64] FLAG: --address="0.0.0.0" Dec 03 14:09:30 crc kubenswrapper[4805]: I1203 14:09:30.529492 4805 flags.go:64] FLAG: --allowed-unsafe-sysctls="[]" Dec 03 14:09:30 crc kubenswrapper[4805]: I1203 14:09:30.529504 4805 flags.go:64] FLAG: --anonymous-auth="true" Dec 03 14:09:30 crc kubenswrapper[4805]: I1203 14:09:30.529512 4805 flags.go:64] FLAG: --application-metrics-count-limit="100" Dec 03 14:09:30 crc kubenswrapper[4805]: I1203 14:09:30.529519 4805 flags.go:64] FLAG: --authentication-token-webhook="false" Dec 03 14:09:30 crc kubenswrapper[4805]: I1203 14:09:30.529525 4805 flags.go:64] FLAG: --authentication-token-webhook-cache-ttl="2m0s" Dec 03 14:09:30 crc kubenswrapper[4805]: I1203 14:09:30.529533 4805 flags.go:64] FLAG: --authorization-mode="AlwaysAllow" Dec 03 14:09:30 crc kubenswrapper[4805]: I1203 14:09:30.529541 4805 flags.go:64] FLAG: --authorization-webhook-cache-authorized-ttl="5m0s" Dec 03 14:09:30 crc kubenswrapper[4805]: I1203 14:09:30.529548 4805 flags.go:64] FLAG: --authorization-webhook-cache-unauthorized-ttl="30s" Dec 03 14:09:30 crc kubenswrapper[4805]: I1203 14:09:30.529555 4805 flags.go:64] FLAG: --boot-id-file="/proc/sys/kernel/random/boot_id" Dec 03 14:09:30 crc kubenswrapper[4805]: I1203 14:09:30.529562 4805 flags.go:64] FLAG: --bootstrap-kubeconfig="/etc/kubernetes/kubeconfig" Dec 03 14:09:30 crc kubenswrapper[4805]: I1203 14:09:30.529569 4805 flags.go:64] FLAG: --cert-dir="/var/lib/kubelet/pki" Dec 03 14:09:30 crc kubenswrapper[4805]: I1203 14:09:30.529575 4805 flags.go:64] FLAG: --cgroup-driver="cgroupfs" Dec 03 14:09:30 crc kubenswrapper[4805]: I1203 14:09:30.529582 4805 flags.go:64] FLAG: --cgroup-root="" Dec 03 14:09:30 crc kubenswrapper[4805]: I1203 14:09:30.529588 4805 flags.go:64] FLAG: --cgroups-per-qos="true" Dec 03 14:09:30 crc kubenswrapper[4805]: I1203 14:09:30.529594 4805 flags.go:64] FLAG: --client-ca-file="" Dec 03 14:09:30 crc kubenswrapper[4805]: I1203 14:09:30.529600 4805 flags.go:64] FLAG: --cloud-config="" Dec 03 14:09:30 crc kubenswrapper[4805]: I1203 14:09:30.529606 4805 flags.go:64] FLAG: --cloud-provider="" Dec 03 14:09:30 crc kubenswrapper[4805]: I1203 14:09:30.529613 4805 flags.go:64] FLAG: --cluster-dns="[]" Dec 03 14:09:30 crc kubenswrapper[4805]: I1203 14:09:30.529620 4805 flags.go:64] FLAG: --cluster-domain="" Dec 03 14:09:30 crc kubenswrapper[4805]: I1203 14:09:30.529627 4805 flags.go:64] FLAG: --config="/etc/kubernetes/kubelet.conf" Dec 03 14:09:30 crc kubenswrapper[4805]: I1203 14:09:30.529633 4805 flags.go:64] FLAG: --config-dir="" Dec 03 14:09:30 crc kubenswrapper[4805]: I1203 14:09:30.529640 4805 flags.go:64] FLAG: --container-hints="/etc/cadvisor/container_hints.json" Dec 03 14:09:30 crc kubenswrapper[4805]: I1203 14:09:30.529646 4805 flags.go:64] FLAG: --container-log-max-files="5" Dec 03 14:09:30 crc kubenswrapper[4805]: I1203 14:09:30.529654 4805 flags.go:64] FLAG: --container-log-max-size="10Mi" Dec 03 14:09:30 crc kubenswrapper[4805]: I1203 14:09:30.529660 4805 flags.go:64] FLAG: --container-runtime-endpoint="/var/run/crio/crio.sock" Dec 03 14:09:30 crc kubenswrapper[4805]: I1203 14:09:30.529668 4805 flags.go:64] FLAG: --containerd="/run/containerd/containerd.sock" Dec 03 14:09:30 crc kubenswrapper[4805]: I1203 14:09:30.529677 4805 flags.go:64] FLAG: --containerd-namespace="k8s.io" Dec 03 14:09:30 crc kubenswrapper[4805]: I1203 14:09:30.529685 4805 flags.go:64] FLAG: --contention-profiling="false" Dec 03 14:09:30 crc kubenswrapper[4805]: I1203 14:09:30.529692 4805 flags.go:64] FLAG: --cpu-cfs-quota="true" Dec 03 14:09:30 crc kubenswrapper[4805]: I1203 14:09:30.529699 4805 flags.go:64] FLAG: --cpu-cfs-quota-period="100ms" Dec 03 14:09:30 crc kubenswrapper[4805]: I1203 14:09:30.529707 4805 flags.go:64] FLAG: --cpu-manager-policy="none" Dec 03 14:09:30 crc kubenswrapper[4805]: I1203 14:09:30.529714 4805 flags.go:64] FLAG: --cpu-manager-policy-options="" Dec 03 14:09:30 crc kubenswrapper[4805]: I1203 14:09:30.529723 4805 flags.go:64] FLAG: --cpu-manager-reconcile-period="10s" Dec 03 14:09:30 crc kubenswrapper[4805]: I1203 14:09:30.529731 4805 flags.go:64] FLAG: --enable-controller-attach-detach="true" Dec 03 14:09:30 crc kubenswrapper[4805]: I1203 14:09:30.529739 4805 flags.go:64] FLAG: --enable-debugging-handlers="true" Dec 03 14:09:30 crc kubenswrapper[4805]: I1203 14:09:30.529749 4805 flags.go:64] FLAG: --enable-load-reader="false" Dec 03 14:09:30 crc kubenswrapper[4805]: I1203 14:09:30.529760 4805 flags.go:64] FLAG: --enable-server="true" Dec 03 14:09:30 crc kubenswrapper[4805]: I1203 14:09:30.529770 4805 flags.go:64] FLAG: --enforce-node-allocatable="[pods]" Dec 03 14:09:30 crc kubenswrapper[4805]: I1203 14:09:30.529782 4805 flags.go:64] FLAG: --event-burst="100" Dec 03 14:09:30 crc kubenswrapper[4805]: I1203 14:09:30.529790 4805 flags.go:64] FLAG: --event-qps="50" Dec 03 14:09:30 crc kubenswrapper[4805]: I1203 14:09:30.529797 4805 flags.go:64] FLAG: --event-storage-age-limit="default=0" Dec 03 14:09:30 crc kubenswrapper[4805]: I1203 14:09:30.529804 4805 flags.go:64] FLAG: --event-storage-event-limit="default=0" Dec 03 14:09:30 crc kubenswrapper[4805]: I1203 14:09:30.529811 4805 flags.go:64] FLAG: --eviction-hard="" Dec 03 14:09:30 crc kubenswrapper[4805]: I1203 14:09:30.529820 4805 flags.go:64] FLAG: --eviction-max-pod-grace-period="0" Dec 03 14:09:30 crc kubenswrapper[4805]: I1203 14:09:30.529828 4805 flags.go:64] FLAG: --eviction-minimum-reclaim="" Dec 03 14:09:30 crc kubenswrapper[4805]: I1203 14:09:30.529835 4805 flags.go:64] FLAG: --eviction-pressure-transition-period="5m0s" Dec 03 14:09:30 crc kubenswrapper[4805]: I1203 14:09:30.529869 4805 flags.go:64] FLAG: --eviction-soft="" Dec 03 14:09:30 crc kubenswrapper[4805]: I1203 14:09:30.529877 4805 flags.go:64] FLAG: --eviction-soft-grace-period="" Dec 03 14:09:30 crc kubenswrapper[4805]: I1203 14:09:30.529885 4805 flags.go:64] FLAG: --exit-on-lock-contention="false" Dec 03 14:09:30 crc kubenswrapper[4805]: I1203 14:09:30.529892 4805 flags.go:64] FLAG: --experimental-allocatable-ignore-eviction="false" Dec 03 14:09:30 crc kubenswrapper[4805]: I1203 14:09:30.529900 4805 flags.go:64] FLAG: --experimental-mounter-path="" Dec 03 14:09:30 crc kubenswrapper[4805]: I1203 14:09:30.529907 4805 flags.go:64] FLAG: --fail-cgroupv1="false" Dec 03 14:09:30 crc kubenswrapper[4805]: I1203 14:09:30.529914 4805 flags.go:64] FLAG: --fail-swap-on="true" Dec 03 14:09:30 crc kubenswrapper[4805]: I1203 14:09:30.529922 4805 flags.go:64] FLAG: --feature-gates="" Dec 03 14:09:30 crc kubenswrapper[4805]: I1203 14:09:30.529937 4805 flags.go:64] FLAG: --file-check-frequency="20s" Dec 03 14:09:30 crc kubenswrapper[4805]: I1203 14:09:30.529944 4805 flags.go:64] FLAG: --global-housekeeping-interval="1m0s" Dec 03 14:09:30 crc kubenswrapper[4805]: I1203 14:09:30.529952 4805 flags.go:64] FLAG: --hairpin-mode="promiscuous-bridge" Dec 03 14:09:30 crc kubenswrapper[4805]: I1203 14:09:30.529960 4805 flags.go:64] FLAG: --healthz-bind-address="127.0.0.1" Dec 03 14:09:30 crc kubenswrapper[4805]: I1203 14:09:30.529967 4805 flags.go:64] FLAG: --healthz-port="10248" Dec 03 14:09:30 crc kubenswrapper[4805]: I1203 14:09:30.529975 4805 flags.go:64] FLAG: --help="false" Dec 03 14:09:30 crc kubenswrapper[4805]: I1203 14:09:30.529983 4805 flags.go:64] FLAG: --hostname-override="" Dec 03 14:09:30 crc kubenswrapper[4805]: I1203 14:09:30.529989 4805 flags.go:64] FLAG: --housekeeping-interval="10s" Dec 03 14:09:30 crc kubenswrapper[4805]: I1203 14:09:30.529997 4805 flags.go:64] FLAG: --http-check-frequency="20s" Dec 03 14:09:30 crc kubenswrapper[4805]: I1203 14:09:30.530004 4805 flags.go:64] FLAG: --image-credential-provider-bin-dir="" Dec 03 14:09:30 crc kubenswrapper[4805]: I1203 14:09:30.530012 4805 flags.go:64] FLAG: --image-credential-provider-config="" Dec 03 14:09:30 crc kubenswrapper[4805]: I1203 14:09:30.530020 4805 flags.go:64] FLAG: --image-gc-high-threshold="85" Dec 03 14:09:30 crc kubenswrapper[4805]: I1203 14:09:30.530030 4805 flags.go:64] FLAG: --image-gc-low-threshold="80" Dec 03 14:09:30 crc kubenswrapper[4805]: I1203 14:09:30.530037 4805 flags.go:64] FLAG: --image-service-endpoint="" Dec 03 14:09:30 crc kubenswrapper[4805]: I1203 14:09:30.530044 4805 flags.go:64] FLAG: --kernel-memcg-notification="false" Dec 03 14:09:30 crc kubenswrapper[4805]: I1203 14:09:30.530052 4805 flags.go:64] FLAG: --kube-api-burst="100" Dec 03 14:09:30 crc kubenswrapper[4805]: I1203 14:09:30.530060 4805 flags.go:64] FLAG: --kube-api-content-type="application/vnd.kubernetes.protobuf" Dec 03 14:09:30 crc kubenswrapper[4805]: I1203 14:09:30.530068 4805 flags.go:64] FLAG: --kube-api-qps="50" Dec 03 14:09:30 crc kubenswrapper[4805]: I1203 14:09:30.530075 4805 flags.go:64] FLAG: --kube-reserved="" Dec 03 14:09:30 crc kubenswrapper[4805]: I1203 14:09:30.530085 4805 flags.go:64] FLAG: --kube-reserved-cgroup="" Dec 03 14:09:30 crc kubenswrapper[4805]: I1203 14:09:30.530092 4805 flags.go:64] FLAG: --kubeconfig="/var/lib/kubelet/kubeconfig" Dec 03 14:09:30 crc kubenswrapper[4805]: I1203 14:09:30.530102 4805 flags.go:64] FLAG: --kubelet-cgroups="" Dec 03 14:09:30 crc kubenswrapper[4805]: I1203 14:09:30.530109 4805 flags.go:64] FLAG: --local-storage-capacity-isolation="true" Dec 03 14:09:30 crc kubenswrapper[4805]: I1203 14:09:30.530117 4805 flags.go:64] FLAG: --lock-file="" Dec 03 14:09:30 crc kubenswrapper[4805]: I1203 14:09:30.530124 4805 flags.go:64] FLAG: --log-cadvisor-usage="false" Dec 03 14:09:30 crc kubenswrapper[4805]: I1203 14:09:30.530132 4805 flags.go:64] FLAG: --log-flush-frequency="5s" Dec 03 14:09:30 crc kubenswrapper[4805]: I1203 14:09:30.530140 4805 flags.go:64] FLAG: --log-json-info-buffer-size="0" Dec 03 14:09:30 crc kubenswrapper[4805]: I1203 14:09:30.530153 4805 flags.go:64] FLAG: --log-json-split-stream="false" Dec 03 14:09:30 crc kubenswrapper[4805]: I1203 14:09:30.530161 4805 flags.go:64] FLAG: --log-text-info-buffer-size="0" Dec 03 14:09:30 crc kubenswrapper[4805]: I1203 14:09:30.530169 4805 flags.go:64] FLAG: --log-text-split-stream="false" Dec 03 14:09:30 crc kubenswrapper[4805]: I1203 14:09:30.530177 4805 flags.go:64] FLAG: --logging-format="text" Dec 03 14:09:30 crc kubenswrapper[4805]: I1203 14:09:30.530185 4805 flags.go:64] FLAG: --machine-id-file="/etc/machine-id,/var/lib/dbus/machine-id" Dec 03 14:09:30 crc kubenswrapper[4805]: I1203 14:09:30.530193 4805 flags.go:64] FLAG: --make-iptables-util-chains="true" Dec 03 14:09:30 crc kubenswrapper[4805]: I1203 14:09:30.530201 4805 flags.go:64] FLAG: --manifest-url="" Dec 03 14:09:30 crc kubenswrapper[4805]: I1203 14:09:30.530208 4805 flags.go:64] FLAG: --manifest-url-header="" Dec 03 14:09:30 crc kubenswrapper[4805]: I1203 14:09:30.530218 4805 flags.go:64] FLAG: --max-housekeeping-interval="15s" Dec 03 14:09:30 crc kubenswrapper[4805]: I1203 14:09:30.530226 4805 flags.go:64] FLAG: --max-open-files="1000000" Dec 03 14:09:30 crc kubenswrapper[4805]: I1203 14:09:30.530235 4805 flags.go:64] FLAG: --max-pods="110" Dec 03 14:09:30 crc kubenswrapper[4805]: I1203 14:09:30.530243 4805 flags.go:64] FLAG: --maximum-dead-containers="-1" Dec 03 14:09:30 crc kubenswrapper[4805]: I1203 14:09:30.530250 4805 flags.go:64] FLAG: --maximum-dead-containers-per-container="1" Dec 03 14:09:30 crc kubenswrapper[4805]: I1203 14:09:30.530257 4805 flags.go:64] FLAG: --memory-manager-policy="None" Dec 03 14:09:30 crc kubenswrapper[4805]: I1203 14:09:30.530265 4805 flags.go:64] FLAG: --minimum-container-ttl-duration="6m0s" Dec 03 14:09:30 crc kubenswrapper[4805]: I1203 14:09:30.530272 4805 flags.go:64] FLAG: --minimum-image-ttl-duration="2m0s" Dec 03 14:09:30 crc kubenswrapper[4805]: I1203 14:09:30.530279 4805 flags.go:64] FLAG: --node-ip="192.168.126.11" Dec 03 14:09:30 crc kubenswrapper[4805]: I1203 14:09:30.530287 4805 flags.go:64] FLAG: --node-labels="node-role.kubernetes.io/control-plane=,node-role.kubernetes.io/master=,node.openshift.io/os_id=rhcos" Dec 03 14:09:30 crc kubenswrapper[4805]: I1203 14:09:30.530303 4805 flags.go:64] FLAG: --node-status-max-images="50" Dec 03 14:09:30 crc kubenswrapper[4805]: I1203 14:09:30.530310 4805 flags.go:64] FLAG: --node-status-update-frequency="10s" Dec 03 14:09:30 crc kubenswrapper[4805]: I1203 14:09:30.530318 4805 flags.go:64] FLAG: --oom-score-adj="-999" Dec 03 14:09:30 crc kubenswrapper[4805]: I1203 14:09:30.530325 4805 flags.go:64] FLAG: --pod-cidr="" Dec 03 14:09:30 crc kubenswrapper[4805]: I1203 14:09:30.530331 4805 flags.go:64] FLAG: --pod-infra-container-image="quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:33549946e22a9ffa738fd94b1345f90921bc8f92fa6137784cb33c77ad806f9d" Dec 03 14:09:30 crc kubenswrapper[4805]: I1203 14:09:30.530343 4805 flags.go:64] FLAG: --pod-manifest-path="" Dec 03 14:09:30 crc kubenswrapper[4805]: I1203 14:09:30.530351 4805 flags.go:64] FLAG: --pod-max-pids="-1" Dec 03 14:09:30 crc kubenswrapper[4805]: I1203 14:09:30.530359 4805 flags.go:64] FLAG: --pods-per-core="0" Dec 03 14:09:30 crc kubenswrapper[4805]: I1203 14:09:30.530365 4805 flags.go:64] FLAG: --port="10250" Dec 03 14:09:30 crc kubenswrapper[4805]: I1203 14:09:30.530372 4805 flags.go:64] FLAG: --protect-kernel-defaults="false" Dec 03 14:09:30 crc kubenswrapper[4805]: I1203 14:09:30.530381 4805 flags.go:64] FLAG: --provider-id="" Dec 03 14:09:30 crc kubenswrapper[4805]: I1203 14:09:30.530388 4805 flags.go:64] FLAG: --qos-reserved="" Dec 03 14:09:30 crc kubenswrapper[4805]: I1203 14:09:30.530395 4805 flags.go:64] FLAG: --read-only-port="10255" Dec 03 14:09:30 crc kubenswrapper[4805]: I1203 14:09:30.530402 4805 flags.go:64] FLAG: --register-node="true" Dec 03 14:09:30 crc kubenswrapper[4805]: I1203 14:09:30.530410 4805 flags.go:64] FLAG: --register-schedulable="true" Dec 03 14:09:30 crc kubenswrapper[4805]: I1203 14:09:30.530417 4805 flags.go:64] FLAG: --register-with-taints="node-role.kubernetes.io/master=:NoSchedule" Dec 03 14:09:30 crc kubenswrapper[4805]: I1203 14:09:30.530430 4805 flags.go:64] FLAG: --registry-burst="10" Dec 03 14:09:30 crc kubenswrapper[4805]: I1203 14:09:30.530437 4805 flags.go:64] FLAG: --registry-qps="5" Dec 03 14:09:30 crc kubenswrapper[4805]: I1203 14:09:30.530444 4805 flags.go:64] FLAG: --reserved-cpus="" Dec 03 14:09:30 crc kubenswrapper[4805]: I1203 14:09:30.530451 4805 flags.go:64] FLAG: --reserved-memory="" Dec 03 14:09:30 crc kubenswrapper[4805]: I1203 14:09:30.530459 4805 flags.go:64] FLAG: --resolv-conf="/etc/resolv.conf" Dec 03 14:09:30 crc kubenswrapper[4805]: I1203 14:09:30.530467 4805 flags.go:64] FLAG: --root-dir="/var/lib/kubelet" Dec 03 14:09:30 crc kubenswrapper[4805]: I1203 14:09:30.530475 4805 flags.go:64] FLAG: --rotate-certificates="false" Dec 03 14:09:30 crc kubenswrapper[4805]: I1203 14:09:30.530482 4805 flags.go:64] FLAG: --rotate-server-certificates="false" Dec 03 14:09:30 crc kubenswrapper[4805]: I1203 14:09:30.530488 4805 flags.go:64] FLAG: --runonce="false" Dec 03 14:09:30 crc kubenswrapper[4805]: I1203 14:09:30.530495 4805 flags.go:64] FLAG: --runtime-cgroups="/system.slice/crio.service" Dec 03 14:09:30 crc kubenswrapper[4805]: I1203 14:09:30.530503 4805 flags.go:64] FLAG: --runtime-request-timeout="2m0s" Dec 03 14:09:30 crc kubenswrapper[4805]: I1203 14:09:30.530510 4805 flags.go:64] FLAG: --seccomp-default="false" Dec 03 14:09:30 crc kubenswrapper[4805]: I1203 14:09:30.530517 4805 flags.go:64] FLAG: --serialize-image-pulls="true" Dec 03 14:09:30 crc kubenswrapper[4805]: I1203 14:09:30.530524 4805 flags.go:64] FLAG: --storage-driver-buffer-duration="1m0s" Dec 03 14:09:30 crc kubenswrapper[4805]: I1203 14:09:30.530531 4805 flags.go:64] FLAG: --storage-driver-db="cadvisor" Dec 03 14:09:30 crc kubenswrapper[4805]: I1203 14:09:30.530539 4805 flags.go:64] FLAG: --storage-driver-host="localhost:8086" Dec 03 14:09:30 crc kubenswrapper[4805]: I1203 14:09:30.530546 4805 flags.go:64] FLAG: --storage-driver-password="root" Dec 03 14:09:30 crc kubenswrapper[4805]: I1203 14:09:30.530553 4805 flags.go:64] FLAG: --storage-driver-secure="false" Dec 03 14:09:30 crc kubenswrapper[4805]: I1203 14:09:30.530560 4805 flags.go:64] FLAG: --storage-driver-table="stats" Dec 03 14:09:30 crc kubenswrapper[4805]: I1203 14:09:30.530567 4805 flags.go:64] FLAG: --storage-driver-user="root" Dec 03 14:09:30 crc kubenswrapper[4805]: I1203 14:09:30.530574 4805 flags.go:64] FLAG: --streaming-connection-idle-timeout="4h0m0s" Dec 03 14:09:30 crc kubenswrapper[4805]: I1203 14:09:30.530581 4805 flags.go:64] FLAG: --sync-frequency="1m0s" Dec 03 14:09:30 crc kubenswrapper[4805]: I1203 14:09:30.530590 4805 flags.go:64] FLAG: --system-cgroups="" Dec 03 14:09:30 crc kubenswrapper[4805]: I1203 14:09:30.530597 4805 flags.go:64] FLAG: --system-reserved="cpu=200m,ephemeral-storage=350Mi,memory=350Mi" Dec 03 14:09:30 crc kubenswrapper[4805]: I1203 14:09:30.530608 4805 flags.go:64] FLAG: --system-reserved-cgroup="" Dec 03 14:09:30 crc kubenswrapper[4805]: I1203 14:09:30.530614 4805 flags.go:64] FLAG: --tls-cert-file="" Dec 03 14:09:30 crc kubenswrapper[4805]: I1203 14:09:30.530621 4805 flags.go:64] FLAG: --tls-cipher-suites="[]" Dec 03 14:09:30 crc kubenswrapper[4805]: I1203 14:09:30.530630 4805 flags.go:64] FLAG: --tls-min-version="" Dec 03 14:09:30 crc kubenswrapper[4805]: I1203 14:09:30.530637 4805 flags.go:64] FLAG: --tls-private-key-file="" Dec 03 14:09:30 crc kubenswrapper[4805]: I1203 14:09:30.530644 4805 flags.go:64] FLAG: --topology-manager-policy="none" Dec 03 14:09:30 crc kubenswrapper[4805]: I1203 14:09:30.530652 4805 flags.go:64] FLAG: --topology-manager-policy-options="" Dec 03 14:09:30 crc kubenswrapper[4805]: I1203 14:09:30.530659 4805 flags.go:64] FLAG: --topology-manager-scope="container" Dec 03 14:09:30 crc kubenswrapper[4805]: I1203 14:09:30.530668 4805 flags.go:64] FLAG: --v="2" Dec 03 14:09:30 crc kubenswrapper[4805]: I1203 14:09:30.530678 4805 flags.go:64] FLAG: --version="false" Dec 03 14:09:30 crc kubenswrapper[4805]: I1203 14:09:30.530687 4805 flags.go:64] FLAG: --vmodule="" Dec 03 14:09:30 crc kubenswrapper[4805]: I1203 14:09:30.530696 4805 flags.go:64] FLAG: --volume-plugin-dir="/etc/kubernetes/kubelet-plugins/volume/exec" Dec 03 14:09:30 crc kubenswrapper[4805]: I1203 14:09:30.530703 4805 flags.go:64] FLAG: --volume-stats-agg-period="1m0s" Dec 03 14:09:30 crc kubenswrapper[4805]: W1203 14:09:30.530944 4805 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Dec 03 14:09:30 crc kubenswrapper[4805]: W1203 14:09:30.530955 4805 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Dec 03 14:09:30 crc kubenswrapper[4805]: W1203 14:09:30.530961 4805 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Dec 03 14:09:30 crc kubenswrapper[4805]: W1203 14:09:30.530968 4805 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Dec 03 14:09:30 crc kubenswrapper[4805]: W1203 14:09:30.530974 4805 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Dec 03 14:09:30 crc kubenswrapper[4805]: W1203 14:09:30.530981 4805 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Dec 03 14:09:30 crc kubenswrapper[4805]: W1203 14:09:30.530988 4805 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Dec 03 14:09:30 crc kubenswrapper[4805]: W1203 14:09:30.530994 4805 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Dec 03 14:09:30 crc kubenswrapper[4805]: W1203 14:09:30.531000 4805 feature_gate.go:330] unrecognized feature gate: Example Dec 03 14:09:30 crc kubenswrapper[4805]: W1203 14:09:30.531008 4805 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Dec 03 14:09:30 crc kubenswrapper[4805]: W1203 14:09:30.531018 4805 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Dec 03 14:09:30 crc kubenswrapper[4805]: W1203 14:09:30.531026 4805 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Dec 03 14:09:30 crc kubenswrapper[4805]: W1203 14:09:30.531032 4805 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Dec 03 14:09:30 crc kubenswrapper[4805]: W1203 14:09:30.531039 4805 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Dec 03 14:09:30 crc kubenswrapper[4805]: W1203 14:09:30.531047 4805 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Dec 03 14:09:30 crc kubenswrapper[4805]: W1203 14:09:30.531055 4805 feature_gate.go:330] unrecognized feature gate: NewOLM Dec 03 14:09:30 crc kubenswrapper[4805]: W1203 14:09:30.531062 4805 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Dec 03 14:09:30 crc kubenswrapper[4805]: W1203 14:09:30.531070 4805 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Dec 03 14:09:30 crc kubenswrapper[4805]: W1203 14:09:30.531078 4805 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Dec 03 14:09:30 crc kubenswrapper[4805]: W1203 14:09:30.531085 4805 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Dec 03 14:09:30 crc kubenswrapper[4805]: W1203 14:09:30.531092 4805 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Dec 03 14:09:30 crc kubenswrapper[4805]: W1203 14:09:30.531099 4805 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Dec 03 14:09:30 crc kubenswrapper[4805]: W1203 14:09:30.531105 4805 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Dec 03 14:09:30 crc kubenswrapper[4805]: W1203 14:09:30.531112 4805 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Dec 03 14:09:30 crc kubenswrapper[4805]: W1203 14:09:30.531120 4805 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Dec 03 14:09:30 crc kubenswrapper[4805]: W1203 14:09:30.531127 4805 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Dec 03 14:09:30 crc kubenswrapper[4805]: W1203 14:09:30.531134 4805 feature_gate.go:330] unrecognized feature gate: SignatureStores Dec 03 14:09:30 crc kubenswrapper[4805]: W1203 14:09:30.531141 4805 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Dec 03 14:09:30 crc kubenswrapper[4805]: W1203 14:09:30.531147 4805 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Dec 03 14:09:30 crc kubenswrapper[4805]: W1203 14:09:30.531155 4805 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Dec 03 14:09:30 crc kubenswrapper[4805]: W1203 14:09:30.531161 4805 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Dec 03 14:09:30 crc kubenswrapper[4805]: W1203 14:09:30.531168 4805 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Dec 03 14:09:30 crc kubenswrapper[4805]: W1203 14:09:30.531174 4805 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Dec 03 14:09:30 crc kubenswrapper[4805]: W1203 14:09:30.531181 4805 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Dec 03 14:09:30 crc kubenswrapper[4805]: W1203 14:09:30.531187 4805 feature_gate.go:330] unrecognized feature gate: GatewayAPI Dec 03 14:09:30 crc kubenswrapper[4805]: W1203 14:09:30.531194 4805 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Dec 03 14:09:30 crc kubenswrapper[4805]: W1203 14:09:30.531200 4805 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Dec 03 14:09:30 crc kubenswrapper[4805]: W1203 14:09:30.531207 4805 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Dec 03 14:09:30 crc kubenswrapper[4805]: W1203 14:09:30.531213 4805 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Dec 03 14:09:30 crc kubenswrapper[4805]: W1203 14:09:30.531220 4805 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Dec 03 14:09:30 crc kubenswrapper[4805]: W1203 14:09:30.531227 4805 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Dec 03 14:09:30 crc kubenswrapper[4805]: W1203 14:09:30.531233 4805 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Dec 03 14:09:30 crc kubenswrapper[4805]: W1203 14:09:30.531240 4805 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Dec 03 14:09:30 crc kubenswrapper[4805]: W1203 14:09:30.531250 4805 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Dec 03 14:09:30 crc kubenswrapper[4805]: W1203 14:09:30.531260 4805 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Dec 03 14:09:30 crc kubenswrapper[4805]: W1203 14:09:30.531269 4805 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Dec 03 14:09:30 crc kubenswrapper[4805]: W1203 14:09:30.531277 4805 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Dec 03 14:09:30 crc kubenswrapper[4805]: W1203 14:09:30.531283 4805 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Dec 03 14:09:30 crc kubenswrapper[4805]: W1203 14:09:30.531290 4805 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Dec 03 14:09:30 crc kubenswrapper[4805]: W1203 14:09:30.531297 4805 feature_gate.go:330] unrecognized feature gate: InsightsConfig Dec 03 14:09:30 crc kubenswrapper[4805]: W1203 14:09:30.531304 4805 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Dec 03 14:09:30 crc kubenswrapper[4805]: W1203 14:09:30.531311 4805 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Dec 03 14:09:30 crc kubenswrapper[4805]: W1203 14:09:30.531317 4805 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Dec 03 14:09:30 crc kubenswrapper[4805]: W1203 14:09:30.531324 4805 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Dec 03 14:09:30 crc kubenswrapper[4805]: W1203 14:09:30.531330 4805 feature_gate.go:330] unrecognized feature gate: PinnedImages Dec 03 14:09:30 crc kubenswrapper[4805]: W1203 14:09:30.531337 4805 feature_gate.go:330] unrecognized feature gate: OVNObservability Dec 03 14:09:30 crc kubenswrapper[4805]: W1203 14:09:30.531345 4805 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Dec 03 14:09:30 crc kubenswrapper[4805]: W1203 14:09:30.531353 4805 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Dec 03 14:09:30 crc kubenswrapper[4805]: W1203 14:09:30.531360 4805 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Dec 03 14:09:30 crc kubenswrapper[4805]: W1203 14:09:30.531366 4805 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Dec 03 14:09:30 crc kubenswrapper[4805]: W1203 14:09:30.531373 4805 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Dec 03 14:09:30 crc kubenswrapper[4805]: W1203 14:09:30.531379 4805 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Dec 03 14:09:30 crc kubenswrapper[4805]: W1203 14:09:30.531386 4805 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Dec 03 14:09:30 crc kubenswrapper[4805]: W1203 14:09:30.531392 4805 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Dec 03 14:09:30 crc kubenswrapper[4805]: W1203 14:09:30.531398 4805 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Dec 03 14:09:30 crc kubenswrapper[4805]: W1203 14:09:30.531406 4805 feature_gate.go:330] unrecognized feature gate: PlatformOperators Dec 03 14:09:30 crc kubenswrapper[4805]: W1203 14:09:30.531412 4805 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Dec 03 14:09:30 crc kubenswrapper[4805]: W1203 14:09:30.531418 4805 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Dec 03 14:09:30 crc kubenswrapper[4805]: W1203 14:09:30.531425 4805 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Dec 03 14:09:30 crc kubenswrapper[4805]: W1203 14:09:30.531432 4805 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Dec 03 14:09:30 crc kubenswrapper[4805]: W1203 14:09:30.531439 4805 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Dec 03 14:09:30 crc kubenswrapper[4805]: I1203 14:09:30.531450 4805 feature_gate.go:386] feature gates: {map[CloudDualStackNodeIPs:true DisableKubeletCloudCredentialProviders:true DynamicResourceAllocation:false EventedPLEG:false KMSv1:true MaxUnavailableStatefulSet:false NodeSwap:false ProcMountType:false RouteExternalCertificate:false ServiceAccountTokenNodeBinding:false TranslateStreamCloseWebsocketRequests:false UserNamespacesPodSecurityStandards:false UserNamespacesSupport:false ValidatingAdmissionPolicy:true VolumeAttributesClass:false]} Dec 03 14:09:30 crc kubenswrapper[4805]: I1203 14:09:30.544077 4805 server.go:491] "Kubelet version" kubeletVersion="v1.31.5" Dec 03 14:09:30 crc kubenswrapper[4805]: I1203 14:09:30.544118 4805 server.go:493] "Golang settings" GOGC="" GOMAXPROCS="" GOTRACEBACK="" Dec 03 14:09:30 crc kubenswrapper[4805]: W1203 14:09:30.544223 4805 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Dec 03 14:09:30 crc kubenswrapper[4805]: W1203 14:09:30.544233 4805 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Dec 03 14:09:30 crc kubenswrapper[4805]: W1203 14:09:30.544241 4805 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Dec 03 14:09:30 crc kubenswrapper[4805]: W1203 14:09:30.544247 4805 feature_gate.go:330] unrecognized feature gate: PlatformOperators Dec 03 14:09:30 crc kubenswrapper[4805]: W1203 14:09:30.544253 4805 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Dec 03 14:09:30 crc kubenswrapper[4805]: W1203 14:09:30.544258 4805 feature_gate.go:330] unrecognized feature gate: InsightsConfig Dec 03 14:09:30 crc kubenswrapper[4805]: W1203 14:09:30.544264 4805 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Dec 03 14:09:30 crc kubenswrapper[4805]: W1203 14:09:30.544270 4805 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Dec 03 14:09:30 crc kubenswrapper[4805]: W1203 14:09:30.544276 4805 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Dec 03 14:09:30 crc kubenswrapper[4805]: W1203 14:09:30.544282 4805 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Dec 03 14:09:30 crc kubenswrapper[4805]: W1203 14:09:30.544287 4805 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Dec 03 14:09:30 crc kubenswrapper[4805]: W1203 14:09:30.544292 4805 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Dec 03 14:09:30 crc kubenswrapper[4805]: W1203 14:09:30.544298 4805 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Dec 03 14:09:30 crc kubenswrapper[4805]: W1203 14:09:30.544304 4805 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Dec 03 14:09:30 crc kubenswrapper[4805]: W1203 14:09:30.544309 4805 feature_gate.go:330] unrecognized feature gate: SignatureStores Dec 03 14:09:30 crc kubenswrapper[4805]: W1203 14:09:30.544314 4805 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Dec 03 14:09:30 crc kubenswrapper[4805]: W1203 14:09:30.544319 4805 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Dec 03 14:09:30 crc kubenswrapper[4805]: W1203 14:09:30.544325 4805 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Dec 03 14:09:30 crc kubenswrapper[4805]: W1203 14:09:30.544330 4805 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Dec 03 14:09:30 crc kubenswrapper[4805]: W1203 14:09:30.544335 4805 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Dec 03 14:09:30 crc kubenswrapper[4805]: W1203 14:09:30.544341 4805 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Dec 03 14:09:30 crc kubenswrapper[4805]: W1203 14:09:30.544346 4805 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Dec 03 14:09:30 crc kubenswrapper[4805]: W1203 14:09:30.544351 4805 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Dec 03 14:09:30 crc kubenswrapper[4805]: W1203 14:09:30.544356 4805 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Dec 03 14:09:30 crc kubenswrapper[4805]: W1203 14:09:30.544364 4805 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Dec 03 14:09:30 crc kubenswrapper[4805]: W1203 14:09:30.544374 4805 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Dec 03 14:09:30 crc kubenswrapper[4805]: W1203 14:09:30.544381 4805 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Dec 03 14:09:30 crc kubenswrapper[4805]: W1203 14:09:30.544387 4805 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Dec 03 14:09:30 crc kubenswrapper[4805]: W1203 14:09:30.544393 4805 feature_gate.go:330] unrecognized feature gate: GatewayAPI Dec 03 14:09:30 crc kubenswrapper[4805]: W1203 14:09:30.544398 4805 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Dec 03 14:09:30 crc kubenswrapper[4805]: W1203 14:09:30.544404 4805 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Dec 03 14:09:30 crc kubenswrapper[4805]: W1203 14:09:30.544409 4805 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Dec 03 14:09:30 crc kubenswrapper[4805]: W1203 14:09:30.544415 4805 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Dec 03 14:09:30 crc kubenswrapper[4805]: W1203 14:09:30.544420 4805 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Dec 03 14:09:30 crc kubenswrapper[4805]: W1203 14:09:30.544428 4805 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Dec 03 14:09:30 crc kubenswrapper[4805]: W1203 14:09:30.544433 4805 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Dec 03 14:09:30 crc kubenswrapper[4805]: W1203 14:09:30.544439 4805 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Dec 03 14:09:30 crc kubenswrapper[4805]: W1203 14:09:30.544445 4805 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Dec 03 14:09:30 crc kubenswrapper[4805]: W1203 14:09:30.544450 4805 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Dec 03 14:09:30 crc kubenswrapper[4805]: W1203 14:09:30.544456 4805 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Dec 03 14:09:30 crc kubenswrapper[4805]: W1203 14:09:30.544461 4805 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Dec 03 14:09:30 crc kubenswrapper[4805]: W1203 14:09:30.544467 4805 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Dec 03 14:09:30 crc kubenswrapper[4805]: W1203 14:09:30.544473 4805 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Dec 03 14:09:30 crc kubenswrapper[4805]: W1203 14:09:30.544478 4805 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Dec 03 14:09:30 crc kubenswrapper[4805]: W1203 14:09:30.544484 4805 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Dec 03 14:09:30 crc kubenswrapper[4805]: W1203 14:09:30.544490 4805 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Dec 03 14:09:30 crc kubenswrapper[4805]: W1203 14:09:30.544496 4805 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Dec 03 14:09:30 crc kubenswrapper[4805]: W1203 14:09:30.544501 4805 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Dec 03 14:09:30 crc kubenswrapper[4805]: W1203 14:09:30.544506 4805 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Dec 03 14:09:30 crc kubenswrapper[4805]: W1203 14:09:30.544512 4805 feature_gate.go:330] unrecognized feature gate: Example Dec 03 14:09:30 crc kubenswrapper[4805]: W1203 14:09:30.544517 4805 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Dec 03 14:09:30 crc kubenswrapper[4805]: W1203 14:09:30.544522 4805 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Dec 03 14:09:30 crc kubenswrapper[4805]: W1203 14:09:30.544527 4805 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Dec 03 14:09:30 crc kubenswrapper[4805]: W1203 14:09:30.544532 4805 feature_gate.go:330] unrecognized feature gate: NewOLM Dec 03 14:09:30 crc kubenswrapper[4805]: W1203 14:09:30.544538 4805 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Dec 03 14:09:30 crc kubenswrapper[4805]: W1203 14:09:30.544543 4805 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Dec 03 14:09:30 crc kubenswrapper[4805]: W1203 14:09:30.544548 4805 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Dec 03 14:09:30 crc kubenswrapper[4805]: W1203 14:09:30.544553 4805 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Dec 03 14:09:30 crc kubenswrapper[4805]: W1203 14:09:30.544558 4805 feature_gate.go:330] unrecognized feature gate: OVNObservability Dec 03 14:09:30 crc kubenswrapper[4805]: W1203 14:09:30.544563 4805 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Dec 03 14:09:30 crc kubenswrapper[4805]: W1203 14:09:30.544570 4805 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Dec 03 14:09:30 crc kubenswrapper[4805]: W1203 14:09:30.544576 4805 feature_gate.go:330] unrecognized feature gate: PinnedImages Dec 03 14:09:30 crc kubenswrapper[4805]: W1203 14:09:30.544581 4805 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Dec 03 14:09:30 crc kubenswrapper[4805]: W1203 14:09:30.544587 4805 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Dec 03 14:09:30 crc kubenswrapper[4805]: W1203 14:09:30.544592 4805 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Dec 03 14:09:30 crc kubenswrapper[4805]: W1203 14:09:30.544598 4805 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Dec 03 14:09:30 crc kubenswrapper[4805]: W1203 14:09:30.544603 4805 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Dec 03 14:09:30 crc kubenswrapper[4805]: W1203 14:09:30.544610 4805 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Dec 03 14:09:30 crc kubenswrapper[4805]: W1203 14:09:30.544616 4805 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Dec 03 14:09:30 crc kubenswrapper[4805]: W1203 14:09:30.544621 4805 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Dec 03 14:09:30 crc kubenswrapper[4805]: W1203 14:09:30.544627 4805 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Dec 03 14:09:30 crc kubenswrapper[4805]: I1203 14:09:30.544636 4805 feature_gate.go:386] feature gates: {map[CloudDualStackNodeIPs:true DisableKubeletCloudCredentialProviders:true DynamicResourceAllocation:false EventedPLEG:false KMSv1:true MaxUnavailableStatefulSet:false NodeSwap:false ProcMountType:false RouteExternalCertificate:false ServiceAccountTokenNodeBinding:false TranslateStreamCloseWebsocketRequests:false UserNamespacesPodSecurityStandards:false UserNamespacesSupport:false ValidatingAdmissionPolicy:true VolumeAttributesClass:false]} Dec 03 14:09:30 crc kubenswrapper[4805]: W1203 14:09:30.544792 4805 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Dec 03 14:09:30 crc kubenswrapper[4805]: W1203 14:09:30.544803 4805 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Dec 03 14:09:30 crc kubenswrapper[4805]: W1203 14:09:30.544809 4805 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Dec 03 14:09:30 crc kubenswrapper[4805]: W1203 14:09:30.544816 4805 feature_gate.go:330] unrecognized feature gate: Example Dec 03 14:09:30 crc kubenswrapper[4805]: W1203 14:09:30.544822 4805 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Dec 03 14:09:30 crc kubenswrapper[4805]: W1203 14:09:30.544827 4805 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Dec 03 14:09:30 crc kubenswrapper[4805]: W1203 14:09:30.544832 4805 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Dec 03 14:09:30 crc kubenswrapper[4805]: W1203 14:09:30.544860 4805 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Dec 03 14:09:30 crc kubenswrapper[4805]: W1203 14:09:30.544866 4805 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Dec 03 14:09:30 crc kubenswrapper[4805]: W1203 14:09:30.544871 4805 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Dec 03 14:09:30 crc kubenswrapper[4805]: W1203 14:09:30.544876 4805 feature_gate.go:330] unrecognized feature gate: SignatureStores Dec 03 14:09:30 crc kubenswrapper[4805]: W1203 14:09:30.544881 4805 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Dec 03 14:09:30 crc kubenswrapper[4805]: W1203 14:09:30.544887 4805 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Dec 03 14:09:30 crc kubenswrapper[4805]: W1203 14:09:30.544892 4805 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Dec 03 14:09:30 crc kubenswrapper[4805]: W1203 14:09:30.544897 4805 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Dec 03 14:09:30 crc kubenswrapper[4805]: W1203 14:09:30.544902 4805 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Dec 03 14:09:30 crc kubenswrapper[4805]: W1203 14:09:30.544907 4805 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Dec 03 14:09:30 crc kubenswrapper[4805]: W1203 14:09:30.544912 4805 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Dec 03 14:09:30 crc kubenswrapper[4805]: W1203 14:09:30.544918 4805 feature_gate.go:330] unrecognized feature gate: PinnedImages Dec 03 14:09:30 crc kubenswrapper[4805]: W1203 14:09:30.544923 4805 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Dec 03 14:09:30 crc kubenswrapper[4805]: W1203 14:09:30.544928 4805 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Dec 03 14:09:30 crc kubenswrapper[4805]: W1203 14:09:30.544935 4805 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Dec 03 14:09:30 crc kubenswrapper[4805]: W1203 14:09:30.544940 4805 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Dec 03 14:09:30 crc kubenswrapper[4805]: W1203 14:09:30.544946 4805 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Dec 03 14:09:30 crc kubenswrapper[4805]: W1203 14:09:30.544951 4805 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Dec 03 14:09:30 crc kubenswrapper[4805]: W1203 14:09:30.544956 4805 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Dec 03 14:09:30 crc kubenswrapper[4805]: W1203 14:09:30.544962 4805 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Dec 03 14:09:30 crc kubenswrapper[4805]: W1203 14:09:30.544967 4805 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Dec 03 14:09:30 crc kubenswrapper[4805]: W1203 14:09:30.544972 4805 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Dec 03 14:09:30 crc kubenswrapper[4805]: W1203 14:09:30.544977 4805 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Dec 03 14:09:30 crc kubenswrapper[4805]: W1203 14:09:30.544983 4805 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Dec 03 14:09:30 crc kubenswrapper[4805]: W1203 14:09:30.544988 4805 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Dec 03 14:09:30 crc kubenswrapper[4805]: W1203 14:09:30.544993 4805 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Dec 03 14:09:30 crc kubenswrapper[4805]: W1203 14:09:30.544998 4805 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Dec 03 14:09:30 crc kubenswrapper[4805]: W1203 14:09:30.545004 4805 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Dec 03 14:09:30 crc kubenswrapper[4805]: W1203 14:09:30.545009 4805 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Dec 03 14:09:30 crc kubenswrapper[4805]: W1203 14:09:30.545015 4805 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Dec 03 14:09:30 crc kubenswrapper[4805]: W1203 14:09:30.545021 4805 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Dec 03 14:09:30 crc kubenswrapper[4805]: W1203 14:09:30.545026 4805 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Dec 03 14:09:30 crc kubenswrapper[4805]: W1203 14:09:30.545032 4805 feature_gate.go:330] unrecognized feature gate: InsightsConfig Dec 03 14:09:30 crc kubenswrapper[4805]: W1203 14:09:30.545037 4805 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Dec 03 14:09:30 crc kubenswrapper[4805]: W1203 14:09:30.545042 4805 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Dec 03 14:09:30 crc kubenswrapper[4805]: W1203 14:09:30.545049 4805 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Dec 03 14:09:30 crc kubenswrapper[4805]: W1203 14:09:30.545056 4805 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Dec 03 14:09:30 crc kubenswrapper[4805]: W1203 14:09:30.545062 4805 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Dec 03 14:09:30 crc kubenswrapper[4805]: W1203 14:09:30.545067 4805 feature_gate.go:330] unrecognized feature gate: OVNObservability Dec 03 14:09:30 crc kubenswrapper[4805]: W1203 14:09:30.545072 4805 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Dec 03 14:09:30 crc kubenswrapper[4805]: W1203 14:09:30.545077 4805 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Dec 03 14:09:30 crc kubenswrapper[4805]: W1203 14:09:30.545083 4805 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Dec 03 14:09:30 crc kubenswrapper[4805]: W1203 14:09:30.545088 4805 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Dec 03 14:09:30 crc kubenswrapper[4805]: W1203 14:09:30.545093 4805 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Dec 03 14:09:30 crc kubenswrapper[4805]: W1203 14:09:30.545098 4805 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Dec 03 14:09:30 crc kubenswrapper[4805]: W1203 14:09:30.545103 4805 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Dec 03 14:09:30 crc kubenswrapper[4805]: W1203 14:09:30.545109 4805 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Dec 03 14:09:30 crc kubenswrapper[4805]: W1203 14:09:30.545114 4805 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Dec 03 14:09:30 crc kubenswrapper[4805]: W1203 14:09:30.545120 4805 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Dec 03 14:09:30 crc kubenswrapper[4805]: W1203 14:09:30.545126 4805 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Dec 03 14:09:30 crc kubenswrapper[4805]: W1203 14:09:30.545132 4805 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Dec 03 14:09:30 crc kubenswrapper[4805]: W1203 14:09:30.545139 4805 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Dec 03 14:09:30 crc kubenswrapper[4805]: W1203 14:09:30.545145 4805 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Dec 03 14:09:30 crc kubenswrapper[4805]: W1203 14:09:30.545152 4805 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Dec 03 14:09:30 crc kubenswrapper[4805]: W1203 14:09:30.545158 4805 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Dec 03 14:09:30 crc kubenswrapper[4805]: W1203 14:09:30.545164 4805 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Dec 03 14:09:30 crc kubenswrapper[4805]: W1203 14:09:30.545169 4805 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Dec 03 14:09:30 crc kubenswrapper[4805]: W1203 14:09:30.545175 4805 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Dec 03 14:09:30 crc kubenswrapper[4805]: W1203 14:09:30.545180 4805 feature_gate.go:330] unrecognized feature gate: GatewayAPI Dec 03 14:09:30 crc kubenswrapper[4805]: W1203 14:09:30.545185 4805 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Dec 03 14:09:30 crc kubenswrapper[4805]: W1203 14:09:30.545190 4805 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Dec 03 14:09:30 crc kubenswrapper[4805]: W1203 14:09:30.545195 4805 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Dec 03 14:09:30 crc kubenswrapper[4805]: W1203 14:09:30.545200 4805 feature_gate.go:330] unrecognized feature gate: NewOLM Dec 03 14:09:30 crc kubenswrapper[4805]: W1203 14:09:30.545205 4805 feature_gate.go:330] unrecognized feature gate: PlatformOperators Dec 03 14:09:30 crc kubenswrapper[4805]: I1203 14:09:30.545213 4805 feature_gate.go:386] feature gates: {map[CloudDualStackNodeIPs:true DisableKubeletCloudCredentialProviders:true DynamicResourceAllocation:false EventedPLEG:false KMSv1:true MaxUnavailableStatefulSet:false NodeSwap:false ProcMountType:false RouteExternalCertificate:false ServiceAccountTokenNodeBinding:false TranslateStreamCloseWebsocketRequests:false UserNamespacesPodSecurityStandards:false UserNamespacesSupport:false ValidatingAdmissionPolicy:true VolumeAttributesClass:false]} Dec 03 14:09:30 crc kubenswrapper[4805]: I1203 14:09:30.545400 4805 server.go:940] "Client rotation is on, will bootstrap in background" Dec 03 14:09:30 crc kubenswrapper[4805]: I1203 14:09:30.548882 4805 bootstrap.go:85] "Current kubeconfig file contents are still valid, no bootstrap necessary" Dec 03 14:09:30 crc kubenswrapper[4805]: I1203 14:09:30.548986 4805 certificate_store.go:130] Loading cert/key pair from "/var/lib/kubelet/pki/kubelet-client-current.pem". Dec 03 14:09:30 crc kubenswrapper[4805]: I1203 14:09:30.549581 4805 server.go:997] "Starting client certificate rotation" Dec 03 14:09:30 crc kubenswrapper[4805]: I1203 14:09:30.549614 4805 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Certificate rotation is enabled Dec 03 14:09:30 crc kubenswrapper[4805]: I1203 14:09:30.549896 4805 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Certificate expiration is 2026-02-24 05:52:08 +0000 UTC, rotation deadline is 2025-11-08 07:24:51.453857043 +0000 UTC Dec 03 14:09:30 crc kubenswrapper[4805]: I1203 14:09:30.550006 4805 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Rotating certificates Dec 03 14:09:30 crc kubenswrapper[4805]: I1203 14:09:30.557797 4805 dynamic_cafile_content.go:123] "Loaded a new CA Bundle and Verifier" name="client-ca-bundle::/etc/kubernetes/kubelet-ca.crt" Dec 03 14:09:30 crc kubenswrapper[4805]: E1203 14:09:30.559301 4805 certificate_manager.go:562] "Unhandled Error" err="kubernetes.io/kube-apiserver-client-kubelet: Failed while requesting a signed certificate from the control plane: cannot create certificate signing request: Post \"https://api-int.crc.testing:6443/apis/certificates.k8s.io/v1/certificatesigningrequests\": dial tcp 38.102.83.2:6443: connect: connection refused" logger="UnhandledError" Dec 03 14:09:30 crc kubenswrapper[4805]: I1203 14:09:30.560354 4805 dynamic_cafile_content.go:161] "Starting controller" name="client-ca-bundle::/etc/kubernetes/kubelet-ca.crt" Dec 03 14:09:30 crc kubenswrapper[4805]: I1203 14:09:30.569005 4805 log.go:25] "Validated CRI v1 runtime API" Dec 03 14:09:30 crc kubenswrapper[4805]: I1203 14:09:30.586700 4805 log.go:25] "Validated CRI v1 image API" Dec 03 14:09:30 crc kubenswrapper[4805]: I1203 14:09:30.589055 4805 server.go:1437] "Using cgroup driver setting received from the CRI runtime" cgroupDriver="systemd" Dec 03 14:09:30 crc kubenswrapper[4805]: I1203 14:09:30.592826 4805 fs.go:133] Filesystem UUIDs: map[0b076daa-c26a-46d2-b3a6-72a8dbc6e257:/dev/vda4 2025-12-03-14-04-42-00:/dev/sr0 7B77-95E7:/dev/vda2 de0497b0-db1b-465a-b278-03db02455c71:/dev/vda3] Dec 03 14:09:30 crc kubenswrapper[4805]: I1203 14:09:30.592928 4805 fs.go:134] Filesystem partitions: map[/dev/shm:{mountpoint:/dev/shm major:0 minor:22 fsType:tmpfs blockSize:0} /dev/vda3:{mountpoint:/boot major:252 minor:3 fsType:ext4 blockSize:0} /dev/vda4:{mountpoint:/var major:252 minor:4 fsType:xfs blockSize:0} /run:{mountpoint:/run major:0 minor:24 fsType:tmpfs blockSize:0} /run/user/1000:{mountpoint:/run/user/1000 major:0 minor:42 fsType:tmpfs blockSize:0} /tmp:{mountpoint:/tmp major:0 minor:30 fsType:tmpfs blockSize:0} /var/lib/etcd:{mountpoint:/var/lib/etcd major:0 minor:43 fsType:tmpfs blockSize:0}] Dec 03 14:09:30 crc kubenswrapper[4805]: I1203 14:09:30.619594 4805 manager.go:217] Machine: {Timestamp:2025-12-03 14:09:30.617390114 +0000 UTC m=+0.280307107 CPUVendorID:AuthenticAMD NumCores:12 NumPhysicalCores:1 NumSockets:12 CpuFrequency:2800000 MemoryCapacity:33654128640 SwapCapacity:0 MemoryByType:map[] NVMInfo:{MemoryModeCapacity:0 AppDirectModeCapacity:0 AvgPowerBudget:0} HugePages:[{PageSize:1048576 NumPages:0} {PageSize:2048 NumPages:0}] MachineID:21801e6708c44f15b81395eb736a7cec SystemUUID:7a729b24-19ef-490e-ab4c-d31c52ed5658 BootID:e4edfbcd-1de1-486e-9c0b-65f9fbba52ad Filesystems:[{Device:/dev/shm DeviceMajor:0 DeviceMinor:22 Capacity:16827064320 Type:vfs Inodes:4108170 HasInodes:true} {Device:/run DeviceMajor:0 DeviceMinor:24 Capacity:6730825728 Type:vfs Inodes:819200 HasInodes:true} {Device:/dev/vda4 DeviceMajor:252 DeviceMinor:4 Capacity:85292941312 Type:vfs Inodes:41679680 HasInodes:true} {Device:/tmp DeviceMajor:0 DeviceMinor:30 Capacity:16827064320 Type:vfs Inodes:1048576 HasInodes:true} {Device:/dev/vda3 DeviceMajor:252 DeviceMinor:3 Capacity:366869504 Type:vfs Inodes:98304 HasInodes:true} {Device:/run/user/1000 DeviceMajor:0 DeviceMinor:42 Capacity:3365412864 Type:vfs Inodes:821634 HasInodes:true} {Device:/var/lib/etcd DeviceMajor:0 DeviceMinor:43 Capacity:1073741824 Type:vfs Inodes:4108170 HasInodes:true}] DiskMap:map[252:0:{Name:vda Major:252 Minor:0 Size:214748364800 Scheduler:none}] NetworkDevices:[{Name:br-ex MacAddress:fa:16:3e:7e:80:1a Speed:0 Mtu:1500} {Name:br-int MacAddress:d6:39:55:2e:22:71 Speed:0 Mtu:1400} {Name:ens3 MacAddress:fa:16:3e:7e:80:1a Speed:-1 Mtu:1500} {Name:ens7 MacAddress:fa:16:3e:df:f7:5d Speed:-1 Mtu:1500} {Name:ens7.20 MacAddress:52:54:00:28:88:77 Speed:-1 Mtu:1496} {Name:ens7.21 MacAddress:52:54:00:40:18:56 Speed:-1 Mtu:1496} {Name:ens7.22 MacAddress:52:54:00:42:27:97 Speed:-1 Mtu:1496} {Name:eth10 MacAddress:e2:22:ac:88:58:90 Speed:0 Mtu:1500} {Name:ovn-k8s-mp0 MacAddress:0a:58:0a:d9:00:02 Speed:0 Mtu:1400} {Name:ovs-system MacAddress:ba:b7:53:fe:ba:4f Speed:0 Mtu:1500}] Topology:[{Id:0 Memory:33654128640 HugePages:[{PageSize:1048576 NumPages:0} {PageSize:2048 NumPages:0}] Cores:[{Id:0 Threads:[0] Caches:[{Id:0 Size:32768 Type:Data Level:1} {Id:0 Size:32768 Type:Instruction Level:1} {Id:0 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:0 Size:16777216 Type:Unified Level:3}] SocketID:0 BookID: DrawerID:} {Id:0 Threads:[1] Caches:[{Id:1 Size:32768 Type:Data Level:1} {Id:1 Size:32768 Type:Instruction Level:1} {Id:1 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:1 Size:16777216 Type:Unified Level:3}] SocketID:1 BookID: DrawerID:} {Id:0 Threads:[10] Caches:[{Id:10 Size:32768 Type:Data Level:1} {Id:10 Size:32768 Type:Instruction Level:1} {Id:10 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:10 Size:16777216 Type:Unified Level:3}] SocketID:10 BookID: DrawerID:} {Id:0 Threads:[11] Caches:[{Id:11 Size:32768 Type:Data Level:1} {Id:11 Size:32768 Type:Instruction Level:1} {Id:11 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:11 Size:16777216 Type:Unified Level:3}] SocketID:11 BookID: DrawerID:} {Id:0 Threads:[2] Caches:[{Id:2 Size:32768 Type:Data Level:1} {Id:2 Size:32768 Type:Instruction Level:1} {Id:2 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:2 Size:16777216 Type:Unified Level:3}] SocketID:2 BookID: DrawerID:} {Id:0 Threads:[3] Caches:[{Id:3 Size:32768 Type:Data Level:1} {Id:3 Size:32768 Type:Instruction Level:1} {Id:3 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:3 Size:16777216 Type:Unified Level:3}] SocketID:3 BookID: DrawerID:} {Id:0 Threads:[4] Caches:[{Id:4 Size:32768 Type:Data Level:1} {Id:4 Size:32768 Type:Instruction Level:1} {Id:4 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:4 Size:16777216 Type:Unified Level:3}] SocketID:4 BookID: DrawerID:} {Id:0 Threads:[5] Caches:[{Id:5 Size:32768 Type:Data Level:1} {Id:5 Size:32768 Type:Instruction Level:1} {Id:5 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:5 Size:16777216 Type:Unified Level:3}] SocketID:5 BookID: DrawerID:} {Id:0 Threads:[6] Caches:[{Id:6 Size:32768 Type:Data Level:1} {Id:6 Size:32768 Type:Instruction Level:1} {Id:6 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:6 Size:16777216 Type:Unified Level:3}] SocketID:6 BookID: DrawerID:} {Id:0 Threads:[7] Caches:[{Id:7 Size:32768 Type:Data Level:1} {Id:7 Size:32768 Type:Instruction Level:1} {Id:7 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:7 Size:16777216 Type:Unified Level:3}] SocketID:7 BookID: DrawerID:} {Id:0 Threads:[8] Caches:[{Id:8 Size:32768 Type:Data Level:1} {Id:8 Size:32768 Type:Instruction Level:1} {Id:8 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:8 Size:16777216 Type:Unified Level:3}] SocketID:8 BookID: DrawerID:} {Id:0 Threads:[9] Caches:[{Id:9 Size:32768 Type:Data Level:1} {Id:9 Size:32768 Type:Instruction Level:1} {Id:9 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:9 Size:16777216 Type:Unified Level:3}] SocketID:9 BookID: DrawerID:}] Caches:[] Distances:[10]}] CloudProvider:Unknown InstanceType:Unknown InstanceID:None} Dec 03 14:09:30 crc kubenswrapper[4805]: I1203 14:09:30.620010 4805 manager_no_libpfm.go:29] cAdvisor is build without cgo and/or libpfm support. Perf event counters are not available. Dec 03 14:09:30 crc kubenswrapper[4805]: I1203 14:09:30.620346 4805 manager.go:233] Version: {KernelVersion:5.14.0-427.50.2.el9_4.x86_64 ContainerOsVersion:Red Hat Enterprise Linux CoreOS 418.94.202502100215-0 DockerVersion: DockerAPIVersion: CadvisorVersion: CadvisorRevision:} Dec 03 14:09:30 crc kubenswrapper[4805]: I1203 14:09:30.621091 4805 swap_util.go:113] "Swap is on" /proc/swaps contents="Filename\t\t\t\tType\t\tSize\t\tUsed\t\tPriority" Dec 03 14:09:30 crc kubenswrapper[4805]: I1203 14:09:30.621396 4805 container_manager_linux.go:267] "Container manager verified user specified cgroup-root exists" cgroupRoot=[] Dec 03 14:09:30 crc kubenswrapper[4805]: I1203 14:09:30.621459 4805 container_manager_linux.go:272] "Creating Container Manager object based on Node Config" nodeConfig={"NodeName":"crc","RuntimeCgroupsName":"/system.slice/crio.service","SystemCgroupsName":"/system.slice","KubeletCgroupsName":"","KubeletOOMScoreAdj":-999,"ContainerRuntime":"","CgroupsPerQOS":true,"CgroupRoot":"/","CgroupDriver":"systemd","KubeletRootDir":"/var/lib/kubelet","ProtectKernelDefaults":true,"KubeReservedCgroupName":"","SystemReservedCgroupName":"","ReservedSystemCPUs":{},"EnforceNodeAllocatable":{"pods":{}},"KubeReserved":null,"SystemReserved":{"cpu":"200m","ephemeral-storage":"350Mi","memory":"350Mi"},"HardEvictionThresholds":[{"Signal":"nodefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null},{"Signal":"imagefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.15},"GracePeriod":0,"MinReclaim":null},{"Signal":"imagefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null},{"Signal":"memory.available","Operator":"LessThan","Value":{"Quantity":"100Mi","Percentage":0},"GracePeriod":0,"MinReclaim":null},{"Signal":"nodefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.1},"GracePeriod":0,"MinReclaim":null}],"QOSReserved":{},"CPUManagerPolicy":"none","CPUManagerPolicyOptions":null,"TopologyManagerScope":"container","CPUManagerReconcilePeriod":10000000000,"ExperimentalMemoryManagerPolicy":"None","ExperimentalMemoryManagerReservedMemory":null,"PodPidsLimit":4096,"EnforceCPULimits":true,"CPUCFSQuotaPeriod":100000000,"TopologyManagerPolicy":"none","TopologyManagerPolicyOptions":null,"CgroupVersion":2} Dec 03 14:09:30 crc kubenswrapper[4805]: I1203 14:09:30.621939 4805 topology_manager.go:138] "Creating topology manager with none policy" Dec 03 14:09:30 crc kubenswrapper[4805]: I1203 14:09:30.621968 4805 container_manager_linux.go:303] "Creating device plugin manager" Dec 03 14:09:30 crc kubenswrapper[4805]: I1203 14:09:30.622314 4805 manager.go:142] "Creating Device Plugin manager" path="/var/lib/kubelet/device-plugins/kubelet.sock" Dec 03 14:09:30 crc kubenswrapper[4805]: I1203 14:09:30.622367 4805 server.go:66] "Creating device plugin registration server" version="v1beta1" socket="/var/lib/kubelet/device-plugins/kubelet.sock" Dec 03 14:09:30 crc kubenswrapper[4805]: I1203 14:09:30.622906 4805 state_mem.go:36] "Initialized new in-memory state store" Dec 03 14:09:30 crc kubenswrapper[4805]: I1203 14:09:30.623064 4805 server.go:1245] "Using root directory" path="/var/lib/kubelet" Dec 03 14:09:30 crc kubenswrapper[4805]: I1203 14:09:30.623879 4805 kubelet.go:418] "Attempting to sync node with API server" Dec 03 14:09:30 crc kubenswrapper[4805]: I1203 14:09:30.623899 4805 kubelet.go:313] "Adding static pod path" path="/etc/kubernetes/manifests" Dec 03 14:09:30 crc kubenswrapper[4805]: I1203 14:09:30.623920 4805 file.go:69] "Watching path" path="/etc/kubernetes/manifests" Dec 03 14:09:30 crc kubenswrapper[4805]: I1203 14:09:30.623932 4805 kubelet.go:324] "Adding apiserver pod source" Dec 03 14:09:30 crc kubenswrapper[4805]: I1203 14:09:30.623943 4805 apiserver.go:42] "Waiting for node sync before watching apiserver pods" Dec 03 14:09:30 crc kubenswrapper[4805]: W1203 14:09:30.626059 4805 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Service: Get "https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0": dial tcp 38.102.83.2:6443: connect: connection refused Dec 03 14:09:30 crc kubenswrapper[4805]: I1203 14:09:30.626129 4805 kuberuntime_manager.go:262] "Container runtime initialized" containerRuntime="cri-o" version="1.31.5-4.rhaos4.18.gitdad78d5.el9" apiVersion="v1" Dec 03 14:09:30 crc kubenswrapper[4805]: E1203 14:09:30.626196 4805 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Service: failed to list *v1.Service: Get \"https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0\": dial tcp 38.102.83.2:6443: connect: connection refused" logger="UnhandledError" Dec 03 14:09:30 crc kubenswrapper[4805]: W1203 14:09:30.626062 4805 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Node: Get "https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0": dial tcp 38.102.83.2:6443: connect: connection refused Dec 03 14:09:30 crc kubenswrapper[4805]: E1203 14:09:30.626293 4805 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Node: failed to list *v1.Node: Get \"https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0\": dial tcp 38.102.83.2:6443: connect: connection refused" logger="UnhandledError" Dec 03 14:09:30 crc kubenswrapper[4805]: I1203 14:09:30.626641 4805 certificate_store.go:130] Loading cert/key pair from "/var/lib/kubelet/pki/kubelet-server-current.pem". Dec 03 14:09:30 crc kubenswrapper[4805]: I1203 14:09:30.628300 4805 kubelet.go:854] "Not starting ClusterTrustBundle informer because we are in static kubelet mode" Dec 03 14:09:30 crc kubenswrapper[4805]: I1203 14:09:30.629162 4805 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/portworx-volume" Dec 03 14:09:30 crc kubenswrapper[4805]: I1203 14:09:30.629204 4805 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/empty-dir" Dec 03 14:09:30 crc kubenswrapper[4805]: I1203 14:09:30.629220 4805 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/git-repo" Dec 03 14:09:30 crc kubenswrapper[4805]: I1203 14:09:30.629233 4805 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/host-path" Dec 03 14:09:30 crc kubenswrapper[4805]: I1203 14:09:30.629254 4805 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/nfs" Dec 03 14:09:30 crc kubenswrapper[4805]: I1203 14:09:30.629267 4805 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/secret" Dec 03 14:09:30 crc kubenswrapper[4805]: I1203 14:09:30.629280 4805 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/iscsi" Dec 03 14:09:30 crc kubenswrapper[4805]: I1203 14:09:30.629301 4805 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/downward-api" Dec 03 14:09:30 crc kubenswrapper[4805]: I1203 14:09:30.629317 4805 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/fc" Dec 03 14:09:30 crc kubenswrapper[4805]: I1203 14:09:30.629331 4805 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/configmap" Dec 03 14:09:30 crc kubenswrapper[4805]: I1203 14:09:30.629349 4805 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/projected" Dec 03 14:09:30 crc kubenswrapper[4805]: I1203 14:09:30.629362 4805 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/local-volume" Dec 03 14:09:30 crc kubenswrapper[4805]: I1203 14:09:30.629578 4805 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/csi" Dec 03 14:09:30 crc kubenswrapper[4805]: I1203 14:09:30.630284 4805 server.go:1280] "Started kubelet" Dec 03 14:09:30 crc kubenswrapper[4805]: I1203 14:09:30.630421 4805 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": dial tcp 38.102.83.2:6443: connect: connection refused Dec 03 14:09:30 crc kubenswrapper[4805]: I1203 14:09:30.630593 4805 ratelimit.go:55] "Setting rate limiting for endpoint" service="podresources" qps=100 burstTokens=10 Dec 03 14:09:30 crc kubenswrapper[4805]: I1203 14:09:30.630673 4805 server.go:163] "Starting to listen" address="0.0.0.0" port=10250 Dec 03 14:09:30 crc kubenswrapper[4805]: I1203 14:09:30.631254 4805 server.go:236] "Starting to serve the podresources API" endpoint="unix:/var/lib/kubelet/pod-resources/kubelet.sock" Dec 03 14:09:30 crc systemd[1]: Started Kubernetes Kubelet. Dec 03 14:09:30 crc kubenswrapper[4805]: I1203 14:09:30.632670 4805 server.go:460] "Adding debug handlers to kubelet server" Dec 03 14:09:30 crc kubenswrapper[4805]: I1203 14:09:30.632803 4805 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate rotation is enabled Dec 03 14:09:30 crc kubenswrapper[4805]: I1203 14:09:30.632864 4805 fs_resource_analyzer.go:67] "Starting FS ResourceAnalyzer" Dec 03 14:09:30 crc kubenswrapper[4805]: I1203 14:09:30.632970 4805 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-29 09:08:37.962473166 +0000 UTC Dec 03 14:09:30 crc kubenswrapper[4805]: I1203 14:09:30.633025 4805 certificate_manager.go:356] kubernetes.io/kubelet-serving: Waiting 618h59m7.329451219s for next certificate rotation Dec 03 14:09:30 crc kubenswrapper[4805]: I1203 14:09:30.633680 4805 volume_manager.go:287] "The desired_state_of_world populator starts" Dec 03 14:09:30 crc kubenswrapper[4805]: I1203 14:09:30.633698 4805 volume_manager.go:289] "Starting Kubelet Volume Manager" Dec 03 14:09:30 crc kubenswrapper[4805]: I1203 14:09:30.633850 4805 desired_state_of_world_populator.go:146] "Desired state populator starts to run" Dec 03 14:09:30 crc kubenswrapper[4805]: E1203 14:09:30.634313 4805 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Dec 03 14:09:30 crc kubenswrapper[4805]: W1203 14:09:30.634637 4805 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSIDriver: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": dial tcp 38.102.83.2:6443: connect: connection refused Dec 03 14:09:30 crc kubenswrapper[4805]: E1203 14:09:30.634810 4805 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: Get \"https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0\": dial tcp 38.102.83.2:6443: connect: connection refused" logger="UnhandledError" Dec 03 14:09:30 crc kubenswrapper[4805]: E1203 14:09:30.634691 4805 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.2:6443: connect: connection refused" interval="200ms" Dec 03 14:09:30 crc kubenswrapper[4805]: E1203 14:09:30.634543 4805 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/default/events\": dial tcp 38.102.83.2:6443: connect: connection refused" event="&Event{ObjectMeta:{crc.187db9d8fe566d82 default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:Starting,Message:Starting kubelet.,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2025-12-03 14:09:30.630253954 +0000 UTC m=+0.293170907,LastTimestamp:2025-12-03 14:09:30.630253954 +0000 UTC m=+0.293170907,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Dec 03 14:09:30 crc kubenswrapper[4805]: I1203 14:09:30.639163 4805 factory.go:55] Registering systemd factory Dec 03 14:09:30 crc kubenswrapper[4805]: I1203 14:09:30.639191 4805 factory.go:221] Registration of the systemd container factory successfully Dec 03 14:09:30 crc kubenswrapper[4805]: I1203 14:09:30.639921 4805 factory.go:153] Registering CRI-O factory Dec 03 14:09:30 crc kubenswrapper[4805]: I1203 14:09:30.640061 4805 factory.go:221] Registration of the crio container factory successfully Dec 03 14:09:30 crc kubenswrapper[4805]: I1203 14:09:30.640250 4805 factory.go:219] Registration of the containerd container factory failed: unable to create containerd client: containerd: cannot unix dial containerd api service: dial unix /run/containerd/containerd.sock: connect: no such file or directory Dec 03 14:09:30 crc kubenswrapper[4805]: I1203 14:09:30.640392 4805 factory.go:103] Registering Raw factory Dec 03 14:09:30 crc kubenswrapper[4805]: I1203 14:09:30.640708 4805 manager.go:1196] Started watching for new ooms in manager Dec 03 14:09:30 crc kubenswrapper[4805]: I1203 14:09:30.646686 4805 manager.go:319] Starting recovery of all containers Dec 03 14:09:30 crc kubenswrapper[4805]: I1203 14:09:30.649713 4805 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7bb08738-c794-4ee8-9972-3a62ca171029" volumeName="kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist" seLinuxMountContext="" Dec 03 14:09:30 crc kubenswrapper[4805]: I1203 14:09:30.649785 4805 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="96b93a3a-6083-4aea-8eab-fe1aa8245ad9" volumeName="kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls" seLinuxMountContext="" Dec 03 14:09:30 crc kubenswrapper[4805]: I1203 14:09:30.649800 4805 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a0128f3a-b052-44ed-a84e-c4c8aaf17c13" volumeName="kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m" seLinuxMountContext="" Dec 03 14:09:30 crc kubenswrapper[4805]: I1203 14:09:30.649814 4805 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca" seLinuxMountContext="" Dec 03 14:09:30 crc kubenswrapper[4805]: I1203 14:09:30.649878 4805 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca" seLinuxMountContext="" Dec 03 14:09:30 crc kubenswrapper[4805]: I1203 14:09:30.649897 4805 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="37a5e44f-9a88-4405-be8a-b645485e7312" volumeName="kubernetes.io/secret/37a5e44f-9a88-4405-be8a-b645485e7312-metrics-tls" seLinuxMountContext="" Dec 03 14:09:30 crc kubenswrapper[4805]: I1203 14:09:30.649910 4805 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="44663579-783b-4372-86d6-acf235a62d72" volumeName="kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc" seLinuxMountContext="" Dec 03 14:09:30 crc kubenswrapper[4805]: I1203 14:09:30.649923 4805 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="57a731c4-ef35-47a8-b875-bfb08a7f8011" volumeName="kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content" seLinuxMountContext="" Dec 03 14:09:30 crc kubenswrapper[4805]: I1203 14:09:30.649940 4805 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" volumeName="kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config" seLinuxMountContext="" Dec 03 14:09:30 crc kubenswrapper[4805]: I1203 14:09:30.649952 4805 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/projected/ef543e1b-8068-4ea3-b32a-61027b32e95d-kube-api-access-s2kz5" seLinuxMountContext="" Dec 03 14:09:30 crc kubenswrapper[4805]: I1203 14:09:30.649987 4805 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls" seLinuxMountContext="" Dec 03 14:09:30 crc kubenswrapper[4805]: I1203 14:09:30.650001 4805 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5b88f790-22fa-440e-b583-365168c0b23d" volumeName="kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn" seLinuxMountContext="" Dec 03 14:09:30 crc kubenswrapper[4805]: I1203 14:09:30.650012 4805 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6731426b-95fe-49ff-bb5f-40441049fde2" volumeName="kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh" seLinuxMountContext="" Dec 03 14:09:30 crc kubenswrapper[4805]: I1203 14:09:30.650027 4805 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls" seLinuxMountContext="" Dec 03 14:09:30 crc kubenswrapper[4805]: I1203 14:09:30.650063 4805 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config" seLinuxMountContext="" Dec 03 14:09:30 crc kubenswrapper[4805]: I1203 14:09:30.650099 4805 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7539238d-5fe0-46ed-884e-1c3b566537ec" volumeName="kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c" seLinuxMountContext="" Dec 03 14:09:30 crc kubenswrapper[4805]: I1203 14:09:30.650113 4805 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="87cf06ed-a83f-41a7-828d-70653580a8cb" volumeName="kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx" seLinuxMountContext="" Dec 03 14:09:30 crc kubenswrapper[4805]: I1203 14:09:30.650125 4805 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config" seLinuxMountContext="" Dec 03 14:09:30 crc kubenswrapper[4805]: I1203 14:09:30.650138 4805 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="01ab3dd5-8196-46d0-ad33-122e2ca51def" volumeName="kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert" seLinuxMountContext="" Dec 03 14:09:30 crc kubenswrapper[4805]: I1203 14:09:30.650153 4805 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca" seLinuxMountContext="" Dec 03 14:09:30 crc kubenswrapper[4805]: I1203 14:09:30.650172 4805 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8" seLinuxMountContext="" Dec 03 14:09:30 crc kubenswrapper[4805]: I1203 14:09:30.650185 4805 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" volumeName="kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg" seLinuxMountContext="" Dec 03 14:09:30 crc kubenswrapper[4805]: I1203 14:09:30.650198 4805 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6312bbd-5731-4ea0-a20f-81d5a57df44a" volumeName="kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert" seLinuxMountContext="" Dec 03 14:09:30 crc kubenswrapper[4805]: I1203 14:09:30.650210 4805 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" volumeName="kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782" seLinuxMountContext="" Dec 03 14:09:30 crc kubenswrapper[4805]: I1203 14:09:30.650222 4805 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" volumeName="kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates" seLinuxMountContext="" Dec 03 14:09:30 crc kubenswrapper[4805]: I1203 14:09:30.650259 4805 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca" seLinuxMountContext="" Dec 03 14:09:30 crc kubenswrapper[4805]: I1203 14:09:30.650275 4805 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images" seLinuxMountContext="" Dec 03 14:09:30 crc kubenswrapper[4805]: I1203 14:09:30.650326 4805 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle" seLinuxMountContext="" Dec 03 14:09:30 crc kubenswrapper[4805]: I1203 14:09:30.650341 4805 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52" seLinuxMountContext="" Dec 03 14:09:30 crc kubenswrapper[4805]: I1203 14:09:30.650355 4805 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token" seLinuxMountContext="" Dec 03 14:09:30 crc kubenswrapper[4805]: I1203 14:09:30.650367 4805 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls" seLinuxMountContext="" Dec 03 14:09:30 crc kubenswrapper[4805]: I1203 14:09:30.650380 4805 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert" seLinuxMountContext="" Dec 03 14:09:30 crc kubenswrapper[4805]: I1203 14:09:30.650393 4805 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="20b0d48f-5fd6-431c-a545-e3c800c7b866" volumeName="kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert" seLinuxMountContext="" Dec 03 14:09:30 crc kubenswrapper[4805]: I1203 14:09:30.650406 4805 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs" seLinuxMountContext="" Dec 03 14:09:30 crc kubenswrapper[4805]: I1203 14:09:30.650417 4805 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="e7e6199b-1264-4501-8953-767f51328d08" volumeName="kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config" seLinuxMountContext="" Dec 03 14:09:30 crc kubenswrapper[4805]: I1203 14:09:30.650429 4805 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="4bb40260-dbaa-4fb0-84df-5e680505d512" volumeName="kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy" seLinuxMountContext="" Dec 03 14:09:30 crc kubenswrapper[4805]: I1203 14:09:30.650441 4805 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="4bb40260-dbaa-4fb0-84df-5e680505d512" volumeName="kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh" seLinuxMountContext="" Dec 03 14:09:30 crc kubenswrapper[4805]: I1203 14:09:30.650454 4805 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" volumeName="kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert" seLinuxMountContext="" Dec 03 14:09:30 crc kubenswrapper[4805]: I1203 14:09:30.650466 4805 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides" seLinuxMountContext="" Dec 03 14:09:30 crc kubenswrapper[4805]: I1203 14:09:30.650479 4805 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="efdd0498-1daa-4136-9a4a-3b948c2293fc" volumeName="kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt" seLinuxMountContext="" Dec 03 14:09:30 crc kubenswrapper[4805]: I1203 14:09:30.650501 4805 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle" seLinuxMountContext="" Dec 03 14:09:30 crc kubenswrapper[4805]: I1203 14:09:30.650517 4805 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="496e6271-fb68-4057-954e-a0d97a4afa3f" volumeName="kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert" seLinuxMountContext="" Dec 03 14:09:30 crc kubenswrapper[4805]: I1203 14:09:30.650529 4805 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies" seLinuxMountContext="" Dec 03 14:09:30 crc kubenswrapper[4805]: I1203 14:09:30.650542 4805 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5225d0e4-402f-4861-b410-819f433b1803" volumeName="kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7" seLinuxMountContext="" Dec 03 14:09:30 crc kubenswrapper[4805]: I1203 14:09:30.650554 4805 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca" seLinuxMountContext="" Dec 03 14:09:30 crc kubenswrapper[4805]: I1203 14:09:30.650567 4805 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images" seLinuxMountContext="" Dec 03 14:09:30 crc kubenswrapper[4805]: I1203 14:09:30.650579 4805 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs" seLinuxMountContext="" Dec 03 14:09:30 crc kubenswrapper[4805]: I1203 14:09:30.650593 4805 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d751cbb-f2e2-430d-9754-c882a5e924a5" volumeName="kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl" seLinuxMountContext="" Dec 03 14:09:30 crc kubenswrapper[4805]: I1203 14:09:30.650608 4805 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="01ab3dd5-8196-46d0-ad33-122e2ca51def" volumeName="kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j" seLinuxMountContext="" Dec 03 14:09:30 crc kubenswrapper[4805]: I1203 14:09:30.650621 4805 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies" seLinuxMountContext="" Dec 03 14:09:30 crc kubenswrapper[4805]: I1203 14:09:30.650665 4805 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config" seLinuxMountContext="" Dec 03 14:09:30 crc kubenswrapper[4805]: I1203 14:09:30.650677 4805 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="efdd0498-1daa-4136-9a4a-3b948c2293fc" volumeName="kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs" seLinuxMountContext="" Dec 03 14:09:30 crc kubenswrapper[4805]: I1203 14:09:30.650695 4805 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets" seLinuxMountContext="" Dec 03 14:09:30 crc kubenswrapper[4805]: I1203 14:09:30.650708 4805 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bd23aa5c-e532-4e53-bccf-e79f130c5ae8" volumeName="kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2" seLinuxMountContext="" Dec 03 14:09:30 crc kubenswrapper[4805]: I1203 14:09:30.650724 4805 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d" volumeName="kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85" seLinuxMountContext="" Dec 03 14:09:30 crc kubenswrapper[4805]: I1203 14:09:30.650740 4805 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" volumeName="kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config" seLinuxMountContext="" Dec 03 14:09:30 crc kubenswrapper[4805]: I1203 14:09:30.650753 4805 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs" seLinuxMountContext="" Dec 03 14:09:30 crc kubenswrapper[4805]: I1203 14:09:30.650768 4805 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca" seLinuxMountContext="" Dec 03 14:09:30 crc kubenswrapper[4805]: I1203 14:09:30.650781 4805 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig" seLinuxMountContext="" Dec 03 14:09:30 crc kubenswrapper[4805]: I1203 14:09:30.650793 4805 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session" seLinuxMountContext="" Dec 03 14:09:30 crc kubenswrapper[4805]: I1203 14:09:30.650809 4805 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="4bb40260-dbaa-4fb0-84df-5e680505d512" volumeName="kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config" seLinuxMountContext="" Dec 03 14:09:30 crc kubenswrapper[4805]: I1203 14:09:30.650820 4805 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="96b93a3a-6083-4aea-8eab-fe1aa8245ad9" volumeName="kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7" seLinuxMountContext="" Dec 03 14:09:30 crc kubenswrapper[4805]: I1203 14:09:30.650853 4805 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6312bbd-5731-4ea0-a20f-81d5a57df44a" volumeName="kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr" seLinuxMountContext="" Dec 03 14:09:30 crc kubenswrapper[4805]: I1203 14:09:30.650869 4805 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b78653f-4ff9-4508-8672-245ed9b561e3" volumeName="kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca" seLinuxMountContext="" Dec 03 14:09:30 crc kubenswrapper[4805]: I1203 14:09:30.650883 4805 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" volumeName="kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca" seLinuxMountContext="" Dec 03 14:09:30 crc kubenswrapper[4805]: I1203 14:09:30.650896 4805 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle" seLinuxMountContext="" Dec 03 14:09:30 crc kubenswrapper[4805]: I1203 14:09:30.650908 4805 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp" seLinuxMountContext="" Dec 03 14:09:30 crc kubenswrapper[4805]: I1203 14:09:30.650921 4805 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert" seLinuxMountContext="" Dec 03 14:09:30 crc kubenswrapper[4805]: I1203 14:09:30.650935 4805 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="57a731c4-ef35-47a8-b875-bfb08a7f8011" volumeName="kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct" seLinuxMountContext="" Dec 03 14:09:30 crc kubenswrapper[4805]: I1203 14:09:30.650949 4805 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" volumeName="kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4" seLinuxMountContext="" Dec 03 14:09:30 crc kubenswrapper[4805]: I1203 14:09:30.650963 4805 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" seLinuxMountContext="" Dec 03 14:09:30 crc kubenswrapper[4805]: I1203 14:09:30.651532 4805 reconstruct.go:144] "Volume is marked device as uncertain and added into the actual state" volumeName="kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" deviceMountPath="/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/globalmount" Dec 03 14:09:30 crc kubenswrapper[4805]: I1203 14:09:30.651575 4805 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config" seLinuxMountContext="" Dec 03 14:09:30 crc kubenswrapper[4805]: I1203 14:09:30.651593 4805 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1d611f23-29be-4491-8495-bee1670e935f" volumeName="kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content" seLinuxMountContext="" Dec 03 14:09:30 crc kubenswrapper[4805]: I1203 14:09:30.651608 4805 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config" seLinuxMountContext="" Dec 03 14:09:30 crc kubenswrapper[4805]: I1203 14:09:30.651621 4805 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token" seLinuxMountContext="" Dec 03 14:09:30 crc kubenswrapper[4805]: I1203 14:09:30.651637 4805 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="fda69060-fa79-4696-b1a6-7980f124bf7c" volumeName="kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config" seLinuxMountContext="" Dec 03 14:09:30 crc kubenswrapper[4805]: I1203 14:09:30.651649 4805 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7539238d-5fe0-46ed-884e-1c3b566537ec" volumeName="kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert" seLinuxMountContext="" Dec 03 14:09:30 crc kubenswrapper[4805]: I1203 14:09:30.651664 4805 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config" seLinuxMountContext="" Dec 03 14:09:30 crc kubenswrapper[4805]: I1203 14:09:30.651677 4805 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5" seLinuxMountContext="" Dec 03 14:09:30 crc kubenswrapper[4805]: I1203 14:09:30.651690 4805 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5225d0e4-402f-4861-b410-819f433b1803" volumeName="kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content" seLinuxMountContext="" Dec 03 14:09:30 crc kubenswrapper[4805]: I1203 14:09:30.651703 4805 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error" seLinuxMountContext="" Dec 03 14:09:30 crc kubenswrapper[4805]: I1203 14:09:30.651715 4805 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" volumeName="kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities" seLinuxMountContext="" Dec 03 14:09:30 crc kubenswrapper[4805]: I1203 14:09:30.651728 4805 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="e7e6199b-1264-4501-8953-767f51328d08" volumeName="kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert" seLinuxMountContext="" Dec 03 14:09:30 crc kubenswrapper[4805]: I1203 14:09:30.651741 4805 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert" seLinuxMountContext="" Dec 03 14:09:30 crc kubenswrapper[4805]: I1203 14:09:30.651754 4805 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv" seLinuxMountContext="" Dec 03 14:09:30 crc kubenswrapper[4805]: I1203 14:09:30.651767 4805 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b574797-001e-440a-8f4e-c0be86edad0f" volumeName="kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88" seLinuxMountContext="" Dec 03 14:09:30 crc kubenswrapper[4805]: I1203 14:09:30.651778 4805 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template" seLinuxMountContext="" Dec 03 14:09:30 crc kubenswrapper[4805]: I1203 14:09:30.651793 4805 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6731426b-95fe-49ff-bb5f-40441049fde2" volumeName="kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls" seLinuxMountContext="" Dec 03 14:09:30 crc kubenswrapper[4805]: I1203 14:09:30.651808 4805 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7539238d-5fe0-46ed-884e-1c3b566537ec" volumeName="kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config" seLinuxMountContext="" Dec 03 14:09:30 crc kubenswrapper[4805]: I1203 14:09:30.651822 4805 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert" seLinuxMountContext="" Dec 03 14:09:30 crc kubenswrapper[4805]: I1203 14:09:30.651852 4805 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="d75a4c96-2883-4a0b-bab2-0fab2b6c0b49" volumeName="kubernetes.io/projected/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-kube-api-access-rczfb" seLinuxMountContext="" Dec 03 14:09:30 crc kubenswrapper[4805]: I1203 14:09:30.651866 4805 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert" seLinuxMountContext="" Dec 03 14:09:30 crc kubenswrapper[4805]: I1203 14:09:30.651879 4805 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert" seLinuxMountContext="" Dec 03 14:09:30 crc kubenswrapper[4805]: I1203 14:09:30.651893 4805 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz" seLinuxMountContext="" Dec 03 14:09:30 crc kubenswrapper[4805]: I1203 14:09:30.651905 4805 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3b6479f0-333b-4a96-9adf-2099afdc2447" volumeName="kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr" seLinuxMountContext="" Dec 03 14:09:30 crc kubenswrapper[4805]: I1203 14:09:30.651917 4805 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" volumeName="kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8" seLinuxMountContext="" Dec 03 14:09:30 crc kubenswrapper[4805]: I1203 14:09:30.651929 4805 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config" seLinuxMountContext="" Dec 03 14:09:30 crc kubenswrapper[4805]: I1203 14:09:30.651942 4805 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert" seLinuxMountContext="" Dec 03 14:09:30 crc kubenswrapper[4805]: I1203 14:09:30.651955 4805 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" volumeName="kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf" seLinuxMountContext="" Dec 03 14:09:30 crc kubenswrapper[4805]: I1203 14:09:30.651967 4805 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b574797-001e-440a-8f4e-c0be86edad0f" volumeName="kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls" seLinuxMountContext="" Dec 03 14:09:30 crc kubenswrapper[4805]: I1203 14:09:30.651980 4805 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1386a44e-36a2-460c-96d0-0359d2b6f0f5" volumeName="kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config" seLinuxMountContext="" Dec 03 14:09:30 crc kubenswrapper[4805]: I1203 14:09:30.651994 4805 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config" seLinuxMountContext="" Dec 03 14:09:30 crc kubenswrapper[4805]: I1203 14:09:30.652008 4805 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7bb08738-c794-4ee8-9972-3a62ca171029" volumeName="kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb" seLinuxMountContext="" Dec 03 14:09:30 crc kubenswrapper[4805]: I1203 14:09:30.652077 4805 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert" seLinuxMountContext="" Dec 03 14:09:30 crc kubenswrapper[4805]: I1203 14:09:30.652102 4805 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-ovnkube-identity-cm" seLinuxMountContext="" Dec 03 14:09:30 crc kubenswrapper[4805]: I1203 14:09:30.652116 4805 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="fda69060-fa79-4696-b1a6-7980f124bf7c" volumeName="kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls" seLinuxMountContext="" Dec 03 14:09:30 crc kubenswrapper[4805]: I1203 14:09:30.652130 4805 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz" seLinuxMountContext="" Dec 03 14:09:30 crc kubenswrapper[4805]: I1203 14:09:30.652146 4805 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config" seLinuxMountContext="" Dec 03 14:09:30 crc kubenswrapper[4805]: I1203 14:09:30.652159 4805 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection" seLinuxMountContext="" Dec 03 14:09:30 crc kubenswrapper[4805]: I1203 14:09:30.652172 4805 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" volumeName="kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert" seLinuxMountContext="" Dec 03 14:09:30 crc kubenswrapper[4805]: I1203 14:09:30.652185 4805 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca" seLinuxMountContext="" Dec 03 14:09:30 crc kubenswrapper[4805]: I1203 14:09:30.652205 4805 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="496e6271-fb68-4057-954e-a0d97a4afa3f" volumeName="kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access" seLinuxMountContext="" Dec 03 14:09:30 crc kubenswrapper[4805]: I1203 14:09:30.652219 4805 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5b88f790-22fa-440e-b583-365168c0b23d" volumeName="kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs" seLinuxMountContext="" Dec 03 14:09:30 crc kubenswrapper[4805]: I1203 14:09:30.652232 4805 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6" seLinuxMountContext="" Dec 03 14:09:30 crc kubenswrapper[4805]: I1203 14:09:30.652246 4805 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls" seLinuxMountContext="" Dec 03 14:09:30 crc kubenswrapper[4805]: I1203 14:09:30.652258 4805 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides" seLinuxMountContext="" Dec 03 14:09:30 crc kubenswrapper[4805]: I1203 14:09:30.652274 4805 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1386a44e-36a2-460c-96d0-0359d2b6f0f5" volumeName="kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert" seLinuxMountContext="" Dec 03 14:09:30 crc kubenswrapper[4805]: I1203 14:09:30.652287 4805 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="37a5e44f-9a88-4405-be8a-b645485e7312" volumeName="kubernetes.io/projected/37a5e44f-9a88-4405-be8a-b645485e7312-kube-api-access-rdwmf" seLinuxMountContext="" Dec 03 14:09:30 crc kubenswrapper[4805]: I1203 14:09:30.652300 4805 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login" seLinuxMountContext="" Dec 03 14:09:30 crc kubenswrapper[4805]: I1203 14:09:30.652313 4805 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" volumeName="kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd" seLinuxMountContext="" Dec 03 14:09:30 crc kubenswrapper[4805]: I1203 14:09:30.652331 4805 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-env-overrides" seLinuxMountContext="" Dec 03 14:09:30 crc kubenswrapper[4805]: I1203 14:09:30.652344 4805 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca" seLinuxMountContext="" Dec 03 14:09:30 crc kubenswrapper[4805]: I1203 14:09:30.652357 4805 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b78653f-4ff9-4508-8672-245ed9b561e3" volumeName="kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access" seLinuxMountContext="" Dec 03 14:09:30 crc kubenswrapper[4805]: I1203 14:09:30.652372 4805 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" volumeName="kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert" seLinuxMountContext="" Dec 03 14:09:30 crc kubenswrapper[4805]: I1203 14:09:30.652384 4805 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca" seLinuxMountContext="" Dec 03 14:09:30 crc kubenswrapper[4805]: I1203 14:09:30.652398 4805 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="25e176fe-21b4-4974-b1ed-c8b94f112a7f" volumeName="kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle" seLinuxMountContext="" Dec 03 14:09:30 crc kubenswrapper[4805]: I1203 14:09:30.652411 4805 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe579f8-e8a6-4643-bce5-a661393c4dde" volumeName="kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp" seLinuxMountContext="" Dec 03 14:09:30 crc kubenswrapper[4805]: I1203 14:09:30.652422 4805 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl" seLinuxMountContext="" Dec 03 14:09:30 crc kubenswrapper[4805]: I1203 14:09:30.652435 4805 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert" seLinuxMountContext="" Dec 03 14:09:30 crc kubenswrapper[4805]: I1203 14:09:30.652448 4805 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="496e6271-fb68-4057-954e-a0d97a4afa3f" volumeName="kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config" seLinuxMountContext="" Dec 03 14:09:30 crc kubenswrapper[4805]: I1203 14:09:30.652459 4805 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5225d0e4-402f-4861-b410-819f433b1803" volumeName="kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities" seLinuxMountContext="" Dec 03 14:09:30 crc kubenswrapper[4805]: I1203 14:09:30.652473 4805 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="57a731c4-ef35-47a8-b875-bfb08a7f8011" volumeName="kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities" seLinuxMountContext="" Dec 03 14:09:30 crc kubenswrapper[4805]: I1203 14:09:30.652489 4805 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe579f8-e8a6-4643-bce5-a661393c4dde" volumeName="kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token" seLinuxMountContext="" Dec 03 14:09:30 crc kubenswrapper[4805]: I1203 14:09:30.652503 4805 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client" seLinuxMountContext="" Dec 03 14:09:30 crc kubenswrapper[4805]: I1203 14:09:30.652517 4805 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="20b0d48f-5fd6-431c-a545-e3c800c7b866" volumeName="kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds" seLinuxMountContext="" Dec 03 14:09:30 crc kubenswrapper[4805]: I1203 14:09:30.652531 4805 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="25e176fe-21b4-4974-b1ed-c8b94f112a7f" volumeName="kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key" seLinuxMountContext="" Dec 03 14:09:30 crc kubenswrapper[4805]: I1203 14:09:30.652546 4805 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert" seLinuxMountContext="" Dec 03 14:09:30 crc kubenswrapper[4805]: I1203 14:09:30.652563 4805 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates" seLinuxMountContext="" Dec 03 14:09:30 crc kubenswrapper[4805]: I1203 14:09:30.652577 4805 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/secret/ef543e1b-8068-4ea3-b32a-61027b32e95d-webhook-cert" seLinuxMountContext="" Dec 03 14:09:30 crc kubenswrapper[4805]: I1203 14:09:30.652589 4805 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" volumeName="kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert" seLinuxMountContext="" Dec 03 14:09:30 crc kubenswrapper[4805]: I1203 14:09:30.652601 4805 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7" seLinuxMountContext="" Dec 03 14:09:30 crc kubenswrapper[4805]: I1203 14:09:30.652613 4805 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config" seLinuxMountContext="" Dec 03 14:09:30 crc kubenswrapper[4805]: I1203 14:09:30.652626 4805 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7bb08738-c794-4ee8-9972-3a62ca171029" volumeName="kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy" seLinuxMountContext="" Dec 03 14:09:30 crc kubenswrapper[4805]: I1203 14:09:30.652638 4805 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3ab1a177-2de0-46d9-b765-d0d0649bb42e" volumeName="kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert" seLinuxMountContext="" Dec 03 14:09:30 crc kubenswrapper[4805]: I1203 14:09:30.652651 4805 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5" seLinuxMountContext="" Dec 03 14:09:30 crc kubenswrapper[4805]: I1203 14:09:30.652663 4805 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe579f8-e8a6-4643-bce5-a661393c4dde" volumeName="kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs" seLinuxMountContext="" Dec 03 14:09:30 crc kubenswrapper[4805]: I1203 14:09:30.652676 4805 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7" seLinuxMountContext="" Dec 03 14:09:30 crc kubenswrapper[4805]: I1203 14:09:30.652689 4805 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" volumeName="kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert" seLinuxMountContext="" Dec 03 14:09:30 crc kubenswrapper[4805]: I1203 14:09:30.652702 4805 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit" seLinuxMountContext="" Dec 03 14:09:30 crc kubenswrapper[4805]: I1203 14:09:30.652715 4805 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1d611f23-29be-4491-8495-bee1670e935f" volumeName="kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz" seLinuxMountContext="" Dec 03 14:09:30 crc kubenswrapper[4805]: I1203 14:09:30.652730 4805 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert" seLinuxMountContext="" Dec 03 14:09:30 crc kubenswrapper[4805]: I1203 14:09:30.652743 4805 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="e7e6199b-1264-4501-8953-767f51328d08" volumeName="kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access" seLinuxMountContext="" Dec 03 14:09:30 crc kubenswrapper[4805]: I1203 14:09:30.652756 4805 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" volumeName="kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp" seLinuxMountContext="" Dec 03 14:09:30 crc kubenswrapper[4805]: I1203 14:09:30.652768 4805 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca" seLinuxMountContext="" Dec 03 14:09:30 crc kubenswrapper[4805]: I1203 14:09:30.652778 4805 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert" seLinuxMountContext="" Dec 03 14:09:30 crc kubenswrapper[4805]: I1203 14:09:30.652790 4805 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data" seLinuxMountContext="" Dec 03 14:09:30 crc kubenswrapper[4805]: I1203 14:09:30.652801 4805 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config" seLinuxMountContext="" Dec 03 14:09:30 crc kubenswrapper[4805]: I1203 14:09:30.652813 4805 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1386a44e-36a2-460c-96d0-0359d2b6f0f5" volumeName="kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access" seLinuxMountContext="" Dec 03 14:09:30 crc kubenswrapper[4805]: I1203 14:09:30.652826 4805 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca" seLinuxMountContext="" Dec 03 14:09:30 crc kubenswrapper[4805]: I1203 14:09:30.652889 4805 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config" seLinuxMountContext="" Dec 03 14:09:30 crc kubenswrapper[4805]: I1203 14:09:30.652904 4805 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn" seLinuxMountContext="" Dec 03 14:09:30 crc kubenswrapper[4805]: I1203 14:09:30.652917 4805 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" volumeName="kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics" seLinuxMountContext="" Dec 03 14:09:30 crc kubenswrapper[4805]: I1203 14:09:30.652929 4805 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="d75a4c96-2883-4a0b-bab2-0fab2b6c0b49" volumeName="kubernetes.io/configmap/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-iptables-alerter-script" seLinuxMountContext="" Dec 03 14:09:30 crc kubenswrapper[4805]: I1203 14:09:30.652944 4805 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk" seLinuxMountContext="" Dec 03 14:09:30 crc kubenswrapper[4805]: I1203 14:09:30.652959 4805 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs" seLinuxMountContext="" Dec 03 14:09:30 crc kubenswrapper[4805]: I1203 14:09:30.652973 4805 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="01ab3dd5-8196-46d0-ad33-122e2ca51def" volumeName="kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config" seLinuxMountContext="" Dec 03 14:09:30 crc kubenswrapper[4805]: I1203 14:09:30.652986 4805 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle" seLinuxMountContext="" Dec 03 14:09:30 crc kubenswrapper[4805]: I1203 14:09:30.653000 4805 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls" seLinuxMountContext="" Dec 03 14:09:30 crc kubenswrapper[4805]: I1203 14:09:30.653015 4805 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="87cf06ed-a83f-41a7-828d-70653580a8cb" volumeName="kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume" seLinuxMountContext="" Dec 03 14:09:30 crc kubenswrapper[4805]: I1203 14:09:30.653028 4805 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted" seLinuxMountContext="" Dec 03 14:09:30 crc kubenswrapper[4805]: I1203 14:09:30.653047 4805 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config" seLinuxMountContext="" Dec 03 14:09:30 crc kubenswrapper[4805]: I1203 14:09:30.653061 4805 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" volumeName="kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh" seLinuxMountContext="" Dec 03 14:09:30 crc kubenswrapper[4805]: I1203 14:09:30.653072 4805 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf" seLinuxMountContext="" Dec 03 14:09:30 crc kubenswrapper[4805]: I1203 14:09:30.653083 4805 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert" seLinuxMountContext="" Dec 03 14:09:30 crc kubenswrapper[4805]: I1203 14:09:30.653097 4805 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b574797-001e-440a-8f4e-c0be86edad0f" volumeName="kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config" seLinuxMountContext="" Dec 03 14:09:30 crc kubenswrapper[4805]: I1203 14:09:30.653111 4805 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" volumeName="kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert" seLinuxMountContext="" Dec 03 14:09:30 crc kubenswrapper[4805]: I1203 14:09:30.653123 4805 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate" seLinuxMountContext="" Dec 03 14:09:30 crc kubenswrapper[4805]: I1203 14:09:30.653137 4805 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="fda69060-fa79-4696-b1a6-7980f124bf7c" volumeName="kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh" seLinuxMountContext="" Dec 03 14:09:30 crc kubenswrapper[4805]: I1203 14:09:30.653149 4805 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca" seLinuxMountContext="" Dec 03 14:09:30 crc kubenswrapper[4805]: I1203 14:09:30.653163 4805 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls" seLinuxMountContext="" Dec 03 14:09:30 crc kubenswrapper[4805]: I1203 14:09:30.653177 4805 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client" seLinuxMountContext="" Dec 03 14:09:30 crc kubenswrapper[4805]: I1203 14:09:30.653190 4805 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config" seLinuxMountContext="" Dec 03 14:09:30 crc kubenswrapper[4805]: I1203 14:09:30.653203 4805 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config" seLinuxMountContext="" Dec 03 14:09:30 crc kubenswrapper[4805]: I1203 14:09:30.653215 4805 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" volumeName="kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca" seLinuxMountContext="" Dec 03 14:09:30 crc kubenswrapper[4805]: I1203 14:09:30.653229 4805 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token" seLinuxMountContext="" Dec 03 14:09:30 crc kubenswrapper[4805]: I1203 14:09:30.653243 4805 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth" seLinuxMountContext="" Dec 03 14:09:30 crc kubenswrapper[4805]: I1203 14:09:30.653255 4805 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf" seLinuxMountContext="" Dec 03 14:09:30 crc kubenswrapper[4805]: I1203 14:09:30.653267 4805 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib" seLinuxMountContext="" Dec 03 14:09:30 crc kubenswrapper[4805]: I1203 14:09:30.653279 4805 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles" seLinuxMountContext="" Dec 03 14:09:30 crc kubenswrapper[4805]: I1203 14:09:30.653291 4805 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb" seLinuxMountContext="" Dec 03 14:09:30 crc kubenswrapper[4805]: I1203 14:09:30.653349 4805 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="87cf06ed-a83f-41a7-828d-70653580a8cb" volumeName="kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls" seLinuxMountContext="" Dec 03 14:09:30 crc kubenswrapper[4805]: I1203 14:09:30.653364 4805 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert" seLinuxMountContext="" Dec 03 14:09:30 crc kubenswrapper[4805]: I1203 14:09:30.653376 4805 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="25e176fe-21b4-4974-b1ed-c8b94f112a7f" volumeName="kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv" seLinuxMountContext="" Dec 03 14:09:30 crc kubenswrapper[4805]: I1203 14:09:30.653389 4805 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3ab1a177-2de0-46d9-b765-d0d0649bb42e" volumeName="kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj" seLinuxMountContext="" Dec 03 14:09:30 crc kubenswrapper[4805]: I1203 14:09:30.653403 4805 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca" seLinuxMountContext="" Dec 03 14:09:30 crc kubenswrapper[4805]: I1203 14:09:30.653428 4805 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz" seLinuxMountContext="" Dec 03 14:09:30 crc kubenswrapper[4805]: I1203 14:09:30.653441 4805 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle" seLinuxMountContext="" Dec 03 14:09:30 crc kubenswrapper[4805]: I1203 14:09:30.653455 4805 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6312bbd-5731-4ea0-a20f-81d5a57df44a" volumeName="kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert" seLinuxMountContext="" Dec 03 14:09:30 crc kubenswrapper[4805]: I1203 14:09:30.653468 4805 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client" seLinuxMountContext="" Dec 03 14:09:30 crc kubenswrapper[4805]: I1203 14:09:30.653481 4805 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b78653f-4ff9-4508-8672-245ed9b561e3" volumeName="kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert" seLinuxMountContext="" Dec 03 14:09:30 crc kubenswrapper[4805]: I1203 14:09:30.653493 4805 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1d611f23-29be-4491-8495-bee1670e935f" volumeName="kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities" seLinuxMountContext="" Dec 03 14:09:30 crc kubenswrapper[4805]: I1203 14:09:30.653506 4805 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49ef4625-1d3a-4a9f-b595-c2433d32326d" volumeName="kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v" seLinuxMountContext="" Dec 03 14:09:30 crc kubenswrapper[4805]: I1203 14:09:30.653519 4805 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a0128f3a-b052-44ed-a84e-c4c8aaf17c13" volumeName="kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls" seLinuxMountContext="" Dec 03 14:09:30 crc kubenswrapper[4805]: I1203 14:09:30.653532 4805 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" volumeName="kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content" seLinuxMountContext="" Dec 03 14:09:30 crc kubenswrapper[4805]: I1203 14:09:30.653543 4805 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle" seLinuxMountContext="" Dec 03 14:09:30 crc kubenswrapper[4805]: I1203 14:09:30.653555 4805 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config" seLinuxMountContext="" Dec 03 14:09:30 crc kubenswrapper[4805]: I1203 14:09:30.653577 4805 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle" seLinuxMountContext="" Dec 03 14:09:30 crc kubenswrapper[4805]: I1203 14:09:30.653589 4805 reconstruct.go:97] "Volume reconstruction finished" Dec 03 14:09:30 crc kubenswrapper[4805]: I1203 14:09:30.653598 4805 reconciler.go:26] "Reconciler: start to sync state" Dec 03 14:09:30 crc kubenswrapper[4805]: I1203 14:09:30.670353 4805 manager.go:324] Recovery completed Dec 03 14:09:30 crc kubenswrapper[4805]: I1203 14:09:30.687557 4805 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 03 14:09:30 crc kubenswrapper[4805]: I1203 14:09:30.688186 4805 kubelet_network_linux.go:50] "Initialized iptables rules." protocol="IPv4" Dec 03 14:09:30 crc kubenswrapper[4805]: I1203 14:09:30.692262 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 14:09:30 crc kubenswrapper[4805]: I1203 14:09:30.692326 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 14:09:30 crc kubenswrapper[4805]: I1203 14:09:30.692343 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 14:09:30 crc kubenswrapper[4805]: I1203 14:09:30.693133 4805 kubelet_network_linux.go:50] "Initialized iptables rules." protocol="IPv6" Dec 03 14:09:30 crc kubenswrapper[4805]: I1203 14:09:30.693174 4805 status_manager.go:217] "Starting to sync pod status with apiserver" Dec 03 14:09:30 crc kubenswrapper[4805]: I1203 14:09:30.693192 4805 kubelet.go:2335] "Starting kubelet main sync loop" Dec 03 14:09:30 crc kubenswrapper[4805]: E1203 14:09:30.693234 4805 kubelet.go:2359] "Skipping pod synchronization" err="[container runtime status check may not have completed yet, PLEG is not healthy: pleg has yet to be successful]" Dec 03 14:09:30 crc kubenswrapper[4805]: I1203 14:09:30.693501 4805 cpu_manager.go:225] "Starting CPU manager" policy="none" Dec 03 14:09:30 crc kubenswrapper[4805]: I1203 14:09:30.693525 4805 cpu_manager.go:226] "Reconciling" reconcilePeriod="10s" Dec 03 14:09:30 crc kubenswrapper[4805]: I1203 14:09:30.693548 4805 state_mem.go:36] "Initialized new in-memory state store" Dec 03 14:09:30 crc kubenswrapper[4805]: W1203 14:09:30.694192 4805 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.RuntimeClass: Get "https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": dial tcp 38.102.83.2:6443: connect: connection refused Dec 03 14:09:30 crc kubenswrapper[4805]: E1203 14:09:30.694295 4805 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: Get \"https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0\": dial tcp 38.102.83.2:6443: connect: connection refused" logger="UnhandledError" Dec 03 14:09:30 crc kubenswrapper[4805]: I1203 14:09:30.704180 4805 policy_none.go:49] "None policy: Start" Dec 03 14:09:30 crc kubenswrapper[4805]: I1203 14:09:30.705132 4805 memory_manager.go:170] "Starting memorymanager" policy="None" Dec 03 14:09:30 crc kubenswrapper[4805]: I1203 14:09:30.705172 4805 state_mem.go:35] "Initializing new in-memory state store" Dec 03 14:09:30 crc kubenswrapper[4805]: E1203 14:09:30.734653 4805 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Dec 03 14:09:30 crc kubenswrapper[4805]: I1203 14:09:30.768160 4805 manager.go:334] "Starting Device Plugin manager" Dec 03 14:09:30 crc kubenswrapper[4805]: I1203 14:09:30.768444 4805 manager.go:513] "Failed to read data from checkpoint" checkpoint="kubelet_internal_checkpoint" err="checkpoint is not found" Dec 03 14:09:30 crc kubenswrapper[4805]: I1203 14:09:30.768472 4805 server.go:79] "Starting device plugin registration server" Dec 03 14:09:30 crc kubenswrapper[4805]: I1203 14:09:30.769077 4805 eviction_manager.go:189] "Eviction manager: starting control loop" Dec 03 14:09:30 crc kubenswrapper[4805]: I1203 14:09:30.769103 4805 container_log_manager.go:189] "Initializing container log rotate workers" workers=1 monitorPeriod="10s" Dec 03 14:09:30 crc kubenswrapper[4805]: I1203 14:09:30.769276 4805 plugin_watcher.go:51] "Plugin Watcher Start" path="/var/lib/kubelet/plugins_registry" Dec 03 14:09:30 crc kubenswrapper[4805]: I1203 14:09:30.769499 4805 plugin_manager.go:116] "The desired_state_of_world populator (plugin watcher) starts" Dec 03 14:09:30 crc kubenswrapper[4805]: I1203 14:09:30.769568 4805 plugin_manager.go:118] "Starting Kubelet Plugin Manager" Dec 03 14:09:30 crc kubenswrapper[4805]: E1203 14:09:30.777943 4805 eviction_manager.go:285] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"crc\" not found" Dec 03 14:09:30 crc kubenswrapper[4805]: I1203 14:09:30.794051 4805 kubelet.go:2421] "SyncLoop ADD" source="file" pods=["openshift-machine-config-operator/kube-rbac-proxy-crio-crc","openshift-etcd/etcd-crc","openshift-kube-apiserver/kube-apiserver-crc","openshift-kube-controller-manager/kube-controller-manager-crc","openshift-kube-scheduler/openshift-kube-scheduler-crc"] Dec 03 14:09:30 crc kubenswrapper[4805]: I1203 14:09:30.794205 4805 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 03 14:09:30 crc kubenswrapper[4805]: I1203 14:09:30.795764 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 14:09:30 crc kubenswrapper[4805]: I1203 14:09:30.795805 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 14:09:30 crc kubenswrapper[4805]: I1203 14:09:30.795819 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 14:09:30 crc kubenswrapper[4805]: I1203 14:09:30.796018 4805 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 03 14:09:30 crc kubenswrapper[4805]: I1203 14:09:30.796356 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Dec 03 14:09:30 crc kubenswrapper[4805]: I1203 14:09:30.796427 4805 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 03 14:09:30 crc kubenswrapper[4805]: I1203 14:09:30.796793 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 14:09:30 crc kubenswrapper[4805]: I1203 14:09:30.796830 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 14:09:30 crc kubenswrapper[4805]: I1203 14:09:30.796871 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 14:09:30 crc kubenswrapper[4805]: I1203 14:09:30.797036 4805 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 03 14:09:30 crc kubenswrapper[4805]: I1203 14:09:30.797259 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd/etcd-crc" Dec 03 14:09:30 crc kubenswrapper[4805]: I1203 14:09:30.797399 4805 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 03 14:09:30 crc kubenswrapper[4805]: I1203 14:09:30.797514 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 14:09:30 crc kubenswrapper[4805]: I1203 14:09:30.797541 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 14:09:30 crc kubenswrapper[4805]: I1203 14:09:30.797551 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 14:09:30 crc kubenswrapper[4805]: I1203 14:09:30.798132 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 14:09:30 crc kubenswrapper[4805]: I1203 14:09:30.798161 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 14:09:30 crc kubenswrapper[4805]: I1203 14:09:30.798171 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 14:09:30 crc kubenswrapper[4805]: I1203 14:09:30.798275 4805 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 03 14:09:30 crc kubenswrapper[4805]: I1203 14:09:30.798455 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 03 14:09:30 crc kubenswrapper[4805]: I1203 14:09:30.798514 4805 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 03 14:09:30 crc kubenswrapper[4805]: I1203 14:09:30.798928 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 14:09:30 crc kubenswrapper[4805]: I1203 14:09:30.798949 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 14:09:30 crc kubenswrapper[4805]: I1203 14:09:30.798958 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 14:09:30 crc kubenswrapper[4805]: I1203 14:09:30.799033 4805 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 03 14:09:30 crc kubenswrapper[4805]: I1203 14:09:30.799036 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 14:09:30 crc kubenswrapper[4805]: I1203 14:09:30.799065 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 14:09:30 crc kubenswrapper[4805]: I1203 14:09:30.799082 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 14:09:30 crc kubenswrapper[4805]: I1203 14:09:30.799141 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 03 14:09:30 crc kubenswrapper[4805]: I1203 14:09:30.799168 4805 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 03 14:09:30 crc kubenswrapper[4805]: I1203 14:09:30.799629 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 14:09:30 crc kubenswrapper[4805]: I1203 14:09:30.799653 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 14:09:30 crc kubenswrapper[4805]: I1203 14:09:30.799663 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 14:09:30 crc kubenswrapper[4805]: I1203 14:09:30.799659 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 14:09:30 crc kubenswrapper[4805]: I1203 14:09:30.799692 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 14:09:30 crc kubenswrapper[4805]: I1203 14:09:30.799711 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 14:09:30 crc kubenswrapper[4805]: I1203 14:09:30.799809 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Dec 03 14:09:30 crc kubenswrapper[4805]: I1203 14:09:30.799850 4805 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 03 14:09:30 crc kubenswrapper[4805]: I1203 14:09:30.799929 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 14:09:30 crc kubenswrapper[4805]: I1203 14:09:30.799966 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 14:09:30 crc kubenswrapper[4805]: I1203 14:09:30.799979 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 14:09:30 crc kubenswrapper[4805]: I1203 14:09:30.800407 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 14:09:30 crc kubenswrapper[4805]: I1203 14:09:30.800434 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 14:09:30 crc kubenswrapper[4805]: I1203 14:09:30.800443 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 14:09:30 crc kubenswrapper[4805]: E1203 14:09:30.839142 4805 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.2:6443: connect: connection refused" interval="400ms" Dec 03 14:09:30 crc kubenswrapper[4805]: I1203 14:09:30.855776 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 03 14:09:30 crc kubenswrapper[4805]: I1203 14:09:30.855871 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 03 14:09:30 crc kubenswrapper[4805]: I1203 14:09:30.855922 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-cert-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Dec 03 14:09:30 crc kubenswrapper[4805]: I1203 14:09:30.855966 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"static-pod-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-static-pod-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 03 14:09:30 crc kubenswrapper[4805]: I1203 14:09:30.856007 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-resource-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 03 14:09:30 crc kubenswrapper[4805]: I1203 14:09:30.856051 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"usr-local-bin\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-usr-local-bin\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 03 14:09:30 crc kubenswrapper[4805]: I1203 14:09:30.856091 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 03 14:09:30 crc kubenswrapper[4805]: I1203 14:09:30.856128 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-cert-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 03 14:09:30 crc kubenswrapper[4805]: I1203 14:09:30.856197 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"data-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-data-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 03 14:09:30 crc kubenswrapper[4805]: I1203 14:09:30.856237 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-log-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 03 14:09:30 crc kubenswrapper[4805]: I1203 14:09:30.856276 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-resource-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 03 14:09:30 crc kubenswrapper[4805]: I1203 14:09:30.856318 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-resource-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Dec 03 14:09:30 crc kubenswrapper[4805]: I1203 14:09:30.856360 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-var-lib-kubelet\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Dec 03 14:09:30 crc kubenswrapper[4805]: I1203 14:09:30.856403 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-kube\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-etc-kube\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Dec 03 14:09:30 crc kubenswrapper[4805]: I1203 14:09:30.856459 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-cert-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 03 14:09:30 crc kubenswrapper[4805]: I1203 14:09:30.870265 4805 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 03 14:09:30 crc kubenswrapper[4805]: I1203 14:09:30.871424 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 14:09:30 crc kubenswrapper[4805]: I1203 14:09:30.871452 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 14:09:30 crc kubenswrapper[4805]: I1203 14:09:30.871464 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 14:09:30 crc kubenswrapper[4805]: I1203 14:09:30.871487 4805 kubelet_node_status.go:76] "Attempting to register node" node="crc" Dec 03 14:09:30 crc kubenswrapper[4805]: E1203 14:09:30.871798 4805 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.2:6443: connect: connection refused" node="crc" Dec 03 14:09:30 crc kubenswrapper[4805]: I1203 14:09:30.957716 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"usr-local-bin\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-usr-local-bin\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 03 14:09:30 crc kubenswrapper[4805]: I1203 14:09:30.957777 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 03 14:09:30 crc kubenswrapper[4805]: I1203 14:09:30.957815 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 03 14:09:30 crc kubenswrapper[4805]: I1203 14:09:30.957888 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-cert-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Dec 03 14:09:30 crc kubenswrapper[4805]: I1203 14:09:30.957971 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"usr-local-bin\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-usr-local-bin\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 03 14:09:30 crc kubenswrapper[4805]: I1203 14:09:30.957983 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 03 14:09:30 crc kubenswrapper[4805]: I1203 14:09:30.958031 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 03 14:09:30 crc kubenswrapper[4805]: I1203 14:09:30.958122 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"static-pod-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-static-pod-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 03 14:09:30 crc kubenswrapper[4805]: I1203 14:09:30.958166 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-resource-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 03 14:09:30 crc kubenswrapper[4805]: I1203 14:09:30.958211 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 03 14:09:30 crc kubenswrapper[4805]: I1203 14:09:30.958253 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-resource-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 03 14:09:30 crc kubenswrapper[4805]: I1203 14:09:30.958256 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-cert-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 03 14:09:30 crc kubenswrapper[4805]: I1203 14:09:30.958205 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-cert-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Dec 03 14:09:30 crc kubenswrapper[4805]: I1203 14:09:30.958324 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 03 14:09:30 crc kubenswrapper[4805]: I1203 14:09:30.958317 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"data-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-data-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 03 14:09:30 crc kubenswrapper[4805]: I1203 14:09:30.958366 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"static-pod-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-static-pod-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 03 14:09:30 crc kubenswrapper[4805]: I1203 14:09:30.958341 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"data-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-data-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 03 14:09:30 crc kubenswrapper[4805]: I1203 14:09:30.958292 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-cert-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 03 14:09:30 crc kubenswrapper[4805]: I1203 14:09:30.958403 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-log-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 03 14:09:30 crc kubenswrapper[4805]: I1203 14:09:30.958432 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-log-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 03 14:09:30 crc kubenswrapper[4805]: I1203 14:09:30.958484 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-resource-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 03 14:09:30 crc kubenswrapper[4805]: I1203 14:09:30.958502 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-resource-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 03 14:09:30 crc kubenswrapper[4805]: I1203 14:09:30.958559 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-resource-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Dec 03 14:09:30 crc kubenswrapper[4805]: I1203 14:09:30.958602 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-var-lib-kubelet\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Dec 03 14:09:30 crc kubenswrapper[4805]: I1203 14:09:30.958636 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-kube\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-etc-kube\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Dec 03 14:09:30 crc kubenswrapper[4805]: I1203 14:09:30.958652 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-var-lib-kubelet\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Dec 03 14:09:30 crc kubenswrapper[4805]: I1203 14:09:30.958670 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-cert-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 03 14:09:30 crc kubenswrapper[4805]: I1203 14:09:30.958724 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-resource-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Dec 03 14:09:30 crc kubenswrapper[4805]: I1203 14:09:30.958732 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-kube\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-etc-kube\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Dec 03 14:09:30 crc kubenswrapper[4805]: I1203 14:09:30.958917 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-cert-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 03 14:09:31 crc kubenswrapper[4805]: I1203 14:09:31.072464 4805 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 03 14:09:31 crc kubenswrapper[4805]: I1203 14:09:31.074198 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 14:09:31 crc kubenswrapper[4805]: I1203 14:09:31.074261 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 14:09:31 crc kubenswrapper[4805]: I1203 14:09:31.074281 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 14:09:31 crc kubenswrapper[4805]: I1203 14:09:31.074316 4805 kubelet_node_status.go:76] "Attempting to register node" node="crc" Dec 03 14:09:31 crc kubenswrapper[4805]: E1203 14:09:31.074926 4805 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.2:6443: connect: connection refused" node="crc" Dec 03 14:09:31 crc kubenswrapper[4805]: I1203 14:09:31.132429 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Dec 03 14:09:31 crc kubenswrapper[4805]: I1203 14:09:31.138951 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd/etcd-crc" Dec 03 14:09:31 crc kubenswrapper[4805]: I1203 14:09:31.161567 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 03 14:09:31 crc kubenswrapper[4805]: W1203 14:09:31.172390 4805 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod2139d3e2895fc6797b9c76a1b4c9886d.slice/crio-ba9dddc3ca74c4e791072ff2e438fd3ab17a59dacfc04e14732d3b8e712216cd WatchSource:0}: Error finding container ba9dddc3ca74c4e791072ff2e438fd3ab17a59dacfc04e14732d3b8e712216cd: Status 404 returned error can't find the container with id ba9dddc3ca74c4e791072ff2e438fd3ab17a59dacfc04e14732d3b8e712216cd Dec 03 14:09:31 crc kubenswrapper[4805]: W1203 14:09:31.173915 4805 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd1b160f5dda77d281dd8e69ec8d817f9.slice/crio-a922e60aadb37bfdb646d6e1d278dc5438495f5ef19a63d7f2c16272c6504ca4 WatchSource:0}: Error finding container a922e60aadb37bfdb646d6e1d278dc5438495f5ef19a63d7f2c16272c6504ca4: Status 404 returned error can't find the container with id a922e60aadb37bfdb646d6e1d278dc5438495f5ef19a63d7f2c16272c6504ca4 Dec 03 14:09:31 crc kubenswrapper[4805]: I1203 14:09:31.191207 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 03 14:09:31 crc kubenswrapper[4805]: I1203 14:09:31.197520 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Dec 03 14:09:31 crc kubenswrapper[4805]: W1203 14:09:31.198755 4805 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf4b27818a5e8e43d0dc095d08835c792.slice/crio-b1899309529239912b0b5cd97f5c227e9326d70aea490633fa6ec7e28d0283a4 WatchSource:0}: Error finding container b1899309529239912b0b5cd97f5c227e9326d70aea490633fa6ec7e28d0283a4: Status 404 returned error can't find the container with id b1899309529239912b0b5cd97f5c227e9326d70aea490633fa6ec7e28d0283a4 Dec 03 14:09:31 crc kubenswrapper[4805]: W1203 14:09:31.215050 4805 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf614b9022728cf315e60c057852e563e.slice/crio-66f15958c1bdb368b6fb12d702292e95498a895aa5bdb9e4bea6a622e4e8e46a WatchSource:0}: Error finding container 66f15958c1bdb368b6fb12d702292e95498a895aa5bdb9e4bea6a622e4e8e46a: Status 404 returned error can't find the container with id 66f15958c1bdb368b6fb12d702292e95498a895aa5bdb9e4bea6a622e4e8e46a Dec 03 14:09:31 crc kubenswrapper[4805]: W1203 14:09:31.219778 4805 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod3dcd261975c3d6b9a6ad6367fd4facd3.slice/crio-fcbbbecddcc8fc765616f8762d2180542d8dca53c6629e36e6dd15383a460740 WatchSource:0}: Error finding container fcbbbecddcc8fc765616f8762d2180542d8dca53c6629e36e6dd15383a460740: Status 404 returned error can't find the container with id fcbbbecddcc8fc765616f8762d2180542d8dca53c6629e36e6dd15383a460740 Dec 03 14:09:31 crc kubenswrapper[4805]: E1203 14:09:31.240070 4805 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.2:6443: connect: connection refused" interval="800ms" Dec 03 14:09:31 crc kubenswrapper[4805]: I1203 14:09:31.475121 4805 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 03 14:09:31 crc kubenswrapper[4805]: I1203 14:09:31.477771 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 14:09:31 crc kubenswrapper[4805]: I1203 14:09:31.477814 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 14:09:31 crc kubenswrapper[4805]: I1203 14:09:31.477825 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 14:09:31 crc kubenswrapper[4805]: I1203 14:09:31.477874 4805 kubelet_node_status.go:76] "Attempting to register node" node="crc" Dec 03 14:09:31 crc kubenswrapper[4805]: E1203 14:09:31.478505 4805 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.2:6443: connect: connection refused" node="crc" Dec 03 14:09:31 crc kubenswrapper[4805]: W1203 14:09:31.520304 4805 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Service: Get "https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0": dial tcp 38.102.83.2:6443: connect: connection refused Dec 03 14:09:31 crc kubenswrapper[4805]: E1203 14:09:31.520401 4805 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Service: failed to list *v1.Service: Get \"https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0\": dial tcp 38.102.83.2:6443: connect: connection refused" logger="UnhandledError" Dec 03 14:09:31 crc kubenswrapper[4805]: W1203 14:09:31.623031 4805 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSIDriver: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": dial tcp 38.102.83.2:6443: connect: connection refused Dec 03 14:09:31 crc kubenswrapper[4805]: E1203 14:09:31.623140 4805 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: Get \"https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0\": dial tcp 38.102.83.2:6443: connect: connection refused" logger="UnhandledError" Dec 03 14:09:31 crc kubenswrapper[4805]: I1203 14:09:31.632340 4805 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": dial tcp 38.102.83.2:6443: connect: connection refused Dec 03 14:09:31 crc kubenswrapper[4805]: I1203 14:09:31.699553 4805 generic.go:334] "Generic (PLEG): container finished" podID="2139d3e2895fc6797b9c76a1b4c9886d" containerID="474f72dac37a808fc4b9832306e76979b151d7221807ff3086013466ed1419c3" exitCode=0 Dec 03 14:09:31 crc kubenswrapper[4805]: I1203 14:09:31.699638 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerDied","Data":"474f72dac37a808fc4b9832306e76979b151d7221807ff3086013466ed1419c3"} Dec 03 14:09:31 crc kubenswrapper[4805]: I1203 14:09:31.699874 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"ba9dddc3ca74c4e791072ff2e438fd3ab17a59dacfc04e14732d3b8e712216cd"} Dec 03 14:09:31 crc kubenswrapper[4805]: I1203 14:09:31.700091 4805 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 03 14:09:31 crc kubenswrapper[4805]: I1203 14:09:31.701333 4805 generic.go:334] "Generic (PLEG): container finished" podID="d1b160f5dda77d281dd8e69ec8d817f9" containerID="0d86b4e22a0aec3b88cf3d60d37796252dc122c2c89dfa9f23123110c7b6e9be" exitCode=0 Dec 03 14:09:31 crc kubenswrapper[4805]: I1203 14:09:31.701432 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" event={"ID":"d1b160f5dda77d281dd8e69ec8d817f9","Type":"ContainerDied","Data":"0d86b4e22a0aec3b88cf3d60d37796252dc122c2c89dfa9f23123110c7b6e9be"} Dec 03 14:09:31 crc kubenswrapper[4805]: I1203 14:09:31.701471 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" event={"ID":"d1b160f5dda77d281dd8e69ec8d817f9","Type":"ContainerStarted","Data":"a922e60aadb37bfdb646d6e1d278dc5438495f5ef19a63d7f2c16272c6504ca4"} Dec 03 14:09:31 crc kubenswrapper[4805]: I1203 14:09:31.701896 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 14:09:31 crc kubenswrapper[4805]: I1203 14:09:31.701937 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 14:09:31 crc kubenswrapper[4805]: I1203 14:09:31.701950 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 14:09:31 crc kubenswrapper[4805]: I1203 14:09:31.702664 4805 generic.go:334] "Generic (PLEG): container finished" podID="3dcd261975c3d6b9a6ad6367fd4facd3" containerID="77c26dae6ccd328a8a1ce4fe6ffe94c78061d8d3ac3727adb80a24c16f4ab533" exitCode=0 Dec 03 14:09:31 crc kubenswrapper[4805]: I1203 14:09:31.702700 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerDied","Data":"77c26dae6ccd328a8a1ce4fe6ffe94c78061d8d3ac3727adb80a24c16f4ab533"} Dec 03 14:09:31 crc kubenswrapper[4805]: I1203 14:09:31.702731 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"fcbbbecddcc8fc765616f8762d2180542d8dca53c6629e36e6dd15383a460740"} Dec 03 14:09:31 crc kubenswrapper[4805]: I1203 14:09:31.702827 4805 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 03 14:09:31 crc kubenswrapper[4805]: I1203 14:09:31.704498 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 14:09:31 crc kubenswrapper[4805]: I1203 14:09:31.704542 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 14:09:31 crc kubenswrapper[4805]: I1203 14:09:31.704557 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 14:09:31 crc kubenswrapper[4805]: I1203 14:09:31.705316 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"8673a29c0e16f4e707d79f0610722ccabb023ff2b4a18d7cf9d305183817d952"} Dec 03 14:09:31 crc kubenswrapper[4805]: I1203 14:09:31.705362 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"66f15958c1bdb368b6fb12d702292e95498a895aa5bdb9e4bea6a622e4e8e46a"} Dec 03 14:09:31 crc kubenswrapper[4805]: I1203 14:09:31.708352 4805 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="af9c940135012c629447528c61743ed1e58708ab57ce76604c6da0479aa459cf" exitCode=0 Dec 03 14:09:31 crc kubenswrapper[4805]: I1203 14:09:31.708385 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerDied","Data":"af9c940135012c629447528c61743ed1e58708ab57ce76604c6da0479aa459cf"} Dec 03 14:09:31 crc kubenswrapper[4805]: I1203 14:09:31.708433 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"b1899309529239912b0b5cd97f5c227e9326d70aea490633fa6ec7e28d0283a4"} Dec 03 14:09:31 crc kubenswrapper[4805]: I1203 14:09:31.708530 4805 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 03 14:09:31 crc kubenswrapper[4805]: I1203 14:09:31.709418 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 14:09:31 crc kubenswrapper[4805]: I1203 14:09:31.709452 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 14:09:31 crc kubenswrapper[4805]: I1203 14:09:31.709465 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 14:09:31 crc kubenswrapper[4805]: I1203 14:09:31.710886 4805 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 03 14:09:31 crc kubenswrapper[4805]: I1203 14:09:31.711690 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 14:09:31 crc kubenswrapper[4805]: I1203 14:09:31.711742 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 14:09:31 crc kubenswrapper[4805]: I1203 14:09:31.711752 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 14:09:31 crc kubenswrapper[4805]: W1203 14:09:31.780317 4805 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Node: Get "https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0": dial tcp 38.102.83.2:6443: connect: connection refused Dec 03 14:09:31 crc kubenswrapper[4805]: E1203 14:09:31.780421 4805 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Node: failed to list *v1.Node: Get \"https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0\": dial tcp 38.102.83.2:6443: connect: connection refused" logger="UnhandledError" Dec 03 14:09:31 crc kubenswrapper[4805]: W1203 14:09:31.898529 4805 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.RuntimeClass: Get "https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": dial tcp 38.102.83.2:6443: connect: connection refused Dec 03 14:09:31 crc kubenswrapper[4805]: E1203 14:09:31.898627 4805 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: Get \"https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0\": dial tcp 38.102.83.2:6443: connect: connection refused" logger="UnhandledError" Dec 03 14:09:32 crc kubenswrapper[4805]: E1203 14:09:32.040718 4805 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.2:6443: connect: connection refused" interval="1.6s" Dec 03 14:09:32 crc kubenswrapper[4805]: I1203 14:09:32.279527 4805 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 03 14:09:32 crc kubenswrapper[4805]: I1203 14:09:32.280743 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 14:09:32 crc kubenswrapper[4805]: I1203 14:09:32.280771 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 14:09:32 crc kubenswrapper[4805]: I1203 14:09:32.280779 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 14:09:32 crc kubenswrapper[4805]: I1203 14:09:32.280798 4805 kubelet_node_status.go:76] "Attempting to register node" node="crc" Dec 03 14:09:32 crc kubenswrapper[4805]: E1203 14:09:32.281164 4805 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.2:6443: connect: connection refused" node="crc" Dec 03 14:09:32 crc kubenswrapper[4805]: I1203 14:09:32.610763 4805 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Rotating certificates Dec 03 14:09:32 crc kubenswrapper[4805]: I1203 14:09:32.712334 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"2fd6d50f839b603a90e5d00ab6e36e832eab2e299b7af1dbb84498c35bf1c08e"} Dec 03 14:09:32 crc kubenswrapper[4805]: I1203 14:09:32.712380 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"66aa257562f8e113fc1cd2936b8aa18e927f90018ad4c11a184c1c56444be648"} Dec 03 14:09:32 crc kubenswrapper[4805]: I1203 14:09:32.712394 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"5ea698a1e0de3b5201e2f9032d834d81d9aab9c2fb17aca582184c7090f1bd2f"} Dec 03 14:09:32 crc kubenswrapper[4805]: I1203 14:09:32.712496 4805 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 03 14:09:32 crc kubenswrapper[4805]: I1203 14:09:32.713360 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 14:09:32 crc kubenswrapper[4805]: I1203 14:09:32.713398 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 14:09:32 crc kubenswrapper[4805]: I1203 14:09:32.713410 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 14:09:32 crc kubenswrapper[4805]: I1203 14:09:32.715398 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"5b9844a11b2b903978d0d464b469e26c9211100da61d018dde9636dca95b684e"} Dec 03 14:09:32 crc kubenswrapper[4805]: I1203 14:09:32.715429 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"35766e7206b1be7a53d13d6890d759cf81dbd999643c12349268e4bf16fbbdb0"} Dec 03 14:09:32 crc kubenswrapper[4805]: I1203 14:09:32.715445 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"abd2d0156113bb5182727a53aac594410ed051c19258beb8a07b808eff0b42a7"} Dec 03 14:09:32 crc kubenswrapper[4805]: I1203 14:09:32.715515 4805 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 03 14:09:32 crc kubenswrapper[4805]: I1203 14:09:32.716345 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 14:09:32 crc kubenswrapper[4805]: I1203 14:09:32.716373 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 14:09:32 crc kubenswrapper[4805]: I1203 14:09:32.716386 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 14:09:32 crc kubenswrapper[4805]: I1203 14:09:32.719351 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"5da55f769e23ad52bac9c77da62c78e70fa34db8ed1a435e4eaf4dad3fd2b68a"} Dec 03 14:09:32 crc kubenswrapper[4805]: I1203 14:09:32.719381 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"bd92a160eb530c5687ebd7cd5b83b41820a2b808a0fd9514e5ace9d237e1bd26"} Dec 03 14:09:32 crc kubenswrapper[4805]: I1203 14:09:32.719394 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"29cd95be3f964415943c873e7914ffcea4abe50dfdf17f8a9cd03fc38df35981"} Dec 03 14:09:32 crc kubenswrapper[4805]: I1203 14:09:32.719403 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"951173b3d3d581275a6f6e43209b5e0611a701ca7a10f81504d723e24a2d21a3"} Dec 03 14:09:32 crc kubenswrapper[4805]: I1203 14:09:32.719411 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"3e5a34636f0a8c4d9179074596d28bc2eb7a504573398c596a637defd829907a"} Dec 03 14:09:32 crc kubenswrapper[4805]: I1203 14:09:32.719469 4805 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 03 14:09:32 crc kubenswrapper[4805]: I1203 14:09:32.720102 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 14:09:32 crc kubenswrapper[4805]: I1203 14:09:32.720126 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 14:09:32 crc kubenswrapper[4805]: I1203 14:09:32.720135 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 14:09:32 crc kubenswrapper[4805]: I1203 14:09:32.721312 4805 generic.go:334] "Generic (PLEG): container finished" podID="2139d3e2895fc6797b9c76a1b4c9886d" containerID="74c8cea1a61c4775a0ad2f4ab4bb8a36437cf23d2ee953311d6ea967e297510a" exitCode=0 Dec 03 14:09:32 crc kubenswrapper[4805]: I1203 14:09:32.721369 4805 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 03 14:09:32 crc kubenswrapper[4805]: I1203 14:09:32.721624 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerDied","Data":"74c8cea1a61c4775a0ad2f4ab4bb8a36437cf23d2ee953311d6ea967e297510a"} Dec 03 14:09:32 crc kubenswrapper[4805]: I1203 14:09:32.721693 4805 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 03 14:09:32 crc kubenswrapper[4805]: I1203 14:09:32.722153 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 14:09:32 crc kubenswrapper[4805]: I1203 14:09:32.722177 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 14:09:32 crc kubenswrapper[4805]: I1203 14:09:32.722184 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 14:09:32 crc kubenswrapper[4805]: I1203 14:09:32.722545 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 14:09:32 crc kubenswrapper[4805]: I1203 14:09:32.722567 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 14:09:32 crc kubenswrapper[4805]: I1203 14:09:32.722576 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 14:09:33 crc kubenswrapper[4805]: I1203 14:09:33.010533 4805 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 03 14:09:33 crc kubenswrapper[4805]: I1203 14:09:33.726308 4805 generic.go:334] "Generic (PLEG): container finished" podID="2139d3e2895fc6797b9c76a1b4c9886d" containerID="a4f2f56a809adc01fba82eba0652d78f58770c132f77b1424f8b17068b1927e3" exitCode=0 Dec 03 14:09:33 crc kubenswrapper[4805]: I1203 14:09:33.726374 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerDied","Data":"a4f2f56a809adc01fba82eba0652d78f58770c132f77b1424f8b17068b1927e3"} Dec 03 14:09:33 crc kubenswrapper[4805]: I1203 14:09:33.726493 4805 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 03 14:09:33 crc kubenswrapper[4805]: I1203 14:09:33.727448 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 14:09:33 crc kubenswrapper[4805]: I1203 14:09:33.727470 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 14:09:33 crc kubenswrapper[4805]: I1203 14:09:33.727478 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 14:09:33 crc kubenswrapper[4805]: I1203 14:09:33.728096 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" event={"ID":"d1b160f5dda77d281dd8e69ec8d817f9","Type":"ContainerStarted","Data":"f3942cab663eda9769efb606b01dba99630de2e30b77a7888c8a57a2a844ee0d"} Dec 03 14:09:33 crc kubenswrapper[4805]: I1203 14:09:33.728135 4805 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 03 14:09:33 crc kubenswrapper[4805]: I1203 14:09:33.728160 4805 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 03 14:09:33 crc kubenswrapper[4805]: I1203 14:09:33.728298 4805 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Dec 03 14:09:33 crc kubenswrapper[4805]: I1203 14:09:33.728368 4805 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 03 14:09:33 crc kubenswrapper[4805]: I1203 14:09:33.728571 4805 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Dec 03 14:09:33 crc kubenswrapper[4805]: I1203 14:09:33.728594 4805 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 03 14:09:33 crc kubenswrapper[4805]: I1203 14:09:33.729524 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 14:09:33 crc kubenswrapper[4805]: I1203 14:09:33.729558 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 14:09:33 crc kubenswrapper[4805]: I1203 14:09:33.729570 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 14:09:33 crc kubenswrapper[4805]: I1203 14:09:33.729974 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 14:09:33 crc kubenswrapper[4805]: I1203 14:09:33.729992 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 14:09:33 crc kubenswrapper[4805]: I1203 14:09:33.730004 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 14:09:33 crc kubenswrapper[4805]: I1203 14:09:33.730090 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 14:09:33 crc kubenswrapper[4805]: I1203 14:09:33.730120 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 14:09:33 crc kubenswrapper[4805]: I1203 14:09:33.729991 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 14:09:33 crc kubenswrapper[4805]: I1203 14:09:33.730138 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 14:09:33 crc kubenswrapper[4805]: I1203 14:09:33.730161 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 14:09:33 crc kubenswrapper[4805]: I1203 14:09:33.730176 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 14:09:33 crc kubenswrapper[4805]: I1203 14:09:33.882049 4805 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 03 14:09:33 crc kubenswrapper[4805]: I1203 14:09:33.883367 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 14:09:33 crc kubenswrapper[4805]: I1203 14:09:33.883440 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 14:09:33 crc kubenswrapper[4805]: I1203 14:09:33.883458 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 14:09:33 crc kubenswrapper[4805]: I1203 14:09:33.883499 4805 kubelet_node_status.go:76] "Attempting to register node" node="crc" Dec 03 14:09:34 crc kubenswrapper[4805]: I1203 14:09:34.154047 4805 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 03 14:09:34 crc kubenswrapper[4805]: I1203 14:09:34.735912 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"2ff1a979387da41b00e95eb9cc8b2c9de966380a12fae687591df3c5b18a4497"} Dec 03 14:09:34 crc kubenswrapper[4805]: I1203 14:09:34.735983 4805 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 03 14:09:34 crc kubenswrapper[4805]: I1203 14:09:34.735996 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"6e5f04faf3b23197bb8ea11d3e574dc86f31e21d798734962001ec8c95de7c84"} Dec 03 14:09:34 crc kubenswrapper[4805]: I1203 14:09:34.736025 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"a837d7d491eaacbe4279b4014bb5fcbe9b514ba5cc82c708227bcd9b25a346e8"} Dec 03 14:09:34 crc kubenswrapper[4805]: I1203 14:09:34.736049 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"7d5b5c85e3de43e7e95111aa967248550607c26200c783251b68696f202f92a1"} Dec 03 14:09:34 crc kubenswrapper[4805]: I1203 14:09:34.736922 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 14:09:34 crc kubenswrapper[4805]: I1203 14:09:34.736952 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 14:09:34 crc kubenswrapper[4805]: I1203 14:09:34.736966 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 14:09:34 crc kubenswrapper[4805]: I1203 14:09:34.807778 4805 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 03 14:09:34 crc kubenswrapper[4805]: I1203 14:09:34.807957 4805 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Dec 03 14:09:34 crc kubenswrapper[4805]: I1203 14:09:34.808002 4805 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 03 14:09:34 crc kubenswrapper[4805]: I1203 14:09:34.809048 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 14:09:34 crc kubenswrapper[4805]: I1203 14:09:34.809120 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 14:09:34 crc kubenswrapper[4805]: I1203 14:09:34.809139 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 14:09:35 crc kubenswrapper[4805]: I1203 14:09:35.744678 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"1e66e1cb18573a14e3c3ae16af4dbd9f8997a6ad38680c3ab4505646ebd48e7c"} Dec 03 14:09:35 crc kubenswrapper[4805]: I1203 14:09:35.745693 4805 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 03 14:09:35 crc kubenswrapper[4805]: I1203 14:09:35.747189 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 14:09:35 crc kubenswrapper[4805]: I1203 14:09:35.747306 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 14:09:35 crc kubenswrapper[4805]: I1203 14:09:35.747342 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 14:09:36 crc kubenswrapper[4805]: I1203 14:09:36.747289 4805 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 03 14:09:36 crc kubenswrapper[4805]: I1203 14:09:36.748523 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 14:09:36 crc kubenswrapper[4805]: I1203 14:09:36.748590 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 14:09:36 crc kubenswrapper[4805]: I1203 14:09:36.748607 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 14:09:37 crc kubenswrapper[4805]: I1203 14:09:37.154879 4805 patch_prober.go:28] interesting pod/kube-controller-manager-crc container/cluster-policy-controller namespace/openshift-kube-controller-manager: Startup probe status=failure output="Get \"https://192.168.126.11:10357/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" start-of-body= Dec 03 14:09:37 crc kubenswrapper[4805]: I1203 14:09:37.154996 4805 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podUID="f614b9022728cf315e60c057852e563e" containerName="cluster-policy-controller" probeResult="failure" output="Get \"https://192.168.126.11:10357/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 03 14:09:38 crc kubenswrapper[4805]: I1203 14:09:38.299796 4805 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 03 14:09:38 crc kubenswrapper[4805]: I1203 14:09:38.300109 4805 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 03 14:09:38 crc kubenswrapper[4805]: I1203 14:09:38.301566 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 14:09:38 crc kubenswrapper[4805]: I1203 14:09:38.301611 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 14:09:38 crc kubenswrapper[4805]: I1203 14:09:38.301621 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 14:09:38 crc kubenswrapper[4805]: I1203 14:09:38.352471 4805 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-etcd/etcd-crc" Dec 03 14:09:38 crc kubenswrapper[4805]: I1203 14:09:38.352682 4805 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 03 14:09:38 crc kubenswrapper[4805]: I1203 14:09:38.353921 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 14:09:38 crc kubenswrapper[4805]: I1203 14:09:38.354001 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 14:09:38 crc kubenswrapper[4805]: I1203 14:09:38.354022 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 14:09:38 crc kubenswrapper[4805]: I1203 14:09:38.841749 4805 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 03 14:09:38 crc kubenswrapper[4805]: I1203 14:09:38.842008 4805 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 03 14:09:38 crc kubenswrapper[4805]: I1203 14:09:38.843207 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 14:09:38 crc kubenswrapper[4805]: I1203 14:09:38.843252 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 14:09:38 crc kubenswrapper[4805]: I1203 14:09:38.843268 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 14:09:39 crc kubenswrapper[4805]: I1203 14:09:39.771297 4805 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Dec 03 14:09:39 crc kubenswrapper[4805]: I1203 14:09:39.771548 4805 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 03 14:09:39 crc kubenswrapper[4805]: I1203 14:09:39.773031 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 14:09:39 crc kubenswrapper[4805]: I1203 14:09:39.773079 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 14:09:39 crc kubenswrapper[4805]: I1203 14:09:39.773090 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 14:09:40 crc kubenswrapper[4805]: I1203 14:09:40.052813 4805 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 03 14:09:40 crc kubenswrapper[4805]: I1203 14:09:40.053320 4805 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 03 14:09:40 crc kubenswrapper[4805]: I1203 14:09:40.054604 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 14:09:40 crc kubenswrapper[4805]: I1203 14:09:40.054673 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 14:09:40 crc kubenswrapper[4805]: I1203 14:09:40.054687 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 14:09:40 crc kubenswrapper[4805]: E1203 14:09:40.778027 4805 eviction_manager.go:285] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"crc\" not found" Dec 03 14:09:42 crc kubenswrapper[4805]: I1203 14:09:42.391270 4805 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 03 14:09:42 crc kubenswrapper[4805]: I1203 14:09:42.391501 4805 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 03 14:09:42 crc kubenswrapper[4805]: I1203 14:09:42.392874 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 14:09:42 crc kubenswrapper[4805]: I1203 14:09:42.392922 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 14:09:42 crc kubenswrapper[4805]: I1203 14:09:42.392940 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 14:09:42 crc kubenswrapper[4805]: I1203 14:09:42.396402 4805 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 03 14:09:42 crc kubenswrapper[4805]: E1203 14:09:42.613371 4805 certificate_manager.go:562] "Unhandled Error" err="kubernetes.io/kube-apiserver-client-kubelet: Failed while requesting a signed certificate from the control plane: cannot create certificate signing request: Post \"https://api-int.crc.testing:6443/apis/certificates.k8s.io/v1/certificatesigningrequests\": net/http: TLS handshake timeout" logger="UnhandledError" Dec 03 14:09:42 crc kubenswrapper[4805]: I1203 14:09:42.633735 4805 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": net/http: TLS handshake timeout Dec 03 14:09:42 crc kubenswrapper[4805]: I1203 14:09:42.762782 4805 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 03 14:09:42 crc kubenswrapper[4805]: I1203 14:09:42.763669 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 14:09:42 crc kubenswrapper[4805]: I1203 14:09:42.763700 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 14:09:42 crc kubenswrapper[4805]: I1203 14:09:42.763710 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 14:09:42 crc kubenswrapper[4805]: I1203 14:09:42.768366 4805 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 03 14:09:43 crc kubenswrapper[4805]: W1203 14:09:43.240952 4805 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Service: Get "https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0": net/http: TLS handshake timeout Dec 03 14:09:43 crc kubenswrapper[4805]: I1203 14:09:43.241055 4805 trace.go:236] Trace[939752321]: "Reflector ListAndWatch" name:k8s.io/client-go/informers/factory.go:160 (03-Dec-2025 14:09:33.239) (total time: 10001ms): Dec 03 14:09:43 crc kubenswrapper[4805]: Trace[939752321]: ---"Objects listed" error:Get "https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0": net/http: TLS handshake timeout 10001ms (14:09:43.240) Dec 03 14:09:43 crc kubenswrapper[4805]: Trace[939752321]: [10.0018197s] [10.0018197s] END Dec 03 14:09:43 crc kubenswrapper[4805]: E1203 14:09:43.241081 4805 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Service: failed to list *v1.Service: Get \"https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0\": net/http: TLS handshake timeout" logger="UnhandledError" Dec 03 14:09:43 crc kubenswrapper[4805]: I1203 14:09:43.435407 4805 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-etcd/etcd-crc" Dec 03 14:09:43 crc kubenswrapper[4805]: I1203 14:09:43.439456 4805 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 03 14:09:43 crc kubenswrapper[4805]: I1203 14:09:43.441071 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 14:09:43 crc kubenswrapper[4805]: I1203 14:09:43.441133 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 14:09:43 crc kubenswrapper[4805]: I1203 14:09:43.441146 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 14:09:43 crc kubenswrapper[4805]: I1203 14:09:43.619238 4805 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver namespace/openshift-kube-apiserver: Startup probe status=failure output="HTTP probe failed with statuscode: 403" start-of-body={"kind":"Status","apiVersion":"v1","metadata":{},"status":"Failure","message":"forbidden: User \"system:anonymous\" cannot get path \"/livez\"","reason":"Forbidden","details":{},"code":403} Dec 03 14:09:43 crc kubenswrapper[4805]: I1203 14:09:43.619300 4805 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" probeResult="failure" output="HTTP probe failed with statuscode: 403" Dec 03 14:09:43 crc kubenswrapper[4805]: I1203 14:09:43.624043 4805 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver namespace/openshift-kube-apiserver: Startup probe status=failure output="HTTP probe failed with statuscode: 403" start-of-body={"kind":"Status","apiVersion":"v1","metadata":{},"status":"Failure","message":"forbidden: User \"system:anonymous\" cannot get path \"/livez\"","reason":"Forbidden","details":{},"code":403} Dec 03 14:09:43 crc kubenswrapper[4805]: I1203 14:09:43.624117 4805 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" probeResult="failure" output="HTTP probe failed with statuscode: 403" Dec 03 14:09:43 crc kubenswrapper[4805]: I1203 14:09:43.764976 4805 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 03 14:09:43 crc kubenswrapper[4805]: I1203 14:09:43.765716 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 14:09:43 crc kubenswrapper[4805]: I1203 14:09:43.765754 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 14:09:43 crc kubenswrapper[4805]: I1203 14:09:43.765765 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 14:09:44 crc kubenswrapper[4805]: I1203 14:09:44.815111 4805 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver namespace/openshift-kube-apiserver: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[+]ping ok Dec 03 14:09:44 crc kubenswrapper[4805]: [+]log ok Dec 03 14:09:44 crc kubenswrapper[4805]: [+]etcd ok Dec 03 14:09:44 crc kubenswrapper[4805]: [+]poststarthook/quota.openshift.io-clusterquotamapping ok Dec 03 14:09:44 crc kubenswrapper[4805]: [+]poststarthook/openshift.io-api-request-count-filter ok Dec 03 14:09:44 crc kubenswrapper[4805]: [+]poststarthook/openshift.io-startkubeinformers ok Dec 03 14:09:44 crc kubenswrapper[4805]: [+]poststarthook/openshift.io-openshift-apiserver-reachable ok Dec 03 14:09:44 crc kubenswrapper[4805]: [+]poststarthook/openshift.io-oauth-apiserver-reachable ok Dec 03 14:09:44 crc kubenswrapper[4805]: [+]poststarthook/start-apiserver-admission-initializer ok Dec 03 14:09:44 crc kubenswrapper[4805]: [+]poststarthook/generic-apiserver-start-informers ok Dec 03 14:09:44 crc kubenswrapper[4805]: [+]poststarthook/priority-and-fairness-config-consumer ok Dec 03 14:09:44 crc kubenswrapper[4805]: [+]poststarthook/priority-and-fairness-filter ok Dec 03 14:09:44 crc kubenswrapper[4805]: [+]poststarthook/storage-object-count-tracker-hook ok Dec 03 14:09:44 crc kubenswrapper[4805]: [+]poststarthook/start-apiextensions-informers ok Dec 03 14:09:44 crc kubenswrapper[4805]: [+]poststarthook/start-apiextensions-controllers ok Dec 03 14:09:44 crc kubenswrapper[4805]: [+]poststarthook/crd-informer-synced ok Dec 03 14:09:44 crc kubenswrapper[4805]: [+]poststarthook/start-system-namespaces-controller ok Dec 03 14:09:44 crc kubenswrapper[4805]: [+]poststarthook/start-cluster-authentication-info-controller ok Dec 03 14:09:44 crc kubenswrapper[4805]: [+]poststarthook/start-kube-apiserver-identity-lease-controller ok Dec 03 14:09:44 crc kubenswrapper[4805]: [+]poststarthook/start-kube-apiserver-identity-lease-garbage-collector ok Dec 03 14:09:44 crc kubenswrapper[4805]: [+]poststarthook/start-legacy-token-tracking-controller ok Dec 03 14:09:44 crc kubenswrapper[4805]: [+]poststarthook/start-service-ip-repair-controllers ok Dec 03 14:09:44 crc kubenswrapper[4805]: [-]poststarthook/rbac/bootstrap-roles failed: reason withheld Dec 03 14:09:44 crc kubenswrapper[4805]: [+]poststarthook/scheduling/bootstrap-system-priority-classes ok Dec 03 14:09:44 crc kubenswrapper[4805]: [+]poststarthook/priority-and-fairness-config-producer ok Dec 03 14:09:44 crc kubenswrapper[4805]: [+]poststarthook/bootstrap-controller ok Dec 03 14:09:44 crc kubenswrapper[4805]: [+]poststarthook/aggregator-reload-proxy-client-cert ok Dec 03 14:09:44 crc kubenswrapper[4805]: [+]poststarthook/start-kube-aggregator-informers ok Dec 03 14:09:44 crc kubenswrapper[4805]: [+]poststarthook/apiservice-status-local-available-controller ok Dec 03 14:09:44 crc kubenswrapper[4805]: [+]poststarthook/apiservice-status-remote-available-controller ok Dec 03 14:09:44 crc kubenswrapper[4805]: [+]poststarthook/apiservice-registration-controller ok Dec 03 14:09:44 crc kubenswrapper[4805]: [+]poststarthook/apiservice-wait-for-first-sync ok Dec 03 14:09:44 crc kubenswrapper[4805]: [+]poststarthook/apiservice-discovery-controller ok Dec 03 14:09:44 crc kubenswrapper[4805]: [+]poststarthook/kube-apiserver-autoregistration ok Dec 03 14:09:44 crc kubenswrapper[4805]: [+]autoregister-completion ok Dec 03 14:09:44 crc kubenswrapper[4805]: [+]poststarthook/apiservice-openapi-controller ok Dec 03 14:09:44 crc kubenswrapper[4805]: [+]poststarthook/apiservice-openapiv3-controller ok Dec 03 14:09:44 crc kubenswrapper[4805]: livez check failed Dec 03 14:09:44 crc kubenswrapper[4805]: I1203 14:09:44.815182 4805 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 03 14:09:46 crc kubenswrapper[4805]: I1203 14:09:46.483560 4805 reflector.go:368] Caches populated for *v1.Service from k8s.io/client-go/informers/factory.go:160 Dec 03 14:09:46 crc kubenswrapper[4805]: I1203 14:09:46.887453 4805 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Rotating certificates Dec 03 14:09:46 crc kubenswrapper[4805]: I1203 14:09:46.901402 4805 reflector.go:368] Caches populated for *v1.CertificateSigningRequest from k8s.io/client-go/tools/watch/informerwatcher.go:146 Dec 03 14:09:47 crc kubenswrapper[4805]: I1203 14:09:47.154601 4805 patch_prober.go:28] interesting pod/kube-controller-manager-crc container/cluster-policy-controller namespace/openshift-kube-controller-manager: Startup probe status=failure output="Get \"https://192.168.126.11:10357/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Dec 03 14:09:47 crc kubenswrapper[4805]: I1203 14:09:47.154694 4805 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podUID="f614b9022728cf315e60c057852e563e" containerName="cluster-policy-controller" probeResult="failure" output="Get \"https://192.168.126.11:10357/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Dec 03 14:09:48 crc kubenswrapper[4805]: I1203 14:09:48.609934 4805 trace.go:236] Trace[661979673]: "Reflector ListAndWatch" name:k8s.io/client-go/informers/factory.go:160 (03-Dec-2025 14:09:34.112) (total time: 14497ms): Dec 03 14:09:48 crc kubenswrapper[4805]: Trace[661979673]: ---"Objects listed" error: 14497ms (14:09:48.609) Dec 03 14:09:48 crc kubenswrapper[4805]: Trace[661979673]: [14.497096828s] [14.497096828s] END Dec 03 14:09:48 crc kubenswrapper[4805]: I1203 14:09:48.609974 4805 reflector.go:368] Caches populated for *v1.RuntimeClass from k8s.io/client-go/informers/factory.go:160 Dec 03 14:09:48 crc kubenswrapper[4805]: E1203 14:09:48.609944 4805 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": context deadline exceeded" interval="3.2s" Dec 03 14:09:48 crc kubenswrapper[4805]: I1203 14:09:48.611990 4805 trace.go:236] Trace[1659649112]: "Reflector ListAndWatch" name:k8s.io/client-go/informers/factory.go:160 (03-Dec-2025 14:09:34.936) (total time: 13675ms): Dec 03 14:09:48 crc kubenswrapper[4805]: Trace[1659649112]: ---"Objects listed" error: 13675ms (14:09:48.611) Dec 03 14:09:48 crc kubenswrapper[4805]: Trace[1659649112]: [13.675944624s] [13.675944624s] END Dec 03 14:09:48 crc kubenswrapper[4805]: I1203 14:09:48.612054 4805 reflector.go:368] Caches populated for *v1.Node from k8s.io/client-go/informers/factory.go:160 Dec 03 14:09:48 crc kubenswrapper[4805]: I1203 14:09:48.613172 4805 reconstruct.go:205] "DevicePaths of reconstructed volumes updated" Dec 03 14:09:48 crc kubenswrapper[4805]: I1203 14:09:48.613362 4805 trace.go:236] Trace[179570795]: "Reflector ListAndWatch" name:k8s.io/client-go/informers/factory.go:160 (03-Dec-2025 14:09:34.098) (total time: 14515ms): Dec 03 14:09:48 crc kubenswrapper[4805]: Trace[179570795]: ---"Objects listed" error: 14515ms (14:09:48.613) Dec 03 14:09:48 crc kubenswrapper[4805]: Trace[179570795]: [14.51506103s] [14.51506103s] END Dec 03 14:09:48 crc kubenswrapper[4805]: I1203 14:09:48.613598 4805 reflector.go:368] Caches populated for *v1.CSIDriver from k8s.io/client-go/informers/factory.go:160 Dec 03 14:09:48 crc kubenswrapper[4805]: E1203 14:09:48.618157 4805 kubelet_node_status.go:99] "Unable to register node with API server" err="nodes \"crc\" is forbidden: autoscaling.openshift.io/ManagedNode infra config cache not synchronized" node="crc" Dec 03 14:09:48 crc kubenswrapper[4805]: I1203 14:09:48.632773 4805 apiserver.go:52] "Watching apiserver" Dec 03 14:09:48 crc kubenswrapper[4805]: I1203 14:09:48.635942 4805 reflector.go:368] Caches populated for *v1.Pod from pkg/kubelet/config/apiserver.go:66 Dec 03 14:09:48 crc kubenswrapper[4805]: I1203 14:09:48.636275 4805 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-network-diagnostics/network-check-target-xd92c","openshift-network-node-identity/network-node-identity-vrzqb","openshift-network-operator/iptables-alerter-4ln5h","openshift-network-operator/network-operator-58b4c7f79c-55gtf","openshift-network-console/networking-console-plugin-85b44fc459-gdk6g","openshift-network-diagnostics/network-check-source-55646444c4-trplf"] Dec 03 14:09:48 crc kubenswrapper[4805]: I1203 14:09:48.636691 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Dec 03 14:09:48 crc kubenswrapper[4805]: I1203 14:09:48.637191 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 14:09:48 crc kubenswrapper[4805]: I1203 14:09:48.637240 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/iptables-alerter-4ln5h" Dec 03 14:09:48 crc kubenswrapper[4805]: I1203 14:09:48.637271 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 14:09:48 crc kubenswrapper[4805]: I1203 14:09:48.637308 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-node-identity/network-node-identity-vrzqb" Dec 03 14:09:48 crc kubenswrapper[4805]: E1203 14:09:48.637360 4805 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 03 14:09:48 crc kubenswrapper[4805]: I1203 14:09:48.637386 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 14:09:48 crc kubenswrapper[4805]: E1203 14:09:48.637460 4805 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 03 14:09:48 crc kubenswrapper[4805]: E1203 14:09:48.637500 4805 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 03 14:09:48 crc kubenswrapper[4805]: I1203 14:09:48.640602 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-operator"/"metrics-tls" Dec 03 14:09:48 crc kubenswrapper[4805]: I1203 14:09:48.640880 4805 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"env-overrides" Dec 03 14:09:48 crc kubenswrapper[4805]: I1203 14:09:48.641312 4805 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"iptables-alerter-script" Dec 03 14:09:48 crc kubenswrapper[4805]: I1203 14:09:48.641394 4805 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"openshift-service-ca.crt" Dec 03 14:09:48 crc kubenswrapper[4805]: I1203 14:09:48.641320 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-node-identity"/"network-node-identity-cert" Dec 03 14:09:48 crc kubenswrapper[4805]: I1203 14:09:48.641522 4805 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"ovnkube-identity-cm" Dec 03 14:09:48 crc kubenswrapper[4805]: I1203 14:09:48.641717 4805 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"openshift-service-ca.crt" Dec 03 14:09:48 crc kubenswrapper[4805]: I1203 14:09:48.643381 4805 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"kube-root-ca.crt" Dec 03 14:09:48 crc kubenswrapper[4805]: I1203 14:09:48.645178 4805 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"kube-root-ca.crt" Dec 03 14:09:48 crc kubenswrapper[4805]: I1203 14:09:48.669391 4805 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver-check-endpoints namespace/openshift-kube-apiserver: Liveness probe status=failure output="Get \"https://192.168.126.11:17697/healthz\": read tcp 192.168.126.11:35664->192.168.126.11:17697: read: connection reset by peer" start-of-body= Dec 03 14:09:48 crc kubenswrapper[4805]: I1203 14:09:48.669443 4805 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" probeResult="failure" output="Get \"https://192.168.126.11:17697/healthz\": read tcp 192.168.126.11:35664->192.168.126.11:17697: read: connection reset by peer" Dec 03 14:09:48 crc kubenswrapper[4805]: I1203 14:09:48.670572 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:48Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:48Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 03 14:09:48 crc kubenswrapper[4805]: I1203 14:09:48.687263 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:48Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:48Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 03 14:09:48 crc kubenswrapper[4805]: I1203 14:09:48.698940 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:48Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:48Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 03 14:09:48 crc kubenswrapper[4805]: I1203 14:09:48.711729 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:48Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:48Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 03 14:09:48 crc kubenswrapper[4805]: I1203 14:09:48.724589 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:48Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:48Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 03 14:09:48 crc kubenswrapper[4805]: I1203 14:09:48.736119 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:48Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:48Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 03 14:09:48 crc kubenswrapper[4805]: I1203 14:09:48.736258 4805 desired_state_of_world_populator.go:154] "Finished populating initial desired state of world" Dec 03 14:09:48 crc kubenswrapper[4805]: I1203 14:09:48.750550 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:48Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:48Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 03 14:09:48 crc kubenswrapper[4805]: I1203 14:09:48.779091 4805 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/0.log" Dec 03 14:09:48 crc kubenswrapper[4805]: I1203 14:09:48.781465 4805 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="5da55f769e23ad52bac9c77da62c78e70fa34db8ed1a435e4eaf4dad3fd2b68a" exitCode=255 Dec 03 14:09:48 crc kubenswrapper[4805]: I1203 14:09:48.781513 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerDied","Data":"5da55f769e23ad52bac9c77da62c78e70fa34db8ed1a435e4eaf4dad3fd2b68a"} Dec 03 14:09:48 crc kubenswrapper[4805]: I1203 14:09:48.799102 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:48Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:48Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 03 14:09:48 crc kubenswrapper[4805]: I1203 14:09:48.814304 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Dec 03 14:09:48 crc kubenswrapper[4805]: I1203 14:09:48.814371 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pcxfs\" (UniqueName: \"kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Dec 03 14:09:48 crc kubenswrapper[4805]: I1203 14:09:48.814406 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 14:09:48 crc kubenswrapper[4805]: I1203 14:09:48.814436 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Dec 03 14:09:48 crc kubenswrapper[4805]: I1203 14:09:48.814469 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities\") pod \"57a731c4-ef35-47a8-b875-bfb08a7f8011\" (UID: \"57a731c4-ef35-47a8-b875-bfb08a7f8011\") " Dec 03 14:09:48 crc kubenswrapper[4805]: I1203 14:09:48.814497 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bf2bz\" (UniqueName: \"kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz\") pod \"1d611f23-29be-4491-8495-bee1670e935f\" (UID: \"1d611f23-29be-4491-8495-bee1670e935f\") " Dec 03 14:09:48 crc kubenswrapper[4805]: I1203 14:09:48.814523 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cfbct\" (UniqueName: \"kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct\") pod \"57a731c4-ef35-47a8-b875-bfb08a7f8011\" (UID: \"57a731c4-ef35-47a8-b875-bfb08a7f8011\") " Dec 03 14:09:48 crc kubenswrapper[4805]: I1203 14:09:48.814553 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Dec 03 14:09:48 crc kubenswrapper[4805]: I1203 14:09:48.814587 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Dec 03 14:09:48 crc kubenswrapper[4805]: I1203 14:09:48.814617 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Dec 03 14:09:48 crc kubenswrapper[4805]: I1203 14:09:48.814642 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Dec 03 14:09:48 crc kubenswrapper[4805]: I1203 14:09:48.814669 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Dec 03 14:09:48 crc kubenswrapper[4805]: I1203 14:09:48.814695 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kfwg7\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 14:09:48 crc kubenswrapper[4805]: I1203 14:09:48.814722 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rnphk\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Dec 03 14:09:48 crc kubenswrapper[4805]: I1203 14:09:48.814747 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca\") pod \"0b78653f-4ff9-4508-8672-245ed9b561e3\" (UID: \"0b78653f-4ff9-4508-8672-245ed9b561e3\") " Dec 03 14:09:48 crc kubenswrapper[4805]: I1203 14:09:48.814772 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert\") pod \"7539238d-5fe0-46ed-884e-1c3b566537ec\" (UID: \"7539238d-5fe0-46ed-884e-1c3b566537ec\") " Dec 03 14:09:48 crc kubenswrapper[4805]: I1203 14:09:48.814799 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lzf88\" (UniqueName: \"kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88\") pod \"0b574797-001e-440a-8f4e-c0be86edad0f\" (UID: \"0b574797-001e-440a-8f4e-c0be86edad0f\") " Dec 03 14:09:48 crc kubenswrapper[4805]: I1203 14:09:48.814838 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert\") pod \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\" (UID: \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\") " Dec 03 14:09:48 crc kubenswrapper[4805]: I1203 14:09:48.814888 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access\") pod \"0b78653f-4ff9-4508-8672-245ed9b561e3\" (UID: \"0b78653f-4ff9-4508-8672-245ed9b561e3\") " Dec 03 14:09:48 crc kubenswrapper[4805]: I1203 14:09:48.814917 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config\") pod \"e7e6199b-1264-4501-8953-767f51328d08\" (UID: \"e7e6199b-1264-4501-8953-767f51328d08\") " Dec 03 14:09:48 crc kubenswrapper[4805]: I1203 14:09:48.814947 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Dec 03 14:09:48 crc kubenswrapper[4805]: I1203 14:09:48.814979 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zkvpv\" (UniqueName: \"kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Dec 03 14:09:48 crc kubenswrapper[4805]: I1203 14:09:48.815004 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Dec 03 14:09:48 crc kubenswrapper[4805]: I1203 14:09:48.815026 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities\") pod \"1d611f23-29be-4491-8495-bee1670e935f\" (UID: \"1d611f23-29be-4491-8495-bee1670e935f\") " Dec 03 14:09:48 crc kubenswrapper[4805]: I1203 14:09:48.815047 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Dec 03 14:09:48 crc kubenswrapper[4805]: I1203 14:09:48.815069 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Dec 03 14:09:48 crc kubenswrapper[4805]: I1203 14:09:48.815089 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Dec 03 14:09:48 crc kubenswrapper[4805]: I1203 14:09:48.815110 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Dec 03 14:09:48 crc kubenswrapper[4805]: I1203 14:09:48.815134 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert\") pod \"496e6271-fb68-4057-954e-a0d97a4afa3f\" (UID: \"496e6271-fb68-4057-954e-a0d97a4afa3f\") " Dec 03 14:09:48 crc kubenswrapper[4805]: I1203 14:09:48.815154 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Dec 03 14:09:48 crc kubenswrapper[4805]: I1203 14:09:48.815183 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Dec 03 14:09:48 crc kubenswrapper[4805]: I1203 14:09:48.815181 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template" (OuterVolumeSpecName: "v4-0-config-system-ocp-branding-template") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-ocp-branding-template". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 14:09:48 crc kubenswrapper[4805]: I1203 14:09:48.815205 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume\") pod \"87cf06ed-a83f-41a7-828d-70653580a8cb\" (UID: \"87cf06ed-a83f-41a7-828d-70653580a8cb\") " Dec 03 14:09:48 crc kubenswrapper[4805]: I1203 14:09:48.815286 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Dec 03 14:09:48 crc kubenswrapper[4805]: I1203 14:09:48.815317 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Dec 03 14:09:48 crc kubenswrapper[4805]: I1203 14:09:48.815345 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Dec 03 14:09:48 crc kubenswrapper[4805]: I1203 14:09:48.815369 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Dec 03 14:09:48 crc kubenswrapper[4805]: I1203 14:09:48.815392 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pj782\" (UniqueName: \"kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782\") pod \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\" (UID: \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\") " Dec 03 14:09:48 crc kubenswrapper[4805]: I1203 14:09:48.815414 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates\") pod \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\" (UID: \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\") " Dec 03 14:09:48 crc kubenswrapper[4805]: I1203 14:09:48.815436 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Dec 03 14:09:48 crc kubenswrapper[4805]: I1203 14:09:48.815539 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 14:09:48 crc kubenswrapper[4805]: I1203 14:09:48.815564 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access\") pod \"e7e6199b-1264-4501-8953-767f51328d08\" (UID: \"e7e6199b-1264-4501-8953-767f51328d08\") " Dec 03 14:09:48 crc kubenswrapper[4805]: I1203 14:09:48.815587 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Dec 03 14:09:48 crc kubenswrapper[4805]: I1203 14:09:48.815611 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x2m85\" (UniqueName: \"kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85\") pod \"cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d\" (UID: \"cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d\") " Dec 03 14:09:48 crc kubenswrapper[4805]: I1203 14:09:48.815633 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 14:09:48 crc kubenswrapper[4805]: I1203 14:09:48.815659 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Dec 03 14:09:48 crc kubenswrapper[4805]: I1203 14:09:48.815683 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-sb6h7\" (UniqueName: \"kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Dec 03 14:09:48 crc kubenswrapper[4805]: I1203 14:09:48.815704 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6g6sz\" (UniqueName: \"kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Dec 03 14:09:48 crc kubenswrapper[4805]: I1203 14:09:48.815731 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Dec 03 14:09:48 crc kubenswrapper[4805]: I1203 14:09:48.815753 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs\") pod \"5fe579f8-e8a6-4643-bce5-a661393c4dde\" (UID: \"5fe579f8-e8a6-4643-bce5-a661393c4dde\") " Dec 03 14:09:48 crc kubenswrapper[4805]: I1203 14:09:48.815752 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls" (OuterVolumeSpecName: "machine-api-operator-tls") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "machine-api-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 14:09:48 crc kubenswrapper[4805]: I1203 14:09:48.815776 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls\") pod \"fda69060-fa79-4696-b1a6-7980f124bf7c\" (UID: \"fda69060-fa79-4696-b1a6-7980f124bf7c\") " Dec 03 14:09:48 crc kubenswrapper[4805]: I1203 14:09:48.815802 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Dec 03 14:09:48 crc kubenswrapper[4805]: I1203 14:09:48.815821 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls\") pod \"87cf06ed-a83f-41a7-828d-70653580a8cb\" (UID: \"87cf06ed-a83f-41a7-828d-70653580a8cb\") " Dec 03 14:09:48 crc kubenswrapper[4805]: I1203 14:09:48.815866 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content\") pod \"5225d0e4-402f-4861-b410-819f433b1803\" (UID: \"5225d0e4-402f-4861-b410-819f433b1803\") " Dec 03 14:09:48 crc kubenswrapper[4805]: I1203 14:09:48.815887 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x4zgh\" (UniqueName: \"kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh\") pod \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\" (UID: \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\") " Dec 03 14:09:48 crc kubenswrapper[4805]: I1203 14:09:48.815908 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Dec 03 14:09:48 crc kubenswrapper[4805]: I1203 14:09:48.815931 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content\") pod \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\" (UID: \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\") " Dec 03 14:09:48 crc kubenswrapper[4805]: I1203 14:09:48.815953 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Dec 03 14:09:48 crc kubenswrapper[4805]: I1203 14:09:48.815975 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Dec 03 14:09:48 crc kubenswrapper[4805]: I1203 14:09:48.815978 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782" (OuterVolumeSpecName: "kube-api-access-pj782") pod "b6cd30de-2eeb-49a2-ab40-9167f4560ff5" (UID: "b6cd30de-2eeb-49a2-ab40-9167f4560ff5"). InnerVolumeSpecName "kube-api-access-pj782". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 14:09:48 crc kubenswrapper[4805]: I1203 14:09:48.816002 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls\") pod \"6731426b-95fe-49ff-bb5f-40441049fde2\" (UID: \"6731426b-95fe-49ff-bb5f-40441049fde2\") " Dec 03 14:09:48 crc kubenswrapper[4805]: I1203 14:09:48.816025 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Dec 03 14:09:48 crc kubenswrapper[4805]: I1203 14:09:48.816045 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Dec 03 14:09:48 crc kubenswrapper[4805]: I1203 14:09:48.816068 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content\") pod \"57a731c4-ef35-47a8-b875-bfb08a7f8011\" (UID: \"57a731c4-ef35-47a8-b875-bfb08a7f8011\") " Dec 03 14:09:48 crc kubenswrapper[4805]: I1203 14:09:48.816091 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Dec 03 14:09:48 crc kubenswrapper[4805]: I1203 14:09:48.816113 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Dec 03 14:09:48 crc kubenswrapper[4805]: I1203 14:09:48.816135 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert\") pod \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\" (UID: \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\") " Dec 03 14:09:48 crc kubenswrapper[4805]: I1203 14:09:48.816138 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume" (OuterVolumeSpecName: "config-volume") pod "87cf06ed-a83f-41a7-828d-70653580a8cb" (UID: "87cf06ed-a83f-41a7-828d-70653580a8cb"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 14:09:48 crc kubenswrapper[4805]: I1203 14:09:48.816159 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist\") pod \"7bb08738-c794-4ee8-9972-3a62ca171029\" (UID: \"7bb08738-c794-4ee8-9972-3a62ca171029\") " Dec 03 14:09:48 crc kubenswrapper[4805]: I1203 14:09:48.816165 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 14:09:48 crc kubenswrapper[4805]: I1203 14:09:48.816185 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config\") pod \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\" (UID: \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\") " Dec 03 14:09:48 crc kubenswrapper[4805]: I1203 14:09:48.816211 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config\") pod \"fda69060-fa79-4696-b1a6-7980f124bf7c\" (UID: \"fda69060-fa79-4696-b1a6-7980f124bf7c\") " Dec 03 14:09:48 crc kubenswrapper[4805]: I1203 14:09:48.816234 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7c4vf\" (UniqueName: \"kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Dec 03 14:09:48 crc kubenswrapper[4805]: I1203 14:09:48.816258 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gf66m\" (UniqueName: \"kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m\") pod \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\" (UID: \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\") " Dec 03 14:09:48 crc kubenswrapper[4805]: I1203 14:09:48.816278 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Dec 03 14:09:48 crc kubenswrapper[4805]: I1203 14:09:48.816299 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9xfj7\" (UniqueName: \"kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7\") pod \"5225d0e4-402f-4861-b410-819f433b1803\" (UID: \"5225d0e4-402f-4861-b410-819f433b1803\") " Dec 03 14:09:48 crc kubenswrapper[4805]: I1203 14:09:48.816320 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w9rds\" (UniqueName: \"kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds\") pod \"20b0d48f-5fd6-431c-a545-e3c800c7b866\" (UID: \"20b0d48f-5fd6-431c-a545-e3c800c7b866\") " Dec 03 14:09:48 crc kubenswrapper[4805]: I1203 14:09:48.816335 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs" (OuterVolumeSpecName: "kube-api-access-pcxfs") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "kube-api-access-pcxfs". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 14:09:48 crc kubenswrapper[4805]: I1203 14:09:48.816338 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config" (OuterVolumeSpecName: "config") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 14:09:48 crc kubenswrapper[4805]: I1203 14:09:48.816344 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Dec 03 14:09:48 crc kubenswrapper[4805]: I1203 14:09:48.816391 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content\") pod \"1d611f23-29be-4491-8495-bee1670e935f\" (UID: \"1d611f23-29be-4491-8495-bee1670e935f\") " Dec 03 14:09:48 crc kubenswrapper[4805]: I1203 14:09:48.816417 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Dec 03 14:09:48 crc kubenswrapper[4805]: I1203 14:09:48.816439 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Dec 03 14:09:48 crc kubenswrapper[4805]: I1203 14:09:48.816463 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 14:09:48 crc kubenswrapper[4805]: I1203 14:09:48.816512 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Dec 03 14:09:48 crc kubenswrapper[4805]: I1203 14:09:48.816521 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls" (OuterVolumeSpecName: "registry-tls") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "registry-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 14:09:48 crc kubenswrapper[4805]: I1203 14:09:48.816535 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls" (OuterVolumeSpecName: "metrics-tls") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "metrics-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 14:09:48 crc kubenswrapper[4805]: I1203 14:09:48.816540 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config\") pod \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\" (UID: \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\") " Dec 03 14:09:48 crc kubenswrapper[4805]: I1203 14:09:48.816579 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 14:09:48 crc kubenswrapper[4805]: I1203 14:09:48.816603 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Dec 03 14:09:48 crc kubenswrapper[4805]: I1203 14:09:48.816629 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Dec 03 14:09:48 crc kubenswrapper[4805]: I1203 14:09:48.816653 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs\") pod \"5b88f790-22fa-440e-b583-365168c0b23d\" (UID: \"5b88f790-22fa-440e-b583-365168c0b23d\") " Dec 03 14:09:48 crc kubenswrapper[4805]: I1203 14:09:48.816664 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle" (OuterVolumeSpecName: "v4-0-config-system-trusted-ca-bundle") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 14:09:48 crc kubenswrapper[4805]: I1203 14:09:48.816676 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xcphl\" (UniqueName: \"kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Dec 03 14:09:48 crc kubenswrapper[4805]: I1203 14:09:48.816700 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Dec 03 14:09:48 crc kubenswrapper[4805]: I1203 14:09:48.816724 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tk88c\" (UniqueName: \"kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c\") pod \"7539238d-5fe0-46ed-884e-1c3b566537ec\" (UID: \"7539238d-5fe0-46ed-884e-1c3b566537ec\") " Dec 03 14:09:48 crc kubenswrapper[4805]: I1203 14:09:48.816755 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config" (OuterVolumeSpecName: "encryption-config") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "encryption-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 14:09:48 crc kubenswrapper[4805]: I1203 14:09:48.816759 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Dec 03 14:09:48 crc kubenswrapper[4805]: I1203 14:09:48.816803 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics\") pod \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\" (UID: \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\") " Dec 03 14:09:48 crc kubenswrapper[4805]: I1203 14:09:48.816831 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert\") pod \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\" (UID: \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\") " Dec 03 14:09:48 crc kubenswrapper[4805]: I1203 14:09:48.816903 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy\") pod \"7bb08738-c794-4ee8-9972-3a62ca171029\" (UID: \"7bb08738-c794-4ee8-9972-3a62ca171029\") " Dec 03 14:09:48 crc kubenswrapper[4805]: I1203 14:09:48.816929 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config\") pod \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\" (UID: \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\") " Dec 03 14:09:48 crc kubenswrapper[4805]: I1203 14:09:48.816952 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Dec 03 14:09:48 crc kubenswrapper[4805]: I1203 14:09:48.816975 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert\") pod \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\" (UID: \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\") " Dec 03 14:09:48 crc kubenswrapper[4805]: I1203 14:09:48.816999 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ngvvp\" (UniqueName: \"kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Dec 03 14:09:48 crc kubenswrapper[4805]: I1203 14:09:48.817009 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config" (OuterVolumeSpecName: "config") pod "1386a44e-36a2-460c-96d0-0359d2b6f0f5" (UID: "1386a44e-36a2-460c-96d0-0359d2b6f0f5"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 14:09:48 crc kubenswrapper[4805]: I1203 14:09:48.817023 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Dec 03 14:09:48 crc kubenswrapper[4805]: I1203 14:09:48.817080 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access\") pod \"496e6271-fb68-4057-954e-a0d97a4afa3f\" (UID: \"496e6271-fb68-4057-954e-a0d97a4afa3f\") " Dec 03 14:09:48 crc kubenswrapper[4805]: I1203 14:09:48.817109 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Dec 03 14:09:48 crc kubenswrapper[4805]: I1203 14:09:48.817134 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls\") pod \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\" (UID: \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\") " Dec 03 14:09:48 crc kubenswrapper[4805]: I1203 14:09:48.817157 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy\") pod \"4bb40260-dbaa-4fb0-84df-5e680505d512\" (UID: \"4bb40260-dbaa-4fb0-84df-5e680505d512\") " Dec 03 14:09:48 crc kubenswrapper[4805]: I1203 14:09:48.817223 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Dec 03 14:09:48 crc kubenswrapper[4805]: I1203 14:09:48.817250 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca\") pod \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\" (UID: \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\") " Dec 03 14:09:48 crc kubenswrapper[4805]: I1203 14:09:48.817281 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6ccd8\" (UniqueName: \"kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Dec 03 14:09:48 crc kubenswrapper[4805]: I1203 14:09:48.817305 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key\") pod \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\" (UID: \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\") " Dec 03 14:09:48 crc kubenswrapper[4805]: I1203 14:09:48.817327 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert\") pod \"0b78653f-4ff9-4508-8672-245ed9b561e3\" (UID: \"0b78653f-4ff9-4508-8672-245ed9b561e3\") " Dec 03 14:09:48 crc kubenswrapper[4805]: I1203 14:09:48.817349 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wxkg8\" (UniqueName: \"kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8\") pod \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\" (UID: \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\") " Dec 03 14:09:48 crc kubenswrapper[4805]: I1203 14:09:48.817370 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Dec 03 14:09:48 crc kubenswrapper[4805]: I1203 14:09:48.817391 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pjr6v\" (UniqueName: \"kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v\") pod \"49ef4625-1d3a-4a9f-b595-c2433d32326d\" (UID: \"49ef4625-1d3a-4a9f-b595-c2433d32326d\") " Dec 03 14:09:48 crc kubenswrapper[4805]: I1203 14:09:48.817413 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xcgwh\" (UniqueName: \"kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh\") pod \"fda69060-fa79-4696-b1a6-7980f124bf7c\" (UID: \"fda69060-fa79-4696-b1a6-7980f124bf7c\") " Dec 03 14:09:48 crc kubenswrapper[4805]: I1203 14:09:48.817434 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Dec 03 14:09:48 crc kubenswrapper[4805]: I1203 14:09:48.817459 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-249nr\" (UniqueName: \"kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr\") pod \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\" (UID: \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\") " Dec 03 14:09:48 crc kubenswrapper[4805]: I1203 14:09:48.817479 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert\") pod \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\" (UID: \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\") " Dec 03 14:09:48 crc kubenswrapper[4805]: I1203 14:09:48.817499 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Dec 03 14:09:48 crc kubenswrapper[4805]: I1203 14:09:48.817076 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config" (OuterVolumeSpecName: "config") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 14:09:48 crc kubenswrapper[4805]: I1203 14:09:48.817521 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Dec 03 14:09:48 crc kubenswrapper[4805]: I1203 14:09:48.817546 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fqsjt\" (UniqueName: \"kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt\") pod \"efdd0498-1daa-4136-9a4a-3b948c2293fc\" (UID: \"efdd0498-1daa-4136-9a4a-3b948c2293fc\") " Dec 03 14:09:48 crc kubenswrapper[4805]: I1203 14:09:48.817568 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Dec 03 14:09:48 crc kubenswrapper[4805]: I1203 14:09:48.817590 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Dec 03 14:09:48 crc kubenswrapper[4805]: I1203 14:09:48.817615 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4d4hj\" (UniqueName: \"kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj\") pod \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\" (UID: \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\") " Dec 03 14:09:48 crc kubenswrapper[4805]: I1203 14:09:48.817639 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert\") pod \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\" (UID: \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\") " Dec 03 14:09:48 crc kubenswrapper[4805]: I1203 14:09:48.817663 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Dec 03 14:09:48 crc kubenswrapper[4805]: I1203 14:09:48.817690 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert\") pod \"e7e6199b-1264-4501-8953-767f51328d08\" (UID: \"e7e6199b-1264-4501-8953-767f51328d08\") " Dec 03 14:09:48 crc kubenswrapper[4805]: I1203 14:09:48.817713 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Dec 03 14:09:48 crc kubenswrapper[4805]: I1203 14:09:48.817735 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Dec 03 14:09:48 crc kubenswrapper[4805]: I1203 14:09:48.817763 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Dec 03 14:09:48 crc kubenswrapper[4805]: I1203 14:09:48.817788 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Dec 03 14:09:48 crc kubenswrapper[4805]: I1203 14:09:48.817840 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Dec 03 14:09:48 crc kubenswrapper[4805]: I1203 14:09:48.817880 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Dec 03 14:09:48 crc kubenswrapper[4805]: I1203 14:09:48.817905 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vt5rc\" (UniqueName: \"kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc\") pod \"44663579-783b-4372-86d6-acf235a62d72\" (UID: \"44663579-783b-4372-86d6-acf235a62d72\") " Dec 03 14:09:48 crc kubenswrapper[4805]: I1203 14:09:48.817931 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config\") pod \"4bb40260-dbaa-4fb0-84df-5e680505d512\" (UID: \"4bb40260-dbaa-4fb0-84df-5e680505d512\") " Dec 03 14:09:48 crc kubenswrapper[4805]: I1203 14:09:48.825246 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:48Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:48Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 03 14:09:48 crc kubenswrapper[4805]: I1203 14:09:48.817217 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv" (OuterVolumeSpecName: "kube-api-access-zkvpv") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "kube-api-access-zkvpv". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 14:09:48 crc kubenswrapper[4805]: I1203 14:09:48.817282 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "496e6271-fb68-4057-954e-a0d97a4afa3f" (UID: "496e6271-fb68-4057-954e-a0d97a4afa3f"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 14:09:48 crc kubenswrapper[4805]: I1203 14:09:48.817279 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert" (OuterVolumeSpecName: "ovn-control-plane-metrics-cert") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "ovn-control-plane-metrics-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 14:09:48 crc kubenswrapper[4805]: I1203 14:09:48.817507 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls" (OuterVolumeSpecName: "metrics-tls") pod "96b93a3a-6083-4aea-8eab-fe1aa8245ad9" (UID: "96b93a3a-6083-4aea-8eab-fe1aa8245ad9"). InnerVolumeSpecName "metrics-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 14:09:48 crc kubenswrapper[4805]: I1203 14:09:48.817554 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7" (OuterVolumeSpecName: "kube-api-access-kfwg7") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "kube-api-access-kfwg7". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 14:09:48 crc kubenswrapper[4805]: I1203 14:09:48.817565 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca" (OuterVolumeSpecName: "etcd-serving-ca") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "etcd-serving-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 14:09:48 crc kubenswrapper[4805]: I1203 14:09:48.817614 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config" (OuterVolumeSpecName: "console-config") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "console-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 14:09:48 crc kubenswrapper[4805]: I1203 14:09:48.817631 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca" (OuterVolumeSpecName: "service-ca") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 14:09:48 crc kubenswrapper[4805]: I1203 14:09:48.817767 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk" (OuterVolumeSpecName: "kube-api-access-rnphk") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "kube-api-access-rnphk". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 14:09:48 crc kubenswrapper[4805]: I1203 14:09:48.817881 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error" (OuterVolumeSpecName: "v4-0-config-user-template-error") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-template-error". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 14:09:48 crc kubenswrapper[4805]: I1203 14:09:48.817972 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities" (OuterVolumeSpecName: "utilities") pod "57a731c4-ef35-47a8-b875-bfb08a7f8011" (UID: "57a731c4-ef35-47a8-b875-bfb08a7f8011"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 14:09:48 crc kubenswrapper[4805]: I1203 14:09:48.818062 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles" (OuterVolumeSpecName: "proxy-ca-bundles") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 14:09:48 crc kubenswrapper[4805]: I1203 14:09:48.818169 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz" (OuterVolumeSpecName: "kube-api-access-bf2bz") pod "1d611f23-29be-4491-8495-bee1670e935f" (UID: "1d611f23-29be-4491-8495-bee1670e935f"). InnerVolumeSpecName "kube-api-access-bf2bz". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 14:09:48 crc kubenswrapper[4805]: I1203 14:09:48.818394 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities" (OuterVolumeSpecName: "utilities") pod "1d611f23-29be-4491-8495-bee1670e935f" (UID: "1d611f23-29be-4491-8495-bee1670e935f"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 14:09:48 crc kubenswrapper[4805]: I1203 14:09:48.818469 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 14:09:48 crc kubenswrapper[4805]: E1203 14:09:48.818569 4805 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-03 14:09:49.318544225 +0000 UTC m=+18.981461348 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 14:09:48 crc kubenswrapper[4805]: I1203 14:09:48.828400 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities\") pod \"5225d0e4-402f-4861-b410-819f433b1803\" (UID: \"5225d0e4-402f-4861-b410-819f433b1803\") " Dec 03 14:09:48 crc kubenswrapper[4805]: I1203 14:09:48.828442 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert\") pod \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\" (UID: \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\") " Dec 03 14:09:48 crc kubenswrapper[4805]: I1203 14:09:48.828471 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Dec 03 14:09:48 crc kubenswrapper[4805]: I1203 14:09:48.828496 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities\") pod \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\" (UID: \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\") " Dec 03 14:09:48 crc kubenswrapper[4805]: I1203 14:09:48.828520 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Dec 03 14:09:48 crc kubenswrapper[4805]: I1203 14:09:48.828551 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w7l8j\" (UniqueName: \"kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j\") pod \"01ab3dd5-8196-46d0-ad33-122e2ca51def\" (UID: \"01ab3dd5-8196-46d0-ad33-122e2ca51def\") " Dec 03 14:09:48 crc kubenswrapper[4805]: I1203 14:09:48.828576 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w4xd4\" (UniqueName: \"kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4\") pod \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\" (UID: \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\") " Dec 03 14:09:48 crc kubenswrapper[4805]: I1203 14:09:48.828600 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qg5z5\" (UniqueName: \"kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Dec 03 14:09:48 crc kubenswrapper[4805]: I1203 14:09:48.828623 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Dec 03 14:09:48 crc kubenswrapper[4805]: I1203 14:09:48.828644 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Dec 03 14:09:48 crc kubenswrapper[4805]: I1203 14:09:48.828669 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Dec 03 14:09:48 crc kubenswrapper[4805]: I1203 14:09:48.828692 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nzwt7\" (UniqueName: \"kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7\") pod \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\" (UID: \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\") " Dec 03 14:09:48 crc kubenswrapper[4805]: I1203 14:09:48.828715 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca\") pod \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\" (UID: \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\") " Dec 03 14:09:48 crc kubenswrapper[4805]: I1203 14:09:48.828739 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Dec 03 14:09:48 crc kubenswrapper[4805]: I1203 14:09:48.828761 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-v47cf\" (UniqueName: \"kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Dec 03 14:09:48 crc kubenswrapper[4805]: I1203 14:09:48.828787 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config\") pod \"7539238d-5fe0-46ed-884e-1c3b566537ec\" (UID: \"7539238d-5fe0-46ed-884e-1c3b566537ec\") " Dec 03 14:09:48 crc kubenswrapper[4805]: I1203 14:09:48.828809 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle\") pod \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\" (UID: \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\") " Dec 03 14:09:48 crc kubenswrapper[4805]: I1203 14:09:48.828829 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config\") pod \"01ab3dd5-8196-46d0-ad33-122e2ca51def\" (UID: \"01ab3dd5-8196-46d0-ad33-122e2ca51def\") " Dec 03 14:09:48 crc kubenswrapper[4805]: I1203 14:09:48.828874 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-d4lsv\" (UniqueName: \"kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv\") pod \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\" (UID: \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\") " Dec 03 14:09:48 crc kubenswrapper[4805]: I1203 14:09:48.828897 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Dec 03 14:09:48 crc kubenswrapper[4805]: I1203 14:09:48.828926 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-279lb\" (UniqueName: \"kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb\") pod \"7bb08738-c794-4ee8-9972-3a62ca171029\" (UID: \"7bb08738-c794-4ee8-9972-3a62ca171029\") " Dec 03 14:09:48 crc kubenswrapper[4805]: I1203 14:09:48.828949 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Dec 03 14:09:48 crc kubenswrapper[4805]: I1203 14:09:48.828977 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Dec 03 14:09:48 crc kubenswrapper[4805]: I1203 14:09:48.829002 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mnrrd\" (UniqueName: \"kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd\") pod \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\" (UID: \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\") " Dec 03 14:09:48 crc kubenswrapper[4805]: I1203 14:09:48.829030 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token\") pod \"5fe579f8-e8a6-4643-bce5-a661393c4dde\" (UID: \"5fe579f8-e8a6-4643-bce5-a661393c4dde\") " Dec 03 14:09:48 crc kubenswrapper[4805]: I1203 14:09:48.829058 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dbsvg\" (UniqueName: \"kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg\") pod \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\" (UID: \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\") " Dec 03 14:09:48 crc kubenswrapper[4805]: I1203 14:09:48.829086 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert\") pod \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\" (UID: \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\") " Dec 03 14:09:48 crc kubenswrapper[4805]: I1203 14:09:48.829110 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fcqwp\" (UniqueName: \"kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp\") pod \"5fe579f8-e8a6-4643-bce5-a661393c4dde\" (UID: \"5fe579f8-e8a6-4643-bce5-a661393c4dde\") " Dec 03 14:09:48 crc kubenswrapper[4805]: I1203 14:09:48.829132 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config\") pod \"496e6271-fb68-4057-954e-a0d97a4afa3f\" (UID: \"496e6271-fb68-4057-954e-a0d97a4afa3f\") " Dec 03 14:09:48 crc kubenswrapper[4805]: I1203 14:09:48.829159 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Dec 03 14:09:48 crc kubenswrapper[4805]: I1203 14:09:48.829181 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs\") pod \"efdd0498-1daa-4136-9a4a-3b948c2293fc\" (UID: \"efdd0498-1daa-4136-9a4a-3b948c2293fc\") " Dec 03 14:09:48 crc kubenswrapper[4805]: I1203 14:09:48.829206 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mg5zb\" (UniqueName: \"kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Dec 03 14:09:48 crc kubenswrapper[4805]: I1203 14:09:48.829232 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Dec 03 14:09:48 crc kubenswrapper[4805]: I1203 14:09:48.829272 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client" (OuterVolumeSpecName: "etcd-client") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "etcd-client". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 14:09:48 crc kubenswrapper[4805]: I1203 14:09:48.829291 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls\") pod \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\" (UID: \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\") " Dec 03 14:09:48 crc kubenswrapper[4805]: I1203 14:09:48.829374 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jkwtn\" (UniqueName: \"kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn\") pod \"5b88f790-22fa-440e-b583-365168c0b23d\" (UID: \"5b88f790-22fa-440e-b583-365168c0b23d\") " Dec 03 14:09:48 crc kubenswrapper[4805]: I1203 14:09:48.829403 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-htfz6\" (UniqueName: \"kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Dec 03 14:09:48 crc kubenswrapper[4805]: I1203 14:09:48.829431 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access\") pod \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\" (UID: \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\") " Dec 03 14:09:48 crc kubenswrapper[4805]: I1203 14:09:48.829456 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Dec 03 14:09:48 crc kubenswrapper[4805]: I1203 14:09:48.829475 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert\") pod \"01ab3dd5-8196-46d0-ad33-122e2ca51def\" (UID: \"01ab3dd5-8196-46d0-ad33-122e2ca51def\") " Dec 03 14:09:48 crc kubenswrapper[4805]: I1203 14:09:48.829492 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-s4n52\" (UniqueName: \"kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Dec 03 14:09:48 crc kubenswrapper[4805]: I1203 14:09:48.829494 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls" (OuterVolumeSpecName: "samples-operator-tls") pod "a0128f3a-b052-44ed-a84e-c4c8aaf17c13" (UID: "a0128f3a-b052-44ed-a84e-c4c8aaf17c13"). InnerVolumeSpecName "samples-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 14:09:48 crc kubenswrapper[4805]: I1203 14:09:48.829510 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jhbk2\" (UniqueName: \"kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2\") pod \"bd23aa5c-e532-4e53-bccf-e79f130c5ae8\" (UID: \"bd23aa5c-e532-4e53-bccf-e79f130c5ae8\") " Dec 03 14:09:48 crc kubenswrapper[4805]: I1203 14:09:48.829530 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Dec 03 14:09:48 crc kubenswrapper[4805]: I1203 14:09:48.829555 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert\") pod \"20b0d48f-5fd6-431c-a545-e3c800c7b866\" (UID: \"20b0d48f-5fd6-431c-a545-e3c800c7b866\") " Dec 03 14:09:48 crc kubenswrapper[4805]: I1203 14:09:48.829573 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Dec 03 14:09:48 crc kubenswrapper[4805]: I1203 14:09:48.829591 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Dec 03 14:09:48 crc kubenswrapper[4805]: I1203 14:09:48.829608 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Dec 03 14:09:48 crc kubenswrapper[4805]: I1203 14:09:48.829625 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 14:09:48 crc kubenswrapper[4805]: I1203 14:09:48.829642 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-d6qdx\" (UniqueName: \"kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx\") pod \"87cf06ed-a83f-41a7-828d-70653580a8cb\" (UID: \"87cf06ed-a83f-41a7-828d-70653580a8cb\") " Dec 03 14:09:48 crc kubenswrapper[4805]: I1203 14:09:48.829658 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Dec 03 14:09:48 crc kubenswrapper[4805]: I1203 14:09:48.829675 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Dec 03 14:09:48 crc kubenswrapper[4805]: I1203 14:09:48.829691 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Dec 03 14:09:48 crc kubenswrapper[4805]: I1203 14:09:48.829710 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config\") pod \"0b574797-001e-440a-8f4e-c0be86edad0f\" (UID: \"0b574797-001e-440a-8f4e-c0be86edad0f\") " Dec 03 14:09:48 crc kubenswrapper[4805]: I1203 14:09:48.829727 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Dec 03 14:09:48 crc kubenswrapper[4805]: I1203 14:09:48.829745 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zgdk5\" (UniqueName: \"kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Dec 03 14:09:48 crc kubenswrapper[4805]: I1203 14:09:48.829765 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Dec 03 14:09:48 crc kubenswrapper[4805]: I1203 14:09:48.829782 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Dec 03 14:09:48 crc kubenswrapper[4805]: I1203 14:09:48.829799 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x7zkh\" (UniqueName: \"kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh\") pod \"6731426b-95fe-49ff-bb5f-40441049fde2\" (UID: \"6731426b-95fe-49ff-bb5f-40441049fde2\") " Dec 03 14:09:48 crc kubenswrapper[4805]: I1203 14:09:48.829815 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lz9wn\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Dec 03 14:09:48 crc kubenswrapper[4805]: I1203 14:09:48.829831 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2d4wz\" (UniqueName: \"kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Dec 03 14:09:48 crc kubenswrapper[4805]: I1203 14:09:48.829892 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qs4fp\" (UniqueName: \"kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp\") pod \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\" (UID: \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\") " Dec 03 14:09:48 crc kubenswrapper[4805]: I1203 14:09:48.829945 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 14:09:48 crc kubenswrapper[4805]: I1203 14:09:48.829963 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2w9zh\" (UniqueName: \"kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh\") pod \"4bb40260-dbaa-4fb0-84df-5e680505d512\" (UID: \"4bb40260-dbaa-4fb0-84df-5e680505d512\") " Dec 03 14:09:48 crc kubenswrapper[4805]: I1203 14:09:48.829981 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls\") pod \"0b574797-001e-440a-8f4e-c0be86edad0f\" (UID: \"0b574797-001e-440a-8f4e-c0be86edad0f\") " Dec 03 14:09:48 crc kubenswrapper[4805]: I1203 14:09:48.829997 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8tdtz\" (UniqueName: \"kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Dec 03 14:09:48 crc kubenswrapper[4805]: I1203 14:09:48.830013 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Dec 03 14:09:48 crc kubenswrapper[4805]: I1203 14:09:48.830032 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert\") pod \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\" (UID: \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\") " Dec 03 14:09:48 crc kubenswrapper[4805]: I1203 14:09:48.830070 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca" (OuterVolumeSpecName: "marketplace-trusted-ca") pod "b6cd30de-2eeb-49a2-ab40-9167f4560ff5" (UID: "b6cd30de-2eeb-49a2-ab40-9167f4560ff5"). InnerVolumeSpecName "marketplace-trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 14:09:48 crc kubenswrapper[4805]: I1203 14:09:48.830075 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 14:09:48 crc kubenswrapper[4805]: I1203 14:09:48.830141 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/37a5e44f-9a88-4405-be8a-b645485e7312-metrics-tls\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Dec 03 14:09:48 crc kubenswrapper[4805]: I1203 14:09:48.830185 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 14:09:48 crc kubenswrapper[4805]: I1203 14:09:48.830219 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rdwmf\" (UniqueName: \"kubernetes.io/projected/37a5e44f-9a88-4405-be8a-b645485e7312-kube-api-access-rdwmf\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Dec 03 14:09:48 crc kubenswrapper[4805]: I1203 14:09:48.830249 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2kz5\" (UniqueName: \"kubernetes.io/projected/ef543e1b-8068-4ea3-b32a-61027b32e95d-kube-api-access-s2kz5\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Dec 03 14:09:48 crc kubenswrapper[4805]: I1203 14:09:48.830280 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 14:09:48 crc kubenswrapper[4805]: I1203 14:09:48.830307 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/ef543e1b-8068-4ea3-b32a-61027b32e95d-webhook-cert\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Dec 03 14:09:48 crc kubenswrapper[4805]: I1203 14:09:48.830338 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rczfb\" (UniqueName: \"kubernetes.io/projected/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-kube-api-access-rczfb\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Dec 03 14:09:48 crc kubenswrapper[4805]: I1203 14:09:48.830368 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-etc-kube\" (UniqueName: \"kubernetes.io/host-path/37a5e44f-9a88-4405-be8a-b645485e7312-host-etc-kube\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Dec 03 14:09:48 crc kubenswrapper[4805]: I1203 14:09:48.830396 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 14:09:48 crc kubenswrapper[4805]: I1203 14:09:48.830424 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-host-slash\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Dec 03 14:09:48 crc kubenswrapper[4805]: I1203 14:09:48.830451 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-env-overrides\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Dec 03 14:09:48 crc kubenswrapper[4805]: I1203 14:09:48.830477 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-identity-cm\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-ovnkube-identity-cm\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Dec 03 14:09:48 crc kubenswrapper[4805]: I1203 14:09:48.830504 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"iptables-alerter-script\" (UniqueName: \"kubernetes.io/configmap/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-iptables-alerter-script\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Dec 03 14:09:48 crc kubenswrapper[4805]: I1203 14:09:48.830528 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls" (OuterVolumeSpecName: "metrics-tls") pod "87cf06ed-a83f-41a7-828d-70653580a8cb" (UID: "87cf06ed-a83f-41a7-828d-70653580a8cb"). InnerVolumeSpecName "metrics-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 14:09:48 crc kubenswrapper[4805]: I1203 14:09:48.830585 4805 reconciler_common.go:293] "Volume detached for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca\") on node \"crc\" DevicePath \"\"" Dec 03 14:09:48 crc kubenswrapper[4805]: I1203 14:09:48.830603 4805 reconciler_common.go:293] "Volume detached for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls\") on node \"crc\" DevicePath \"\"" Dec 03 14:09:48 crc kubenswrapper[4805]: I1203 14:09:48.830619 4805 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 03 14:09:48 crc kubenswrapper[4805]: I1203 14:09:48.830635 4805 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pcxfs\" (UniqueName: \"kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs\") on node \"crc\" DevicePath \"\"" Dec 03 14:09:48 crc kubenswrapper[4805]: I1203 14:09:48.830648 4805 reconciler_common.go:293] "Volume detached for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls\") on node \"crc\" DevicePath \"\"" Dec 03 14:09:48 crc kubenswrapper[4805]: I1203 14:09:48.830662 4805 reconciler_common.go:293] "Volume detached for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config\") on node \"crc\" DevicePath \"\"" Dec 03 14:09:48 crc kubenswrapper[4805]: I1203 14:09:48.830675 4805 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template\") on node \"crc\" DevicePath \"\"" Dec 03 14:09:48 crc kubenswrapper[4805]: I1203 14:09:48.830691 4805 reconciler_common.go:293] "Volume detached for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls\") on node \"crc\" DevicePath \"\"" Dec 03 14:09:48 crc kubenswrapper[4805]: I1203 14:09:48.830705 4805 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config\") on node \"crc\" DevicePath \"\"" Dec 03 14:09:48 crc kubenswrapper[4805]: I1203 14:09:48.830717 4805 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config\") on node \"crc\" DevicePath \"\"" Dec 03 14:09:48 crc kubenswrapper[4805]: I1203 14:09:48.830731 4805 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 14:09:48 crc kubenswrapper[4805]: I1203 14:09:48.830745 4805 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume\") on node \"crc\" DevicePath \"\"" Dec 03 14:09:48 crc kubenswrapper[4805]: I1203 14:09:48.830759 4805 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pj782\" (UniqueName: \"kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782\") on node \"crc\" DevicePath \"\"" Dec 03 14:09:48 crc kubenswrapper[4805]: I1203 14:09:48.830785 4805 reconciler_common.go:293] "Volume detached for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client\") on node \"crc\" DevicePath \"\"" Dec 03 14:09:48 crc kubenswrapper[4805]: I1203 14:09:48.830798 4805 reconciler_common.go:293] "Volume detached for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls\") on node \"crc\" DevicePath \"\"" Dec 03 14:09:48 crc kubenswrapper[4805]: I1203 14:09:48.830811 4805 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config\") on node \"crc\" DevicePath \"\"" Dec 03 14:09:48 crc kubenswrapper[4805]: I1203 14:09:48.831071 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib" (OuterVolumeSpecName: "ovnkube-script-lib") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "ovnkube-script-lib". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 14:09:48 crc kubenswrapper[4805]: I1203 14:09:48.831793 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert" (OuterVolumeSpecName: "profile-collector-cert") pod "b6312bbd-5731-4ea0-a20f-81d5a57df44a" (UID: "b6312bbd-5731-4ea0-a20f-81d5a57df44a"). InnerVolumeSpecName "profile-collector-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 14:09:48 crc kubenswrapper[4805]: I1203 14:09:48.831874 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config" (OuterVolumeSpecName: "config") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 14:09:48 crc kubenswrapper[4805]: I1203 14:09:48.832256 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth" (OuterVolumeSpecName: "stats-auth") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "stats-auth". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 14:09:48 crc kubenswrapper[4805]: I1203 14:09:48.832341 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate" (OuterVolumeSpecName: "default-certificate") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "default-certificate". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 14:09:48 crc kubenswrapper[4805]: I1203 14:09:48.833595 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 14:09:48 crc kubenswrapper[4805]: I1203 14:09:48.833771 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn" (OuterVolumeSpecName: "kube-api-access-jkwtn") pod "5b88f790-22fa-440e-b583-365168c0b23d" (UID: "5b88f790-22fa-440e-b583-365168c0b23d"). InnerVolumeSpecName "kube-api-access-jkwtn". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 14:09:48 crc kubenswrapper[4805]: I1203 14:09:48.835190 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key" (OuterVolumeSpecName: "signing-key") pod "25e176fe-21b4-4974-b1ed-c8b94f112a7f" (UID: "25e176fe-21b4-4974-b1ed-c8b94f112a7f"). InnerVolumeSpecName "signing-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 14:09:48 crc kubenswrapper[4805]: I1203 14:09:48.835517 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "0b78653f-4ff9-4508-8672-245ed9b561e3" (UID: "0b78653f-4ff9-4508-8672-245ed9b561e3"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 14:09:48 crc kubenswrapper[4805]: I1203 14:09:48.835590 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls" (OuterVolumeSpecName: "machine-approver-tls") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "machine-approver-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 14:09:48 crc kubenswrapper[4805]: I1203 14:09:48.835689 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs" (OuterVolumeSpecName: "metrics-certs") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "metrics-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 14:09:48 crc kubenswrapper[4805]: I1203 14:09:48.835869 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8" (OuterVolumeSpecName: "kube-api-access-wxkg8") pod "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" (UID: "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59"). InnerVolumeSpecName "kube-api-access-wxkg8". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 14:09:48 crc kubenswrapper[4805]: I1203 14:09:48.841058 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "01ab3dd5-8196-46d0-ad33-122e2ca51def" (UID: "01ab3dd5-8196-46d0-ad33-122e2ca51def"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 14:09:48 crc kubenswrapper[4805]: I1203 14:09:48.841286 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle" (OuterVolumeSpecName: "service-ca-bundle") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "service-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 14:09:48 crc kubenswrapper[4805]: I1203 14:09:48.841303 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52" (OuterVolumeSpecName: "kube-api-access-s4n52") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "kube-api-access-s4n52". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 14:09:48 crc kubenswrapper[4805]: I1203 14:09:48.818622 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "7539238d-5fe0-46ed-884e-1c3b566537ec" (UID: "7539238d-5fe0-46ed-884e-1c3b566537ec"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 14:09:48 crc kubenswrapper[4805]: I1203 14:09:48.818766 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca" (OuterVolumeSpecName: "image-import-ca") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "image-import-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 14:09:48 crc kubenswrapper[4805]: I1203 14:09:48.841366 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs" (OuterVolumeSpecName: "metrics-certs") pod "5b88f790-22fa-440e-b583-365168c0b23d" (UID: "5b88f790-22fa-440e-b583-365168c0b23d"). InnerVolumeSpecName "metrics-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 14:09:48 crc kubenswrapper[4805]: I1203 14:09:48.818831 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 14:09:48 crc kubenswrapper[4805]: I1203 14:09:48.819005 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88" (OuterVolumeSpecName: "kube-api-access-lzf88") pod "0b574797-001e-440a-8f4e-c0be86edad0f" (UID: "0b574797-001e-440a-8f4e-c0be86edad0f"). InnerVolumeSpecName "kube-api-access-lzf88". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 14:09:48 crc kubenswrapper[4805]: I1203 14:09:48.818449 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct" (OuterVolumeSpecName: "kube-api-access-cfbct") pod "57a731c4-ef35-47a8-b875-bfb08a7f8011" (UID: "57a731c4-ef35-47a8-b875-bfb08a7f8011"). InnerVolumeSpecName "kube-api-access-cfbct". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 14:09:48 crc kubenswrapper[4805]: I1203 14:09:48.819180 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca" (OuterVolumeSpecName: "service-ca") pod "0b78653f-4ff9-4508-8672-245ed9b561e3" (UID: "0b78653f-4ff9-4508-8672-245ed9b561e3"). InnerVolumeSpecName "service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 14:09:48 crc kubenswrapper[4805]: I1203 14:09:48.819325 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "1386a44e-36a2-460c-96d0-0359d2b6f0f5" (UID: "1386a44e-36a2-460c-96d0-0359d2b6f0f5"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 14:09:48 crc kubenswrapper[4805]: I1203 14:09:48.819549 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session" (OuterVolumeSpecName: "v4-0-config-system-session") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-session". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 14:09:48 crc kubenswrapper[4805]: I1203 14:09:48.819575 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "0b78653f-4ff9-4508-8672-245ed9b561e3" (UID: "0b78653f-4ff9-4508-8672-245ed9b561e3"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 14:09:48 crc kubenswrapper[4805]: I1203 14:09:48.819648 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert" (OuterVolumeSpecName: "apiservice-cert") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "apiservice-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 14:09:48 crc kubenswrapper[4805]: I1203 14:09:48.819659 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets" (OuterVolumeSpecName: "installation-pull-secrets") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "installation-pull-secrets". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 14:09:48 crc kubenswrapper[4805]: I1203 14:09:48.819827 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "e7e6199b-1264-4501-8953-767f51328d08" (UID: "e7e6199b-1264-4501-8953-767f51328d08"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 14:09:48 crc kubenswrapper[4805]: I1203 14:09:48.819870 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config" (OuterVolumeSpecName: "config") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 14:09:48 crc kubenswrapper[4805]: I1203 14:09:48.823316 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config" (OuterVolumeSpecName: "auth-proxy-config") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 14:09:48 crc kubenswrapper[4805]: I1203 14:09:48.824526 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca" (OuterVolumeSpecName: "client-ca") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 14:09:48 crc kubenswrapper[4805]: I1203 14:09:48.824989 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config" (OuterVolumeSpecName: "config") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 14:09:48 crc kubenswrapper[4805]: I1203 14:09:48.825251 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds" (OuterVolumeSpecName: "kube-api-access-w9rds") pod "20b0d48f-5fd6-431c-a545-e3c800c7b866" (UID: "20b0d48f-5fd6-431c-a545-e3c800c7b866"). InnerVolumeSpecName "kube-api-access-w9rds". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 14:09:48 crc kubenswrapper[4805]: I1203 14:09:48.825567 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "496e6271-fb68-4057-954e-a0d97a4afa3f" (UID: "496e6271-fb68-4057-954e-a0d97a4afa3f"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 14:09:48 crc kubenswrapper[4805]: I1203 14:09:48.826821 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates" (OuterVolumeSpecName: "registry-certificates") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "registry-certificates". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 14:09:48 crc kubenswrapper[4805]: I1203 14:09:48.827517 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config" (OuterVolumeSpecName: "config") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 14:09:48 crc kubenswrapper[4805]: I1203 14:09:48.841483 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6" (OuterVolumeSpecName: "kube-api-access-htfz6") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "kube-api-access-htfz6". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 14:09:48 crc kubenswrapper[4805]: I1203 14:09:48.841523 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v" (OuterVolumeSpecName: "kube-api-access-pjr6v") pod "49ef4625-1d3a-4a9f-b595-c2433d32326d" (UID: "49ef4625-1d3a-4a9f-b595-c2433d32326d"). InnerVolumeSpecName "kube-api-access-pjr6v". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 14:09:48 crc kubenswrapper[4805]: I1203 14:09:48.841526 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2" (OuterVolumeSpecName: "kube-api-access-jhbk2") pod "bd23aa5c-e532-4e53-bccf-e79f130c5ae8" (UID: "bd23aa5c-e532-4e53-bccf-e79f130c5ae8"). InnerVolumeSpecName "kube-api-access-jhbk2". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 14:09:48 crc kubenswrapper[4805]: I1203 14:09:48.841701 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit" (OuterVolumeSpecName: "audit") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "audit". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 14:09:48 crc kubenswrapper[4805]: I1203 14:09:48.841713 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7" (OuterVolumeSpecName: "kube-api-access-sb6h7") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "kube-api-access-sb6h7". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 14:09:48 crc kubenswrapper[4805]: I1203 14:09:48.841871 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh" (OuterVolumeSpecName: "kube-api-access-xcgwh") pod "fda69060-fa79-4696-b1a6-7980f124bf7c" (UID: "fda69060-fa79-4696-b1a6-7980f124bf7c"). InnerVolumeSpecName "kube-api-access-xcgwh". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 14:09:48 crc kubenswrapper[4805]: I1203 14:09:48.842037 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz" (OuterVolumeSpecName: "kube-api-access-6g6sz") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "kube-api-access-6g6sz". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 14:09:48 crc kubenswrapper[4805]: I1203 14:09:48.858481 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr" (OuterVolumeSpecName: "kube-api-access-249nr") pod "b6312bbd-5731-4ea0-a20f-81d5a57df44a" (UID: "b6312bbd-5731-4ea0-a20f-81d5a57df44a"). InnerVolumeSpecName "kube-api-access-249nr". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 14:09:48 crc kubenswrapper[4805]: I1203 14:09:48.858668 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config" (OuterVolumeSpecName: "config") pod "e7e6199b-1264-4501-8953-767f51328d08" (UID: "e7e6199b-1264-4501-8953-767f51328d08"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 14:09:48 crc kubenswrapper[4805]: I1203 14:09:48.859080 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config" (OuterVolumeSpecName: "console-oauth-config") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "console-oauth-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 14:09:48 crc kubenswrapper[4805]: I1203 14:09:48.859426 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" (UID: "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 14:09:48 crc kubenswrapper[4805]: I1203 14:09:48.860051 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert" (OuterVolumeSpecName: "oauth-serving-cert") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "oauth-serving-cert". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 14:09:48 crc kubenswrapper[4805]: I1203 14:09:48.860247 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config" (OuterVolumeSpecName: "multus-daemon-config") pod "4bb40260-dbaa-4fb0-84df-5e680505d512" (UID: "4bb40260-dbaa-4fb0-84df-5e680505d512"). InnerVolumeSpecName "multus-daemon-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 14:09:48 crc kubenswrapper[4805]: I1203 14:09:48.860318 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca" (OuterVolumeSpecName: "client-ca") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 14:09:48 crc kubenswrapper[4805]: I1203 14:09:48.860624 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config" (OuterVolumeSpecName: "auth-proxy-config") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 14:09:48 crc kubenswrapper[4805]: I1203 14:09:48.860651 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh" (OuterVolumeSpecName: "kube-api-access-x4zgh") pod "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" (UID: "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d"). InnerVolumeSpecName "kube-api-access-x4zgh". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 14:09:48 crc kubenswrapper[4805]: I1203 14:09:48.860770 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 14:09:48 crc kubenswrapper[4805]: I1203 14:09:48.860841 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls" (OuterVolumeSpecName: "proxy-tls") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "proxy-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 14:09:48 crc kubenswrapper[4805]: E1203 14:09:48.860961 4805 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 03 14:09:48 crc kubenswrapper[4805]: E1203 14:09:48.860999 4805 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 03 14:09:48 crc kubenswrapper[4805]: E1203 14:09:48.861011 4805 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 03 14:09:48 crc kubenswrapper[4805]: I1203 14:09:48.861047 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7" (OuterVolumeSpecName: "kube-api-access-9xfj7") pod "5225d0e4-402f-4861-b410-819f433b1803" (UID: "5225d0e4-402f-4861-b410-819f433b1803"). InnerVolumeSpecName "kube-api-access-9xfj7". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 14:09:48 crc kubenswrapper[4805]: E1203 14:09:48.861068 4805 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-12-03 14:09:49.361050976 +0000 UTC m=+19.023967889 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 03 14:09:48 crc kubenswrapper[4805]: I1203 14:09:48.861148 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 14:09:48 crc kubenswrapper[4805]: I1203 14:09:48.861142 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:48Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:48Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 03 14:09:48 crc kubenswrapper[4805]: I1203 14:09:48.861262 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client" (OuterVolumeSpecName: "etcd-client") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "etcd-client". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 14:09:48 crc kubenswrapper[4805]: I1203 14:09:48.861625 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj" (OuterVolumeSpecName: "kube-api-access-4d4hj") pod "3ab1a177-2de0-46d9-b765-d0d0649bb42e" (UID: "3ab1a177-2de0-46d9-b765-d0d0649bb42e"). InnerVolumeSpecName "kube-api-access-4d4hj". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 14:09:48 crc kubenswrapper[4805]: I1203 14:09:48.861902 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 14:09:48 crc kubenswrapper[4805]: I1203 14:09:48.818537 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates" (OuterVolumeSpecName: "available-featuregates") pod "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" (UID: "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d"). InnerVolumeSpecName "available-featuregates". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 14:09:48 crc kubenswrapper[4805]: I1203 14:09:48.861933 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert" (OuterVolumeSpecName: "srv-cert") pod "b6312bbd-5731-4ea0-a20f-81d5a57df44a" (UID: "b6312bbd-5731-4ea0-a20f-81d5a57df44a"). InnerVolumeSpecName "srv-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 14:09:48 crc kubenswrapper[4805]: I1203 14:09:48.862032 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls" (OuterVolumeSpecName: "proxy-tls") pod "fda69060-fa79-4696-b1a6-7980f124bf7c" (UID: "fda69060-fa79-4696-b1a6-7980f124bf7c"). InnerVolumeSpecName "proxy-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 14:09:48 crc kubenswrapper[4805]: I1203 14:09:48.862115 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config" (OuterVolumeSpecName: "config") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 14:09:48 crc kubenswrapper[4805]: I1203 14:09:48.862288 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities" (OuterVolumeSpecName: "utilities") pod "5225d0e4-402f-4861-b410-819f433b1803" (UID: "5225d0e4-402f-4861-b410-819f433b1803"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 14:09:48 crc kubenswrapper[4805]: I1203 14:09:48.862617 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7" (OuterVolumeSpecName: "kube-api-access-nzwt7") pod "96b93a3a-6083-4aea-8eab-fe1aa8245ad9" (UID: "96b93a3a-6083-4aea-8eab-fe1aa8245ad9"). InnerVolumeSpecName "kube-api-access-nzwt7". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 14:09:48 crc kubenswrapper[4805]: I1203 14:09:48.863180 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy" (OuterVolumeSpecName: "cni-binary-copy") pod "7bb08738-c794-4ee8-9972-3a62ca171029" (UID: "7bb08738-c794-4ee8-9972-3a62ca171029"). InnerVolumeSpecName "cni-binary-copy". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 14:09:48 crc kubenswrapper[4805]: I1203 14:09:48.863365 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl" (OuterVolumeSpecName: "kube-api-access-xcphl") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "kube-api-access-xcphl". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 14:09:48 crc kubenswrapper[4805]: I1203 14:09:48.863477 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy" (OuterVolumeSpecName: "cni-binary-copy") pod "4bb40260-dbaa-4fb0-84df-5e680505d512" (UID: "4bb40260-dbaa-4fb0-84df-5e680505d512"). InnerVolumeSpecName "cni-binary-copy". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 14:09:48 crc kubenswrapper[4805]: I1203 14:09:48.863504 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca" (OuterVolumeSpecName: "serviceca") pod "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" (UID: "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59"). InnerVolumeSpecName "serviceca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 14:09:48 crc kubenswrapper[4805]: I1203 14:09:48.866358 4805 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/kube-apiserver-crc"] Dec 03 14:09:48 crc kubenswrapper[4805]: I1203 14:09:48.866387 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config" (OuterVolumeSpecName: "config") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 14:09:48 crc kubenswrapper[4805]: I1203 14:09:48.866904 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities" (OuterVolumeSpecName: "utilities") pod "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" (UID: "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 14:09:48 crc kubenswrapper[4805]: I1203 14:09:48.867029 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf" (OuterVolumeSpecName: "kube-api-access-7c4vf") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "kube-api-access-7c4vf". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 14:09:48 crc kubenswrapper[4805]: I1203 14:09:48.867069 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 14:09:48 crc kubenswrapper[4805]: I1203 14:09:48.867084 4805 scope.go:117] "RemoveContainer" containerID="5da55f769e23ad52bac9c77da62c78e70fa34db8ed1a435e4eaf4dad3fd2b68a" Dec 03 14:09:48 crc kubenswrapper[4805]: I1203 14:09:48.867527 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m" (OuterVolumeSpecName: "kube-api-access-gf66m") pod "a0128f3a-b052-44ed-a84e-c4c8aaf17c13" (UID: "a0128f3a-b052-44ed-a84e-c4c8aaf17c13"). InnerVolumeSpecName "kube-api-access-gf66m". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 14:09:48 crc kubenswrapper[4805]: I1203 14:09:48.867627 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 14:09:48 crc kubenswrapper[4805]: I1203 14:09:48.867809 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j" (OuterVolumeSpecName: "kube-api-access-w7l8j") pod "01ab3dd5-8196-46d0-ad33-122e2ca51def" (UID: "01ab3dd5-8196-46d0-ad33-122e2ca51def"). InnerVolumeSpecName "kube-api-access-w7l8j". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 14:09:48 crc kubenswrapper[4805]: I1203 14:09:48.868204 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 14:09:48 crc kubenswrapper[4805]: I1203 14:09:48.868590 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv" (OuterVolumeSpecName: "kube-api-access-d4lsv") pod "25e176fe-21b4-4974-b1ed-c8b94f112a7f" (UID: "25e176fe-21b4-4974-b1ed-c8b94f112a7f"). InnerVolumeSpecName "kube-api-access-d4lsv". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 14:09:48 crc kubenswrapper[4805]: I1203 14:09:48.869521 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle" (OuterVolumeSpecName: "signing-cabundle") pod "25e176fe-21b4-4974-b1ed-c8b94f112a7f" (UID: "25e176fe-21b4-4974-b1ed-c8b94f112a7f"). InnerVolumeSpecName "signing-cabundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 14:09:48 crc kubenswrapper[4805]: I1203 14:09:48.870027 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config" (OuterVolumeSpecName: "config") pod "01ab3dd5-8196-46d0-ad33-122e2ca51def" (UID: "01ab3dd5-8196-46d0-ad33-122e2ca51def"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 14:09:48 crc kubenswrapper[4805]: I1203 14:09:48.870312 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config" (OuterVolumeSpecName: "ovnkube-config") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "ovnkube-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 14:09:48 crc kubenswrapper[4805]: I1203 14:09:48.872758 4805 swap_util.go:74] "error creating dir to test if tmpfs noswap is enabled. Assuming not supported" mount path="" error="stat /var/lib/kubelet/plugins/kubernetes.io/empty-dir: no such file or directory" Dec 03 14:09:48 crc kubenswrapper[4805]: I1203 14:09:48.873175 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "1d611f23-29be-4491-8495-bee1670e935f" (UID: "1d611f23-29be-4491-8495-bee1670e935f"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 14:09:48 crc kubenswrapper[4805]: I1203 14:09:48.873879 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb" (OuterVolumeSpecName: "kube-api-access-mg5zb") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "kube-api-access-mg5zb". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 14:09:48 crc kubenswrapper[4805]: I1203 14:09:48.874173 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn" (OuterVolumeSpecName: "kube-api-access-lz9wn") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "kube-api-access-lz9wn". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 14:09:48 crc kubenswrapper[4805]: I1203 14:09:48.874281 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "57a731c4-ef35-47a8-b875-bfb08a7f8011" (UID: "57a731c4-ef35-47a8-b875-bfb08a7f8011"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 14:09:48 crc kubenswrapper[4805]: E1203 14:09:48.874429 4805 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Dec 03 14:09:48 crc kubenswrapper[4805]: E1203 14:09:48.874534 4805 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-03 14:09:49.374477045 +0000 UTC m=+19.037393968 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Dec 03 14:09:48 crc kubenswrapper[4805]: I1203 14:09:48.874532 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 14:09:48 crc kubenswrapper[4805]: I1203 14:09:48.874757 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle" (OuterVolumeSpecName: "service-ca-bundle") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "service-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 14:09:48 crc kubenswrapper[4805]: I1203 14:09:48.874595 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config" (OuterVolumeSpecName: "mcc-auth-proxy-config") pod "0b574797-001e-440a-8f4e-c0be86edad0f" (UID: "0b574797-001e-440a-8f4e-c0be86edad0f"). InnerVolumeSpecName "mcc-auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 14:09:48 crc kubenswrapper[4805]: I1203 14:09:48.874946 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 14:09:48 crc kubenswrapper[4805]: I1203 14:09:48.875103 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies" (OuterVolumeSpecName: "audit-policies") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "audit-policies". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 14:09:48 crc kubenswrapper[4805]: I1203 14:09:48.875176 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx" (OuterVolumeSpecName: "kube-api-access-d6qdx") pod "87cf06ed-a83f-41a7-828d-70653580a8cb" (UID: "87cf06ed-a83f-41a7-828d-70653580a8cb"). InnerVolumeSpecName "kube-api-access-d6qdx". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 14:09:48 crc kubenswrapper[4805]: I1203 14:09:48.875210 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz" (OuterVolumeSpecName: "kube-api-access-8tdtz") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "kube-api-access-8tdtz". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 14:09:48 crc kubenswrapper[4805]: I1203 14:09:48.875285 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config" (OuterVolumeSpecName: "config") pod "496e6271-fb68-4057-954e-a0d97a4afa3f" (UID: "496e6271-fb68-4057-954e-a0d97a4afa3f"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 14:09:48 crc kubenswrapper[4805]: I1203 14:09:48.875446 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls" (OuterVolumeSpecName: "proxy-tls") pod "0b574797-001e-440a-8f4e-c0be86edad0f" (UID: "0b574797-001e-440a-8f4e-c0be86edad0f"). InnerVolumeSpecName "proxy-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 14:09:48 crc kubenswrapper[4805]: I1203 14:09:48.875641 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5" (OuterVolumeSpecName: "kube-api-access-zgdk5") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "kube-api-access-zgdk5". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 14:09:48 crc kubenswrapper[4805]: I1203 14:09:48.875659 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides" (OuterVolumeSpecName: "env-overrides") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "env-overrides". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 14:09:48 crc kubenswrapper[4805]: E1203 14:09:48.875828 4805 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 03 14:09:48 crc kubenswrapper[4805]: E1203 14:09:48.875922 4805 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-03 14:09:49.375904285 +0000 UTC m=+19.038821258 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 03 14:09:48 crc kubenswrapper[4805]: I1203 14:09:48.876065 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images" (OuterVolumeSpecName: "images") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "images". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 14:09:48 crc kubenswrapper[4805]: I1203 14:09:48.876320 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert" (OuterVolumeSpecName: "package-server-manager-serving-cert") pod "3ab1a177-2de0-46d9-b765-d0d0649bb42e" (UID: "3ab1a177-2de0-46d9-b765-d0d0649bb42e"). InnerVolumeSpecName "package-server-manager-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 14:09:48 crc kubenswrapper[4805]: I1203 14:09:48.877127 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls" (OuterVolumeSpecName: "control-plane-machine-set-operator-tls") pod "6731426b-95fe-49ff-bb5f-40441049fde2" (UID: "6731426b-95fe-49ff-bb5f-40441049fde2"). InnerVolumeSpecName "control-plane-machine-set-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 14:09:48 crc kubenswrapper[4805]: I1203 14:09:48.877979 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh" (OuterVolumeSpecName: "kube-api-access-x7zkh") pod "6731426b-95fe-49ff-bb5f-40441049fde2" (UID: "6731426b-95fe-49ff-bb5f-40441049fde2"). InnerVolumeSpecName "kube-api-access-x7zkh". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 14:09:48 crc kubenswrapper[4805]: I1203 14:09:48.878216 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-env-overrides\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Dec 03 14:09:48 crc kubenswrapper[4805]: I1203 14:09:48.879341 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh" (OuterVolumeSpecName: "kube-api-access-2w9zh") pod "4bb40260-dbaa-4fb0-84df-5e680505d512" (UID: "4bb40260-dbaa-4fb0-84df-5e680505d512"). InnerVolumeSpecName "kube-api-access-2w9zh". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 14:09:48 crc kubenswrapper[4805]: I1203 14:09:48.880118 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "1386a44e-36a2-460c-96d0-0359d2b6f0f5" (UID: "1386a44e-36a2-460c-96d0-0359d2b6f0f5"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 14:09:48 crc kubenswrapper[4805]: I1203 14:09:48.880254 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8" (OuterVolumeSpecName: "kube-api-access-6ccd8") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "kube-api-access-6ccd8". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 14:09:48 crc kubenswrapper[4805]: I1203 14:09:48.880403 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca" (OuterVolumeSpecName: "etcd-ca") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "etcd-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 14:09:48 crc kubenswrapper[4805]: I1203 14:09:48.880443 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images" (OuterVolumeSpecName: "images") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "images". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 14:09:48 crc kubenswrapper[4805]: I1203 14:09:48.880473 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config" (OuterVolumeSpecName: "config") pod "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" (UID: "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 14:09:48 crc kubenswrapper[4805]: I1203 14:09:48.881604 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs" (OuterVolumeSpecName: "tmpfs") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "tmpfs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 14:09:48 crc kubenswrapper[4805]: I1203 14:09:48.882862 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85" (OuterVolumeSpecName: "kube-api-access-x2m85") pod "cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d" (UID: "cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d"). InnerVolumeSpecName "kube-api-access-x2m85". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 14:09:48 crc kubenswrapper[4805]: I1203 14:09:48.883079 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca" (OuterVolumeSpecName: "etcd-service-ca") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "etcd-service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 14:09:48 crc kubenswrapper[4805]: I1203 14:09:48.884007 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config" (OuterVolumeSpecName: "ovnkube-config") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "ovnkube-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 14:09:48 crc kubenswrapper[4805]: I1203 14:09:48.884224 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert" (OuterVolumeSpecName: "webhook-cert") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "webhook-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 14:09:48 crc kubenswrapper[4805]: I1203 14:09:48.885298 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"iptables-alerter-script\" (UniqueName: \"kubernetes.io/configmap/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-iptables-alerter-script\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Dec 03 14:09:48 crc kubenswrapper[4805]: I1203 14:09:48.887950 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/ef543e1b-8068-4ea3-b32a-61027b32e95d-webhook-cert\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Dec 03 14:09:48 crc kubenswrapper[4805]: E1203 14:09:48.888460 4805 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 03 14:09:48 crc kubenswrapper[4805]: E1203 14:09:48.888486 4805 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 03 14:09:48 crc kubenswrapper[4805]: E1203 14:09:48.888499 4805 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 03 14:09:48 crc kubenswrapper[4805]: E1203 14:09:48.888544 4805 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-12-03 14:09:49.388530262 +0000 UTC m=+19.051447185 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 03 14:09:48 crc kubenswrapper[4805]: I1203 14:09:48.887957 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/37a5e44f-9a88-4405-be8a-b645485e7312-metrics-tls\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Dec 03 14:09:48 crc kubenswrapper[4805]: I1203 14:09:48.889275 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client" (OuterVolumeSpecName: "etcd-client") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "etcd-client". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 14:09:48 crc kubenswrapper[4805]: I1203 14:09:48.889281 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs" (OuterVolumeSpecName: "webhook-certs") pod "efdd0498-1daa-4136-9a4a-3b948c2293fc" (UID: "efdd0498-1daa-4136-9a4a-3b948c2293fc"). InnerVolumeSpecName "webhook-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 14:09:48 crc kubenswrapper[4805]: I1203 14:09:48.889518 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt" (OuterVolumeSpecName: "kube-api-access-fqsjt") pod "efdd0498-1daa-4136-9a4a-3b948c2293fc" (UID: "efdd0498-1daa-4136-9a4a-3b948c2293fc"). InnerVolumeSpecName "kube-api-access-fqsjt". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 14:09:48 crc kubenswrapper[4805]: I1203 14:09:48.889548 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics" (OuterVolumeSpecName: "marketplace-operator-metrics") pod "b6cd30de-2eeb-49a2-ab40-9167f4560ff5" (UID: "b6cd30de-2eeb-49a2-ab40-9167f4560ff5"). InnerVolumeSpecName "marketplace-operator-metrics". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 14:09:48 crc kubenswrapper[4805]: I1203 14:09:48.889551 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 14:09:48 crc kubenswrapper[4805]: I1203 14:09:48.889601 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs" (OuterVolumeSpecName: "certs") pod "5fe579f8-e8a6-4643-bce5-a661393c4dde" (UID: "5fe579f8-e8a6-4643-bce5-a661393c4dde"). InnerVolumeSpecName "certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 14:09:48 crc kubenswrapper[4805]: I1203 14:09:48.889710 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert" (OuterVolumeSpecName: "cert") pod "20b0d48f-5fd6-431c-a545-e3c800c7b866" (UID: "20b0d48f-5fd6-431c-a545-e3c800c7b866"). InnerVolumeSpecName "cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 14:09:48 crc kubenswrapper[4805]: I1203 14:09:48.889991 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist" (OuterVolumeSpecName: "cni-sysctl-allowlist") pod "7bb08738-c794-4ee8-9972-3a62ca171029" (UID: "7bb08738-c794-4ee8-9972-3a62ca171029"). InnerVolumeSpecName "cni-sysctl-allowlist". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 14:09:48 crc kubenswrapper[4805]: I1203 14:09:48.890168 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca" (OuterVolumeSpecName: "etcd-serving-ca") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "etcd-serving-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 14:09:48 crc kubenswrapper[4805]: I1203 14:09:48.890172 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:48Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:48Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 03 14:09:48 crc kubenswrapper[4805]: I1203 14:09:48.890639 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert" (OuterVolumeSpecName: "profile-collector-cert") pod "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" (UID: "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9"). InnerVolumeSpecName "profile-collector-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 14:09:48 crc kubenswrapper[4805]: I1203 14:09:48.890878 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert" (OuterVolumeSpecName: "srv-cert") pod "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" (UID: "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9"). InnerVolumeSpecName "srv-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 14:09:48 crc kubenswrapper[4805]: I1203 14:09:48.892288 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c" (OuterVolumeSpecName: "kube-api-access-tk88c") pod "7539238d-5fe0-46ed-884e-1c3b566537ec" (UID: "7539238d-5fe0-46ed-884e-1c3b566537ec"). InnerVolumeSpecName "kube-api-access-tk88c". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 14:09:48 crc kubenswrapper[4805]: I1203 14:09:48.893892 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg" (OuterVolumeSpecName: "kube-api-access-dbsvg") pod "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" (UID: "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9"). InnerVolumeSpecName "kube-api-access-dbsvg". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 14:09:48 crc kubenswrapper[4805]: I1203 14:09:48.893956 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config" (OuterVolumeSpecName: "encryption-config") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "encryption-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 14:09:48 crc kubenswrapper[4805]: I1203 14:09:48.893988 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 14:09:48 crc kubenswrapper[4805]: I1203 14:09:48.894277 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" (UID: "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 14:09:48 crc kubenswrapper[4805]: I1203 14:09:48.894453 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc" (OuterVolumeSpecName: "kube-api-access-vt5rc") pod "44663579-783b-4372-86d6-acf235a62d72" (UID: "44663579-783b-4372-86d6-acf235a62d72"). InnerVolumeSpecName "kube-api-access-vt5rc". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 14:09:48 crc kubenswrapper[4805]: I1203 14:09:48.894777 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" (UID: "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 14:09:48 crc kubenswrapper[4805]: I1203 14:09:48.894999 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4" (OuterVolumeSpecName: "kube-api-access-w4xd4") pod "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" (UID: "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b"). InnerVolumeSpecName "kube-api-access-w4xd4". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 14:09:48 crc kubenswrapper[4805]: I1203 14:09:48.895169 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp" (OuterVolumeSpecName: "kube-api-access-ngvvp") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "kube-api-access-ngvvp". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 14:09:48 crc kubenswrapper[4805]: I1203 14:09:48.895299 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config" (OuterVolumeSpecName: "config") pod "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" (UID: "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 14:09:48 crc kubenswrapper[4805]: I1203 14:09:48.895327 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "e7e6199b-1264-4501-8953-767f51328d08" (UID: "e7e6199b-1264-4501-8953-767f51328d08"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 14:09:48 crc kubenswrapper[4805]: I1203 14:09:48.895347 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config" (OuterVolumeSpecName: "config") pod "7539238d-5fe0-46ed-884e-1c3b566537ec" (UID: "7539238d-5fe0-46ed-884e-1c3b566537ec"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 14:09:48 crc kubenswrapper[4805]: I1203 14:09:48.897977 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp" (OuterVolumeSpecName: "kube-api-access-qs4fp") pod "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" (UID: "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c"). InnerVolumeSpecName "kube-api-access-qs4fp". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 14:09:48 crc kubenswrapper[4805]: I1203 14:09:48.898090 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb" (OuterVolumeSpecName: "kube-api-access-279lb") pod "7bb08738-c794-4ee8-9972-3a62ca171029" (UID: "7bb08738-c794-4ee8-9972-3a62ca171029"). InnerVolumeSpecName "kube-api-access-279lb". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 14:09:48 crc kubenswrapper[4805]: I1203 14:09:48.899069 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz" (OuterVolumeSpecName: "kube-api-access-2d4wz") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "kube-api-access-2d4wz". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 14:09:48 crc kubenswrapper[4805]: I1203 14:09:48.901353 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rczfb\" (UniqueName: \"kubernetes.io/projected/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-kube-api-access-rczfb\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Dec 03 14:09:48 crc kubenswrapper[4805]: I1203 14:09:48.903656 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rdwmf\" (UniqueName: \"kubernetes.io/projected/37a5e44f-9a88-4405-be8a-b645485e7312-kube-api-access-rdwmf\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Dec 03 14:09:48 crc kubenswrapper[4805]: I1203 14:09:48.904874 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" (UID: "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 14:09:48 crc kubenswrapper[4805]: I1203 14:09:48.906026 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 14:09:48 crc kubenswrapper[4805]: I1203 14:09:48.909186 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection" (OuterVolumeSpecName: "v4-0-config-user-template-provider-selection") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-template-provider-selection". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 14:09:48 crc kubenswrapper[4805]: I1203 14:09:48.909242 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert" (OuterVolumeSpecName: "v4-0-config-system-serving-cert") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 14:09:48 crc kubenswrapper[4805]: I1203 14:09:48.885252 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-identity-cm\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-ovnkube-identity-cm\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Dec 03 14:09:48 crc kubenswrapper[4805]: I1203 14:09:48.913748 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs" (OuterVolumeSpecName: "v4-0-config-system-router-certs") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-router-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 14:09:48 crc kubenswrapper[4805]: I1203 14:09:48.918983 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf" (OuterVolumeSpecName: "kube-api-access-v47cf") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "kube-api-access-v47cf". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 14:09:48 crc kubenswrapper[4805]: I1203 14:09:48.919832 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data" (OuterVolumeSpecName: "v4-0-config-user-idp-0-file-data") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-idp-0-file-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 14:09:48 crc kubenswrapper[4805]: I1203 14:09:48.920067 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert" (OuterVolumeSpecName: "ovn-node-metrics-cert") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "ovn-node-metrics-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 14:09:48 crc kubenswrapper[4805]: I1203 14:09:48.920134 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd" (OuterVolumeSpecName: "kube-api-access-mnrrd") pod "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" (UID: "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d"). InnerVolumeSpecName "kube-api-access-mnrrd". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 14:09:48 crc kubenswrapper[4805]: I1203 14:09:48.920364 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca" (OuterVolumeSpecName: "v4-0-config-system-service-ca") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 14:09:48 crc kubenswrapper[4805]: I1203 14:09:48.920596 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides" (OuterVolumeSpecName: "env-overrides") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "env-overrides". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 14:09:48 crc kubenswrapper[4805]: I1203 14:09:48.920616 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies" (OuterVolumeSpecName: "audit-policies") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "audit-policies". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 14:09:48 crc kubenswrapper[4805]: I1203 14:09:48.920716 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config" (OuterVolumeSpecName: "mcd-auth-proxy-config") pod "fda69060-fa79-4696-b1a6-7980f124bf7c" (UID: "fda69060-fa79-4696-b1a6-7980f124bf7c"). InnerVolumeSpecName "mcd-auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 14:09:48 crc kubenswrapper[4805]: I1203 14:09:48.921226 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:48Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:48Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 03 14:09:48 crc kubenswrapper[4805]: I1203 14:09:48.923973 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s2kz5\" (UniqueName: \"kubernetes.io/projected/ef543e1b-8068-4ea3-b32a-61027b32e95d-kube-api-access-s2kz5\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Dec 03 14:09:48 crc kubenswrapper[4805]: I1203 14:09:48.928121 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 14:09:48 crc kubenswrapper[4805]: I1203 14:09:48.928407 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls" (OuterVolumeSpecName: "image-registry-operator-tls") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "image-registry-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 14:09:48 crc kubenswrapper[4805]: I1203 14:09:48.928396 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5" (OuterVolumeSpecName: "kube-api-access-qg5z5") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "kube-api-access-qg5z5". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 14:09:48 crc kubenswrapper[4805]: I1203 14:09:48.928545 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp" (OuterVolumeSpecName: "kube-api-access-fcqwp") pod "5fe579f8-e8a6-4643-bce5-a661393c4dde" (UID: "5fe579f8-e8a6-4643-bce5-a661393c4dde"). InnerVolumeSpecName "kube-api-access-fcqwp". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 14:09:48 crc kubenswrapper[4805]: I1203 14:09:48.928534 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert" (OuterVolumeSpecName: "console-serving-cert") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "console-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 14:09:48 crc kubenswrapper[4805]: I1203 14:09:48.928641 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token" (OuterVolumeSpecName: "node-bootstrap-token") pod "5fe579f8-e8a6-4643-bce5-a661393c4dde" (UID: "5fe579f8-e8a6-4643-bce5-a661393c4dde"). InnerVolumeSpecName "node-bootstrap-token". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 14:09:48 crc kubenswrapper[4805]: I1203 14:09:48.929107 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 14:09:48 crc kubenswrapper[4805]: I1203 14:09:48.929133 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig" (OuterVolumeSpecName: "v4-0-config-system-cliconfig") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-cliconfig". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 14:09:48 crc kubenswrapper[4805]: I1203 14:09:48.930025 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login" (OuterVolumeSpecName: "v4-0-config-user-template-login") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-template-login". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 14:09:48 crc kubenswrapper[4805]: I1203 14:09:48.930325 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 14:09:48 crc kubenswrapper[4805]: I1203 14:09:48.931203 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-etc-kube\" (UniqueName: \"kubernetes.io/host-path/37a5e44f-9a88-4405-be8a-b645485e7312-host-etc-kube\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Dec 03 14:09:48 crc kubenswrapper[4805]: I1203 14:09:48.931238 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-host-slash\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Dec 03 14:09:48 crc kubenswrapper[4805]: I1203 14:09:48.935407 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-etc-kube\" (UniqueName: \"kubernetes.io/host-path/37a5e44f-9a88-4405-be8a-b645485e7312-host-etc-kube\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Dec 03 14:09:48 crc kubenswrapper[4805]: I1203 14:09:48.935697 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-host-slash\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Dec 03 14:09:48 crc kubenswrapper[4805]: I1203 14:09:48.939343 4805 reconciler_common.go:293] "Volume detached for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca\") on node \"crc\" DevicePath \"\"" Dec 03 14:09:48 crc kubenswrapper[4805]: I1203 14:09:48.939401 4805 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-249nr\" (UniqueName: \"kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr\") on node \"crc\" DevicePath \"\"" Dec 03 14:09:48 crc kubenswrapper[4805]: I1203 14:09:48.939421 4805 reconciler_common.go:293] "Volume detached for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert\") on node \"crc\" DevicePath \"\"" Dec 03 14:09:48 crc kubenswrapper[4805]: I1203 14:09:48.939441 4805 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config\") on node \"crc\" DevicePath \"\"" Dec 03 14:09:48 crc kubenswrapper[4805]: I1203 14:09:48.939453 4805 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 03 14:09:48 crc kubenswrapper[4805]: I1203 14:09:48.939467 4805 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xcgwh\" (UniqueName: \"kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh\") on node \"crc\" DevicePath \"\"" Dec 03 14:09:48 crc kubenswrapper[4805]: I1203 14:09:48.939478 4805 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fqsjt\" (UniqueName: \"kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt\") on node \"crc\" DevicePath \"\"" Dec 03 14:09:48 crc kubenswrapper[4805]: I1203 14:09:48.939494 4805 reconciler_common.go:293] "Volume detached for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 03 14:09:48 crc kubenswrapper[4805]: I1203 14:09:48.939506 4805 reconciler_common.go:293] "Volume detached for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client\") on node \"crc\" DevicePath \"\"" Dec 03 14:09:48 crc kubenswrapper[4805]: I1203 14:09:48.939518 4805 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4d4hj\" (UniqueName: \"kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj\") on node \"crc\" DevicePath \"\"" Dec 03 14:09:48 crc kubenswrapper[4805]: I1203 14:09:48.939529 4805 reconciler_common.go:293] "Volume detached for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert\") on node \"crc\" DevicePath \"\"" Dec 03 14:09:48 crc kubenswrapper[4805]: I1203 14:09:48.939544 4805 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs\") on node \"crc\" DevicePath \"\"" Dec 03 14:09:48 crc kubenswrapper[4805]: I1203 14:09:48.939557 4805 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 03 14:09:48 crc kubenswrapper[4805]: I1203 14:09:48.939569 4805 reconciler_common.go:293] "Volume detached for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls\") on node \"crc\" DevicePath \"\"" Dec 03 14:09:48 crc kubenswrapper[4805]: I1203 14:09:48.939583 4805 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 03 14:09:48 crc kubenswrapper[4805]: I1203 14:09:48.939595 4805 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection\") on node \"crc\" DevicePath \"\"" Dec 03 14:09:48 crc kubenswrapper[4805]: I1203 14:09:48.939607 4805 reconciler_common.go:293] "Volume detached for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert\") on node \"crc\" DevicePath \"\"" Dec 03 14:09:48 crc kubenswrapper[4805]: I1203 14:09:48.939621 4805 reconciler_common.go:293] "Volume detached for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client\") on node \"crc\" DevicePath \"\"" Dec 03 14:09:48 crc kubenswrapper[4805]: I1203 14:09:48.939635 4805 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 14:09:48 crc kubenswrapper[4805]: I1203 14:09:48.939647 4805 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vt5rc\" (UniqueName: \"kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc\") on node \"crc\" DevicePath \"\"" Dec 03 14:09:48 crc kubenswrapper[4805]: I1203 14:09:48.939658 4805 reconciler_common.go:293] "Volume detached for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config\") on node \"crc\" DevicePath \"\"" Dec 03 14:09:48 crc kubenswrapper[4805]: I1203 14:09:48.939672 4805 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities\") on node \"crc\" DevicePath \"\"" Dec 03 14:09:48 crc kubenswrapper[4805]: I1203 14:09:48.939688 4805 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 03 14:09:48 crc kubenswrapper[4805]: I1203 14:09:48.939700 4805 reconciler_common.go:293] "Volume detached for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides\") on node \"crc\" DevicePath \"\"" Dec 03 14:09:48 crc kubenswrapper[4805]: I1203 14:09:48.939711 4805 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities\") on node \"crc\" DevicePath \"\"" Dec 03 14:09:48 crc kubenswrapper[4805]: I1203 14:09:48.939724 4805 reconciler_common.go:293] "Volume detached for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config\") on node \"crc\" DevicePath \"\"" Dec 03 14:09:48 crc kubenswrapper[4805]: I1203 14:09:48.939739 4805 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w7l8j\" (UniqueName: \"kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j\") on node \"crc\" DevicePath \"\"" Dec 03 14:09:48 crc kubenswrapper[4805]: I1203 14:09:48.939750 4805 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w4xd4\" (UniqueName: \"kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4\") on node \"crc\" DevicePath \"\"" Dec 03 14:09:48 crc kubenswrapper[4805]: I1203 14:09:48.939762 4805 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qg5z5\" (UniqueName: \"kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5\") on node \"crc\" DevicePath \"\"" Dec 03 14:09:48 crc kubenswrapper[4805]: I1203 14:09:48.939776 4805 reconciler_common.go:293] "Volume detached for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config\") on node \"crc\" DevicePath \"\"" Dec 03 14:09:48 crc kubenswrapper[4805]: I1203 14:09:48.939787 4805 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 03 14:09:48 crc kubenswrapper[4805]: I1203 14:09:48.939798 4805 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca\") on node \"crc\" DevicePath \"\"" Dec 03 14:09:48 crc kubenswrapper[4805]: I1203 14:09:48.939809 4805 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nzwt7\" (UniqueName: \"kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7\") on node \"crc\" DevicePath \"\"" Dec 03 14:09:48 crc kubenswrapper[4805]: I1203 14:09:48.939823 4805 reconciler_common.go:293] "Volume detached for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca\") on node \"crc\" DevicePath \"\"" Dec 03 14:09:48 crc kubenswrapper[4805]: I1203 14:09:48.939837 4805 reconciler_common.go:293] "Volume detached for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca\") on node \"crc\" DevicePath \"\"" Dec 03 14:09:48 crc kubenswrapper[4805]: I1203 14:09:48.939866 4805 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-v47cf\" (UniqueName: \"kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf\") on node \"crc\" DevicePath \"\"" Dec 03 14:09:48 crc kubenswrapper[4805]: I1203 14:09:48.939877 4805 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config\") on node \"crc\" DevicePath \"\"" Dec 03 14:09:48 crc kubenswrapper[4805]: I1203 14:09:48.939891 4805 reconciler_common.go:293] "Volume detached for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle\") on node \"crc\" DevicePath \"\"" Dec 03 14:09:48 crc kubenswrapper[4805]: I1203 14:09:48.939902 4805 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config\") on node \"crc\" DevicePath \"\"" Dec 03 14:09:48 crc kubenswrapper[4805]: I1203 14:09:48.939915 4805 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-d4lsv\" (UniqueName: \"kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv\") on node \"crc\" DevicePath \"\"" Dec 03 14:09:48 crc kubenswrapper[4805]: I1203 14:09:48.939929 4805 reconciler_common.go:293] "Volume detached for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies\") on node \"crc\" DevicePath \"\"" Dec 03 14:09:48 crc kubenswrapper[4805]: I1203 14:09:48.939942 4805 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-279lb\" (UniqueName: \"kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb\") on node \"crc\" DevicePath \"\"" Dec 03 14:09:48 crc kubenswrapper[4805]: I1203 14:09:48.939953 4805 reconciler_common.go:293] "Volume detached for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca\") on node \"crc\" DevicePath \"\"" Dec 03 14:09:48 crc kubenswrapper[4805]: I1203 14:09:48.939963 4805 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token\") on node \"crc\" DevicePath \"\"" Dec 03 14:09:48 crc kubenswrapper[4805]: I1203 14:09:48.939977 4805 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mnrrd\" (UniqueName: \"kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd\") on node \"crc\" DevicePath \"\"" Dec 03 14:09:48 crc kubenswrapper[4805]: I1203 14:09:48.939987 4805 reconciler_common.go:293] "Volume detached for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token\") on node \"crc\" DevicePath \"\"" Dec 03 14:09:48 crc kubenswrapper[4805]: I1203 14:09:48.939999 4805 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dbsvg\" (UniqueName: \"kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg\") on node \"crc\" DevicePath \"\"" Dec 03 14:09:48 crc kubenswrapper[4805]: I1203 14:09:48.940010 4805 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 03 14:09:48 crc kubenswrapper[4805]: I1203 14:09:48.940025 4805 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fcqwp\" (UniqueName: \"kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp\") on node \"crc\" DevicePath \"\"" Dec 03 14:09:48 crc kubenswrapper[4805]: I1203 14:09:48.940050 4805 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config\") on node \"crc\" DevicePath \"\"" Dec 03 14:09:48 crc kubenswrapper[4805]: I1203 14:09:48.940061 4805 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token\") on node \"crc\" DevicePath \"\"" Dec 03 14:09:48 crc kubenswrapper[4805]: I1203 14:09:48.940072 4805 reconciler_common.go:293] "Volume detached for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs\") on node \"crc\" DevicePath \"\"" Dec 03 14:09:48 crc kubenswrapper[4805]: I1203 14:09:48.940086 4805 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mg5zb\" (UniqueName: \"kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb\") on node \"crc\" DevicePath \"\"" Dec 03 14:09:48 crc kubenswrapper[4805]: I1203 14:09:48.940097 4805 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca\") on node \"crc\" DevicePath \"\"" Dec 03 14:09:48 crc kubenswrapper[4805]: I1203 14:09:48.940108 4805 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jkwtn\" (UniqueName: \"kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn\") on node \"crc\" DevicePath \"\"" Dec 03 14:09:48 crc kubenswrapper[4805]: I1203 14:09:48.940150 4805 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-htfz6\" (UniqueName: \"kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6\") on node \"crc\" DevicePath \"\"" Dec 03 14:09:48 crc kubenswrapper[4805]: I1203 14:09:48.940161 4805 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access\") on node \"crc\" DevicePath \"\"" Dec 03 14:09:48 crc kubenswrapper[4805]: I1203 14:09:48.940171 4805 reconciler_common.go:293] "Volume detached for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs\") on node \"crc\" DevicePath \"\"" Dec 03 14:09:48 crc kubenswrapper[4805]: I1203 14:09:48.940181 4805 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 03 14:09:48 crc kubenswrapper[4805]: I1203 14:09:48.940195 4805 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-s4n52\" (UniqueName: \"kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52\") on node \"crc\" DevicePath \"\"" Dec 03 14:09:48 crc kubenswrapper[4805]: I1203 14:09:48.940205 4805 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jhbk2\" (UniqueName: \"kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2\") on node \"crc\" DevicePath \"\"" Dec 03 14:09:48 crc kubenswrapper[4805]: I1203 14:09:48.940217 4805 reconciler_common.go:293] "Volume detached for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Dec 03 14:09:48 crc kubenswrapper[4805]: I1203 14:09:48.940228 4805 reconciler_common.go:293] "Volume detached for volume \"cert\" (UniqueName: \"kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert\") on node \"crc\" DevicePath \"\"" Dec 03 14:09:48 crc kubenswrapper[4805]: I1203 14:09:48.940243 4805 reconciler_common.go:293] "Volume detached for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs\") on node \"crc\" DevicePath \"\"" Dec 03 14:09:48 crc kubenswrapper[4805]: I1203 14:09:48.940254 4805 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 03 14:09:48 crc kubenswrapper[4805]: I1203 14:09:48.940266 4805 reconciler_common.go:293] "Volume detached for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 03 14:09:48 crc kubenswrapper[4805]: I1203 14:09:48.940277 4805 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca\") on node \"crc\" DevicePath \"\"" Dec 03 14:09:48 crc kubenswrapper[4805]: I1203 14:09:48.940292 4805 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-d6qdx\" (UniqueName: \"kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx\") on node \"crc\" DevicePath \"\"" Dec 03 14:09:48 crc kubenswrapper[4805]: I1203 14:09:48.940303 4805 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca\") on node \"crc\" DevicePath \"\"" Dec 03 14:09:48 crc kubenswrapper[4805]: I1203 14:09:48.940314 4805 reconciler_common.go:293] "Volume detached for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 14:09:48 crc kubenswrapper[4805]: I1203 14:09:48.940328 4805 reconciler_common.go:293] "Volume detached for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies\") on node \"crc\" DevicePath \"\"" Dec 03 14:09:48 crc kubenswrapper[4805]: I1203 14:09:48.940340 4805 reconciler_common.go:293] "Volume detached for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Dec 03 14:09:48 crc kubenswrapper[4805]: I1203 14:09:48.940352 4805 reconciler_common.go:293] "Volume detached for volume \"images\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images\") on node \"crc\" DevicePath \"\"" Dec 03 14:09:48 crc kubenswrapper[4805]: I1203 14:09:48.940365 4805 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zgdk5\" (UniqueName: \"kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5\") on node \"crc\" DevicePath \"\"" Dec 03 14:09:48 crc kubenswrapper[4805]: I1203 14:09:48.940382 4805 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data\") on node \"crc\" DevicePath \"\"" Dec 03 14:09:48 crc kubenswrapper[4805]: I1203 14:09:48.940398 4805 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca\") on node \"crc\" DevicePath \"\"" Dec 03 14:09:48 crc kubenswrapper[4805]: I1203 14:09:48.940410 4805 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x7zkh\" (UniqueName: \"kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh\") on node \"crc\" DevicePath \"\"" Dec 03 14:09:48 crc kubenswrapper[4805]: I1203 14:09:48.940422 4805 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lz9wn\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn\") on node \"crc\" DevicePath \"\"" Dec 03 14:09:48 crc kubenswrapper[4805]: I1203 14:09:48.940439 4805 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2d4wz\" (UniqueName: \"kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz\") on node \"crc\" DevicePath \"\"" Dec 03 14:09:48 crc kubenswrapper[4805]: I1203 14:09:48.940451 4805 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qs4fp\" (UniqueName: \"kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp\") on node \"crc\" DevicePath \"\"" Dec 03 14:09:48 crc kubenswrapper[4805]: I1203 14:09:48.940463 4805 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2w9zh\" (UniqueName: \"kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh\") on node \"crc\" DevicePath \"\"" Dec 03 14:09:48 crc kubenswrapper[4805]: I1203 14:09:48.940477 4805 reconciler_common.go:293] "Volume detached for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls\") on node \"crc\" DevicePath \"\"" Dec 03 14:09:48 crc kubenswrapper[4805]: I1203 14:09:48.940488 4805 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8tdtz\" (UniqueName: \"kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz\") on node \"crc\" DevicePath \"\"" Dec 03 14:09:48 crc kubenswrapper[4805]: I1203 14:09:48.940498 4805 reconciler_common.go:293] "Volume detached for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides\") on node \"crc\" DevicePath \"\"" Dec 03 14:09:48 crc kubenswrapper[4805]: I1203 14:09:48.940510 4805 reconciler_common.go:293] "Volume detached for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 03 14:09:48 crc kubenswrapper[4805]: I1203 14:09:48.940524 4805 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities\") on node \"crc\" DevicePath \"\"" Dec 03 14:09:48 crc kubenswrapper[4805]: I1203 14:09:48.940536 4805 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bf2bz\" (UniqueName: \"kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz\") on node \"crc\" DevicePath \"\"" Dec 03 14:09:48 crc kubenswrapper[4805]: I1203 14:09:48.940549 4805 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cfbct\" (UniqueName: \"kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct\") on node \"crc\" DevicePath \"\"" Dec 03 14:09:48 crc kubenswrapper[4805]: I1203 14:09:48.940560 4805 reconciler_common.go:293] "Volume detached for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert\") on node \"crc\" DevicePath \"\"" Dec 03 14:09:48 crc kubenswrapper[4805]: I1203 14:09:48.940574 4805 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kfwg7\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7\") on node \"crc\" DevicePath \"\"" Dec 03 14:09:48 crc kubenswrapper[4805]: I1203 14:09:48.940589 4805 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rnphk\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk\") on node \"crc\" DevicePath \"\"" Dec 03 14:09:48 crc kubenswrapper[4805]: I1203 14:09:48.940601 4805 reconciler_common.go:293] "Volume detached for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca\") on node \"crc\" DevicePath \"\"" Dec 03 14:09:48 crc kubenswrapper[4805]: I1203 14:09:48.940611 4805 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 03 14:09:48 crc kubenswrapper[4805]: I1203 14:09:48.940626 4805 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lzf88\" (UniqueName: \"kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88\") on node \"crc\" DevicePath \"\"" Dec 03 14:09:48 crc kubenswrapper[4805]: I1203 14:09:48.940636 4805 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 03 14:09:48 crc kubenswrapper[4805]: I1203 14:09:48.940646 4805 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access\") on node \"crc\" DevicePath \"\"" Dec 03 14:09:48 crc kubenswrapper[4805]: I1203 14:09:48.940662 4805 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config\") on node \"crc\" DevicePath \"\"" Dec 03 14:09:48 crc kubenswrapper[4805]: I1203 14:09:48.940674 4805 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zkvpv\" (UniqueName: \"kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv\") on node \"crc\" DevicePath \"\"" Dec 03 14:09:48 crc kubenswrapper[4805]: I1203 14:09:48.940685 4805 reconciler_common.go:293] "Volume detached for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca\") on node \"crc\" DevicePath \"\"" Dec 03 14:09:48 crc kubenswrapper[4805]: I1203 14:09:48.940696 4805 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities\") on node \"crc\" DevicePath \"\"" Dec 03 14:09:48 crc kubenswrapper[4805]: I1203 14:09:48.940711 4805 reconciler_common.go:293] "Volume detached for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Dec 03 14:09:48 crc kubenswrapper[4805]: I1203 14:09:48.940722 4805 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session\") on node \"crc\" DevicePath \"\"" Dec 03 14:09:48 crc kubenswrapper[4805]: I1203 14:09:48.940734 4805 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config\") on node \"crc\" DevicePath \"\"" Dec 03 14:09:48 crc kubenswrapper[4805]: I1203 14:09:48.940744 4805 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config\") on node \"crc\" DevicePath \"\"" Dec 03 14:09:48 crc kubenswrapper[4805]: I1203 14:09:48.940758 4805 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 03 14:09:48 crc kubenswrapper[4805]: I1203 14:09:48.940774 4805 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config\") on node \"crc\" DevicePath \"\"" Dec 03 14:09:48 crc kubenswrapper[4805]: I1203 14:09:48.940786 4805 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error\") on node \"crc\" DevicePath \"\"" Dec 03 14:09:48 crc kubenswrapper[4805]: I1203 14:09:48.940801 4805 reconciler_common.go:293] "Volume detached for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib\") on node \"crc\" DevicePath \"\"" Dec 03 14:09:48 crc kubenswrapper[4805]: I1203 14:09:48.940812 4805 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config\") on node \"crc\" DevicePath \"\"" Dec 03 14:09:48 crc kubenswrapper[4805]: I1203 14:09:48.940822 4805 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig\") on node \"crc\" DevicePath \"\"" Dec 03 14:09:48 crc kubenswrapper[4805]: I1203 14:09:48.940839 4805 reconciler_common.go:293] "Volume detached for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls\") on node \"crc\" DevicePath \"\"" Dec 03 14:09:48 crc kubenswrapper[4805]: I1203 14:09:48.940883 4805 reconciler_common.go:293] "Volume detached for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates\") on node \"crc\" DevicePath \"\"" Dec 03 14:09:48 crc kubenswrapper[4805]: I1203 14:09:48.940894 4805 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 14:09:48 crc kubenswrapper[4805]: I1203 14:09:48.940905 4805 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access\") on node \"crc\" DevicePath \"\"" Dec 03 14:09:48 crc kubenswrapper[4805]: I1203 14:09:48.940915 4805 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 03 14:09:48 crc kubenswrapper[4805]: I1203 14:09:48.940930 4805 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x2m85\" (UniqueName: \"kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85\") on node \"crc\" DevicePath \"\"" Dec 03 14:09:48 crc kubenswrapper[4805]: I1203 14:09:48.940940 4805 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token\") on node \"crc\" DevicePath \"\"" Dec 03 14:09:48 crc kubenswrapper[4805]: I1203 14:09:48.940952 4805 reconciler_common.go:293] "Volume detached for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 14:09:48 crc kubenswrapper[4805]: I1203 14:09:48.940963 4805 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-sb6h7\" (UniqueName: \"kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7\") on node \"crc\" DevicePath \"\"" Dec 03 14:09:48 crc kubenswrapper[4805]: I1203 14:09:48.940978 4805 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6g6sz\" (UniqueName: \"kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz\") on node \"crc\" DevicePath \"\"" Dec 03 14:09:48 crc kubenswrapper[4805]: I1203 14:09:48.940990 4805 reconciler_common.go:293] "Volume detached for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config\") on node \"crc\" DevicePath \"\"" Dec 03 14:09:48 crc kubenswrapper[4805]: I1203 14:09:48.941003 4805 reconciler_common.go:293] "Volume detached for volume \"certs\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs\") on node \"crc\" DevicePath \"\"" Dec 03 14:09:48 crc kubenswrapper[4805]: I1203 14:09:48.941017 4805 reconciler_common.go:293] "Volume detached for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls\") on node \"crc\" DevicePath \"\"" Dec 03 14:09:48 crc kubenswrapper[4805]: I1203 14:09:48.941034 4805 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login\") on node \"crc\" DevicePath \"\"" Dec 03 14:09:48 crc kubenswrapper[4805]: I1203 14:09:48.941045 4805 reconciler_common.go:293] "Volume detached for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls\") on node \"crc\" DevicePath \"\"" Dec 03 14:09:48 crc kubenswrapper[4805]: I1203 14:09:48.941057 4805 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x4zgh\" (UniqueName: \"kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh\") on node \"crc\" DevicePath \"\"" Dec 03 14:09:48 crc kubenswrapper[4805]: I1203 14:09:48.941072 4805 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config\") on node \"crc\" DevicePath \"\"" Dec 03 14:09:48 crc kubenswrapper[4805]: I1203 14:09:48.941082 4805 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 03 14:09:48 crc kubenswrapper[4805]: I1203 14:09:48.941095 4805 reconciler_common.go:293] "Volume detached for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls\") on node \"crc\" DevicePath \"\"" Dec 03 14:09:48 crc kubenswrapper[4805]: I1203 14:09:48.941107 4805 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 14:09:48 crc kubenswrapper[4805]: I1203 14:09:48.941124 4805 reconciler_common.go:293] "Volume detached for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate\") on node \"crc\" DevicePath \"\"" Dec 03 14:09:48 crc kubenswrapper[4805]: I1203 14:09:48.941136 4805 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 03 14:09:48 crc kubenswrapper[4805]: I1203 14:09:48.941146 4805 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 03 14:09:48 crc kubenswrapper[4805]: I1203 14:09:48.941162 4805 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 03 14:09:48 crc kubenswrapper[4805]: I1203 14:09:48.941173 4805 reconciler_common.go:293] "Volume detached for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config\") on node \"crc\" DevicePath \"\"" Dec 03 14:09:48 crc kubenswrapper[4805]: I1203 14:09:48.941183 4805 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 03 14:09:48 crc kubenswrapper[4805]: I1203 14:09:48.941194 4805 reconciler_common.go:293] "Volume detached for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist\") on node \"crc\" DevicePath \"\"" Dec 03 14:09:48 crc kubenswrapper[4805]: I1203 14:09:48.941215 4805 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config\") on node \"crc\" DevicePath \"\"" Dec 03 14:09:48 crc kubenswrapper[4805]: I1203 14:09:48.941227 4805 reconciler_common.go:293] "Volume detached for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Dec 03 14:09:48 crc kubenswrapper[4805]: I1203 14:09:48.941238 4805 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7c4vf\" (UniqueName: \"kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf\") on node \"crc\" DevicePath \"\"" Dec 03 14:09:48 crc kubenswrapper[4805]: I1203 14:09:48.941249 4805 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gf66m\" (UniqueName: \"kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m\") on node \"crc\" DevicePath \"\"" Dec 03 14:09:48 crc kubenswrapper[4805]: I1203 14:09:48.941262 4805 reconciler_common.go:293] "Volume detached for volume \"images\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images\") on node \"crc\" DevicePath \"\"" Dec 03 14:09:48 crc kubenswrapper[4805]: I1203 14:09:48.941273 4805 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9xfj7\" (UniqueName: \"kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7\") on node \"crc\" DevicePath \"\"" Dec 03 14:09:48 crc kubenswrapper[4805]: I1203 14:09:48.941283 4805 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w9rds\" (UniqueName: \"kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds\") on node \"crc\" DevicePath \"\"" Dec 03 14:09:48 crc kubenswrapper[4805]: I1203 14:09:48.941296 4805 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 03 14:09:48 crc kubenswrapper[4805]: I1203 14:09:48.941310 4805 reconciler_common.go:293] "Volume detached for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert\") on node \"crc\" DevicePath \"\"" Dec 03 14:09:48 crc kubenswrapper[4805]: I1203 14:09:48.941321 4805 reconciler_common.go:293] "Volume detached for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca\") on node \"crc\" DevicePath \"\"" Dec 03 14:09:48 crc kubenswrapper[4805]: I1203 14:09:48.941332 4805 reconciler_common.go:293] "Volume detached for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets\") on node \"crc\" DevicePath \"\"" Dec 03 14:09:48 crc kubenswrapper[4805]: I1203 14:09:48.941557 4805 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca\") on node \"crc\" DevicePath \"\"" Dec 03 14:09:48 crc kubenswrapper[4805]: I1203 14:09:48.941576 4805 reconciler_common.go:293] "Volume detached for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates\") on node \"crc\" DevicePath \"\"" Dec 03 14:09:48 crc kubenswrapper[4805]: I1203 14:09:48.942748 4805 reconciler_common.go:293] "Volume detached for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit\") on node \"crc\" DevicePath \"\"" Dec 03 14:09:48 crc kubenswrapper[4805]: I1203 14:09:48.942776 4805 reconciler_common.go:293] "Volume detached for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls\") on node \"crc\" DevicePath \"\"" Dec 03 14:09:48 crc kubenswrapper[4805]: I1203 14:09:48.942791 4805 reconciler_common.go:293] "Volume detached for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs\") on node \"crc\" DevicePath \"\"" Dec 03 14:09:48 crc kubenswrapper[4805]: I1203 14:09:48.942802 4805 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xcphl\" (UniqueName: \"kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl\") on node \"crc\" DevicePath \"\"" Dec 03 14:09:48 crc kubenswrapper[4805]: I1203 14:09:48.942814 4805 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 14:09:48 crc kubenswrapper[4805]: I1203 14:09:48.942824 4805 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tk88c\" (UniqueName: \"kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c\") on node \"crc\" DevicePath \"\"" Dec 03 14:09:48 crc kubenswrapper[4805]: I1203 14:09:48.942834 4805 reconciler_common.go:293] "Volume detached for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config\") on node \"crc\" DevicePath \"\"" Dec 03 14:09:48 crc kubenswrapper[4805]: I1203 14:09:48.943007 4805 reconciler_common.go:293] "Volume detached for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics\") on node \"crc\" DevicePath \"\"" Dec 03 14:09:48 crc kubenswrapper[4805]: I1203 14:09:48.943020 4805 reconciler_common.go:293] "Volume detached for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert\") on node \"crc\" DevicePath \"\"" Dec 03 14:09:48 crc kubenswrapper[4805]: I1203 14:09:48.943030 4805 reconciler_common.go:293] "Volume detached for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy\") on node \"crc\" DevicePath \"\"" Dec 03 14:09:48 crc kubenswrapper[4805]: I1203 14:09:48.943041 4805 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config\") on node \"crc\" DevicePath \"\"" Dec 03 14:09:48 crc kubenswrapper[4805]: I1203 14:09:48.943051 4805 reconciler_common.go:293] "Volume detached for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth\") on node \"crc\" DevicePath \"\"" Dec 03 14:09:48 crc kubenswrapper[4805]: I1203 14:09:48.943061 4805 reconciler_common.go:293] "Volume detached for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert\") on node \"crc\" DevicePath \"\"" Dec 03 14:09:48 crc kubenswrapper[4805]: I1203 14:09:48.943071 4805 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ngvvp\" (UniqueName: \"kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp\") on node \"crc\" DevicePath \"\"" Dec 03 14:09:48 crc kubenswrapper[4805]: I1203 14:09:48.943081 4805 reconciler_common.go:293] "Volume detached for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca\") on node \"crc\" DevicePath \"\"" Dec 03 14:09:48 crc kubenswrapper[4805]: I1203 14:09:48.943092 4805 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access\") on node \"crc\" DevicePath \"\"" Dec 03 14:09:48 crc kubenswrapper[4805]: I1203 14:09:48.943101 4805 reconciler_common.go:293] "Volume detached for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles\") on node \"crc\" DevicePath \"\"" Dec 03 14:09:48 crc kubenswrapper[4805]: I1203 14:09:48.943111 4805 reconciler_common.go:293] "Volume detached for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls\") on node \"crc\" DevicePath \"\"" Dec 03 14:09:48 crc kubenswrapper[4805]: I1203 14:09:48.943121 4805 reconciler_common.go:293] "Volume detached for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy\") on node \"crc\" DevicePath \"\"" Dec 03 14:09:48 crc kubenswrapper[4805]: I1203 14:09:48.943131 4805 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca\") on node \"crc\" DevicePath \"\"" Dec 03 14:09:48 crc kubenswrapper[4805]: I1203 14:09:48.943141 4805 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6ccd8\" (UniqueName: \"kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8\") on node \"crc\" DevicePath \"\"" Dec 03 14:09:48 crc kubenswrapper[4805]: I1203 14:09:48.943151 4805 reconciler_common.go:293] "Volume detached for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key\") on node \"crc\" DevicePath \"\"" Dec 03 14:09:48 crc kubenswrapper[4805]: I1203 14:09:48.943162 4805 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 03 14:09:48 crc kubenswrapper[4805]: I1203 14:09:48.943171 4805 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wxkg8\" (UniqueName: \"kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8\") on node \"crc\" DevicePath \"\"" Dec 03 14:09:48 crc kubenswrapper[4805]: I1203 14:09:48.943182 4805 reconciler_common.go:293] "Volume detached for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert\") on node \"crc\" DevicePath \"\"" Dec 03 14:09:48 crc kubenswrapper[4805]: I1203 14:09:48.943191 4805 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pjr6v\" (UniqueName: \"kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v\") on node \"crc\" DevicePath \"\"" Dec 03 14:09:48 crc kubenswrapper[4805]: I1203 14:09:48.943355 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:48Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:48Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 03 14:09:48 crc kubenswrapper[4805]: I1203 14:09:48.948911 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "5225d0e4-402f-4861-b410-819f433b1803" (UID: "5225d0e4-402f-4861-b410-819f433b1803"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 14:09:48 crc kubenswrapper[4805]: I1203 14:09:48.951210 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Dec 03 14:09:48 crc kubenswrapper[4805]: I1203 14:09:48.953948 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted" (OuterVolumeSpecName: "ca-trust-extracted") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "ca-trust-extracted". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 14:09:48 crc kubenswrapper[4805]: I1203 14:09:48.962376 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/iptables-alerter-4ln5h" Dec 03 14:09:48 crc kubenswrapper[4805]: W1203 14:09:48.964409 4805 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod37a5e44f_9a88_4405_be8a_b645485e7312.slice/crio-598a2b5f550eec235d1eed0accc3c9e96673c32d096f59b3e0485ecd0febf3a1 WatchSource:0}: Error finding container 598a2b5f550eec235d1eed0accc3c9e96673c32d096f59b3e0485ecd0febf3a1: Status 404 returned error can't find the container with id 598a2b5f550eec235d1eed0accc3c9e96673c32d096f59b3e0485ecd0febf3a1 Dec 03 14:09:48 crc kubenswrapper[4805]: I1203 14:09:48.968512 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-node-identity/network-node-identity-vrzqb" Dec 03 14:09:49 crc kubenswrapper[4805]: I1203 14:09:49.043906 4805 reconciler_common.go:293] "Volume detached for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted\") on node \"crc\" DevicePath \"\"" Dec 03 14:09:49 crc kubenswrapper[4805]: I1203 14:09:49.043938 4805 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 03 14:09:49 crc kubenswrapper[4805]: I1203 14:09:49.347378 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 14:09:49 crc kubenswrapper[4805]: E1203 14:09:49.347592 4805 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-03 14:09:50.347565493 +0000 UTC m=+20.010482416 (durationBeforeRetry 1s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 14:09:49 crc kubenswrapper[4805]: I1203 14:09:49.448018 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 14:09:49 crc kubenswrapper[4805]: I1203 14:09:49.448069 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 14:09:49 crc kubenswrapper[4805]: I1203 14:09:49.448111 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 14:09:49 crc kubenswrapper[4805]: I1203 14:09:49.448138 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 14:09:49 crc kubenswrapper[4805]: E1203 14:09:49.448192 4805 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Dec 03 14:09:49 crc kubenswrapper[4805]: E1203 14:09:49.448241 4805 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 03 14:09:49 crc kubenswrapper[4805]: E1203 14:09:49.448254 4805 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 03 14:09:49 crc kubenswrapper[4805]: E1203 14:09:49.448264 4805 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 03 14:09:49 crc kubenswrapper[4805]: E1203 14:09:49.448272 4805 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 03 14:09:49 crc kubenswrapper[4805]: E1203 14:09:49.448280 4805 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 03 14:09:49 crc kubenswrapper[4805]: E1203 14:09:49.448280 4805 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 03 14:09:49 crc kubenswrapper[4805]: E1203 14:09:49.448282 4805 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-03 14:09:50.448262926 +0000 UTC m=+20.111179909 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Dec 03 14:09:49 crc kubenswrapper[4805]: E1203 14:09:49.448246 4805 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 03 14:09:49 crc kubenswrapper[4805]: E1203 14:09:49.448320 4805 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-12-03 14:09:50.448307267 +0000 UTC m=+20.111224200 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 03 14:09:49 crc kubenswrapper[4805]: E1203 14:09:49.448336 4805 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-12-03 14:09:50.448328858 +0000 UTC m=+20.111245881 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 03 14:09:49 crc kubenswrapper[4805]: E1203 14:09:49.448354 4805 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-03 14:09:50.448348139 +0000 UTC m=+20.111265152 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 03 14:09:49 crc kubenswrapper[4805]: I1203 14:09:49.524522 4805 csr.go:261] certificate signing request csr-g2ngt is approved, waiting to be issued Dec 03 14:09:49 crc kubenswrapper[4805]: I1203 14:09:49.531898 4805 csr.go:257] certificate signing request csr-g2ngt is issued Dec 03 14:09:49 crc kubenswrapper[4805]: I1203 14:09:49.785291 4805 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/0.log" Dec 03 14:09:49 crc kubenswrapper[4805]: I1203 14:09:49.786664 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"e64730957bbbb2feb4d506e0ed9731a0594bf66fa1e8fcdb8ec944ded6308eed"} Dec 03 14:09:49 crc kubenswrapper[4805]: I1203 14:09:49.786993 4805 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 03 14:09:49 crc kubenswrapper[4805]: I1203 14:09:49.788443 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" event={"ID":"ef543e1b-8068-4ea3-b32a-61027b32e95d","Type":"ContainerStarted","Data":"541ace7ccbb66dbd191c3ca8c1e4479e3edbf80dfc67326bea8aaee81fdcfc3c"} Dec 03 14:09:49 crc kubenswrapper[4805]: I1203 14:09:49.788469 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" event={"ID":"ef543e1b-8068-4ea3-b32a-61027b32e95d","Type":"ContainerStarted","Data":"821e3e9a902aacecf6c41e8b8d8aa1ac951ce2fbb6505b7148b16348a3754e55"} Dec 03 14:09:49 crc kubenswrapper[4805]: I1203 14:09:49.788482 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" event={"ID":"ef543e1b-8068-4ea3-b32a-61027b32e95d","Type":"ContainerStarted","Data":"ef6042257f8c30b6052faf65e8ca85f2af84c34c414ca1b3c4bb052f188f5732"} Dec 03 14:09:49 crc kubenswrapper[4805]: I1203 14:09:49.789479 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" event={"ID":"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49","Type":"ContainerStarted","Data":"79191558252dac7d03aed3e41d9d284b466dbc010629f1a3121e2d7ad23b02b8"} Dec 03 14:09:49 crc kubenswrapper[4805]: I1203 14:09:49.790621 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" event={"ID":"37a5e44f-9a88-4405-be8a-b645485e7312","Type":"ContainerStarted","Data":"38339331a3209d6787d1def81b72f4a8514bf91080a139cad6d8ac5cc6bef10e"} Dec 03 14:09:49 crc kubenswrapper[4805]: I1203 14:09:49.790653 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" event={"ID":"37a5e44f-9a88-4405-be8a-b645485e7312","Type":"ContainerStarted","Data":"598a2b5f550eec235d1eed0accc3c9e96673c32d096f59b3e0485ecd0febf3a1"} Dec 03 14:09:49 crc kubenswrapper[4805]: I1203 14:09:49.804881 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:48Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:48Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T14:09:49Z is after 2025-08-24T17:21:41Z" Dec 03 14:09:49 crc kubenswrapper[4805]: I1203 14:09:49.814157 4805 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 03 14:09:49 crc kubenswrapper[4805]: I1203 14:09:49.817910 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:48Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:48Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T14:09:49Z is after 2025-08-24T17:21:41Z" Dec 03 14:09:49 crc kubenswrapper[4805]: I1203 14:09:49.831539 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:48Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:48Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T14:09:49Z is after 2025-08-24T17:21:41Z" Dec 03 14:09:49 crc kubenswrapper[4805]: I1203 14:09:49.850756 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:48Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:48Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T14:09:49Z is after 2025-08-24T17:21:41Z" Dec 03 14:09:49 crc kubenswrapper[4805]: I1203 14:09:49.869040 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ced8291d-0ce5-4e74-9eaf-2bf033774039\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:30Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:30Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3e5a34636f0a8c4d9179074596d28bc2eb7a504573398c596a637defd829907a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://29cd95be3f964415943c873e7914ffcea4abe50dfdf17f8a9cd03fc38df35981\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://951173b3d3d581275a6f6e43209b5e0611a701ca7a10f81504d723e24a2d21a3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e64730957bbbb2feb4d506e0ed9731a0594bf66fa1e8fcdb8ec944ded6308eed\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5da55f769e23ad52bac9c77da62c78e70fa34db8ed1a435e4eaf4dad3fd2b68a\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-03T14:09:48Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1203 14:09:42.992179 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1203 14:09:42.993233 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1830926909/tls.crt::/tmp/serving-cert-1830926909/tls.key\\\\\\\"\\\\nI1203 14:09:48.622403 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1203 14:09:48.626476 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1203 14:09:48.626527 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1203 14:09:48.626574 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1203 14:09:48.626582 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1203 14:09:48.633990 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1203 14:09:48.634024 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 14:09:48.634029 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 14:09:48.634034 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1203 14:09:48.634038 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1203 14:09:48.634042 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1203 14:09:48.634045 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1203 14:09:48.634344 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1203 14:09:48.644497 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T14:09:32Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bd92a160eb530c5687ebd7cd5b83b41820a2b808a0fd9514e5ace9d237e1bd26\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:32Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://af9c940135012c629447528c61743ed1e58708ab57ce76604c6da0479aa459cf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://af9c940135012c629447528c61743ed1e58708ab57ce76604c6da0479aa459cf\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T14:09:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T14:09:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T14:09:30Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T14:09:49Z is after 2025-08-24T17:21:41Z" Dec 03 14:09:49 crc kubenswrapper[4805]: I1203 14:09:49.904887 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:48Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:48Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T14:09:49Z is after 2025-08-24T17:21:41Z" Dec 03 14:09:49 crc kubenswrapper[4805]: I1203 14:09:49.919763 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:48Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:48Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T14:09:49Z is after 2025-08-24T17:21:41Z" Dec 03 14:09:49 crc kubenswrapper[4805]: I1203 14:09:49.940205 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:48Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:48Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T14:09:49Z is after 2025-08-24T17:21:41Z" Dec 03 14:09:49 crc kubenswrapper[4805]: I1203 14:09:49.963724 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:48Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:48Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T14:09:49Z is after 2025-08-24T17:21:41Z" Dec 03 14:09:49 crc kubenswrapper[4805]: I1203 14:09:49.978081 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:49Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:49Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://541ace7ccbb66dbd191c3ca8c1e4479e3edbf80dfc67326bea8aaee81fdcfc3c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://821e3e9a902aacecf6c41e8b8d8aa1ac951ce2fbb6505b7148b16348a3754e55\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T14:09:49Z is after 2025-08-24T17:21:41Z" Dec 03 14:09:49 crc kubenswrapper[4805]: I1203 14:09:49.997234 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ced8291d-0ce5-4e74-9eaf-2bf033774039\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:30Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:30Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3e5a34636f0a8c4d9179074596d28bc2eb7a504573398c596a637defd829907a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://29cd95be3f964415943c873e7914ffcea4abe50dfdf17f8a9cd03fc38df35981\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://951173b3d3d581275a6f6e43209b5e0611a701ca7a10f81504d723e24a2d21a3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e64730957bbbb2feb4d506e0ed9731a0594bf66fa1e8fcdb8ec944ded6308eed\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5da55f769e23ad52bac9c77da62c78e70fa34db8ed1a435e4eaf4dad3fd2b68a\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-03T14:09:48Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1203 14:09:42.992179 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1203 14:09:42.993233 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1830926909/tls.crt::/tmp/serving-cert-1830926909/tls.key\\\\\\\"\\\\nI1203 14:09:48.622403 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1203 14:09:48.626476 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1203 14:09:48.626527 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1203 14:09:48.626574 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1203 14:09:48.626582 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1203 14:09:48.633990 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1203 14:09:48.634024 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 14:09:48.634029 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 14:09:48.634034 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1203 14:09:48.634038 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1203 14:09:48.634042 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1203 14:09:48.634045 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1203 14:09:48.634344 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1203 14:09:48.644497 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T14:09:32Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bd92a160eb530c5687ebd7cd5b83b41820a2b808a0fd9514e5ace9d237e1bd26\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:32Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://af9c940135012c629447528c61743ed1e58708ab57ce76604c6da0479aa459cf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://af9c940135012c629447528c61743ed1e58708ab57ce76604c6da0479aa459cf\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T14:09:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T14:09:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T14:09:30Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T14:09:49Z is after 2025-08-24T17:21:41Z" Dec 03 14:09:50 crc kubenswrapper[4805]: I1203 14:09:50.009115 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:49Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:49Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://38339331a3209d6787d1def81b72f4a8514bf91080a139cad6d8ac5cc6bef10e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T14:09:50Z is after 2025-08-24T17:21:41Z" Dec 03 14:09:50 crc kubenswrapper[4805]: I1203 14:09:50.032453 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:48Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:48Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T14:09:50Z is after 2025-08-24T17:21:41Z" Dec 03 14:09:50 crc kubenswrapper[4805]: I1203 14:09:50.045730 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:48Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:48Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T14:09:50Z is after 2025-08-24T17:21:41Z" Dec 03 14:09:50 crc kubenswrapper[4805]: I1203 14:09:50.355749 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 14:09:50 crc kubenswrapper[4805]: E1203 14:09:50.355992 4805 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-03 14:09:52.355960611 +0000 UTC m=+22.018877534 (durationBeforeRetry 2s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 14:09:50 crc kubenswrapper[4805]: I1203 14:09:50.432360 4805 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/multus-k4mhs"] Dec 03 14:09:50 crc kubenswrapper[4805]: I1203 14:09:50.432624 4805 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-dns/node-resolver-pdzgk"] Dec 03 14:09:50 crc kubenswrapper[4805]: I1203 14:09:50.432708 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-k4mhs" Dec 03 14:09:50 crc kubenswrapper[4805]: I1203 14:09:50.432760 4805 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-daemon-gskh4"] Dec 03 14:09:50 crc kubenswrapper[4805]: I1203 14:09:50.433010 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-daemon-gskh4" Dec 03 14:09:50 crc kubenswrapper[4805]: I1203 14:09:50.433096 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/node-resolver-pdzgk" Dec 03 14:09:50 crc kubenswrapper[4805]: I1203 14:09:50.434208 4805 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/multus-additional-cni-plugins-bpk49"] Dec 03 14:09:50 crc kubenswrapper[4805]: I1203 14:09:50.435000 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-additional-cni-plugins-bpk49" Dec 03 14:09:50 crc kubenswrapper[4805]: I1203 14:09:50.438461 4805 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"kube-root-ca.crt" Dec 03 14:09:50 crc kubenswrapper[4805]: I1203 14:09:50.438597 4805 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"openshift-service-ca.crt" Dec 03 14:09:50 crc kubenswrapper[4805]: I1203 14:09:50.438614 4805 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"cni-copy-resources" Dec 03 14:09:50 crc kubenswrapper[4805]: I1203 14:09:50.438673 4805 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"openshift-service-ca.crt" Dec 03 14:09:50 crc kubenswrapper[4805]: I1203 14:09:50.438785 4805 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"openshift-service-ca.crt" Dec 03 14:09:50 crc kubenswrapper[4805]: I1203 14:09:50.438810 4805 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"kube-rbac-proxy" Dec 03 14:09:50 crc kubenswrapper[4805]: I1203 14:09:50.438829 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"proxy-tls" Dec 03 14:09:50 crc kubenswrapper[4805]: I1203 14:09:50.438996 4805 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"kube-root-ca.crt" Dec 03 14:09:50 crc kubenswrapper[4805]: I1203 14:09:50.439151 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-ancillary-tools-dockercfg-vnmsz" Dec 03 14:09:50 crc kubenswrapper[4805]: I1203 14:09:50.439201 4805 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"default-cni-sysctl-allowlist" Dec 03 14:09:50 crc kubenswrapper[4805]: I1203 14:09:50.439261 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"node-resolver-dockercfg-kz9s7" Dec 03 14:09:50 crc kubenswrapper[4805]: I1203 14:09:50.439378 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-daemon-dockercfg-r5tcq" Dec 03 14:09:50 crc kubenswrapper[4805]: I1203 14:09:50.439444 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"default-dockercfg-2q5b6" Dec 03 14:09:50 crc kubenswrapper[4805]: I1203 14:09:50.439491 4805 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"kube-root-ca.crt" Dec 03 14:09:50 crc kubenswrapper[4805]: I1203 14:09:50.439593 4805 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"multus-daemon-config" Dec 03 14:09:50 crc kubenswrapper[4805]: I1203 14:09:50.452073 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:48Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:48Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T14:09:50Z is after 2025-08-24T17:21:41Z" Dec 03 14:09:50 crc kubenswrapper[4805]: I1203 14:09:50.456498 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ggn4n\" (UniqueName: \"kubernetes.io/projected/860769de-6a5b-4cb9-a752-264e7e2e785f-kube-api-access-ggn4n\") pod \"node-resolver-pdzgk\" (UID: \"860769de-6a5b-4cb9-a752-264e7e2e785f\") " pod="openshift-dns/node-resolver-pdzgk" Dec 03 14:09:50 crc kubenswrapper[4805]: I1203 14:09:50.456551 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 14:09:50 crc kubenswrapper[4805]: I1203 14:09:50.456580 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-bin\" (UniqueName: \"kubernetes.io/host-path/c09c1cdc-f461-45c9-8444-5b8764bccabc-host-var-lib-cni-bin\") pod \"multus-k4mhs\" (UID: \"c09c1cdc-f461-45c9-8444-5b8764bccabc\") " pod="openshift-multus/multus-k4mhs" Dec 03 14:09:50 crc kubenswrapper[4805]: I1203 14:09:50.456604 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/c09c1cdc-f461-45c9-8444-5b8764bccabc-multus-daemon-config\") pod \"multus-k4mhs\" (UID: \"c09c1cdc-f461-45c9-8444-5b8764bccabc\") " pod="openshift-multus/multus-k4mhs" Dec 03 14:09:50 crc kubenswrapper[4805]: I1203 14:09:50.456628 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/6098937f-e3f6-45e8-a647-4994a79cd711-mcd-auth-proxy-config\") pod \"machine-config-daemon-gskh4\" (UID: \"6098937f-e3f6-45e8-a647-4994a79cd711\") " pod="openshift-machine-config-operator/machine-config-daemon-gskh4" Dec 03 14:09:50 crc kubenswrapper[4805]: I1203 14:09:50.456652 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/de00a81f-82e1-4223-9614-f98606b9a8ef-system-cni-dir\") pod \"multus-additional-cni-plugins-bpk49\" (UID: \"de00a81f-82e1-4223-9614-f98606b9a8ef\") " pod="openshift-multus/multus-additional-cni-plugins-bpk49" Dec 03 14:09:50 crc kubenswrapper[4805]: E1203 14:09:50.456664 4805 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Dec 03 14:09:50 crc kubenswrapper[4805]: I1203 14:09:50.456727 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/c09c1cdc-f461-45c9-8444-5b8764bccabc-host-run-netns\") pod \"multus-k4mhs\" (UID: \"c09c1cdc-f461-45c9-8444-5b8764bccabc\") " pod="openshift-multus/multus-k4mhs" Dec 03 14:09:50 crc kubenswrapper[4805]: E1203 14:09:50.456780 4805 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-03 14:09:52.456757107 +0000 UTC m=+22.119674130 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Dec 03 14:09:50 crc kubenswrapper[4805]: I1203 14:09:50.456820 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 14:09:50 crc kubenswrapper[4805]: I1203 14:09:50.456865 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-socket-dir-parent\" (UniqueName: \"kubernetes.io/host-path/c09c1cdc-f461-45c9-8444-5b8764bccabc-multus-socket-dir-parent\") pod \"multus-k4mhs\" (UID: \"c09c1cdc-f461-45c9-8444-5b8764bccabc\") " pod="openshift-multus/multus-k4mhs" Dec 03 14:09:50 crc kubenswrapper[4805]: I1203 14:09:50.456892 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tuning-conf-dir\" (UniqueName: \"kubernetes.io/host-path/de00a81f-82e1-4223-9614-f98606b9a8ef-tuning-conf-dir\") pod \"multus-additional-cni-plugins-bpk49\" (UID: \"de00a81f-82e1-4223-9614-f98606b9a8ef\") " pod="openshift-multus/multus-additional-cni-plugins-bpk49" Dec 03 14:09:50 crc kubenswrapper[4805]: I1203 14:09:50.456938 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/c09c1cdc-f461-45c9-8444-5b8764bccabc-system-cni-dir\") pod \"multus-k4mhs\" (UID: \"c09c1cdc-f461-45c9-8444-5b8764bccabc\") " pod="openshift-multus/multus-k4mhs" Dec 03 14:09:50 crc kubenswrapper[4805]: E1203 14:09:50.456940 4805 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 03 14:09:50 crc kubenswrapper[4805]: I1203 14:09:50.456962 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-multus\" (UniqueName: \"kubernetes.io/host-path/c09c1cdc-f461-45c9-8444-5b8764bccabc-host-var-lib-cni-multus\") pod \"multus-k4mhs\" (UID: \"c09c1cdc-f461-45c9-8444-5b8764bccabc\") " pod="openshift-multus/multus-k4mhs" Dec 03 14:09:50 crc kubenswrapper[4805]: I1203 14:09:50.456994 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nwrw9\" (UniqueName: \"kubernetes.io/projected/6098937f-e3f6-45e8-a647-4994a79cd711-kube-api-access-nwrw9\") pod \"machine-config-daemon-gskh4\" (UID: \"6098937f-e3f6-45e8-a647-4994a79cd711\") " pod="openshift-machine-config-operator/machine-config-daemon-gskh4" Dec 03 14:09:50 crc kubenswrapper[4805]: E1203 14:09:50.457023 4805 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-03 14:09:52.457002134 +0000 UTC m=+22.119919147 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 03 14:09:50 crc kubenswrapper[4805]: I1203 14:09:50.457052 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"hostroot\" (UniqueName: \"kubernetes.io/host-path/c09c1cdc-f461-45c9-8444-5b8764bccabc-hostroot\") pod \"multus-k4mhs\" (UID: \"c09c1cdc-f461-45c9-8444-5b8764bccabc\") " pod="openshift-multus/multus-k4mhs" Dec 03 14:09:50 crc kubenswrapper[4805]: I1203 14:09:50.457091 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-25d7t\" (UniqueName: \"kubernetes.io/projected/de00a81f-82e1-4223-9614-f98606b9a8ef-kube-api-access-25d7t\") pod \"multus-additional-cni-plugins-bpk49\" (UID: \"de00a81f-82e1-4223-9614-f98606b9a8ef\") " pod="openshift-multus/multus-additional-cni-plugins-bpk49" Dec 03 14:09:50 crc kubenswrapper[4805]: I1203 14:09:50.457127 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/6098937f-e3f6-45e8-a647-4994a79cd711-proxy-tls\") pod \"machine-config-daemon-gskh4\" (UID: \"6098937f-e3f6-45e8-a647-4994a79cd711\") " pod="openshift-machine-config-operator/machine-config-daemon-gskh4" Dec 03 14:09:50 crc kubenswrapper[4805]: I1203 14:09:50.457156 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/c09c1cdc-f461-45c9-8444-5b8764bccabc-os-release\") pod \"multus-k4mhs\" (UID: \"c09c1cdc-f461-45c9-8444-5b8764bccabc\") " pod="openshift-multus/multus-k4mhs" Dec 03 14:09:50 crc kubenswrapper[4805]: I1203 14:09:50.457185 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"hosts-file\" (UniqueName: \"kubernetes.io/host-path/860769de-6a5b-4cb9-a752-264e7e2e785f-hosts-file\") pod \"node-resolver-pdzgk\" (UID: \"860769de-6a5b-4cb9-a752-264e7e2e785f\") " pod="openshift-dns/node-resolver-pdzgk" Dec 03 14:09:50 crc kubenswrapper[4805]: I1203 14:09:50.457253 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-cni-dir\" (UniqueName: \"kubernetes.io/host-path/c09c1cdc-f461-45c9-8444-5b8764bccabc-multus-cni-dir\") pod \"multus-k4mhs\" (UID: \"c09c1cdc-f461-45c9-8444-5b8764bccabc\") " pod="openshift-multus/multus-k4mhs" Dec 03 14:09:50 crc kubenswrapper[4805]: I1203 14:09:50.457287 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/de00a81f-82e1-4223-9614-f98606b9a8ef-cni-sysctl-allowlist\") pod \"multus-additional-cni-plugins-bpk49\" (UID: \"de00a81f-82e1-4223-9614-f98606b9a8ef\") " pod="openshift-multus/multus-additional-cni-plugins-bpk49" Dec 03 14:09:50 crc kubenswrapper[4805]: I1203 14:09:50.457329 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 14:09:50 crc kubenswrapper[4805]: I1203 14:09:50.457361 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rootfs\" (UniqueName: \"kubernetes.io/host-path/6098937f-e3f6-45e8-a647-4994a79cd711-rootfs\") pod \"machine-config-daemon-gskh4\" (UID: \"6098937f-e3f6-45e8-a647-4994a79cd711\") " pod="openshift-machine-config-operator/machine-config-daemon-gskh4" Dec 03 14:09:50 crc kubenswrapper[4805]: I1203 14:09:50.457399 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/de00a81f-82e1-4223-9614-f98606b9a8ef-cnibin\") pod \"multus-additional-cni-plugins-bpk49\" (UID: \"de00a81f-82e1-4223-9614-f98606b9a8ef\") " pod="openshift-multus/multus-additional-cni-plugins-bpk49" Dec 03 14:09:50 crc kubenswrapper[4805]: I1203 14:09:50.457428 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/de00a81f-82e1-4223-9614-f98606b9a8ef-os-release\") pod \"multus-additional-cni-plugins-bpk49\" (UID: \"de00a81f-82e1-4223-9614-f98606b9a8ef\") " pod="openshift-multus/multus-additional-cni-plugins-bpk49" Dec 03 14:09:50 crc kubenswrapper[4805]: E1203 14:09:50.457440 4805 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 03 14:09:50 crc kubenswrapper[4805]: E1203 14:09:50.457461 4805 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 03 14:09:50 crc kubenswrapper[4805]: I1203 14:09:50.457463 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/c09c1cdc-f461-45c9-8444-5b8764bccabc-cni-binary-copy\") pod \"multus-k4mhs\" (UID: \"c09c1cdc-f461-45c9-8444-5b8764bccabc\") " pod="openshift-multus/multus-k4mhs" Dec 03 14:09:50 crc kubenswrapper[4805]: E1203 14:09:50.457474 4805 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 03 14:09:50 crc kubenswrapper[4805]: I1203 14:09:50.457495 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qdcd7\" (UniqueName: \"kubernetes.io/projected/c09c1cdc-f461-45c9-8444-5b8764bccabc-kube-api-access-qdcd7\") pod \"multus-k4mhs\" (UID: \"c09c1cdc-f461-45c9-8444-5b8764bccabc\") " pod="openshift-multus/multus-k4mhs" Dec 03 14:09:50 crc kubenswrapper[4805]: E1203 14:09:50.457518 4805 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-12-03 14:09:52.457501457 +0000 UTC m=+22.120418470 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 03 14:09:50 crc kubenswrapper[4805]: I1203 14:09:50.457542 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/de00a81f-82e1-4223-9614-f98606b9a8ef-cni-binary-copy\") pod \"multus-additional-cni-plugins-bpk49\" (UID: \"de00a81f-82e1-4223-9614-f98606b9a8ef\") " pod="openshift-multus/multus-additional-cni-plugins-bpk49" Dec 03 14:09:50 crc kubenswrapper[4805]: I1203 14:09:50.457590 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-k8s-cni-cncf-io\" (UniqueName: \"kubernetes.io/host-path/c09c1cdc-f461-45c9-8444-5b8764bccabc-host-run-k8s-cni-cncf-io\") pod \"multus-k4mhs\" (UID: \"c09c1cdc-f461-45c9-8444-5b8764bccabc\") " pod="openshift-multus/multus-k4mhs" Dec 03 14:09:50 crc kubenswrapper[4805]: I1203 14:09:50.457617 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/c09c1cdc-f461-45c9-8444-5b8764bccabc-host-var-lib-kubelet\") pod \"multus-k4mhs\" (UID: \"c09c1cdc-f461-45c9-8444-5b8764bccabc\") " pod="openshift-multus/multus-k4mhs" Dec 03 14:09:50 crc kubenswrapper[4805]: I1203 14:09:50.457637 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-multus-certs\" (UniqueName: \"kubernetes.io/host-path/c09c1cdc-f461-45c9-8444-5b8764bccabc-host-run-multus-certs\") pod \"multus-k4mhs\" (UID: \"c09c1cdc-f461-45c9-8444-5b8764bccabc\") " pod="openshift-multus/multus-k4mhs" Dec 03 14:09:50 crc kubenswrapper[4805]: I1203 14:09:50.457670 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 14:09:50 crc kubenswrapper[4805]: I1203 14:09:50.457708 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/c09c1cdc-f461-45c9-8444-5b8764bccabc-cnibin\") pod \"multus-k4mhs\" (UID: \"c09c1cdc-f461-45c9-8444-5b8764bccabc\") " pod="openshift-multus/multus-k4mhs" Dec 03 14:09:50 crc kubenswrapper[4805]: I1203 14:09:50.457781 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-conf-dir\" (UniqueName: \"kubernetes.io/host-path/c09c1cdc-f461-45c9-8444-5b8764bccabc-multus-conf-dir\") pod \"multus-k4mhs\" (UID: \"c09c1cdc-f461-45c9-8444-5b8764bccabc\") " pod="openshift-multus/multus-k4mhs" Dec 03 14:09:50 crc kubenswrapper[4805]: I1203 14:09:50.457810 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-kubernetes\" (UniqueName: \"kubernetes.io/host-path/c09c1cdc-f461-45c9-8444-5b8764bccabc-etc-kubernetes\") pod \"multus-k4mhs\" (UID: \"c09c1cdc-f461-45c9-8444-5b8764bccabc\") " pod="openshift-multus/multus-k4mhs" Dec 03 14:09:50 crc kubenswrapper[4805]: E1203 14:09:50.457888 4805 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 03 14:09:50 crc kubenswrapper[4805]: E1203 14:09:50.457914 4805 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 03 14:09:50 crc kubenswrapper[4805]: E1203 14:09:50.457932 4805 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 03 14:09:50 crc kubenswrapper[4805]: E1203 14:09:50.457981 4805 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-12-03 14:09:52.45796479 +0000 UTC m=+22.120881803 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 03 14:09:50 crc kubenswrapper[4805]: I1203 14:09:50.471782 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-k4mhs" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c09c1cdc-f461-45c9-8444-5b8764bccabc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:50Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:50Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:50Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qdcd7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T14:09:50Z\\\"}}\" for pod \"openshift-multus\"/\"multus-k4mhs\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T14:09:50Z is after 2025-08-24T17:21:41Z" Dec 03 14:09:50 crc kubenswrapper[4805]: I1203 14:09:50.488684 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:48Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:48Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T14:09:50Z is after 2025-08-24T17:21:41Z" Dec 03 14:09:50 crc kubenswrapper[4805]: I1203 14:09:50.516903 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:49Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:49Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://541ace7ccbb66dbd191c3ca8c1e4479e3edbf80dfc67326bea8aaee81fdcfc3c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://821e3e9a902aacecf6c41e8b8d8aa1ac951ce2fbb6505b7148b16348a3754e55\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T14:09:50Z is after 2025-08-24T17:21:41Z" Dec 03 14:09:50 crc kubenswrapper[4805]: I1203 14:09:50.533246 4805 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Certificate expiration is 2026-12-03 14:04:49 +0000 UTC, rotation deadline is 2026-10-11 05:45:01.94572616 +0000 UTC Dec 03 14:09:50 crc kubenswrapper[4805]: I1203 14:09:50.533298 4805 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Waiting 7479h35m11.412434486s for next certificate rotation Dec 03 14:09:50 crc kubenswrapper[4805]: I1203 14:09:50.547358 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-gskh4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6098937f-e3f6-45e8-a647-4994a79cd711\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:50Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:50Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:50Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nwrw9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nwrw9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T14:09:50Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-gskh4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T14:09:50Z is after 2025-08-24T17:21:41Z" Dec 03 14:09:50 crc kubenswrapper[4805]: I1203 14:09:50.550484 4805 transport.go:147] "Certificate rotation detected, shutting down client connections to start using new credentials" Dec 03 14:09:50 crc kubenswrapper[4805]: W1203 14:09:50.550649 4805 reflector.go:484] object-"openshift-machine-config-operator"/"machine-config-daemon-dockercfg-r5tcq": watch of *v1.Secret ended with: very short watch: object-"openshift-machine-config-operator"/"machine-config-daemon-dockercfg-r5tcq": Unexpected watch close - watch lasted less than a second and no items received Dec 03 14:09:50 crc kubenswrapper[4805]: W1203 14:09:50.550677 4805 reflector.go:484] object-"openshift-machine-config-operator"/"kube-rbac-proxy": watch of *v1.ConfigMap ended with: very short watch: object-"openshift-machine-config-operator"/"kube-rbac-proxy": Unexpected watch close - watch lasted less than a second and no items received Dec 03 14:09:50 crc kubenswrapper[4805]: W1203 14:09:50.550712 4805 reflector.go:484] object-"openshift-multus"/"multus-daemon-config": watch of *v1.ConfigMap ended with: very short watch: object-"openshift-multus"/"multus-daemon-config": Unexpected watch close - watch lasted less than a second and no items received Dec 03 14:09:50 crc kubenswrapper[4805]: W1203 14:09:50.550733 4805 reflector.go:484] object-"openshift-dns"/"openshift-service-ca.crt": watch of *v1.ConfigMap ended with: very short watch: object-"openshift-dns"/"openshift-service-ca.crt": Unexpected watch close - watch lasted less than a second and no items received Dec 03 14:09:50 crc kubenswrapper[4805]: W1203 14:09:50.550754 4805 reflector.go:484] object-"openshift-machine-config-operator"/"openshift-service-ca.crt": watch of *v1.ConfigMap ended with: very short watch: object-"openshift-machine-config-operator"/"openshift-service-ca.crt": Unexpected watch close - watch lasted less than a second and no items received Dec 03 14:09:50 crc kubenswrapper[4805]: W1203 14:09:50.550759 4805 reflector.go:484] object-"openshift-machine-config-operator"/"proxy-tls": watch of *v1.Secret ended with: very short watch: object-"openshift-machine-config-operator"/"proxy-tls": Unexpected watch close - watch lasted less than a second and no items received Dec 03 14:09:50 crc kubenswrapper[4805]: W1203 14:09:50.550762 4805 reflector.go:484] object-"openshift-dns"/"kube-root-ca.crt": watch of *v1.ConfigMap ended with: very short watch: object-"openshift-dns"/"kube-root-ca.crt": Unexpected watch close - watch lasted less than a second and no items received Dec 03 14:09:50 crc kubenswrapper[4805]: W1203 14:09:50.550769 4805 reflector.go:484] object-"openshift-multus"/"multus-ancillary-tools-dockercfg-vnmsz": watch of *v1.Secret ended with: very short watch: object-"openshift-multus"/"multus-ancillary-tools-dockercfg-vnmsz": Unexpected watch close - watch lasted less than a second and no items received Dec 03 14:09:50 crc kubenswrapper[4805]: W1203 14:09:50.550785 4805 reflector.go:484] object-"openshift-multus"/"cni-copy-resources": watch of *v1.ConfigMap ended with: very short watch: object-"openshift-multus"/"cni-copy-resources": Unexpected watch close - watch lasted less than a second and no items received Dec 03 14:09:50 crc kubenswrapper[4805]: W1203 14:09:50.550877 4805 reflector.go:484] object-"openshift-multus"/"kube-root-ca.crt": watch of *v1.ConfigMap ended with: very short watch: object-"openshift-multus"/"kube-root-ca.crt": Unexpected watch close - watch lasted less than a second and no items received Dec 03 14:09:50 crc kubenswrapper[4805]: W1203 14:09:50.550921 4805 reflector.go:484] object-"openshift-multus"/"default-dockercfg-2q5b6": watch of *v1.Secret ended with: very short watch: object-"openshift-multus"/"default-dockercfg-2q5b6": Unexpected watch close - watch lasted less than a second and no items received Dec 03 14:09:50 crc kubenswrapper[4805]: W1203 14:09:50.550931 4805 reflector.go:484] object-"openshift-multus"/"openshift-service-ca.crt": watch of *v1.ConfigMap ended with: very short watch: object-"openshift-multus"/"openshift-service-ca.crt": Unexpected watch close - watch lasted less than a second and no items received Dec 03 14:09:50 crc kubenswrapper[4805]: W1203 14:09:50.550972 4805 reflector.go:484] object-"openshift-machine-config-operator"/"kube-root-ca.crt": watch of *v1.ConfigMap ended with: very short watch: object-"openshift-machine-config-operator"/"kube-root-ca.crt": Unexpected watch close - watch lasted less than a second and no items received Dec 03 14:09:50 crc kubenswrapper[4805]: W1203 14:09:50.551120 4805 reflector.go:484] object-"openshift-dns"/"node-resolver-dockercfg-kz9s7": watch of *v1.Secret ended with: very short watch: object-"openshift-dns"/"node-resolver-dockercfg-kz9s7": Unexpected watch close - watch lasted less than a second and no items received Dec 03 14:09:50 crc kubenswrapper[4805]: W1203 14:09:50.550994 4805 reflector.go:484] object-"openshift-multus"/"default-cni-sysctl-allowlist": watch of *v1.ConfigMap ended with: very short watch: object-"openshift-multus"/"default-cni-sysctl-allowlist": Unexpected watch close - watch lasted less than a second and no items received Dec 03 14:09:50 crc kubenswrapper[4805]: I1203 14:09:50.558446 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tuning-conf-dir\" (UniqueName: \"kubernetes.io/host-path/de00a81f-82e1-4223-9614-f98606b9a8ef-tuning-conf-dir\") pod \"multus-additional-cni-plugins-bpk49\" (UID: \"de00a81f-82e1-4223-9614-f98606b9a8ef\") " pod="openshift-multus/multus-additional-cni-plugins-bpk49" Dec 03 14:09:50 crc kubenswrapper[4805]: I1203 14:09:50.558616 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/c09c1cdc-f461-45c9-8444-5b8764bccabc-system-cni-dir\") pod \"multus-k4mhs\" (UID: \"c09c1cdc-f461-45c9-8444-5b8764bccabc\") " pod="openshift-multus/multus-k4mhs" Dec 03 14:09:50 crc kubenswrapper[4805]: I1203 14:09:50.558641 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-multus\" (UniqueName: \"kubernetes.io/host-path/c09c1cdc-f461-45c9-8444-5b8764bccabc-host-var-lib-cni-multus\") pod \"multus-k4mhs\" (UID: \"c09c1cdc-f461-45c9-8444-5b8764bccabc\") " pod="openshift-multus/multus-k4mhs" Dec 03 14:09:50 crc kubenswrapper[4805]: I1203 14:09:50.558661 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nwrw9\" (UniqueName: \"kubernetes.io/projected/6098937f-e3f6-45e8-a647-4994a79cd711-kube-api-access-nwrw9\") pod \"machine-config-daemon-gskh4\" (UID: \"6098937f-e3f6-45e8-a647-4994a79cd711\") " pod="openshift-machine-config-operator/machine-config-daemon-gskh4" Dec 03 14:09:50 crc kubenswrapper[4805]: I1203 14:09:50.558681 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"hostroot\" (UniqueName: \"kubernetes.io/host-path/c09c1cdc-f461-45c9-8444-5b8764bccabc-hostroot\") pod \"multus-k4mhs\" (UID: \"c09c1cdc-f461-45c9-8444-5b8764bccabc\") " pod="openshift-multus/multus-k4mhs" Dec 03 14:09:50 crc kubenswrapper[4805]: I1203 14:09:50.558700 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-25d7t\" (UniqueName: \"kubernetes.io/projected/de00a81f-82e1-4223-9614-f98606b9a8ef-kube-api-access-25d7t\") pod \"multus-additional-cni-plugins-bpk49\" (UID: \"de00a81f-82e1-4223-9614-f98606b9a8ef\") " pod="openshift-multus/multus-additional-cni-plugins-bpk49" Dec 03 14:09:50 crc kubenswrapper[4805]: I1203 14:09:50.558719 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/6098937f-e3f6-45e8-a647-4994a79cd711-proxy-tls\") pod \"machine-config-daemon-gskh4\" (UID: \"6098937f-e3f6-45e8-a647-4994a79cd711\") " pod="openshift-machine-config-operator/machine-config-daemon-gskh4" Dec 03 14:09:50 crc kubenswrapper[4805]: I1203 14:09:50.558738 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/c09c1cdc-f461-45c9-8444-5b8764bccabc-os-release\") pod \"multus-k4mhs\" (UID: \"c09c1cdc-f461-45c9-8444-5b8764bccabc\") " pod="openshift-multus/multus-k4mhs" Dec 03 14:09:50 crc kubenswrapper[4805]: I1203 14:09:50.558758 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"hosts-file\" (UniqueName: \"kubernetes.io/host-path/860769de-6a5b-4cb9-a752-264e7e2e785f-hosts-file\") pod \"node-resolver-pdzgk\" (UID: \"860769de-6a5b-4cb9-a752-264e7e2e785f\") " pod="openshift-dns/node-resolver-pdzgk" Dec 03 14:09:50 crc kubenswrapper[4805]: I1203 14:09:50.558778 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-cni-dir\" (UniqueName: \"kubernetes.io/host-path/c09c1cdc-f461-45c9-8444-5b8764bccabc-multus-cni-dir\") pod \"multus-k4mhs\" (UID: \"c09c1cdc-f461-45c9-8444-5b8764bccabc\") " pod="openshift-multus/multus-k4mhs" Dec 03 14:09:50 crc kubenswrapper[4805]: I1203 14:09:50.558801 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/de00a81f-82e1-4223-9614-f98606b9a8ef-os-release\") pod \"multus-additional-cni-plugins-bpk49\" (UID: \"de00a81f-82e1-4223-9614-f98606b9a8ef\") " pod="openshift-multus/multus-additional-cni-plugins-bpk49" Dec 03 14:09:50 crc kubenswrapper[4805]: I1203 14:09:50.558866 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/de00a81f-82e1-4223-9614-f98606b9a8ef-cni-sysctl-allowlist\") pod \"multus-additional-cni-plugins-bpk49\" (UID: \"de00a81f-82e1-4223-9614-f98606b9a8ef\") " pod="openshift-multus/multus-additional-cni-plugins-bpk49" Dec 03 14:09:50 crc kubenswrapper[4805]: I1203 14:09:50.558872 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"hostroot\" (UniqueName: \"kubernetes.io/host-path/c09c1cdc-f461-45c9-8444-5b8764bccabc-hostroot\") pod \"multus-k4mhs\" (UID: \"c09c1cdc-f461-45c9-8444-5b8764bccabc\") " pod="openshift-multus/multus-k4mhs" Dec 03 14:09:50 crc kubenswrapper[4805]: I1203 14:09:50.558898 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rootfs\" (UniqueName: \"kubernetes.io/host-path/6098937f-e3f6-45e8-a647-4994a79cd711-rootfs\") pod \"machine-config-daemon-gskh4\" (UID: \"6098937f-e3f6-45e8-a647-4994a79cd711\") " pod="openshift-machine-config-operator/machine-config-daemon-gskh4" Dec 03 14:09:50 crc kubenswrapper[4805]: I1203 14:09:50.558935 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rootfs\" (UniqueName: \"kubernetes.io/host-path/6098937f-e3f6-45e8-a647-4994a79cd711-rootfs\") pod \"machine-config-daemon-gskh4\" (UID: \"6098937f-e3f6-45e8-a647-4994a79cd711\") " pod="openshift-machine-config-operator/machine-config-daemon-gskh4" Dec 03 14:09:50 crc kubenswrapper[4805]: I1203 14:09:50.558949 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/de00a81f-82e1-4223-9614-f98606b9a8ef-cnibin\") pod \"multus-additional-cni-plugins-bpk49\" (UID: \"de00a81f-82e1-4223-9614-f98606b9a8ef\") " pod="openshift-multus/multus-additional-cni-plugins-bpk49" Dec 03 14:09:50 crc kubenswrapper[4805]: I1203 14:09:50.558979 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/c09c1cdc-f461-45c9-8444-5b8764bccabc-cni-binary-copy\") pod \"multus-k4mhs\" (UID: \"c09c1cdc-f461-45c9-8444-5b8764bccabc\") " pod="openshift-multus/multus-k4mhs" Dec 03 14:09:50 crc kubenswrapper[4805]: I1203 14:09:50.559002 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qdcd7\" (UniqueName: \"kubernetes.io/projected/c09c1cdc-f461-45c9-8444-5b8764bccabc-kube-api-access-qdcd7\") pod \"multus-k4mhs\" (UID: \"c09c1cdc-f461-45c9-8444-5b8764bccabc\") " pod="openshift-multus/multus-k4mhs" Dec 03 14:09:50 crc kubenswrapper[4805]: I1203 14:09:50.559025 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/de00a81f-82e1-4223-9614-f98606b9a8ef-cni-binary-copy\") pod \"multus-additional-cni-plugins-bpk49\" (UID: \"de00a81f-82e1-4223-9614-f98606b9a8ef\") " pod="openshift-multus/multus-additional-cni-plugins-bpk49" Dec 03 14:09:50 crc kubenswrapper[4805]: I1203 14:09:50.559050 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-k8s-cni-cncf-io\" (UniqueName: \"kubernetes.io/host-path/c09c1cdc-f461-45c9-8444-5b8764bccabc-host-run-k8s-cni-cncf-io\") pod \"multus-k4mhs\" (UID: \"c09c1cdc-f461-45c9-8444-5b8764bccabc\") " pod="openshift-multus/multus-k4mhs" Dec 03 14:09:50 crc kubenswrapper[4805]: I1203 14:09:50.559073 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/c09c1cdc-f461-45c9-8444-5b8764bccabc-host-var-lib-kubelet\") pod \"multus-k4mhs\" (UID: \"c09c1cdc-f461-45c9-8444-5b8764bccabc\") " pod="openshift-multus/multus-k4mhs" Dec 03 14:09:50 crc kubenswrapper[4805]: I1203 14:09:50.559099 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-multus-certs\" (UniqueName: \"kubernetes.io/host-path/c09c1cdc-f461-45c9-8444-5b8764bccabc-host-run-multus-certs\") pod \"multus-k4mhs\" (UID: \"c09c1cdc-f461-45c9-8444-5b8764bccabc\") " pod="openshift-multus/multus-k4mhs" Dec 03 14:09:50 crc kubenswrapper[4805]: I1203 14:09:50.559120 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-kubernetes\" (UniqueName: \"kubernetes.io/host-path/c09c1cdc-f461-45c9-8444-5b8764bccabc-etc-kubernetes\") pod \"multus-k4mhs\" (UID: \"c09c1cdc-f461-45c9-8444-5b8764bccabc\") " pod="openshift-multus/multus-k4mhs" Dec 03 14:09:50 crc kubenswrapper[4805]: I1203 14:09:50.559158 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/c09c1cdc-f461-45c9-8444-5b8764bccabc-cnibin\") pod \"multus-k4mhs\" (UID: \"c09c1cdc-f461-45c9-8444-5b8764bccabc\") " pod="openshift-multus/multus-k4mhs" Dec 03 14:09:50 crc kubenswrapper[4805]: I1203 14:09:50.559180 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-conf-dir\" (UniqueName: \"kubernetes.io/host-path/c09c1cdc-f461-45c9-8444-5b8764bccabc-multus-conf-dir\") pod \"multus-k4mhs\" (UID: \"c09c1cdc-f461-45c9-8444-5b8764bccabc\") " pod="openshift-multus/multus-k4mhs" Dec 03 14:09:50 crc kubenswrapper[4805]: I1203 14:09:50.559204 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/c09c1cdc-f461-45c9-8444-5b8764bccabc-multus-daemon-config\") pod \"multus-k4mhs\" (UID: \"c09c1cdc-f461-45c9-8444-5b8764bccabc\") " pod="openshift-multus/multus-k4mhs" Dec 03 14:09:50 crc kubenswrapper[4805]: I1203 14:09:50.559225 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ggn4n\" (UniqueName: \"kubernetes.io/projected/860769de-6a5b-4cb9-a752-264e7e2e785f-kube-api-access-ggn4n\") pod \"node-resolver-pdzgk\" (UID: \"860769de-6a5b-4cb9-a752-264e7e2e785f\") " pod="openshift-dns/node-resolver-pdzgk" Dec 03 14:09:50 crc kubenswrapper[4805]: I1203 14:09:50.559259 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-bin\" (UniqueName: \"kubernetes.io/host-path/c09c1cdc-f461-45c9-8444-5b8764bccabc-host-var-lib-cni-bin\") pod \"multus-k4mhs\" (UID: \"c09c1cdc-f461-45c9-8444-5b8764bccabc\") " pod="openshift-multus/multus-k4mhs" Dec 03 14:09:50 crc kubenswrapper[4805]: I1203 14:09:50.559309 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/6098937f-e3f6-45e8-a647-4994a79cd711-mcd-auth-proxy-config\") pod \"machine-config-daemon-gskh4\" (UID: \"6098937f-e3f6-45e8-a647-4994a79cd711\") " pod="openshift-machine-config-operator/machine-config-daemon-gskh4" Dec 03 14:09:50 crc kubenswrapper[4805]: I1203 14:09:50.559334 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/de00a81f-82e1-4223-9614-f98606b9a8ef-system-cni-dir\") pod \"multus-additional-cni-plugins-bpk49\" (UID: \"de00a81f-82e1-4223-9614-f98606b9a8ef\") " pod="openshift-multus/multus-additional-cni-plugins-bpk49" Dec 03 14:09:50 crc kubenswrapper[4805]: I1203 14:09:50.559355 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-socket-dir-parent\" (UniqueName: \"kubernetes.io/host-path/c09c1cdc-f461-45c9-8444-5b8764bccabc-multus-socket-dir-parent\") pod \"multus-k4mhs\" (UID: \"c09c1cdc-f461-45c9-8444-5b8764bccabc\") " pod="openshift-multus/multus-k4mhs" Dec 03 14:09:50 crc kubenswrapper[4805]: I1203 14:09:50.559376 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/c09c1cdc-f461-45c9-8444-5b8764bccabc-host-run-netns\") pod \"multus-k4mhs\" (UID: \"c09c1cdc-f461-45c9-8444-5b8764bccabc\") " pod="openshift-multus/multus-k4mhs" Dec 03 14:09:50 crc kubenswrapper[4805]: I1203 14:09:50.559457 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/c09c1cdc-f461-45c9-8444-5b8764bccabc-host-run-netns\") pod \"multus-k4mhs\" (UID: \"c09c1cdc-f461-45c9-8444-5b8764bccabc\") " pod="openshift-multus/multus-k4mhs" Dec 03 14:09:50 crc kubenswrapper[4805]: I1203 14:09:50.559519 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/c09c1cdc-f461-45c9-8444-5b8764bccabc-system-cni-dir\") pod \"multus-k4mhs\" (UID: \"c09c1cdc-f461-45c9-8444-5b8764bccabc\") " pod="openshift-multus/multus-k4mhs" Dec 03 14:09:50 crc kubenswrapper[4805]: I1203 14:09:50.559553 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-multus\" (UniqueName: \"kubernetes.io/host-path/c09c1cdc-f461-45c9-8444-5b8764bccabc-host-var-lib-cni-multus\") pod \"multus-k4mhs\" (UID: \"c09c1cdc-f461-45c9-8444-5b8764bccabc\") " pod="openshift-multus/multus-k4mhs" Dec 03 14:09:50 crc kubenswrapper[4805]: I1203 14:09:50.560397 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-kubernetes\" (UniqueName: \"kubernetes.io/host-path/c09c1cdc-f461-45c9-8444-5b8764bccabc-etc-kubernetes\") pod \"multus-k4mhs\" (UID: \"c09c1cdc-f461-45c9-8444-5b8764bccabc\") " pod="openshift-multus/multus-k4mhs" Dec 03 14:09:50 crc kubenswrapper[4805]: I1203 14:09:50.560414 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-bin\" (UniqueName: \"kubernetes.io/host-path/c09c1cdc-f461-45c9-8444-5b8764bccabc-host-var-lib-cni-bin\") pod \"multus-k4mhs\" (UID: \"c09c1cdc-f461-45c9-8444-5b8764bccabc\") " pod="openshift-multus/multus-k4mhs" Dec 03 14:09:50 crc kubenswrapper[4805]: I1203 14:09:50.560410 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tuning-conf-dir\" (UniqueName: \"kubernetes.io/host-path/de00a81f-82e1-4223-9614-f98606b9a8ef-tuning-conf-dir\") pod \"multus-additional-cni-plugins-bpk49\" (UID: \"de00a81f-82e1-4223-9614-f98606b9a8ef\") " pod="openshift-multus/multus-additional-cni-plugins-bpk49" Dec 03 14:09:50 crc kubenswrapper[4805]: I1203 14:09:50.560882 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/de00a81f-82e1-4223-9614-f98606b9a8ef-system-cni-dir\") pod \"multus-additional-cni-plugins-bpk49\" (UID: \"de00a81f-82e1-4223-9614-f98606b9a8ef\") " pod="openshift-multus/multus-additional-cni-plugins-bpk49" Dec 03 14:09:50 crc kubenswrapper[4805]: I1203 14:09:50.560885 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/6098937f-e3f6-45e8-a647-4994a79cd711-mcd-auth-proxy-config\") pod \"machine-config-daemon-gskh4\" (UID: \"6098937f-e3f6-45e8-a647-4994a79cd711\") " pod="openshift-machine-config-operator/machine-config-daemon-gskh4" Dec 03 14:09:50 crc kubenswrapper[4805]: I1203 14:09:50.561050 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/c09c1cdc-f461-45c9-8444-5b8764bccabc-cnibin\") pod \"multus-k4mhs\" (UID: \"c09c1cdc-f461-45c9-8444-5b8764bccabc\") " pod="openshift-multus/multus-k4mhs" Dec 03 14:09:50 crc kubenswrapper[4805]: I1203 14:09:50.561062 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-socket-dir-parent\" (UniqueName: \"kubernetes.io/host-path/c09c1cdc-f461-45c9-8444-5b8764bccabc-multus-socket-dir-parent\") pod \"multus-k4mhs\" (UID: \"c09c1cdc-f461-45c9-8444-5b8764bccabc\") " pod="openshift-multus/multus-k4mhs" Dec 03 14:09:50 crc kubenswrapper[4805]: I1203 14:09:50.561076 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-multus-certs\" (UniqueName: \"kubernetes.io/host-path/c09c1cdc-f461-45c9-8444-5b8764bccabc-host-run-multus-certs\") pod \"multus-k4mhs\" (UID: \"c09c1cdc-f461-45c9-8444-5b8764bccabc\") " pod="openshift-multus/multus-k4mhs" Dec 03 14:09:50 crc kubenswrapper[4805]: I1203 14:09:50.561097 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-k8s-cni-cncf-io\" (UniqueName: \"kubernetes.io/host-path/c09c1cdc-f461-45c9-8444-5b8764bccabc-host-run-k8s-cni-cncf-io\") pod \"multus-k4mhs\" (UID: \"c09c1cdc-f461-45c9-8444-5b8764bccabc\") " pod="openshift-multus/multus-k4mhs" Dec 03 14:09:50 crc kubenswrapper[4805]: I1203 14:09:50.561114 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/c09c1cdc-f461-45c9-8444-5b8764bccabc-host-var-lib-kubelet\") pod \"multus-k4mhs\" (UID: \"c09c1cdc-f461-45c9-8444-5b8764bccabc\") " pod="openshift-multus/multus-k4mhs" Dec 03 14:09:50 crc kubenswrapper[4805]: I1203 14:09:50.561112 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"hosts-file\" (UniqueName: \"kubernetes.io/host-path/860769de-6a5b-4cb9-a752-264e7e2e785f-hosts-file\") pod \"node-resolver-pdzgk\" (UID: \"860769de-6a5b-4cb9-a752-264e7e2e785f\") " pod="openshift-dns/node-resolver-pdzgk" Dec 03 14:09:50 crc kubenswrapper[4805]: I1203 14:09:50.561131 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/de00a81f-82e1-4223-9614-f98606b9a8ef-cni-binary-copy\") pod \"multus-additional-cni-plugins-bpk49\" (UID: \"de00a81f-82e1-4223-9614-f98606b9a8ef\") " pod="openshift-multus/multus-additional-cni-plugins-bpk49" Dec 03 14:09:50 crc kubenswrapper[4805]: I1203 14:09:50.561142 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-conf-dir\" (UniqueName: \"kubernetes.io/host-path/c09c1cdc-f461-45c9-8444-5b8764bccabc-multus-conf-dir\") pod \"multus-k4mhs\" (UID: \"c09c1cdc-f461-45c9-8444-5b8764bccabc\") " pod="openshift-multus/multus-k4mhs" Dec 03 14:09:50 crc kubenswrapper[4805]: I1203 14:09:50.561164 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/de00a81f-82e1-4223-9614-f98606b9a8ef-os-release\") pod \"multus-additional-cni-plugins-bpk49\" (UID: \"de00a81f-82e1-4223-9614-f98606b9a8ef\") " pod="openshift-multus/multus-additional-cni-plugins-bpk49" Dec 03 14:09:50 crc kubenswrapper[4805]: I1203 14:09:50.561223 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/de00a81f-82e1-4223-9614-f98606b9a8ef-cnibin\") pod \"multus-additional-cni-plugins-bpk49\" (UID: \"de00a81f-82e1-4223-9614-f98606b9a8ef\") " pod="openshift-multus/multus-additional-cni-plugins-bpk49" Dec 03 14:09:50 crc kubenswrapper[4805]: I1203 14:09:50.561249 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/c09c1cdc-f461-45c9-8444-5b8764bccabc-os-release\") pod \"multus-k4mhs\" (UID: \"c09c1cdc-f461-45c9-8444-5b8764bccabc\") " pod="openshift-multus/multus-k4mhs" Dec 03 14:09:50 crc kubenswrapper[4805]: I1203 14:09:50.561257 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-cni-dir\" (UniqueName: \"kubernetes.io/host-path/c09c1cdc-f461-45c9-8444-5b8764bccabc-multus-cni-dir\") pod \"multus-k4mhs\" (UID: \"c09c1cdc-f461-45c9-8444-5b8764bccabc\") " pod="openshift-multus/multus-k4mhs" Dec 03 14:09:50 crc kubenswrapper[4805]: I1203 14:09:50.561595 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/c09c1cdc-f461-45c9-8444-5b8764bccabc-multus-daemon-config\") pod \"multus-k4mhs\" (UID: \"c09c1cdc-f461-45c9-8444-5b8764bccabc\") " pod="openshift-multus/multus-k4mhs" Dec 03 14:09:50 crc kubenswrapper[4805]: I1203 14:09:50.562504 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/c09c1cdc-f461-45c9-8444-5b8764bccabc-cni-binary-copy\") pod \"multus-k4mhs\" (UID: \"c09c1cdc-f461-45c9-8444-5b8764bccabc\") " pod="openshift-multus/multus-k4mhs" Dec 03 14:09:50 crc kubenswrapper[4805]: I1203 14:09:50.565523 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/de00a81f-82e1-4223-9614-f98606b9a8ef-cni-sysctl-allowlist\") pod \"multus-additional-cni-plugins-bpk49\" (UID: \"de00a81f-82e1-4223-9614-f98606b9a8ef\") " pod="openshift-multus/multus-additional-cni-plugins-bpk49" Dec 03 14:09:50 crc kubenswrapper[4805]: I1203 14:09:50.579775 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/6098937f-e3f6-45e8-a647-4994a79cd711-proxy-tls\") pod \"machine-config-daemon-gskh4\" (UID: \"6098937f-e3f6-45e8-a647-4994a79cd711\") " pod="openshift-machine-config-operator/machine-config-daemon-gskh4" Dec 03 14:09:50 crc kubenswrapper[4805]: I1203 14:09:50.586479 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nwrw9\" (UniqueName: \"kubernetes.io/projected/6098937f-e3f6-45e8-a647-4994a79cd711-kube-api-access-nwrw9\") pod \"machine-config-daemon-gskh4\" (UID: \"6098937f-e3f6-45e8-a647-4994a79cd711\") " pod="openshift-machine-config-operator/machine-config-daemon-gskh4" Dec 03 14:09:50 crc kubenswrapper[4805]: I1203 14:09:50.599475 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qdcd7\" (UniqueName: \"kubernetes.io/projected/c09c1cdc-f461-45c9-8444-5b8764bccabc-kube-api-access-qdcd7\") pod \"multus-k4mhs\" (UID: \"c09c1cdc-f461-45c9-8444-5b8764bccabc\") " pod="openshift-multus/multus-k4mhs" Dec 03 14:09:50 crc kubenswrapper[4805]: I1203 14:09:50.606270 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ggn4n\" (UniqueName: \"kubernetes.io/projected/860769de-6a5b-4cb9-a752-264e7e2e785f-kube-api-access-ggn4n\") pod \"node-resolver-pdzgk\" (UID: \"860769de-6a5b-4cb9-a752-264e7e2e785f\") " pod="openshift-dns/node-resolver-pdzgk" Dec 03 14:09:50 crc kubenswrapper[4805]: I1203 14:09:50.610521 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-25d7t\" (UniqueName: \"kubernetes.io/projected/de00a81f-82e1-4223-9614-f98606b9a8ef-kube-api-access-25d7t\") pod \"multus-additional-cni-plugins-bpk49\" (UID: \"de00a81f-82e1-4223-9614-f98606b9a8ef\") " pod="openshift-multus/multus-additional-cni-plugins-bpk49" Dec 03 14:09:50 crc kubenswrapper[4805]: I1203 14:09:50.693770 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 14:09:50 crc kubenswrapper[4805]: I1203 14:09:50.693783 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 14:09:50 crc kubenswrapper[4805]: I1203 14:09:50.693824 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 14:09:50 crc kubenswrapper[4805]: E1203 14:09:50.693919 4805 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 03 14:09:50 crc kubenswrapper[4805]: E1203 14:09:50.694021 4805 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 03 14:09:50 crc kubenswrapper[4805]: E1203 14:09:50.694101 4805 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 03 14:09:50 crc kubenswrapper[4805]: I1203 14:09:50.698111 4805 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="01ab3dd5-8196-46d0-ad33-122e2ca51def" path="/var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes" Dec 03 14:09:50 crc kubenswrapper[4805]: I1203 14:09:50.699009 4805 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" path="/var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes" Dec 03 14:09:50 crc kubenswrapper[4805]: I1203 14:09:50.700216 4805 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="09efc573-dbb6-4249-bd59-9b87aba8dd28" path="/var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes" Dec 03 14:09:50 crc kubenswrapper[4805]: I1203 14:09:50.700944 4805 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0b574797-001e-440a-8f4e-c0be86edad0f" path="/var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes" Dec 03 14:09:50 crc kubenswrapper[4805]: I1203 14:09:50.701986 4805 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0b78653f-4ff9-4508-8672-245ed9b561e3" path="/var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes" Dec 03 14:09:50 crc kubenswrapper[4805]: I1203 14:09:50.702587 4805 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1386a44e-36a2-460c-96d0-0359d2b6f0f5" path="/var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes" Dec 03 14:09:50 crc kubenswrapper[4805]: I1203 14:09:50.703331 4805 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1bf7eb37-55a3-4c65-b768-a94c82151e69" path="/var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes" Dec 03 14:09:50 crc kubenswrapper[4805]: I1203 14:09:50.704362 4805 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1d611f23-29be-4491-8495-bee1670e935f" path="/var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes" Dec 03 14:09:50 crc kubenswrapper[4805]: I1203 14:09:50.705117 4805 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="20b0d48f-5fd6-431c-a545-e3c800c7b866" path="/var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/volumes" Dec 03 14:09:50 crc kubenswrapper[4805]: I1203 14:09:50.706115 4805 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" path="/var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes" Dec 03 14:09:50 crc kubenswrapper[4805]: I1203 14:09:50.706680 4805 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="22c825df-677d-4ca6-82db-3454ed06e783" path="/var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes" Dec 03 14:09:50 crc kubenswrapper[4805]: I1203 14:09:50.707804 4805 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="25e176fe-21b4-4974-b1ed-c8b94f112a7f" path="/var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes" Dec 03 14:09:50 crc kubenswrapper[4805]: I1203 14:09:50.708512 4805 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" path="/var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes" Dec 03 14:09:50 crc kubenswrapper[4805]: I1203 14:09:50.709166 4805 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="31d8b7a1-420e-4252-a5b7-eebe8a111292" path="/var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes" Dec 03 14:09:50 crc kubenswrapper[4805]: I1203 14:09:50.710155 4805 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3ab1a177-2de0-46d9-b765-d0d0649bb42e" path="/var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/volumes" Dec 03 14:09:50 crc kubenswrapper[4805]: I1203 14:09:50.710700 4805 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" path="/var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes" Dec 03 14:09:50 crc kubenswrapper[4805]: I1203 14:09:50.711804 4805 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="43509403-f426-496e-be36-56cef71462f5" path="/var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes" Dec 03 14:09:50 crc kubenswrapper[4805]: I1203 14:09:50.712241 4805 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="44663579-783b-4372-86d6-acf235a62d72" path="/var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/volumes" Dec 03 14:09:50 crc kubenswrapper[4805]: I1203 14:09:50.712897 4805 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="496e6271-fb68-4057-954e-a0d97a4afa3f" path="/var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes" Dec 03 14:09:50 crc kubenswrapper[4805]: I1203 14:09:50.714413 4805 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" path="/var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes" Dec 03 14:09:50 crc kubenswrapper[4805]: I1203 14:09:50.715145 4805 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="49ef4625-1d3a-4a9f-b595-c2433d32326d" path="/var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/volumes" Dec 03 14:09:50 crc kubenswrapper[4805]: I1203 14:09:50.716032 4805 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4bb40260-dbaa-4fb0-84df-5e680505d512" path="/var/lib/kubelet/pods/4bb40260-dbaa-4fb0-84df-5e680505d512/volumes" Dec 03 14:09:50 crc kubenswrapper[4805]: I1203 14:09:50.716449 4805 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5225d0e4-402f-4861-b410-819f433b1803" path="/var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes" Dec 03 14:09:50 crc kubenswrapper[4805]: I1203 14:09:50.717171 4805 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5441d097-087c-4d9a-baa8-b210afa90fc9" path="/var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes" Dec 03 14:09:50 crc kubenswrapper[4805]: I1203 14:09:50.717713 4805 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="57a731c4-ef35-47a8-b875-bfb08a7f8011" path="/var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes" Dec 03 14:09:50 crc kubenswrapper[4805]: I1203 14:09:50.718488 4805 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5b88f790-22fa-440e-b583-365168c0b23d" path="/var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/volumes" Dec 03 14:09:50 crc kubenswrapper[4805]: I1203 14:09:50.719232 4805 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5fe579f8-e8a6-4643-bce5-a661393c4dde" path="/var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/volumes" Dec 03 14:09:50 crc kubenswrapper[4805]: I1203 14:09:50.719685 4805 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6402fda4-df10-493c-b4e5-d0569419652d" path="/var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes" Dec 03 14:09:50 crc kubenswrapper[4805]: I1203 14:09:50.720261 4805 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6509e943-70c6-444c-bc41-48a544e36fbd" path="/var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes" Dec 03 14:09:50 crc kubenswrapper[4805]: I1203 14:09:50.720728 4805 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6731426b-95fe-49ff-bb5f-40441049fde2" path="/var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/volumes" Dec 03 14:09:50 crc kubenswrapper[4805]: I1203 14:09:50.722240 4805 kubelet_volumes.go:152] "Cleaned up orphaned volume subpath from pod" podUID="6ea678ab-3438-413e-bfe3-290ae7725660" path="/var/lib/kubelet/pods/6ea678ab-3438-413e-bfe3-290ae7725660/volume-subpaths/run-systemd/ovnkube-controller/6" Dec 03 14:09:50 crc kubenswrapper[4805]: I1203 14:09:50.722339 4805 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6ea678ab-3438-413e-bfe3-290ae7725660" path="/var/lib/kubelet/pods/6ea678ab-3438-413e-bfe3-290ae7725660/volumes" Dec 03 14:09:50 crc kubenswrapper[4805]: I1203 14:09:50.723780 4805 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7539238d-5fe0-46ed-884e-1c3b566537ec" path="/var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes" Dec 03 14:09:50 crc kubenswrapper[4805]: I1203 14:09:50.724484 4805 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7583ce53-e0fe-4a16-9e4d-50516596a136" path="/var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes" Dec 03 14:09:50 crc kubenswrapper[4805]: I1203 14:09:50.725050 4805 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7bb08738-c794-4ee8-9972-3a62ca171029" path="/var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes" Dec 03 14:09:50 crc kubenswrapper[4805]: I1203 14:09:50.726314 4805 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="87cf06ed-a83f-41a7-828d-70653580a8cb" path="/var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes" Dec 03 14:09:50 crc kubenswrapper[4805]: I1203 14:09:50.727098 4805 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" path="/var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes" Dec 03 14:09:50 crc kubenswrapper[4805]: I1203 14:09:50.727620 4805 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="925f1c65-6136-48ba-85aa-3a3b50560753" path="/var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes" Dec 03 14:09:50 crc kubenswrapper[4805]: I1203 14:09:50.728253 4805 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="96b93a3a-6083-4aea-8eab-fe1aa8245ad9" path="/var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/volumes" Dec 03 14:09:50 crc kubenswrapper[4805]: I1203 14:09:50.729101 4805 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9d4552c7-cd75-42dd-8880-30dd377c49a4" path="/var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes" Dec 03 14:09:50 crc kubenswrapper[4805]: I1203 14:09:50.729570 4805 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a0128f3a-b052-44ed-a84e-c4c8aaf17c13" path="/var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/volumes" Dec 03 14:09:50 crc kubenswrapper[4805]: I1203 14:09:50.730152 4805 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a31745f5-9847-4afe-82a5-3161cc66ca93" path="/var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes" Dec 03 14:09:50 crc kubenswrapper[4805]: I1203 14:09:50.730817 4805 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" path="/var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes" Dec 03 14:09:50 crc kubenswrapper[4805]: I1203 14:09:50.732390 4805 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b6312bbd-5731-4ea0-a20f-81d5a57df44a" path="/var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/volumes" Dec 03 14:09:50 crc kubenswrapper[4805]: I1203 14:09:50.733089 4805 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" path="/var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes" Dec 03 14:09:50 crc kubenswrapper[4805]: I1203 14:09:50.734260 4805 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" path="/var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes" Dec 03 14:09:50 crc kubenswrapper[4805]: I1203 14:09:50.735092 4805 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bd23aa5c-e532-4e53-bccf-e79f130c5ae8" path="/var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/volumes" Dec 03 14:09:50 crc kubenswrapper[4805]: I1203 14:09:50.735884 4805 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bf126b07-da06-4140-9a57-dfd54fc6b486" path="/var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes" Dec 03 14:09:50 crc kubenswrapper[4805]: I1203 14:09:50.736333 4805 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c03ee662-fb2f-4fc4-a2c1-af487c19d254" path="/var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes" Dec 03 14:09:50 crc kubenswrapper[4805]: I1203 14:09:50.736813 4805 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d" path="/var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/volumes" Dec 03 14:09:50 crc kubenswrapper[4805]: I1203 14:09:50.737307 4805 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e7e6199b-1264-4501-8953-767f51328d08" path="/var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes" Dec 03 14:09:50 crc kubenswrapper[4805]: I1203 14:09:50.738001 4805 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="efdd0498-1daa-4136-9a4a-3b948c2293fc" path="/var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/volumes" Dec 03 14:09:50 crc kubenswrapper[4805]: I1203 14:09:50.738690 4805 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" path="/var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/volumes" Dec 03 14:09:50 crc kubenswrapper[4805]: I1203 14:09:50.739299 4805 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="fda69060-fa79-4696-b1a6-7980f124bf7c" path="/var/lib/kubelet/pods/fda69060-fa79-4696-b1a6-7980f124bf7c/volumes" Dec 03 14:09:50 crc kubenswrapper[4805]: I1203 14:09:50.747352 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-k4mhs" Dec 03 14:09:50 crc kubenswrapper[4805]: I1203 14:09:50.759416 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-daemon-gskh4" Dec 03 14:09:50 crc kubenswrapper[4805]: W1203 14:09:50.761863 4805 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podc09c1cdc_f461_45c9_8444_5b8764bccabc.slice/crio-efd07ec69fed3dd0a968ad09995640e3b6ab518e66313adaed0e3007bd936418 WatchSource:0}: Error finding container efd07ec69fed3dd0a968ad09995640e3b6ab518e66313adaed0e3007bd936418: Status 404 returned error can't find the container with id efd07ec69fed3dd0a968ad09995640e3b6ab518e66313adaed0e3007bd936418 Dec 03 14:09:50 crc kubenswrapper[4805]: I1203 14:09:50.768982 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/node-resolver-pdzgk" Dec 03 14:09:50 crc kubenswrapper[4805]: W1203 14:09:50.769637 4805 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod6098937f_e3f6_45e8_a647_4994a79cd711.slice/crio-b8784b8802c77a2032c76addbcaa4dd02eb6428de88e2fdcc4e341f451920258 WatchSource:0}: Error finding container b8784b8802c77a2032c76addbcaa4dd02eb6428de88e2fdcc4e341f451920258: Status 404 returned error can't find the container with id b8784b8802c77a2032c76addbcaa4dd02eb6428de88e2fdcc4e341f451920258 Dec 03 14:09:50 crc kubenswrapper[4805]: I1203 14:09:50.776178 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-additional-cni-plugins-bpk49" Dec 03 14:09:50 crc kubenswrapper[4805]: W1203 14:09:50.783906 4805 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod860769de_6a5b_4cb9_a752_264e7e2e785f.slice/crio-e1ba7951cfae2eec07fe4798080fd2e88594bd4b628acf056e2afa89efd5a5b9 WatchSource:0}: Error finding container e1ba7951cfae2eec07fe4798080fd2e88594bd4b628acf056e2afa89efd5a5b9: Status 404 returned error can't find the container with id e1ba7951cfae2eec07fe4798080fd2e88594bd4b628acf056e2afa89efd5a5b9 Dec 03 14:09:50 crc kubenswrapper[4805]: I1203 14:09:50.793712 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-gskh4" event={"ID":"6098937f-e3f6-45e8-a647-4994a79cd711","Type":"ContainerStarted","Data":"b8784b8802c77a2032c76addbcaa4dd02eb6428de88e2fdcc4e341f451920258"} Dec 03 14:09:50 crc kubenswrapper[4805]: I1203 14:09:50.794694 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-k4mhs" event={"ID":"c09c1cdc-f461-45c9-8444-5b8764bccabc","Type":"ContainerStarted","Data":"efd07ec69fed3dd0a968ad09995640e3b6ab518e66313adaed0e3007bd936418"} Dec 03 14:09:50 crc kubenswrapper[4805]: I1203 14:09:50.795767 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/node-resolver-pdzgk" event={"ID":"860769de-6a5b-4cb9-a752-264e7e2e785f","Type":"ContainerStarted","Data":"e1ba7951cfae2eec07fe4798080fd2e88594bd4b628acf056e2afa89efd5a5b9"} Dec 03 14:09:50 crc kubenswrapper[4805]: I1203 14:09:50.801375 4805 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 03 14:09:50 crc kubenswrapper[4805]: W1203 14:09:50.802556 4805 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podde00a81f_82e1_4223_9614_f98606b9a8ef.slice/crio-d24573710ca84801bf2d9f2139eb66525226fee9d5a8ed9bac92ec635401a7f3 WatchSource:0}: Error finding container d24573710ca84801bf2d9f2139eb66525226fee9d5a8ed9bac92ec635401a7f3: Status 404 returned error can't find the container with id d24573710ca84801bf2d9f2139eb66525226fee9d5a8ed9bac92ec635401a7f3 Dec 03 14:09:50 crc kubenswrapper[4805]: I1203 14:09:50.830239 4805 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-psjpm"] Dec 03 14:09:50 crc kubenswrapper[4805]: I1203 14:09:50.831640 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-psjpm" Dec 03 14:09:50 crc kubenswrapper[4805]: I1203 14:09:50.835657 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-node-metrics-cert" Dec 03 14:09:50 crc kubenswrapper[4805]: I1203 14:09:50.836215 4805 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"ovnkube-config" Dec 03 14:09:50 crc kubenswrapper[4805]: I1203 14:09:50.836397 4805 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"env-overrides" Dec 03 14:09:50 crc kubenswrapper[4805]: I1203 14:09:50.836743 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-kubernetes-node-dockercfg-pwtwl" Dec 03 14:09:50 crc kubenswrapper[4805]: I1203 14:09:50.837105 4805 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"kube-root-ca.crt" Dec 03 14:09:50 crc kubenswrapper[4805]: I1203 14:09:50.837189 4805 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"ovnkube-script-lib" Dec 03 14:09:50 crc kubenswrapper[4805]: I1203 14:09:50.837271 4805 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"openshift-service-ca.crt" Dec 03 14:09:50 crc kubenswrapper[4805]: I1203 14:09:50.861943 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/deeec30d-cad2-4dfd-adb6-64f6646581be-ovnkube-script-lib\") pod \"ovnkube-node-psjpm\" (UID: \"deeec30d-cad2-4dfd-adb6-64f6646581be\") " pod="openshift-ovn-kubernetes/ovnkube-node-psjpm" Dec 03 14:09:50 crc kubenswrapper[4805]: I1203 14:09:50.861986 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/deeec30d-cad2-4dfd-adb6-64f6646581be-systemd-units\") pod \"ovnkube-node-psjpm\" (UID: \"deeec30d-cad2-4dfd-adb6-64f6646581be\") " pod="openshift-ovn-kubernetes/ovnkube-node-psjpm" Dec 03 14:09:50 crc kubenswrapper[4805]: I1203 14:09:50.862011 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/deeec30d-cad2-4dfd-adb6-64f6646581be-etc-openvswitch\") pod \"ovnkube-node-psjpm\" (UID: \"deeec30d-cad2-4dfd-adb6-64f6646581be\") " pod="openshift-ovn-kubernetes/ovnkube-node-psjpm" Dec 03 14:09:50 crc kubenswrapper[4805]: I1203 14:09:50.862032 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/deeec30d-cad2-4dfd-adb6-64f6646581be-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-psjpm\" (UID: \"deeec30d-cad2-4dfd-adb6-64f6646581be\") " pod="openshift-ovn-kubernetes/ovnkube-node-psjpm" Dec 03 14:09:50 crc kubenswrapper[4805]: I1203 14:09:50.862090 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wdlzx\" (UniqueName: \"kubernetes.io/projected/deeec30d-cad2-4dfd-adb6-64f6646581be-kube-api-access-wdlzx\") pod \"ovnkube-node-psjpm\" (UID: \"deeec30d-cad2-4dfd-adb6-64f6646581be\") " pod="openshift-ovn-kubernetes/ovnkube-node-psjpm" Dec 03 14:09:50 crc kubenswrapper[4805]: I1203 14:09:50.862146 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/deeec30d-cad2-4dfd-adb6-64f6646581be-host-kubelet\") pod \"ovnkube-node-psjpm\" (UID: \"deeec30d-cad2-4dfd-adb6-64f6646581be\") " pod="openshift-ovn-kubernetes/ovnkube-node-psjpm" Dec 03 14:09:50 crc kubenswrapper[4805]: I1203 14:09:50.862170 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/deeec30d-cad2-4dfd-adb6-64f6646581be-host-cni-netd\") pod \"ovnkube-node-psjpm\" (UID: \"deeec30d-cad2-4dfd-adb6-64f6646581be\") " pod="openshift-ovn-kubernetes/ovnkube-node-psjpm" Dec 03 14:09:50 crc kubenswrapper[4805]: I1203 14:09:50.862191 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/deeec30d-cad2-4dfd-adb6-64f6646581be-env-overrides\") pod \"ovnkube-node-psjpm\" (UID: \"deeec30d-cad2-4dfd-adb6-64f6646581be\") " pod="openshift-ovn-kubernetes/ovnkube-node-psjpm" Dec 03 14:09:50 crc kubenswrapper[4805]: I1203 14:09:50.862227 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/deeec30d-cad2-4dfd-adb6-64f6646581be-var-lib-openvswitch\") pod \"ovnkube-node-psjpm\" (UID: \"deeec30d-cad2-4dfd-adb6-64f6646581be\") " pod="openshift-ovn-kubernetes/ovnkube-node-psjpm" Dec 03 14:09:50 crc kubenswrapper[4805]: I1203 14:09:50.862246 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/deeec30d-cad2-4dfd-adb6-64f6646581be-run-ovn\") pod \"ovnkube-node-psjpm\" (UID: \"deeec30d-cad2-4dfd-adb6-64f6646581be\") " pod="openshift-ovn-kubernetes/ovnkube-node-psjpm" Dec 03 14:09:50 crc kubenswrapper[4805]: I1203 14:09:50.862266 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/deeec30d-cad2-4dfd-adb6-64f6646581be-host-run-ovn-kubernetes\") pod \"ovnkube-node-psjpm\" (UID: \"deeec30d-cad2-4dfd-adb6-64f6646581be\") " pod="openshift-ovn-kubernetes/ovnkube-node-psjpm" Dec 03 14:09:50 crc kubenswrapper[4805]: I1203 14:09:50.862285 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/deeec30d-cad2-4dfd-adb6-64f6646581be-log-socket\") pod \"ovnkube-node-psjpm\" (UID: \"deeec30d-cad2-4dfd-adb6-64f6646581be\") " pod="openshift-ovn-kubernetes/ovnkube-node-psjpm" Dec 03 14:09:50 crc kubenswrapper[4805]: I1203 14:09:50.862304 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/deeec30d-cad2-4dfd-adb6-64f6646581be-host-run-netns\") pod \"ovnkube-node-psjpm\" (UID: \"deeec30d-cad2-4dfd-adb6-64f6646581be\") " pod="openshift-ovn-kubernetes/ovnkube-node-psjpm" Dec 03 14:09:50 crc kubenswrapper[4805]: I1203 14:09:50.862323 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/deeec30d-cad2-4dfd-adb6-64f6646581be-ovn-node-metrics-cert\") pod \"ovnkube-node-psjpm\" (UID: \"deeec30d-cad2-4dfd-adb6-64f6646581be\") " pod="openshift-ovn-kubernetes/ovnkube-node-psjpm" Dec 03 14:09:50 crc kubenswrapper[4805]: I1203 14:09:50.862354 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/deeec30d-cad2-4dfd-adb6-64f6646581be-run-systemd\") pod \"ovnkube-node-psjpm\" (UID: \"deeec30d-cad2-4dfd-adb6-64f6646581be\") " pod="openshift-ovn-kubernetes/ovnkube-node-psjpm" Dec 03 14:09:50 crc kubenswrapper[4805]: I1203 14:09:50.862375 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/deeec30d-cad2-4dfd-adb6-64f6646581be-run-openvswitch\") pod \"ovnkube-node-psjpm\" (UID: \"deeec30d-cad2-4dfd-adb6-64f6646581be\") " pod="openshift-ovn-kubernetes/ovnkube-node-psjpm" Dec 03 14:09:50 crc kubenswrapper[4805]: I1203 14:09:50.862405 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/deeec30d-cad2-4dfd-adb6-64f6646581be-host-cni-bin\") pod \"ovnkube-node-psjpm\" (UID: \"deeec30d-cad2-4dfd-adb6-64f6646581be\") " pod="openshift-ovn-kubernetes/ovnkube-node-psjpm" Dec 03 14:09:50 crc kubenswrapper[4805]: I1203 14:09:50.862423 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/deeec30d-cad2-4dfd-adb6-64f6646581be-ovnkube-config\") pod \"ovnkube-node-psjpm\" (UID: \"deeec30d-cad2-4dfd-adb6-64f6646581be\") " pod="openshift-ovn-kubernetes/ovnkube-node-psjpm" Dec 03 14:09:50 crc kubenswrapper[4805]: I1203 14:09:50.862471 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/deeec30d-cad2-4dfd-adb6-64f6646581be-host-slash\") pod \"ovnkube-node-psjpm\" (UID: \"deeec30d-cad2-4dfd-adb6-64f6646581be\") " pod="openshift-ovn-kubernetes/ovnkube-node-psjpm" Dec 03 14:09:50 crc kubenswrapper[4805]: I1203 14:09:50.862491 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/deeec30d-cad2-4dfd-adb6-64f6646581be-node-log\") pod \"ovnkube-node-psjpm\" (UID: \"deeec30d-cad2-4dfd-adb6-64f6646581be\") " pod="openshift-ovn-kubernetes/ovnkube-node-psjpm" Dec 03 14:09:50 crc kubenswrapper[4805]: I1203 14:09:50.962869 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/deeec30d-cad2-4dfd-adb6-64f6646581be-host-run-netns\") pod \"ovnkube-node-psjpm\" (UID: \"deeec30d-cad2-4dfd-adb6-64f6646581be\") " pod="openshift-ovn-kubernetes/ovnkube-node-psjpm" Dec 03 14:09:50 crc kubenswrapper[4805]: I1203 14:09:50.963224 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/deeec30d-cad2-4dfd-adb6-64f6646581be-ovn-node-metrics-cert\") pod \"ovnkube-node-psjpm\" (UID: \"deeec30d-cad2-4dfd-adb6-64f6646581be\") " pod="openshift-ovn-kubernetes/ovnkube-node-psjpm" Dec 03 14:09:50 crc kubenswrapper[4805]: I1203 14:09:50.963249 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/deeec30d-cad2-4dfd-adb6-64f6646581be-run-openvswitch\") pod \"ovnkube-node-psjpm\" (UID: \"deeec30d-cad2-4dfd-adb6-64f6646581be\") " pod="openshift-ovn-kubernetes/ovnkube-node-psjpm" Dec 03 14:09:50 crc kubenswrapper[4805]: I1203 14:09:50.963281 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/deeec30d-cad2-4dfd-adb6-64f6646581be-run-systemd\") pod \"ovnkube-node-psjpm\" (UID: \"deeec30d-cad2-4dfd-adb6-64f6646581be\") " pod="openshift-ovn-kubernetes/ovnkube-node-psjpm" Dec 03 14:09:50 crc kubenswrapper[4805]: I1203 14:09:50.963301 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/deeec30d-cad2-4dfd-adb6-64f6646581be-host-cni-bin\") pod \"ovnkube-node-psjpm\" (UID: \"deeec30d-cad2-4dfd-adb6-64f6646581be\") " pod="openshift-ovn-kubernetes/ovnkube-node-psjpm" Dec 03 14:09:50 crc kubenswrapper[4805]: I1203 14:09:50.963319 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/deeec30d-cad2-4dfd-adb6-64f6646581be-ovnkube-config\") pod \"ovnkube-node-psjpm\" (UID: \"deeec30d-cad2-4dfd-adb6-64f6646581be\") " pod="openshift-ovn-kubernetes/ovnkube-node-psjpm" Dec 03 14:09:50 crc kubenswrapper[4805]: I1203 14:09:50.963352 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/deeec30d-cad2-4dfd-adb6-64f6646581be-host-slash\") pod \"ovnkube-node-psjpm\" (UID: \"deeec30d-cad2-4dfd-adb6-64f6646581be\") " pod="openshift-ovn-kubernetes/ovnkube-node-psjpm" Dec 03 14:09:50 crc kubenswrapper[4805]: I1203 14:09:50.963375 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/deeec30d-cad2-4dfd-adb6-64f6646581be-node-log\") pod \"ovnkube-node-psjpm\" (UID: \"deeec30d-cad2-4dfd-adb6-64f6646581be\") " pod="openshift-ovn-kubernetes/ovnkube-node-psjpm" Dec 03 14:09:50 crc kubenswrapper[4805]: I1203 14:09:50.963396 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/deeec30d-cad2-4dfd-adb6-64f6646581be-ovnkube-script-lib\") pod \"ovnkube-node-psjpm\" (UID: \"deeec30d-cad2-4dfd-adb6-64f6646581be\") " pod="openshift-ovn-kubernetes/ovnkube-node-psjpm" Dec 03 14:09:50 crc kubenswrapper[4805]: I1203 14:09:50.963418 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/deeec30d-cad2-4dfd-adb6-64f6646581be-systemd-units\") pod \"ovnkube-node-psjpm\" (UID: \"deeec30d-cad2-4dfd-adb6-64f6646581be\") " pod="openshift-ovn-kubernetes/ovnkube-node-psjpm" Dec 03 14:09:50 crc kubenswrapper[4805]: I1203 14:09:50.963441 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/deeec30d-cad2-4dfd-adb6-64f6646581be-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-psjpm\" (UID: \"deeec30d-cad2-4dfd-adb6-64f6646581be\") " pod="openshift-ovn-kubernetes/ovnkube-node-psjpm" Dec 03 14:09:50 crc kubenswrapper[4805]: I1203 14:09:50.963515 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/deeec30d-cad2-4dfd-adb6-64f6646581be-etc-openvswitch\") pod \"ovnkube-node-psjpm\" (UID: \"deeec30d-cad2-4dfd-adb6-64f6646581be\") " pod="openshift-ovn-kubernetes/ovnkube-node-psjpm" Dec 03 14:09:50 crc kubenswrapper[4805]: I1203 14:09:50.963538 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wdlzx\" (UniqueName: \"kubernetes.io/projected/deeec30d-cad2-4dfd-adb6-64f6646581be-kube-api-access-wdlzx\") pod \"ovnkube-node-psjpm\" (UID: \"deeec30d-cad2-4dfd-adb6-64f6646581be\") " pod="openshift-ovn-kubernetes/ovnkube-node-psjpm" Dec 03 14:09:50 crc kubenswrapper[4805]: I1203 14:09:50.963562 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/deeec30d-cad2-4dfd-adb6-64f6646581be-host-kubelet\") pod \"ovnkube-node-psjpm\" (UID: \"deeec30d-cad2-4dfd-adb6-64f6646581be\") " pod="openshift-ovn-kubernetes/ovnkube-node-psjpm" Dec 03 14:09:50 crc kubenswrapper[4805]: I1203 14:09:50.963582 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/deeec30d-cad2-4dfd-adb6-64f6646581be-var-lib-openvswitch\") pod \"ovnkube-node-psjpm\" (UID: \"deeec30d-cad2-4dfd-adb6-64f6646581be\") " pod="openshift-ovn-kubernetes/ovnkube-node-psjpm" Dec 03 14:09:50 crc kubenswrapper[4805]: I1203 14:09:50.963603 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/deeec30d-cad2-4dfd-adb6-64f6646581be-run-ovn\") pod \"ovnkube-node-psjpm\" (UID: \"deeec30d-cad2-4dfd-adb6-64f6646581be\") " pod="openshift-ovn-kubernetes/ovnkube-node-psjpm" Dec 03 14:09:50 crc kubenswrapper[4805]: I1203 14:09:50.963623 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/deeec30d-cad2-4dfd-adb6-64f6646581be-host-cni-netd\") pod \"ovnkube-node-psjpm\" (UID: \"deeec30d-cad2-4dfd-adb6-64f6646581be\") " pod="openshift-ovn-kubernetes/ovnkube-node-psjpm" Dec 03 14:09:50 crc kubenswrapper[4805]: I1203 14:09:50.963648 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/deeec30d-cad2-4dfd-adb6-64f6646581be-env-overrides\") pod \"ovnkube-node-psjpm\" (UID: \"deeec30d-cad2-4dfd-adb6-64f6646581be\") " pod="openshift-ovn-kubernetes/ovnkube-node-psjpm" Dec 03 14:09:50 crc kubenswrapper[4805]: I1203 14:09:50.963670 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/deeec30d-cad2-4dfd-adb6-64f6646581be-log-socket\") pod \"ovnkube-node-psjpm\" (UID: \"deeec30d-cad2-4dfd-adb6-64f6646581be\") " pod="openshift-ovn-kubernetes/ovnkube-node-psjpm" Dec 03 14:09:50 crc kubenswrapper[4805]: I1203 14:09:50.963692 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/deeec30d-cad2-4dfd-adb6-64f6646581be-host-run-ovn-kubernetes\") pod \"ovnkube-node-psjpm\" (UID: \"deeec30d-cad2-4dfd-adb6-64f6646581be\") " pod="openshift-ovn-kubernetes/ovnkube-node-psjpm" Dec 03 14:09:50 crc kubenswrapper[4805]: I1203 14:09:50.963769 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/deeec30d-cad2-4dfd-adb6-64f6646581be-host-run-ovn-kubernetes\") pod \"ovnkube-node-psjpm\" (UID: \"deeec30d-cad2-4dfd-adb6-64f6646581be\") " pod="openshift-ovn-kubernetes/ovnkube-node-psjpm" Dec 03 14:09:50 crc kubenswrapper[4805]: I1203 14:09:50.963818 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/deeec30d-cad2-4dfd-adb6-64f6646581be-host-run-netns\") pod \"ovnkube-node-psjpm\" (UID: \"deeec30d-cad2-4dfd-adb6-64f6646581be\") " pod="openshift-ovn-kubernetes/ovnkube-node-psjpm" Dec 03 14:09:50 crc kubenswrapper[4805]: I1203 14:09:50.964248 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/deeec30d-cad2-4dfd-adb6-64f6646581be-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-psjpm\" (UID: \"deeec30d-cad2-4dfd-adb6-64f6646581be\") " pod="openshift-ovn-kubernetes/ovnkube-node-psjpm" Dec 03 14:09:50 crc kubenswrapper[4805]: I1203 14:09:50.964296 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/deeec30d-cad2-4dfd-adb6-64f6646581be-run-openvswitch\") pod \"ovnkube-node-psjpm\" (UID: \"deeec30d-cad2-4dfd-adb6-64f6646581be\") " pod="openshift-ovn-kubernetes/ovnkube-node-psjpm" Dec 03 14:09:50 crc kubenswrapper[4805]: I1203 14:09:50.964330 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/deeec30d-cad2-4dfd-adb6-64f6646581be-run-systemd\") pod \"ovnkube-node-psjpm\" (UID: \"deeec30d-cad2-4dfd-adb6-64f6646581be\") " pod="openshift-ovn-kubernetes/ovnkube-node-psjpm" Dec 03 14:09:50 crc kubenswrapper[4805]: I1203 14:09:50.964391 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/deeec30d-cad2-4dfd-adb6-64f6646581be-host-cni-bin\") pod \"ovnkube-node-psjpm\" (UID: \"deeec30d-cad2-4dfd-adb6-64f6646581be\") " pod="openshift-ovn-kubernetes/ovnkube-node-psjpm" Dec 03 14:09:50 crc kubenswrapper[4805]: I1203 14:09:50.964532 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/deeec30d-cad2-4dfd-adb6-64f6646581be-var-lib-openvswitch\") pod \"ovnkube-node-psjpm\" (UID: \"deeec30d-cad2-4dfd-adb6-64f6646581be\") " pod="openshift-ovn-kubernetes/ovnkube-node-psjpm" Dec 03 14:09:50 crc kubenswrapper[4805]: I1203 14:09:50.964607 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/deeec30d-cad2-4dfd-adb6-64f6646581be-etc-openvswitch\") pod \"ovnkube-node-psjpm\" (UID: \"deeec30d-cad2-4dfd-adb6-64f6646581be\") " pod="openshift-ovn-kubernetes/ovnkube-node-psjpm" Dec 03 14:09:50 crc kubenswrapper[4805]: I1203 14:09:50.964743 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/deeec30d-cad2-4dfd-adb6-64f6646581be-host-cni-netd\") pod \"ovnkube-node-psjpm\" (UID: \"deeec30d-cad2-4dfd-adb6-64f6646581be\") " pod="openshift-ovn-kubernetes/ovnkube-node-psjpm" Dec 03 14:09:50 crc kubenswrapper[4805]: I1203 14:09:50.964788 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/deeec30d-cad2-4dfd-adb6-64f6646581be-run-ovn\") pod \"ovnkube-node-psjpm\" (UID: \"deeec30d-cad2-4dfd-adb6-64f6646581be\") " pod="openshift-ovn-kubernetes/ovnkube-node-psjpm" Dec 03 14:09:50 crc kubenswrapper[4805]: I1203 14:09:50.964819 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/deeec30d-cad2-4dfd-adb6-64f6646581be-node-log\") pod \"ovnkube-node-psjpm\" (UID: \"deeec30d-cad2-4dfd-adb6-64f6646581be\") " pod="openshift-ovn-kubernetes/ovnkube-node-psjpm" Dec 03 14:09:50 crc kubenswrapper[4805]: I1203 14:09:50.964869 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/deeec30d-cad2-4dfd-adb6-64f6646581be-host-slash\") pod \"ovnkube-node-psjpm\" (UID: \"deeec30d-cad2-4dfd-adb6-64f6646581be\") " pod="openshift-ovn-kubernetes/ovnkube-node-psjpm" Dec 03 14:09:50 crc kubenswrapper[4805]: I1203 14:09:50.964902 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/deeec30d-cad2-4dfd-adb6-64f6646581be-host-kubelet\") pod \"ovnkube-node-psjpm\" (UID: \"deeec30d-cad2-4dfd-adb6-64f6646581be\") " pod="openshift-ovn-kubernetes/ovnkube-node-psjpm" Dec 03 14:09:50 crc kubenswrapper[4805]: I1203 14:09:50.965062 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/deeec30d-cad2-4dfd-adb6-64f6646581be-ovnkube-config\") pod \"ovnkube-node-psjpm\" (UID: \"deeec30d-cad2-4dfd-adb6-64f6646581be\") " pod="openshift-ovn-kubernetes/ovnkube-node-psjpm" Dec 03 14:09:50 crc kubenswrapper[4805]: I1203 14:09:50.965110 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/deeec30d-cad2-4dfd-adb6-64f6646581be-systemd-units\") pod \"ovnkube-node-psjpm\" (UID: \"deeec30d-cad2-4dfd-adb6-64f6646581be\") " pod="openshift-ovn-kubernetes/ovnkube-node-psjpm" Dec 03 14:09:50 crc kubenswrapper[4805]: I1203 14:09:50.965147 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/deeec30d-cad2-4dfd-adb6-64f6646581be-log-socket\") pod \"ovnkube-node-psjpm\" (UID: \"deeec30d-cad2-4dfd-adb6-64f6646581be\") " pod="openshift-ovn-kubernetes/ovnkube-node-psjpm" Dec 03 14:09:50 crc kubenswrapper[4805]: I1203 14:09:50.965287 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/deeec30d-cad2-4dfd-adb6-64f6646581be-env-overrides\") pod \"ovnkube-node-psjpm\" (UID: \"deeec30d-cad2-4dfd-adb6-64f6646581be\") " pod="openshift-ovn-kubernetes/ovnkube-node-psjpm" Dec 03 14:09:50 crc kubenswrapper[4805]: I1203 14:09:50.968152 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/deeec30d-cad2-4dfd-adb6-64f6646581be-ovnkube-script-lib\") pod \"ovnkube-node-psjpm\" (UID: \"deeec30d-cad2-4dfd-adb6-64f6646581be\") " pod="openshift-ovn-kubernetes/ovnkube-node-psjpm" Dec 03 14:09:50 crc kubenswrapper[4805]: I1203 14:09:50.973905 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/deeec30d-cad2-4dfd-adb6-64f6646581be-ovn-node-metrics-cert\") pod \"ovnkube-node-psjpm\" (UID: \"deeec30d-cad2-4dfd-adb6-64f6646581be\") " pod="openshift-ovn-kubernetes/ovnkube-node-psjpm" Dec 03 14:09:50 crc kubenswrapper[4805]: I1203 14:09:50.988347 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wdlzx\" (UniqueName: \"kubernetes.io/projected/deeec30d-cad2-4dfd-adb6-64f6646581be-kube-api-access-wdlzx\") pod \"ovnkube-node-psjpm\" (UID: \"deeec30d-cad2-4dfd-adb6-64f6646581be\") " pod="openshift-ovn-kubernetes/ovnkube-node-psjpm" Dec 03 14:09:51 crc kubenswrapper[4805]: I1203 14:09:51.153892 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-psjpm" Dec 03 14:09:51 crc kubenswrapper[4805]: I1203 14:09:51.400774 4805 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"kube-root-ca.crt" Dec 03 14:09:51 crc kubenswrapper[4805]: I1203 14:09:51.522571 4805 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"openshift-service-ca.crt" Dec 03 14:09:51 crc kubenswrapper[4805]: I1203 14:09:51.563274 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-pdzgk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"860769de-6a5b-4cb9-a752-264e7e2e785f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:50Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:50Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:50Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ggn4n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T14:09:50Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-pdzgk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T14:09:51Z is after 2025-08-24T17:21:41Z" Dec 03 14:09:51 crc kubenswrapper[4805]: I1203 14:09:51.565209 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-daemon-dockercfg-r5tcq" Dec 03 14:09:51 crc kubenswrapper[4805]: I1203 14:09:51.577650 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:49Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:49Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://38339331a3209d6787d1def81b72f4a8514bf91080a139cad6d8ac5cc6bef10e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T14:09:51Z is after 2025-08-24T17:21:41Z" Dec 03 14:09:51 crc kubenswrapper[4805]: I1203 14:09:51.580486 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"proxy-tls" Dec 03 14:09:51 crc kubenswrapper[4805]: I1203 14:09:51.591118 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:48Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:48Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T14:09:51Z is after 2025-08-24T17:21:41Z" Dec 03 14:09:51 crc kubenswrapper[4805]: I1203 14:09:51.601404 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:48Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:48Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T14:09:51Z is after 2025-08-24T17:21:41Z" Dec 03 14:09:51 crc kubenswrapper[4805]: I1203 14:09:51.618189 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ced8291d-0ce5-4e74-9eaf-2bf033774039\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:30Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:30Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3e5a34636f0a8c4d9179074596d28bc2eb7a504573398c596a637defd829907a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://29cd95be3f964415943c873e7914ffcea4abe50dfdf17f8a9cd03fc38df35981\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://951173b3d3d581275a6f6e43209b5e0611a701ca7a10f81504d723e24a2d21a3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e64730957bbbb2feb4d506e0ed9731a0594bf66fa1e8fcdb8ec944ded6308eed\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5da55f769e23ad52bac9c77da62c78e70fa34db8ed1a435e4eaf4dad3fd2b68a\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-03T14:09:48Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1203 14:09:42.992179 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1203 14:09:42.993233 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1830926909/tls.crt::/tmp/serving-cert-1830926909/tls.key\\\\\\\"\\\\nI1203 14:09:48.622403 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1203 14:09:48.626476 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1203 14:09:48.626527 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1203 14:09:48.626574 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1203 14:09:48.626582 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1203 14:09:48.633990 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1203 14:09:48.634024 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 14:09:48.634029 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 14:09:48.634034 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1203 14:09:48.634038 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1203 14:09:48.634042 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1203 14:09:48.634045 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1203 14:09:48.634344 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1203 14:09:48.644497 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T14:09:32Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bd92a160eb530c5687ebd7cd5b83b41820a2b808a0fd9514e5ace9d237e1bd26\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:32Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://af9c940135012c629447528c61743ed1e58708ab57ce76604c6da0479aa459cf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://af9c940135012c629447528c61743ed1e58708ab57ce76604c6da0479aa459cf\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T14:09:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T14:09:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T14:09:30Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T14:09:51Z is after 2025-08-24T17:21:41Z" Dec 03 14:09:51 crc kubenswrapper[4805]: I1203 14:09:51.635260 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:48Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:48Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T14:09:51Z is after 2025-08-24T17:21:41Z" Dec 03 14:09:51 crc kubenswrapper[4805]: I1203 14:09:51.649190 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-k4mhs" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c09c1cdc-f461-45c9-8444-5b8764bccabc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:50Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:50Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:50Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qdcd7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T14:09:50Z\\\"}}\" for pod \"openshift-multus\"/\"multus-k4mhs\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T14:09:51Z is after 2025-08-24T17:21:41Z" Dec 03 14:09:51 crc kubenswrapper[4805]: I1203 14:09:51.655831 4805 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"default-cni-sysctl-allowlist" Dec 03 14:09:51 crc kubenswrapper[4805]: I1203 14:09:51.664252 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-bpk49" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"de00a81f-82e1-4223-9614-f98606b9a8ef\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:50Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:50Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:50Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:50Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-25d7t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-25d7t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-25d7t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-25d7t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-25d7t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-25d7t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-25d7t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T14:09:50Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-bpk49\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T14:09:51Z is after 2025-08-24T17:21:41Z" Dec 03 14:09:51 crc kubenswrapper[4805]: I1203 14:09:51.679660 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:48Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:48Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T14:09:51Z is after 2025-08-24T17:21:41Z" Dec 03 14:09:51 crc kubenswrapper[4805]: I1203 14:09:51.696630 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:49Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:49Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://541ace7ccbb66dbd191c3ca8c1e4479e3edbf80dfc67326bea8aaee81fdcfc3c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://821e3e9a902aacecf6c41e8b8d8aa1ac951ce2fbb6505b7148b16348a3754e55\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T14:09:51Z is after 2025-08-24T17:21:41Z" Dec 03 14:09:51 crc kubenswrapper[4805]: I1203 14:09:51.708990 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-gskh4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6098937f-e3f6-45e8-a647-4994a79cd711\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:50Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:50Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:50Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nwrw9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nwrw9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T14:09:50Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-gskh4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T14:09:51Z is after 2025-08-24T17:21:41Z" Dec 03 14:09:51 crc kubenswrapper[4805]: I1203 14:09:51.738332 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-psjpm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"deeec30d-cad2-4dfd-adb6-64f6646581be\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:50Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:50Z\\\",\\\"message\\\":\\\"containers with incomplete status: [kubecfg-setup]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:50Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:50Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wdlzx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wdlzx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wdlzx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wdlzx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wdlzx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wdlzx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wdlzx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wdlzx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wdlzx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T14:09:50Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-psjpm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T14:09:51Z is after 2025-08-24T17:21:41Z" Dec 03 14:09:51 crc kubenswrapper[4805]: I1203 14:09:51.747785 4805 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"kube-root-ca.crt" Dec 03 14:09:51 crc kubenswrapper[4805]: I1203 14:09:51.749507 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-pdzgk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"860769de-6a5b-4cb9-a752-264e7e2e785f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:50Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:50Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:50Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ggn4n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T14:09:50Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-pdzgk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T14:09:51Z is after 2025-08-24T17:21:41Z" Dec 03 14:09:51 crc kubenswrapper[4805]: I1203 14:09:51.766948 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:48Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:48Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T14:09:51Z is after 2025-08-24T17:21:41Z" Dec 03 14:09:51 crc kubenswrapper[4805]: I1203 14:09:51.787415 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ced8291d-0ce5-4e74-9eaf-2bf033774039\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:30Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:30Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3e5a34636f0a8c4d9179074596d28bc2eb7a504573398c596a637defd829907a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://29cd95be3f964415943c873e7914ffcea4abe50dfdf17f8a9cd03fc38df35981\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://951173b3d3d581275a6f6e43209b5e0611a701ca7a10f81504d723e24a2d21a3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e64730957bbbb2feb4d506e0ed9731a0594bf66fa1e8fcdb8ec944ded6308eed\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5da55f769e23ad52bac9c77da62c78e70fa34db8ed1a435e4eaf4dad3fd2b68a\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-03T14:09:48Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1203 14:09:42.992179 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1203 14:09:42.993233 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1830926909/tls.crt::/tmp/serving-cert-1830926909/tls.key\\\\\\\"\\\\nI1203 14:09:48.622403 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1203 14:09:48.626476 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1203 14:09:48.626527 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1203 14:09:48.626574 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1203 14:09:48.626582 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1203 14:09:48.633990 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1203 14:09:48.634024 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 14:09:48.634029 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 14:09:48.634034 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1203 14:09:48.634038 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1203 14:09:48.634042 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1203 14:09:48.634045 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1203 14:09:48.634344 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1203 14:09:48.644497 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T14:09:32Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bd92a160eb530c5687ebd7cd5b83b41820a2b808a0fd9514e5ace9d237e1bd26\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:32Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://af9c940135012c629447528c61743ed1e58708ab57ce76604c6da0479aa459cf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://af9c940135012c629447528c61743ed1e58708ab57ce76604c6da0479aa459cf\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T14:09:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T14:09:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T14:09:30Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T14:09:51Z is after 2025-08-24T17:21:41Z" Dec 03 14:09:51 crc kubenswrapper[4805]: I1203 14:09:51.799236 4805 generic.go:334] "Generic (PLEG): container finished" podID="deeec30d-cad2-4dfd-adb6-64f6646581be" containerID="2f875252fbe6a673a72fc9c56bb96899f62732305ad77b3108e3b244bfa91ecb" exitCode=0 Dec 03 14:09:51 crc kubenswrapper[4805]: I1203 14:09:51.799310 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-psjpm" event={"ID":"deeec30d-cad2-4dfd-adb6-64f6646581be","Type":"ContainerDied","Data":"2f875252fbe6a673a72fc9c56bb96899f62732305ad77b3108e3b244bfa91ecb"} Dec 03 14:09:51 crc kubenswrapper[4805]: I1203 14:09:51.799338 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-psjpm" event={"ID":"deeec30d-cad2-4dfd-adb6-64f6646581be","Type":"ContainerStarted","Data":"fd34043a1f52c16209aad4766079aff91088a34a1a1e9a27a022a8c8083c8d86"} Dec 03 14:09:51 crc kubenswrapper[4805]: I1203 14:09:51.800749 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/node-resolver-pdzgk" event={"ID":"860769de-6a5b-4cb9-a752-264e7e2e785f","Type":"ContainerStarted","Data":"5e2f0a0cc21505d609f81918365d5e4f5d21959ac12ab147d6a51601c9de4c14"} Dec 03 14:09:51 crc kubenswrapper[4805]: I1203 14:09:51.802565 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-gskh4" event={"ID":"6098937f-e3f6-45e8-a647-4994a79cd711","Type":"ContainerStarted","Data":"28d5653f26ec61647808b3a733ff7eac70aa69ac184660f24147e8cfaf0eb802"} Dec 03 14:09:51 crc kubenswrapper[4805]: I1203 14:09:51.802586 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-gskh4" event={"ID":"6098937f-e3f6-45e8-a647-4994a79cd711","Type":"ContainerStarted","Data":"9ad6b68a42fd33ee167e67873c6eed3f11cdf6dc71d22e952c021436eb967dd9"} Dec 03 14:09:51 crc kubenswrapper[4805]: I1203 14:09:51.805698 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:49Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:49Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://38339331a3209d6787d1def81b72f4a8514bf91080a139cad6d8ac5cc6bef10e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T14:09:51Z is after 2025-08-24T17:21:41Z" Dec 03 14:09:51 crc kubenswrapper[4805]: I1203 14:09:51.811461 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-k4mhs" event={"ID":"c09c1cdc-f461-45c9-8444-5b8764bccabc","Type":"ContainerStarted","Data":"3354c1a44f891182d04116fc76234a2b337d50567d032a3584f902a9516387cc"} Dec 03 14:09:51 crc kubenswrapper[4805]: I1203 14:09:51.818458 4805 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 03 14:09:51 crc kubenswrapper[4805]: I1203 14:09:51.819439 4805 generic.go:334] "Generic (PLEG): container finished" podID="de00a81f-82e1-4223-9614-f98606b9a8ef" containerID="7741e0e70416ebe094bfabc6009e98d7d4c0fbcba1fc82df13b9ba40774b456a" exitCode=0 Dec 03 14:09:51 crc kubenswrapper[4805]: I1203 14:09:51.819533 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-bpk49" event={"ID":"de00a81f-82e1-4223-9614-f98606b9a8ef","Type":"ContainerDied","Data":"7741e0e70416ebe094bfabc6009e98d7d4c0fbcba1fc82df13b9ba40774b456a"} Dec 03 14:09:51 crc kubenswrapper[4805]: I1203 14:09:51.819575 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-bpk49" event={"ID":"de00a81f-82e1-4223-9614-f98606b9a8ef","Type":"ContainerStarted","Data":"d24573710ca84801bf2d9f2139eb66525226fee9d5a8ed9bac92ec635401a7f3"} Dec 03 14:09:51 crc kubenswrapper[4805]: I1203 14:09:51.820049 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 14:09:51 crc kubenswrapper[4805]: I1203 14:09:51.820075 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 14:09:51 crc kubenswrapper[4805]: I1203 14:09:51.820085 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 14:09:51 crc kubenswrapper[4805]: I1203 14:09:51.820201 4805 kubelet_node_status.go:76] "Attempting to register node" node="crc" Dec 03 14:09:51 crc kubenswrapper[4805]: I1203 14:09:51.821728 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:48Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:48Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T14:09:51Z is after 2025-08-24T17:21:41Z" Dec 03 14:09:51 crc kubenswrapper[4805]: I1203 14:09:51.822012 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" event={"ID":"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49","Type":"ContainerStarted","Data":"b3123d3fa879097db88ed2f3cfb05631f440c24c5405503b2de3546f12e2e899"} Dec 03 14:09:51 crc kubenswrapper[4805]: I1203 14:09:51.826760 4805 kubelet_node_status.go:115] "Node was previously registered" node="crc" Dec 03 14:09:51 crc kubenswrapper[4805]: I1203 14:09:51.827050 4805 kubelet_node_status.go:79] "Successfully registered node" node="crc" Dec 03 14:09:51 crc kubenswrapper[4805]: I1203 14:09:51.827919 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 14:09:51 crc kubenswrapper[4805]: I1203 14:09:51.827947 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 14:09:51 crc kubenswrapper[4805]: I1203 14:09:51.827954 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 14:09:51 crc kubenswrapper[4805]: I1203 14:09:51.827968 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 14:09:51 crc kubenswrapper[4805]: I1203 14:09:51.827981 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T14:09:51Z","lastTransitionTime":"2025-12-03T14:09:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 14:09:51 crc kubenswrapper[4805]: I1203 14:09:51.835217 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ced8291d-0ce5-4e74-9eaf-2bf033774039\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:30Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:30Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3e5a34636f0a8c4d9179074596d28bc2eb7a504573398c596a637defd829907a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://29cd95be3f964415943c873e7914ffcea4abe50dfdf17f8a9cd03fc38df35981\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://951173b3d3d581275a6f6e43209b5e0611a701ca7a10f81504d723e24a2d21a3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e64730957bbbb2feb4d506e0ed9731a0594bf66fa1e8fcdb8ec944ded6308eed\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5da55f769e23ad52bac9c77da62c78e70fa34db8ed1a435e4eaf4dad3fd2b68a\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-03T14:09:48Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1203 14:09:42.992179 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1203 14:09:42.993233 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1830926909/tls.crt::/tmp/serving-cert-1830926909/tls.key\\\\\\\"\\\\nI1203 14:09:48.622403 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1203 14:09:48.626476 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1203 14:09:48.626527 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1203 14:09:48.626574 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1203 14:09:48.626582 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1203 14:09:48.633990 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1203 14:09:48.634024 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 14:09:48.634029 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 14:09:48.634034 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1203 14:09:48.634038 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1203 14:09:48.634042 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1203 14:09:48.634045 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1203 14:09:48.634344 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1203 14:09:48.644497 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T14:09:32Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bd92a160eb530c5687ebd7cd5b83b41820a2b808a0fd9514e5ace9d237e1bd26\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:32Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://af9c940135012c629447528c61743ed1e58708ab57ce76604c6da0479aa459cf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://af9c940135012c629447528c61743ed1e58708ab57ce76604c6da0479aa459cf\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T14:09:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T14:09:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T14:09:30Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T14:09:51Z is after 2025-08-24T17:21:41Z" Dec 03 14:09:51 crc kubenswrapper[4805]: E1203 14:09:51.844183 4805 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T14:09:51Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:51Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T14:09:51Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:51Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T14:09:51Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:51Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T14:09:51Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:51Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"e4edfbcd-1de1-486e-9c0b-65f9fbba52ad\\\",\\\"systemUUID\\\":\\\"7a729b24-19ef-490e-ab4c-d31c52ed5658\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T14:09:51Z is after 2025-08-24T17:21:41Z" Dec 03 14:09:51 crc kubenswrapper[4805]: I1203 14:09:51.846970 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:49Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:49Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://38339331a3209d6787d1def81b72f4a8514bf91080a139cad6d8ac5cc6bef10e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T14:09:51Z is after 2025-08-24T17:21:41Z" Dec 03 14:09:51 crc kubenswrapper[4805]: I1203 14:09:51.850123 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 14:09:51 crc kubenswrapper[4805]: I1203 14:09:51.850183 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 14:09:51 crc kubenswrapper[4805]: I1203 14:09:51.850201 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 14:09:51 crc kubenswrapper[4805]: I1203 14:09:51.850224 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 14:09:51 crc kubenswrapper[4805]: I1203 14:09:51.850237 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T14:09:51Z","lastTransitionTime":"2025-12-03T14:09:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 14:09:51 crc kubenswrapper[4805]: I1203 14:09:51.859497 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:48Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:48Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T14:09:51Z is after 2025-08-24T17:21:41Z" Dec 03 14:09:51 crc kubenswrapper[4805]: I1203 14:09:51.863054 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-ancillary-tools-dockercfg-vnmsz" Dec 03 14:09:51 crc kubenswrapper[4805]: E1203 14:09:51.864516 4805 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T14:09:51Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:51Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T14:09:51Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:51Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T14:09:51Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:51Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T14:09:51Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:51Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"e4edfbcd-1de1-486e-9c0b-65f9fbba52ad\\\",\\\"systemUUID\\\":\\\"7a729b24-19ef-490e-ab4c-d31c52ed5658\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T14:09:51Z is after 2025-08-24T17:21:41Z" Dec 03 14:09:51 crc kubenswrapper[4805]: I1203 14:09:51.865786 4805 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"kube-rbac-proxy" Dec 03 14:09:51 crc kubenswrapper[4805]: I1203 14:09:51.867751 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 14:09:51 crc kubenswrapper[4805]: I1203 14:09:51.867803 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 14:09:51 crc kubenswrapper[4805]: I1203 14:09:51.867818 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 14:09:51 crc kubenswrapper[4805]: I1203 14:09:51.867858 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 14:09:51 crc kubenswrapper[4805]: I1203 14:09:51.867873 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T14:09:51Z","lastTransitionTime":"2025-12-03T14:09:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 14:09:51 crc kubenswrapper[4805]: I1203 14:09:51.872674 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:48Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:48Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T14:09:51Z is after 2025-08-24T17:21:41Z" Dec 03 14:09:51 crc kubenswrapper[4805]: I1203 14:09:51.873458 4805 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"multus-daemon-config" Dec 03 14:09:51 crc kubenswrapper[4805]: E1203 14:09:51.879809 4805 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T14:09:51Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:51Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T14:09:51Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:51Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T14:09:51Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:51Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T14:09:51Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:51Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"e4edfbcd-1de1-486e-9c0b-65f9fbba52ad\\\",\\\"systemUUID\\\":\\\"7a729b24-19ef-490e-ab4c-d31c52ed5658\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T14:09:51Z is after 2025-08-24T17:21:41Z" Dec 03 14:09:51 crc kubenswrapper[4805]: I1203 14:09:51.883177 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"node-resolver-dockercfg-kz9s7" Dec 03 14:09:51 crc kubenswrapper[4805]: I1203 14:09:51.883392 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 14:09:51 crc kubenswrapper[4805]: I1203 14:09:51.883433 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 14:09:51 crc kubenswrapper[4805]: I1203 14:09:51.883444 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 14:09:51 crc kubenswrapper[4805]: I1203 14:09:51.883462 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 14:09:51 crc kubenswrapper[4805]: I1203 14:09:51.883545 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T14:09:51Z","lastTransitionTime":"2025-12-03T14:09:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 14:09:51 crc kubenswrapper[4805]: I1203 14:09:51.893330 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:48Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:48Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T14:09:51Z is after 2025-08-24T17:21:41Z" Dec 03 14:09:51 crc kubenswrapper[4805]: E1203 14:09:51.900250 4805 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T14:09:51Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:51Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T14:09:51Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:51Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T14:09:51Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:51Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T14:09:51Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:51Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"e4edfbcd-1de1-486e-9c0b-65f9fbba52ad\\\",\\\"systemUUID\\\":\\\"7a729b24-19ef-490e-ab4c-d31c52ed5658\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T14:09:51Z is after 2025-08-24T17:21:41Z" Dec 03 14:09:51 crc kubenswrapper[4805]: I1203 14:09:51.904350 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 14:09:51 crc kubenswrapper[4805]: I1203 14:09:51.904388 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 14:09:51 crc kubenswrapper[4805]: I1203 14:09:51.904402 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 14:09:51 crc kubenswrapper[4805]: I1203 14:09:51.904419 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 14:09:51 crc kubenswrapper[4805]: I1203 14:09:51.904431 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T14:09:51Z","lastTransitionTime":"2025-12-03T14:09:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 14:09:51 crc kubenswrapper[4805]: I1203 14:09:51.907588 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-k4mhs" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c09c1cdc-f461-45c9-8444-5b8764bccabc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3354c1a44f891182d04116fc76234a2b337d50567d032a3584f902a9516387cc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qdcd7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T14:09:50Z\\\"}}\" for pod \"openshift-multus\"/\"multus-k4mhs\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T14:09:51Z is after 2025-08-24T17:21:41Z" Dec 03 14:09:51 crc kubenswrapper[4805]: E1203 14:09:51.923440 4805 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T14:09:51Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:51Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T14:09:51Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:51Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T14:09:51Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:51Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T14:09:51Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:51Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"e4edfbcd-1de1-486e-9c0b-65f9fbba52ad\\\",\\\"systemUUID\\\":\\\"7a729b24-19ef-490e-ab4c-d31c52ed5658\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T14:09:51Z is after 2025-08-24T17:21:41Z" Dec 03 14:09:51 crc kubenswrapper[4805]: E1203 14:09:51.923620 4805 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Dec 03 14:09:51 crc kubenswrapper[4805]: I1203 14:09:51.924185 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-bpk49" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"de00a81f-82e1-4223-9614-f98606b9a8ef\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:50Z\\\",\\\"message\\\":\\\"containers with incomplete status: [cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:50Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:50Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-25d7t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7741e0e70416ebe094bfabc6009e98d7d4c0fbcba1fc82df13b9ba40774b456a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7741e0e70416ebe094bfabc6009e98d7d4c0fbcba1fc82df13b9ba40774b456a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T14:09:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T14:09:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-25d7t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-25d7t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-25d7t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-25d7t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-25d7t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-25d7t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T14:09:50Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-bpk49\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T14:09:51Z is after 2025-08-24T17:21:41Z" Dec 03 14:09:51 crc kubenswrapper[4805]: I1203 14:09:51.926798 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 14:09:51 crc kubenswrapper[4805]: I1203 14:09:51.926877 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 14:09:51 crc kubenswrapper[4805]: I1203 14:09:51.926885 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 14:09:51 crc kubenswrapper[4805]: I1203 14:09:51.926899 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 14:09:51 crc kubenswrapper[4805]: I1203 14:09:51.926909 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T14:09:51Z","lastTransitionTime":"2025-12-03T14:09:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 14:09:51 crc kubenswrapper[4805]: I1203 14:09:51.935548 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:48Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:48Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T14:09:51Z is after 2025-08-24T17:21:41Z" Dec 03 14:09:51 crc kubenswrapper[4805]: I1203 14:09:51.948463 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:49Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:49Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://541ace7ccbb66dbd191c3ca8c1e4479e3edbf80dfc67326bea8aaee81fdcfc3c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://821e3e9a902aacecf6c41e8b8d8aa1ac951ce2fbb6505b7148b16348a3754e55\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T14:09:51Z is after 2025-08-24T17:21:41Z" Dec 03 14:09:51 crc kubenswrapper[4805]: I1203 14:09:51.959177 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-gskh4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6098937f-e3f6-45e8-a647-4994a79cd711\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://28d5653f26ec61647808b3a733ff7eac70aa69ac184660f24147e8cfaf0eb802\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nwrw9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9ad6b68a42fd33ee167e67873c6eed3f11cdf6dc71d22e952c021436eb967dd9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nwrw9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T14:09:50Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-gskh4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T14:09:51Z is after 2025-08-24T17:21:41Z" Dec 03 14:09:51 crc kubenswrapper[4805]: I1203 14:09:51.975991 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-psjpm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"deeec30d-cad2-4dfd-adb6-64f6646581be\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:50Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:50Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wdlzx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wdlzx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wdlzx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wdlzx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wdlzx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wdlzx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wdlzx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wdlzx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2f875252fbe6a673a72fc9c56bb96899f62732305ad77b3108e3b244bfa91ecb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2f875252fbe6a673a72fc9c56bb96899f62732305ad77b3108e3b244bfa91ecb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T14:09:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T14:09:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wdlzx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T14:09:50Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-psjpm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T14:09:51Z is after 2025-08-24T17:21:41Z" Dec 03 14:09:51 crc kubenswrapper[4805]: I1203 14:09:51.983008 4805 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/node-ca-bdtdg"] Dec 03 14:09:51 crc kubenswrapper[4805]: I1203 14:09:51.983180 4805 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"openshift-service-ca.crt" Dec 03 14:09:51 crc kubenswrapper[4805]: I1203 14:09:51.983397 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/node-ca-bdtdg" Dec 03 14:09:51 crc kubenswrapper[4805]: I1203 14:09:51.984986 4805 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"openshift-service-ca.crt" Dec 03 14:09:51 crc kubenswrapper[4805]: I1203 14:09:51.985522 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"node-ca-dockercfg-4777p" Dec 03 14:09:51 crc kubenswrapper[4805]: I1203 14:09:51.985701 4805 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"kube-root-ca.crt" Dec 03 14:09:51 crc kubenswrapper[4805]: I1203 14:09:51.986033 4805 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"image-registry-certificates" Dec 03 14:09:51 crc kubenswrapper[4805]: I1203 14:09:51.990187 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-pdzgk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"860769de-6a5b-4cb9-a752-264e7e2e785f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5e2f0a0cc21505d609f81918365d5e4f5d21959ac12ab147d6a51601c9de4c14\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ggn4n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T14:09:50Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-pdzgk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T14:09:51Z is after 2025-08-24T17:21:41Z" Dec 03 14:09:52 crc kubenswrapper[4805]: I1203 14:09:52.000763 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-pdzgk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"860769de-6a5b-4cb9-a752-264e7e2e785f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5e2f0a0cc21505d609f81918365d5e4f5d21959ac12ab147d6a51601c9de4c14\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ggn4n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T14:09:50Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-pdzgk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T14:09:51Z is after 2025-08-24T17:21:41Z" Dec 03 14:09:52 crc kubenswrapper[4805]: I1203 14:09:52.007259 4805 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"cni-copy-resources" Dec 03 14:09:52 crc kubenswrapper[4805]: I1203 14:09:52.012091 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:49Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:49Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://38339331a3209d6787d1def81b72f4a8514bf91080a139cad6d8ac5cc6bef10e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T14:09:52Z is after 2025-08-24T17:21:41Z" Dec 03 14:09:52 crc kubenswrapper[4805]: I1203 14:09:52.024328 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:48Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:48Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T14:09:52Z is after 2025-08-24T17:21:41Z" Dec 03 14:09:52 crc kubenswrapper[4805]: I1203 14:09:52.029822 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 14:09:52 crc kubenswrapper[4805]: I1203 14:09:52.029892 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 14:09:52 crc kubenswrapper[4805]: I1203 14:09:52.029909 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 14:09:52 crc kubenswrapper[4805]: I1203 14:09:52.029929 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 14:09:52 crc kubenswrapper[4805]: I1203 14:09:52.029940 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T14:09:52Z","lastTransitionTime":"2025-12-03T14:09:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 14:09:52 crc kubenswrapper[4805]: I1203 14:09:52.032764 4805 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"kube-root-ca.crt" Dec 03 14:09:52 crc kubenswrapper[4805]: I1203 14:09:52.037791 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:51Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:51Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b3123d3fa879097db88ed2f3cfb05631f440c24c5405503b2de3546f12e2e899\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T14:09:52Z is after 2025-08-24T17:21:41Z" Dec 03 14:09:52 crc kubenswrapper[4805]: I1203 14:09:52.048730 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-bdtdg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"53a0e9e3-b6d8-44b7-a32b-aa3450442865\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:51Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:51Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:51Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9ww54\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T14:09:51Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-bdtdg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T14:09:52Z is after 2025-08-24T17:21:41Z" Dec 03 14:09:52 crc kubenswrapper[4805]: I1203 14:09:52.053715 4805 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"openshift-service-ca.crt" Dec 03 14:09:52 crc kubenswrapper[4805]: I1203 14:09:52.070202 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ced8291d-0ce5-4e74-9eaf-2bf033774039\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:30Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:30Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3e5a34636f0a8c4d9179074596d28bc2eb7a504573398c596a637defd829907a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://29cd95be3f964415943c873e7914ffcea4abe50dfdf17f8a9cd03fc38df35981\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://951173b3d3d581275a6f6e43209b5e0611a701ca7a10f81504d723e24a2d21a3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e64730957bbbb2feb4d506e0ed9731a0594bf66fa1e8fcdb8ec944ded6308eed\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5da55f769e23ad52bac9c77da62c78e70fa34db8ed1a435e4eaf4dad3fd2b68a\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-03T14:09:48Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1203 14:09:42.992179 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1203 14:09:42.993233 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1830926909/tls.crt::/tmp/serving-cert-1830926909/tls.key\\\\\\\"\\\\nI1203 14:09:48.622403 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1203 14:09:48.626476 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1203 14:09:48.626527 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1203 14:09:48.626574 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1203 14:09:48.626582 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1203 14:09:48.633990 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1203 14:09:48.634024 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 14:09:48.634029 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 14:09:48.634034 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1203 14:09:48.634038 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1203 14:09:48.634042 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1203 14:09:48.634045 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1203 14:09:48.634344 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1203 14:09:48.644497 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T14:09:32Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bd92a160eb530c5687ebd7cd5b83b41820a2b808a0fd9514e5ace9d237e1bd26\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:32Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://af9c940135012c629447528c61743ed1e58708ab57ce76604c6da0479aa459cf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://af9c940135012c629447528c61743ed1e58708ab57ce76604c6da0479aa459cf\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T14:09:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T14:09:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T14:09:30Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T14:09:52Z is after 2025-08-24T17:21:41Z" Dec 03 14:09:52 crc kubenswrapper[4805]: I1203 14:09:52.073879 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/53a0e9e3-b6d8-44b7-a32b-aa3450442865-host\") pod \"node-ca-bdtdg\" (UID: \"53a0e9e3-b6d8-44b7-a32b-aa3450442865\") " pod="openshift-image-registry/node-ca-bdtdg" Dec 03 14:09:52 crc kubenswrapper[4805]: I1203 14:09:52.073921 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9ww54\" (UniqueName: \"kubernetes.io/projected/53a0e9e3-b6d8-44b7-a32b-aa3450442865-kube-api-access-9ww54\") pod \"node-ca-bdtdg\" (UID: \"53a0e9e3-b6d8-44b7-a32b-aa3450442865\") " pod="openshift-image-registry/node-ca-bdtdg" Dec 03 14:09:52 crc kubenswrapper[4805]: I1203 14:09:52.073946 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/53a0e9e3-b6d8-44b7-a32b-aa3450442865-serviceca\") pod \"node-ca-bdtdg\" (UID: \"53a0e9e3-b6d8-44b7-a32b-aa3450442865\") " pod="openshift-image-registry/node-ca-bdtdg" Dec 03 14:09:52 crc kubenswrapper[4805]: I1203 14:09:52.084683 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-bpk49" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"de00a81f-82e1-4223-9614-f98606b9a8ef\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:50Z\\\",\\\"message\\\":\\\"containers with incomplete status: [cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:50Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:50Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-25d7t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7741e0e70416ebe094bfabc6009e98d7d4c0fbcba1fc82df13b9ba40774b456a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7741e0e70416ebe094bfabc6009e98d7d4c0fbcba1fc82df13b9ba40774b456a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T14:09:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T14:09:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-25d7t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-25d7t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-25d7t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-25d7t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-25d7t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-25d7t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T14:09:50Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-bpk49\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T14:09:52Z is after 2025-08-24T17:21:41Z" Dec 03 14:09:52 crc kubenswrapper[4805]: I1203 14:09:52.100340 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:48Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:48Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T14:09:52Z is after 2025-08-24T17:21:41Z" Dec 03 14:09:52 crc kubenswrapper[4805]: I1203 14:09:52.116610 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-k4mhs" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c09c1cdc-f461-45c9-8444-5b8764bccabc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3354c1a44f891182d04116fc76234a2b337d50567d032a3584f902a9516387cc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qdcd7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T14:09:50Z\\\"}}\" for pod \"openshift-multus\"/\"multus-k4mhs\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T14:09:52Z is after 2025-08-24T17:21:41Z" Dec 03 14:09:52 crc kubenswrapper[4805]: I1203 14:09:52.119530 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"default-dockercfg-2q5b6" Dec 03 14:09:52 crc kubenswrapper[4805]: I1203 14:09:52.132572 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 14:09:52 crc kubenswrapper[4805]: I1203 14:09:52.132612 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 14:09:52 crc kubenswrapper[4805]: I1203 14:09:52.132623 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 14:09:52 crc kubenswrapper[4805]: I1203 14:09:52.132640 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 14:09:52 crc kubenswrapper[4805]: I1203 14:09:52.132651 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T14:09:52Z","lastTransitionTime":"2025-12-03T14:09:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 14:09:52 crc kubenswrapper[4805]: I1203 14:09:52.134599 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-psjpm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"deeec30d-cad2-4dfd-adb6-64f6646581be\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:50Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:50Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wdlzx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wdlzx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wdlzx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wdlzx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wdlzx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wdlzx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wdlzx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wdlzx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2f875252fbe6a673a72fc9c56bb96899f62732305ad77b3108e3b244bfa91ecb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2f875252fbe6a673a72fc9c56bb96899f62732305ad77b3108e3b244bfa91ecb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T14:09:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T14:09:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wdlzx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T14:09:50Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-psjpm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T14:09:52Z is after 2025-08-24T17:21:41Z" Dec 03 14:09:52 crc kubenswrapper[4805]: I1203 14:09:52.145769 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:48Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:48Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T14:09:52Z is after 2025-08-24T17:21:41Z" Dec 03 14:09:52 crc kubenswrapper[4805]: I1203 14:09:52.157132 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:49Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:49Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://541ace7ccbb66dbd191c3ca8c1e4479e3edbf80dfc67326bea8aaee81fdcfc3c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://821e3e9a902aacecf6c41e8b8d8aa1ac951ce2fbb6505b7148b16348a3754e55\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T14:09:52Z is after 2025-08-24T17:21:41Z" Dec 03 14:09:52 crc kubenswrapper[4805]: I1203 14:09:52.168508 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-gskh4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6098937f-e3f6-45e8-a647-4994a79cd711\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://28d5653f26ec61647808b3a733ff7eac70aa69ac184660f24147e8cfaf0eb802\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nwrw9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9ad6b68a42fd33ee167e67873c6eed3f11cdf6dc71d22e952c021436eb967dd9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nwrw9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T14:09:50Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-gskh4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T14:09:52Z is after 2025-08-24T17:21:41Z" Dec 03 14:09:52 crc kubenswrapper[4805]: I1203 14:09:52.174875 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/53a0e9e3-b6d8-44b7-a32b-aa3450442865-host\") pod \"node-ca-bdtdg\" (UID: \"53a0e9e3-b6d8-44b7-a32b-aa3450442865\") " pod="openshift-image-registry/node-ca-bdtdg" Dec 03 14:09:52 crc kubenswrapper[4805]: I1203 14:09:52.174917 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/53a0e9e3-b6d8-44b7-a32b-aa3450442865-serviceca\") pod \"node-ca-bdtdg\" (UID: \"53a0e9e3-b6d8-44b7-a32b-aa3450442865\") " pod="openshift-image-registry/node-ca-bdtdg" Dec 03 14:09:52 crc kubenswrapper[4805]: I1203 14:09:52.174932 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9ww54\" (UniqueName: \"kubernetes.io/projected/53a0e9e3-b6d8-44b7-a32b-aa3450442865-kube-api-access-9ww54\") pod \"node-ca-bdtdg\" (UID: \"53a0e9e3-b6d8-44b7-a32b-aa3450442865\") " pod="openshift-image-registry/node-ca-bdtdg" Dec 03 14:09:52 crc kubenswrapper[4805]: I1203 14:09:52.174954 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/53a0e9e3-b6d8-44b7-a32b-aa3450442865-host\") pod \"node-ca-bdtdg\" (UID: \"53a0e9e3-b6d8-44b7-a32b-aa3450442865\") " pod="openshift-image-registry/node-ca-bdtdg" Dec 03 14:09:52 crc kubenswrapper[4805]: I1203 14:09:52.176064 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/53a0e9e3-b6d8-44b7-a32b-aa3450442865-serviceca\") pod \"node-ca-bdtdg\" (UID: \"53a0e9e3-b6d8-44b7-a32b-aa3450442865\") " pod="openshift-image-registry/node-ca-bdtdg" Dec 03 14:09:52 crc kubenswrapper[4805]: I1203 14:09:52.197024 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9ww54\" (UniqueName: \"kubernetes.io/projected/53a0e9e3-b6d8-44b7-a32b-aa3450442865-kube-api-access-9ww54\") pod \"node-ca-bdtdg\" (UID: \"53a0e9e3-b6d8-44b7-a32b-aa3450442865\") " pod="openshift-image-registry/node-ca-bdtdg" Dec 03 14:09:52 crc kubenswrapper[4805]: I1203 14:09:52.234885 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 14:09:52 crc kubenswrapper[4805]: I1203 14:09:52.235114 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 14:09:52 crc kubenswrapper[4805]: I1203 14:09:52.235124 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 14:09:52 crc kubenswrapper[4805]: I1203 14:09:52.235136 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 14:09:52 crc kubenswrapper[4805]: I1203 14:09:52.235146 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T14:09:52Z","lastTransitionTime":"2025-12-03T14:09:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 14:09:52 crc kubenswrapper[4805]: I1203 14:09:52.298645 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/node-ca-bdtdg" Dec 03 14:09:52 crc kubenswrapper[4805]: W1203 14:09:52.311833 4805 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod53a0e9e3_b6d8_44b7_a32b_aa3450442865.slice/crio-2b3b83784b7416db5adda7de699affab83e457566616b4a50b419e100aa07a78 WatchSource:0}: Error finding container 2b3b83784b7416db5adda7de699affab83e457566616b4a50b419e100aa07a78: Status 404 returned error can't find the container with id 2b3b83784b7416db5adda7de699affab83e457566616b4a50b419e100aa07a78 Dec 03 14:09:52 crc kubenswrapper[4805]: I1203 14:09:52.338093 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 14:09:52 crc kubenswrapper[4805]: I1203 14:09:52.338141 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 14:09:52 crc kubenswrapper[4805]: I1203 14:09:52.338149 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 14:09:52 crc kubenswrapper[4805]: I1203 14:09:52.338169 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 14:09:52 crc kubenswrapper[4805]: I1203 14:09:52.338179 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T14:09:52Z","lastTransitionTime":"2025-12-03T14:09:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 14:09:52 crc kubenswrapper[4805]: I1203 14:09:52.376780 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 14:09:52 crc kubenswrapper[4805]: E1203 14:09:52.377005 4805 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-03 14:09:56.376972895 +0000 UTC m=+26.039889818 (durationBeforeRetry 4s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 14:09:52 crc kubenswrapper[4805]: I1203 14:09:52.440273 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 14:09:52 crc kubenswrapper[4805]: I1203 14:09:52.440318 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 14:09:52 crc kubenswrapper[4805]: I1203 14:09:52.440331 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 14:09:52 crc kubenswrapper[4805]: I1203 14:09:52.440344 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 14:09:52 crc kubenswrapper[4805]: I1203 14:09:52.440354 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T14:09:52Z","lastTransitionTime":"2025-12-03T14:09:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 14:09:52 crc kubenswrapper[4805]: I1203 14:09:52.477907 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 14:09:52 crc kubenswrapper[4805]: I1203 14:09:52.477972 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 14:09:52 crc kubenswrapper[4805]: I1203 14:09:52.478004 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 14:09:52 crc kubenswrapper[4805]: I1203 14:09:52.478030 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 14:09:52 crc kubenswrapper[4805]: E1203 14:09:52.478111 4805 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Dec 03 14:09:52 crc kubenswrapper[4805]: E1203 14:09:52.478114 4805 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 03 14:09:52 crc kubenswrapper[4805]: E1203 14:09:52.478177 4805 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-03 14:09:56.478161962 +0000 UTC m=+26.141078885 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Dec 03 14:09:52 crc kubenswrapper[4805]: E1203 14:09:52.478207 4805 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-03 14:09:56.478185842 +0000 UTC m=+26.141102855 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 03 14:09:52 crc kubenswrapper[4805]: E1203 14:09:52.478287 4805 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 03 14:09:52 crc kubenswrapper[4805]: E1203 14:09:52.478292 4805 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 03 14:09:52 crc kubenswrapper[4805]: E1203 14:09:52.478344 4805 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 03 14:09:52 crc kubenswrapper[4805]: E1203 14:09:52.478366 4805 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 03 14:09:52 crc kubenswrapper[4805]: E1203 14:09:52.478442 4805 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-12-03 14:09:56.478416949 +0000 UTC m=+26.141333902 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 03 14:09:52 crc kubenswrapper[4805]: E1203 14:09:52.478303 4805 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 03 14:09:52 crc kubenswrapper[4805]: E1203 14:09:52.478482 4805 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 03 14:09:52 crc kubenswrapper[4805]: E1203 14:09:52.478521 4805 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-12-03 14:09:56.478509191 +0000 UTC m=+26.141426154 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 03 14:09:52 crc kubenswrapper[4805]: I1203 14:09:52.542176 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 14:09:52 crc kubenswrapper[4805]: I1203 14:09:52.542373 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 14:09:52 crc kubenswrapper[4805]: I1203 14:09:52.542466 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 14:09:52 crc kubenswrapper[4805]: I1203 14:09:52.542537 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 14:09:52 crc kubenswrapper[4805]: I1203 14:09:52.542597 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T14:09:52Z","lastTransitionTime":"2025-12-03T14:09:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 14:09:52 crc kubenswrapper[4805]: I1203 14:09:52.645235 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 14:09:52 crc kubenswrapper[4805]: I1203 14:09:52.645279 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 14:09:52 crc kubenswrapper[4805]: I1203 14:09:52.645290 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 14:09:52 crc kubenswrapper[4805]: I1203 14:09:52.645308 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 14:09:52 crc kubenswrapper[4805]: I1203 14:09:52.645320 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T14:09:52Z","lastTransitionTime":"2025-12-03T14:09:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 14:09:52 crc kubenswrapper[4805]: I1203 14:09:52.694759 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 14:09:52 crc kubenswrapper[4805]: E1203 14:09:52.695027 4805 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 03 14:09:52 crc kubenswrapper[4805]: I1203 14:09:52.695319 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 14:09:52 crc kubenswrapper[4805]: I1203 14:09:52.695413 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 14:09:52 crc kubenswrapper[4805]: E1203 14:09:52.695483 4805 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 03 14:09:52 crc kubenswrapper[4805]: E1203 14:09:52.695568 4805 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 03 14:09:52 crc kubenswrapper[4805]: I1203 14:09:52.747257 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 14:09:52 crc kubenswrapper[4805]: I1203 14:09:52.747492 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 14:09:52 crc kubenswrapper[4805]: I1203 14:09:52.747554 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 14:09:52 crc kubenswrapper[4805]: I1203 14:09:52.747621 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 14:09:52 crc kubenswrapper[4805]: I1203 14:09:52.747682 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T14:09:52Z","lastTransitionTime":"2025-12-03T14:09:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 14:09:52 crc kubenswrapper[4805]: I1203 14:09:52.827636 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-psjpm" event={"ID":"deeec30d-cad2-4dfd-adb6-64f6646581be","Type":"ContainerStarted","Data":"6ce80ccafb3262a11fbe9923317a9d38b61e5dc5b3e7e7ae136f143d6b643fd7"} Dec 03 14:09:52 crc kubenswrapper[4805]: I1203 14:09:52.827698 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-psjpm" event={"ID":"deeec30d-cad2-4dfd-adb6-64f6646581be","Type":"ContainerStarted","Data":"47d76962beb8c0f196d50af96fb3ec611097823ca80a1c8bd718f2d18840128c"} Dec 03 14:09:52 crc kubenswrapper[4805]: I1203 14:09:52.832625 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-bpk49" event={"ID":"de00a81f-82e1-4223-9614-f98606b9a8ef","Type":"ContainerStarted","Data":"ee3499c02eb52bbf08c447d2433d435a74dfb70670dc4cbaac44f5fbff9e8a29"} Dec 03 14:09:52 crc kubenswrapper[4805]: I1203 14:09:52.834017 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/node-ca-bdtdg" event={"ID":"53a0e9e3-b6d8-44b7-a32b-aa3450442865","Type":"ContainerStarted","Data":"2b3b83784b7416db5adda7de699affab83e457566616b4a50b419e100aa07a78"} Dec 03 14:09:52 crc kubenswrapper[4805]: I1203 14:09:52.848490 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:48Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:48Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T14:09:52Z is after 2025-08-24T17:21:41Z" Dec 03 14:09:52 crc kubenswrapper[4805]: I1203 14:09:52.849678 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 14:09:52 crc kubenswrapper[4805]: I1203 14:09:52.849714 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 14:09:52 crc kubenswrapper[4805]: I1203 14:09:52.849725 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 14:09:52 crc kubenswrapper[4805]: I1203 14:09:52.849741 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 14:09:52 crc kubenswrapper[4805]: I1203 14:09:52.849754 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T14:09:52Z","lastTransitionTime":"2025-12-03T14:09:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 14:09:52 crc kubenswrapper[4805]: I1203 14:09:52.861179 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:49Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:49Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://541ace7ccbb66dbd191c3ca8c1e4479e3edbf80dfc67326bea8aaee81fdcfc3c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://821e3e9a902aacecf6c41e8b8d8aa1ac951ce2fbb6505b7148b16348a3754e55\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T14:09:52Z is after 2025-08-24T17:21:41Z" Dec 03 14:09:52 crc kubenswrapper[4805]: I1203 14:09:52.870178 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-gskh4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6098937f-e3f6-45e8-a647-4994a79cd711\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://28d5653f26ec61647808b3a733ff7eac70aa69ac184660f24147e8cfaf0eb802\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nwrw9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9ad6b68a42fd33ee167e67873c6eed3f11cdf6dc71d22e952c021436eb967dd9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nwrw9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T14:09:50Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-gskh4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T14:09:52Z is after 2025-08-24T17:21:41Z" Dec 03 14:09:52 crc kubenswrapper[4805]: I1203 14:09:52.893643 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-psjpm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"deeec30d-cad2-4dfd-adb6-64f6646581be\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:50Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:50Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wdlzx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wdlzx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wdlzx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wdlzx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wdlzx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wdlzx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wdlzx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wdlzx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2f875252fbe6a673a72fc9c56bb96899f62732305ad77b3108e3b244bfa91ecb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2f875252fbe6a673a72fc9c56bb96899f62732305ad77b3108e3b244bfa91ecb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T14:09:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T14:09:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wdlzx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T14:09:50Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-psjpm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T14:09:52Z is after 2025-08-24T17:21:41Z" Dec 03 14:09:52 crc kubenswrapper[4805]: I1203 14:09:52.904178 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-pdzgk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"860769de-6a5b-4cb9-a752-264e7e2e785f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5e2f0a0cc21505d609f81918365d5e4f5d21959ac12ab147d6a51601c9de4c14\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ggn4n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T14:09:50Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-pdzgk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T14:09:52Z is after 2025-08-24T17:21:41Z" Dec 03 14:09:52 crc kubenswrapper[4805]: I1203 14:09:52.918091 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:51Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:51Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b3123d3fa879097db88ed2f3cfb05631f440c24c5405503b2de3546f12e2e899\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T14:09:52Z is after 2025-08-24T17:21:41Z" Dec 03 14:09:52 crc kubenswrapper[4805]: I1203 14:09:52.928486 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-bdtdg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"53a0e9e3-b6d8-44b7-a32b-aa3450442865\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:51Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:51Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:51Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9ww54\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T14:09:51Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-bdtdg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T14:09:52Z is after 2025-08-24T17:21:41Z" Dec 03 14:09:52 crc kubenswrapper[4805]: I1203 14:09:52.941871 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ced8291d-0ce5-4e74-9eaf-2bf033774039\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:30Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:30Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3e5a34636f0a8c4d9179074596d28bc2eb7a504573398c596a637defd829907a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://29cd95be3f964415943c873e7914ffcea4abe50dfdf17f8a9cd03fc38df35981\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://951173b3d3d581275a6f6e43209b5e0611a701ca7a10f81504d723e24a2d21a3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e64730957bbbb2feb4d506e0ed9731a0594bf66fa1e8fcdb8ec944ded6308eed\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5da55f769e23ad52bac9c77da62c78e70fa34db8ed1a435e4eaf4dad3fd2b68a\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-03T14:09:48Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1203 14:09:42.992179 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1203 14:09:42.993233 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1830926909/tls.crt::/tmp/serving-cert-1830926909/tls.key\\\\\\\"\\\\nI1203 14:09:48.622403 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1203 14:09:48.626476 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1203 14:09:48.626527 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1203 14:09:48.626574 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1203 14:09:48.626582 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1203 14:09:48.633990 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1203 14:09:48.634024 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 14:09:48.634029 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 14:09:48.634034 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1203 14:09:48.634038 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1203 14:09:48.634042 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1203 14:09:48.634045 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1203 14:09:48.634344 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1203 14:09:48.644497 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T14:09:32Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bd92a160eb530c5687ebd7cd5b83b41820a2b808a0fd9514e5ace9d237e1bd26\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:32Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://af9c940135012c629447528c61743ed1e58708ab57ce76604c6da0479aa459cf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://af9c940135012c629447528c61743ed1e58708ab57ce76604c6da0479aa459cf\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T14:09:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T14:09:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T14:09:30Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T14:09:52Z is after 2025-08-24T17:21:41Z" Dec 03 14:09:52 crc kubenswrapper[4805]: I1203 14:09:52.951600 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 14:09:52 crc kubenswrapper[4805]: I1203 14:09:52.951638 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 14:09:52 crc kubenswrapper[4805]: I1203 14:09:52.951651 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 14:09:52 crc kubenswrapper[4805]: I1203 14:09:52.951668 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 14:09:52 crc kubenswrapper[4805]: I1203 14:09:52.951680 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T14:09:52Z","lastTransitionTime":"2025-12-03T14:09:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 14:09:52 crc kubenswrapper[4805]: I1203 14:09:52.953236 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:49Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:49Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://38339331a3209d6787d1def81b72f4a8514bf91080a139cad6d8ac5cc6bef10e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T14:09:52Z is after 2025-08-24T17:21:41Z" Dec 03 14:09:52 crc kubenswrapper[4805]: I1203 14:09:52.964434 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:48Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:48Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T14:09:52Z is after 2025-08-24T17:21:41Z" Dec 03 14:09:52 crc kubenswrapper[4805]: I1203 14:09:52.975549 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:48Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:48Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T14:09:52Z is after 2025-08-24T17:21:41Z" Dec 03 14:09:52 crc kubenswrapper[4805]: I1203 14:09:52.987621 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-k4mhs" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c09c1cdc-f461-45c9-8444-5b8764bccabc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3354c1a44f891182d04116fc76234a2b337d50567d032a3584f902a9516387cc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qdcd7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T14:09:50Z\\\"}}\" for pod \"openshift-multus\"/\"multus-k4mhs\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T14:09:52Z is after 2025-08-24T17:21:41Z" Dec 03 14:09:53 crc kubenswrapper[4805]: I1203 14:09:53.002033 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-bpk49" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"de00a81f-82e1-4223-9614-f98606b9a8ef\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:50Z\\\",\\\"message\\\":\\\"containers with incomplete status: [cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:50Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:50Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-25d7t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7741e0e70416ebe094bfabc6009e98d7d4c0fbcba1fc82df13b9ba40774b456a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7741e0e70416ebe094bfabc6009e98d7d4c0fbcba1fc82df13b9ba40774b456a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T14:09:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T14:09:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-25d7t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ee3499c02eb52bbf08c447d2433d435a74dfb70670dc4cbaac44f5fbff9e8a29\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-25d7t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-25d7t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-25d7t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-25d7t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-25d7t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T14:09:50Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-bpk49\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T14:09:53Z is after 2025-08-24T17:21:41Z" Dec 03 14:09:53 crc kubenswrapper[4805]: I1203 14:09:53.054353 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 14:09:53 crc kubenswrapper[4805]: I1203 14:09:53.054400 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 14:09:53 crc kubenswrapper[4805]: I1203 14:09:53.054410 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 14:09:53 crc kubenswrapper[4805]: I1203 14:09:53.054426 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 14:09:53 crc kubenswrapper[4805]: I1203 14:09:53.054437 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T14:09:53Z","lastTransitionTime":"2025-12-03T14:09:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 14:09:53 crc kubenswrapper[4805]: I1203 14:09:53.156731 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 14:09:53 crc kubenswrapper[4805]: I1203 14:09:53.156767 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 14:09:53 crc kubenswrapper[4805]: I1203 14:09:53.156785 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 14:09:53 crc kubenswrapper[4805]: I1203 14:09:53.156800 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 14:09:53 crc kubenswrapper[4805]: I1203 14:09:53.156811 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T14:09:53Z","lastTransitionTime":"2025-12-03T14:09:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 14:09:53 crc kubenswrapper[4805]: I1203 14:09:53.258821 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 14:09:53 crc kubenswrapper[4805]: I1203 14:09:53.258869 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 14:09:53 crc kubenswrapper[4805]: I1203 14:09:53.258882 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 14:09:53 crc kubenswrapper[4805]: I1203 14:09:53.258897 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 14:09:53 crc kubenswrapper[4805]: I1203 14:09:53.258907 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T14:09:53Z","lastTransitionTime":"2025-12-03T14:09:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 14:09:53 crc kubenswrapper[4805]: I1203 14:09:53.360865 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 14:09:53 crc kubenswrapper[4805]: I1203 14:09:53.361230 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 14:09:53 crc kubenswrapper[4805]: I1203 14:09:53.361245 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 14:09:53 crc kubenswrapper[4805]: I1203 14:09:53.361262 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 14:09:53 crc kubenswrapper[4805]: I1203 14:09:53.361274 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T14:09:53Z","lastTransitionTime":"2025-12-03T14:09:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 14:09:53 crc kubenswrapper[4805]: I1203 14:09:53.460465 4805 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-etcd/etcd-crc" Dec 03 14:09:53 crc kubenswrapper[4805]: I1203 14:09:53.462937 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 14:09:53 crc kubenswrapper[4805]: I1203 14:09:53.462977 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 14:09:53 crc kubenswrapper[4805]: I1203 14:09:53.462989 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 14:09:53 crc kubenswrapper[4805]: I1203 14:09:53.463006 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 14:09:53 crc kubenswrapper[4805]: I1203 14:09:53.463018 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T14:09:53Z","lastTransitionTime":"2025-12-03T14:09:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 14:09:53 crc kubenswrapper[4805]: I1203 14:09:53.472951 4805 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-etcd/etcd-crc" Dec 03 14:09:53 crc kubenswrapper[4805]: I1203 14:09:53.476179 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ced8291d-0ce5-4e74-9eaf-2bf033774039\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:30Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:30Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3e5a34636f0a8c4d9179074596d28bc2eb7a504573398c596a637defd829907a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://29cd95be3f964415943c873e7914ffcea4abe50dfdf17f8a9cd03fc38df35981\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://951173b3d3d581275a6f6e43209b5e0611a701ca7a10f81504d723e24a2d21a3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e64730957bbbb2feb4d506e0ed9731a0594bf66fa1e8fcdb8ec944ded6308eed\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5da55f769e23ad52bac9c77da62c78e70fa34db8ed1a435e4eaf4dad3fd2b68a\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-03T14:09:48Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1203 14:09:42.992179 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1203 14:09:42.993233 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1830926909/tls.crt::/tmp/serving-cert-1830926909/tls.key\\\\\\\"\\\\nI1203 14:09:48.622403 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1203 14:09:48.626476 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1203 14:09:48.626527 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1203 14:09:48.626574 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1203 14:09:48.626582 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1203 14:09:48.633990 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1203 14:09:48.634024 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 14:09:48.634029 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 14:09:48.634034 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1203 14:09:48.634038 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1203 14:09:48.634042 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1203 14:09:48.634045 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1203 14:09:48.634344 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1203 14:09:48.644497 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T14:09:32Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bd92a160eb530c5687ebd7cd5b83b41820a2b808a0fd9514e5ace9d237e1bd26\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:32Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://af9c940135012c629447528c61743ed1e58708ab57ce76604c6da0479aa459cf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://af9c940135012c629447528c61743ed1e58708ab57ce76604c6da0479aa459cf\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T14:09:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T14:09:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T14:09:30Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T14:09:53Z is after 2025-08-24T17:21:41Z" Dec 03 14:09:53 crc kubenswrapper[4805]: I1203 14:09:53.478828 4805 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-etcd/etcd-crc"] Dec 03 14:09:53 crc kubenswrapper[4805]: I1203 14:09:53.493269 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:49Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:49Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://38339331a3209d6787d1def81b72f4a8514bf91080a139cad6d8ac5cc6bef10e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T14:09:53Z is after 2025-08-24T17:21:41Z" Dec 03 14:09:53 crc kubenswrapper[4805]: I1203 14:09:53.507380 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:48Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:48Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T14:09:53Z is after 2025-08-24T17:21:41Z" Dec 03 14:09:53 crc kubenswrapper[4805]: I1203 14:09:53.519360 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:51Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:51Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b3123d3fa879097db88ed2f3cfb05631f440c24c5405503b2de3546f12e2e899\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T14:09:53Z is after 2025-08-24T17:21:41Z" Dec 03 14:09:53 crc kubenswrapper[4805]: I1203 14:09:53.529939 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-bdtdg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"53a0e9e3-b6d8-44b7-a32b-aa3450442865\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:51Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:51Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:51Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9ww54\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T14:09:51Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-bdtdg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T14:09:53Z is after 2025-08-24T17:21:41Z" Dec 03 14:09:53 crc kubenswrapper[4805]: I1203 14:09:53.541694 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:48Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:48Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T14:09:53Z is after 2025-08-24T17:21:41Z" Dec 03 14:09:53 crc kubenswrapper[4805]: I1203 14:09:53.556387 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-k4mhs" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c09c1cdc-f461-45c9-8444-5b8764bccabc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3354c1a44f891182d04116fc76234a2b337d50567d032a3584f902a9516387cc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qdcd7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T14:09:50Z\\\"}}\" for pod \"openshift-multus\"/\"multus-k4mhs\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T14:09:53Z is after 2025-08-24T17:21:41Z" Dec 03 14:09:53 crc kubenswrapper[4805]: I1203 14:09:53.565495 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 14:09:53 crc kubenswrapper[4805]: I1203 14:09:53.565524 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 14:09:53 crc kubenswrapper[4805]: I1203 14:09:53.565532 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 14:09:53 crc kubenswrapper[4805]: I1203 14:09:53.565546 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 14:09:53 crc kubenswrapper[4805]: I1203 14:09:53.565557 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T14:09:53Z","lastTransitionTime":"2025-12-03T14:09:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 14:09:53 crc kubenswrapper[4805]: I1203 14:09:53.568768 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-bpk49" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"de00a81f-82e1-4223-9614-f98606b9a8ef\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:50Z\\\",\\\"message\\\":\\\"containers with incomplete status: [cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:50Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:50Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-25d7t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7741e0e70416ebe094bfabc6009e98d7d4c0fbcba1fc82df13b9ba40774b456a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7741e0e70416ebe094bfabc6009e98d7d4c0fbcba1fc82df13b9ba40774b456a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T14:09:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T14:09:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-25d7t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ee3499c02eb52bbf08c447d2433d435a74dfb70670dc4cbaac44f5fbff9e8a29\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-25d7t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-25d7t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-25d7t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-25d7t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-25d7t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T14:09:50Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-bpk49\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T14:09:53Z is after 2025-08-24T17:21:41Z" Dec 03 14:09:53 crc kubenswrapper[4805]: I1203 14:09:53.579343 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:48Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:48Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T14:09:53Z is after 2025-08-24T17:21:41Z" Dec 03 14:09:53 crc kubenswrapper[4805]: I1203 14:09:53.595069 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:49Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:49Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://541ace7ccbb66dbd191c3ca8c1e4479e3edbf80dfc67326bea8aaee81fdcfc3c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://821e3e9a902aacecf6c41e8b8d8aa1ac951ce2fbb6505b7148b16348a3754e55\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T14:09:53Z is after 2025-08-24T17:21:41Z" Dec 03 14:09:53 crc kubenswrapper[4805]: I1203 14:09:53.606912 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-gskh4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6098937f-e3f6-45e8-a647-4994a79cd711\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://28d5653f26ec61647808b3a733ff7eac70aa69ac184660f24147e8cfaf0eb802\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nwrw9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9ad6b68a42fd33ee167e67873c6eed3f11cdf6dc71d22e952c021436eb967dd9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nwrw9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T14:09:50Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-gskh4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T14:09:53Z is after 2025-08-24T17:21:41Z" Dec 03 14:09:53 crc kubenswrapper[4805]: I1203 14:09:53.630099 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-psjpm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"deeec30d-cad2-4dfd-adb6-64f6646581be\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:50Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:50Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wdlzx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wdlzx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wdlzx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wdlzx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wdlzx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wdlzx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wdlzx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wdlzx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2f875252fbe6a673a72fc9c56bb96899f62732305ad77b3108e3b244bfa91ecb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2f875252fbe6a673a72fc9c56bb96899f62732305ad77b3108e3b244bfa91ecb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T14:09:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T14:09:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wdlzx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T14:09:50Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-psjpm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T14:09:53Z is after 2025-08-24T17:21:41Z" Dec 03 14:09:53 crc kubenswrapper[4805]: I1203 14:09:53.643155 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-pdzgk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"860769de-6a5b-4cb9-a752-264e7e2e785f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5e2f0a0cc21505d609f81918365d5e4f5d21959ac12ab147d6a51601c9de4c14\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ggn4n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T14:09:50Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-pdzgk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T14:09:53Z is after 2025-08-24T17:21:41Z" Dec 03 14:09:53 crc kubenswrapper[4805]: I1203 14:09:53.653470 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-pdzgk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"860769de-6a5b-4cb9-a752-264e7e2e785f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5e2f0a0cc21505d609f81918365d5e4f5d21959ac12ab147d6a51601c9de4c14\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ggn4n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T14:09:50Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-pdzgk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T14:09:53Z is after 2025-08-24T17:21:41Z" Dec 03 14:09:53 crc kubenswrapper[4805]: I1203 14:09:53.667934 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 14:09:53 crc kubenswrapper[4805]: I1203 14:09:53.667978 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 14:09:53 crc kubenswrapper[4805]: I1203 14:09:53.667989 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 14:09:53 crc kubenswrapper[4805]: I1203 14:09:53.668005 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 14:09:53 crc kubenswrapper[4805]: I1203 14:09:53.668016 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T14:09:53Z","lastTransitionTime":"2025-12-03T14:09:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 14:09:53 crc kubenswrapper[4805]: I1203 14:09:53.669181 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ced8291d-0ce5-4e74-9eaf-2bf033774039\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:30Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:30Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3e5a34636f0a8c4d9179074596d28bc2eb7a504573398c596a637defd829907a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://29cd95be3f964415943c873e7914ffcea4abe50dfdf17f8a9cd03fc38df35981\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://951173b3d3d581275a6f6e43209b5e0611a701ca7a10f81504d723e24a2d21a3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e64730957bbbb2feb4d506e0ed9731a0594bf66fa1e8fcdb8ec944ded6308eed\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5da55f769e23ad52bac9c77da62c78e70fa34db8ed1a435e4eaf4dad3fd2b68a\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-03T14:09:48Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1203 14:09:42.992179 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1203 14:09:42.993233 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1830926909/tls.crt::/tmp/serving-cert-1830926909/tls.key\\\\\\\"\\\\nI1203 14:09:48.622403 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1203 14:09:48.626476 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1203 14:09:48.626527 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1203 14:09:48.626574 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1203 14:09:48.626582 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1203 14:09:48.633990 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1203 14:09:48.634024 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 14:09:48.634029 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 14:09:48.634034 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1203 14:09:48.634038 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1203 14:09:48.634042 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1203 14:09:48.634045 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1203 14:09:48.634344 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1203 14:09:48.644497 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T14:09:32Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bd92a160eb530c5687ebd7cd5b83b41820a2b808a0fd9514e5ace9d237e1bd26\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:32Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://af9c940135012c629447528c61743ed1e58708ab57ce76604c6da0479aa459cf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://af9c940135012c629447528c61743ed1e58708ab57ce76604c6da0479aa459cf\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T14:09:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T14:09:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T14:09:30Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T14:09:53Z is after 2025-08-24T17:21:41Z" Dec 03 14:09:53 crc kubenswrapper[4805]: I1203 14:09:53.681617 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:49Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:49Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://38339331a3209d6787d1def81b72f4a8514bf91080a139cad6d8ac5cc6bef10e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T14:09:53Z is after 2025-08-24T17:21:41Z" Dec 03 14:09:53 crc kubenswrapper[4805]: I1203 14:09:53.692978 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:48Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:48Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T14:09:53Z is after 2025-08-24T17:21:41Z" Dec 03 14:09:53 crc kubenswrapper[4805]: I1203 14:09:53.705076 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:51Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:51Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b3123d3fa879097db88ed2f3cfb05631f440c24c5405503b2de3546f12e2e899\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T14:09:53Z is after 2025-08-24T17:21:41Z" Dec 03 14:09:53 crc kubenswrapper[4805]: I1203 14:09:53.714697 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-bdtdg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"53a0e9e3-b6d8-44b7-a32b-aa3450442865\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:51Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:51Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:51Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9ww54\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T14:09:51Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-bdtdg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T14:09:53Z is after 2025-08-24T17:21:41Z" Dec 03 14:09:53 crc kubenswrapper[4805]: I1203 14:09:53.732949 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b6d36671-7173-4e9c-843f-005b4dcf576c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a837d7d491eaacbe4279b4014bb5fcbe9b514ba5cc82c708227bcd9b25a346e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6e5f04faf3b23197bb8ea11d3e574dc86f31e21d798734962001ec8c95de7c84\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2ff1a979387da41b00e95eb9cc8b2c9de966380a12fae687591df3c5b18a4497\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1e66e1cb18573a14e3c3ae16af4dbd9f8997a6ad38680c3ab4505646ebd48e7c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7d5b5c85e3de43e7e95111aa967248550607c26200c783251b68696f202f92a1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://474f72dac37a808fc4b9832306e76979b151d7221807ff3086013466ed1419c3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://474f72dac37a808fc4b9832306e76979b151d7221807ff3086013466ed1419c3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T14:09:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T14:09:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://74c8cea1a61c4775a0ad2f4ab4bb8a36437cf23d2ee953311d6ea967e297510a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://74c8cea1a61c4775a0ad2f4ab4bb8a36437cf23d2ee953311d6ea967e297510a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T14:09:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T14:09:31Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://a4f2f56a809adc01fba82eba0652d78f58770c132f77b1424f8b17068b1927e3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a4f2f56a809adc01fba82eba0652d78f58770c132f77b1424f8b17068b1927e3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T14:09:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T14:09:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T14:09:30Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T14:09:53Z is after 2025-08-24T17:21:41Z" Dec 03 14:09:53 crc kubenswrapper[4805]: I1203 14:09:53.743861 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:48Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:48Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T14:09:53Z is after 2025-08-24T17:21:41Z" Dec 03 14:09:53 crc kubenswrapper[4805]: I1203 14:09:53.754178 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-k4mhs" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c09c1cdc-f461-45c9-8444-5b8764bccabc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3354c1a44f891182d04116fc76234a2b337d50567d032a3584f902a9516387cc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qdcd7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T14:09:50Z\\\"}}\" for pod \"openshift-multus\"/\"multus-k4mhs\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T14:09:53Z is after 2025-08-24T17:21:41Z" Dec 03 14:09:53 crc kubenswrapper[4805]: I1203 14:09:53.765275 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-bpk49" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"de00a81f-82e1-4223-9614-f98606b9a8ef\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:50Z\\\",\\\"message\\\":\\\"containers with incomplete status: [cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:50Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:50Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-25d7t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7741e0e70416ebe094bfabc6009e98d7d4c0fbcba1fc82df13b9ba40774b456a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7741e0e70416ebe094bfabc6009e98d7d4c0fbcba1fc82df13b9ba40774b456a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T14:09:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T14:09:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-25d7t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ee3499c02eb52bbf08c447d2433d435a74dfb70670dc4cbaac44f5fbff9e8a29\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-25d7t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-25d7t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-25d7t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-25d7t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-25d7t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T14:09:50Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-bpk49\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T14:09:53Z is after 2025-08-24T17:21:41Z" Dec 03 14:09:53 crc kubenswrapper[4805]: I1203 14:09:53.769358 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 14:09:53 crc kubenswrapper[4805]: I1203 14:09:53.769388 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 14:09:53 crc kubenswrapper[4805]: I1203 14:09:53.769400 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 14:09:53 crc kubenswrapper[4805]: I1203 14:09:53.769417 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 14:09:53 crc kubenswrapper[4805]: I1203 14:09:53.769428 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T14:09:53Z","lastTransitionTime":"2025-12-03T14:09:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 14:09:53 crc kubenswrapper[4805]: I1203 14:09:53.776983 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:49Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:49Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://541ace7ccbb66dbd191c3ca8c1e4479e3edbf80dfc67326bea8aaee81fdcfc3c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://821e3e9a902aacecf6c41e8b8d8aa1ac951ce2fbb6505b7148b16348a3754e55\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T14:09:53Z is after 2025-08-24T17:21:41Z" Dec 03 14:09:53 crc kubenswrapper[4805]: I1203 14:09:53.788583 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-gskh4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6098937f-e3f6-45e8-a647-4994a79cd711\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://28d5653f26ec61647808b3a733ff7eac70aa69ac184660f24147e8cfaf0eb802\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nwrw9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9ad6b68a42fd33ee167e67873c6eed3f11cdf6dc71d22e952c021436eb967dd9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nwrw9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T14:09:50Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-gskh4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T14:09:53Z is after 2025-08-24T17:21:41Z" Dec 03 14:09:53 crc kubenswrapper[4805]: I1203 14:09:53.804552 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-psjpm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"deeec30d-cad2-4dfd-adb6-64f6646581be\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:50Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:50Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wdlzx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wdlzx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wdlzx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wdlzx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wdlzx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wdlzx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wdlzx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wdlzx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2f875252fbe6a673a72fc9c56bb96899f62732305ad77b3108e3b244bfa91ecb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2f875252fbe6a673a72fc9c56bb96899f62732305ad77b3108e3b244bfa91ecb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T14:09:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T14:09:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wdlzx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T14:09:50Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-psjpm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T14:09:53Z is after 2025-08-24T17:21:41Z" Dec 03 14:09:53 crc kubenswrapper[4805]: I1203 14:09:53.821820 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:48Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:48Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T14:09:53Z is after 2025-08-24T17:21:41Z" Dec 03 14:09:53 crc kubenswrapper[4805]: I1203 14:09:53.839441 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/node-ca-bdtdg" event={"ID":"53a0e9e3-b6d8-44b7-a32b-aa3450442865","Type":"ContainerStarted","Data":"f557e17ad105d6972d3c1480656c6073f67e28c0f5773aa365549951f4d161c7"} Dec 03 14:09:53 crc kubenswrapper[4805]: I1203 14:09:53.844200 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-psjpm" event={"ID":"deeec30d-cad2-4dfd-adb6-64f6646581be","Type":"ContainerStarted","Data":"cb8bf05ef093eb0727a066d4bbbdf9af9422867cd4dadacafc6eeb016867a61e"} Dec 03 14:09:53 crc kubenswrapper[4805]: I1203 14:09:53.844250 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-psjpm" event={"ID":"deeec30d-cad2-4dfd-adb6-64f6646581be","Type":"ContainerStarted","Data":"0304467102ed053628c0d3e06548c74b6492336b6fd044246508d1e8bdb17bc3"} Dec 03 14:09:53 crc kubenswrapper[4805]: I1203 14:09:53.844270 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-psjpm" event={"ID":"deeec30d-cad2-4dfd-adb6-64f6646581be","Type":"ContainerStarted","Data":"bd67c6de01a8d196a1f51d4d90640c9e69de6f05d5e201b25e6eaa5c33440ea8"} Dec 03 14:09:53 crc kubenswrapper[4805]: I1203 14:09:53.844287 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-psjpm" event={"ID":"deeec30d-cad2-4dfd-adb6-64f6646581be","Type":"ContainerStarted","Data":"4b8d00cdf0f37f2e5cd4e952cf85d3de8d72c43fdf786088fd9f26ed17b1e21f"} Dec 03 14:09:53 crc kubenswrapper[4805]: I1203 14:09:53.846522 4805 generic.go:334] "Generic (PLEG): container finished" podID="de00a81f-82e1-4223-9614-f98606b9a8ef" containerID="ee3499c02eb52bbf08c447d2433d435a74dfb70670dc4cbaac44f5fbff9e8a29" exitCode=0 Dec 03 14:09:53 crc kubenswrapper[4805]: I1203 14:09:53.846591 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-bpk49" event={"ID":"de00a81f-82e1-4223-9614-f98606b9a8ef","Type":"ContainerDied","Data":"ee3499c02eb52bbf08c447d2433d435a74dfb70670dc4cbaac44f5fbff9e8a29"} Dec 03 14:09:53 crc kubenswrapper[4805]: I1203 14:09:53.859981 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:48Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:48Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T14:09:53Z is after 2025-08-24T17:21:41Z" Dec 03 14:09:53 crc kubenswrapper[4805]: I1203 14:09:53.871575 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 14:09:53 crc kubenswrapper[4805]: I1203 14:09:53.871609 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 14:09:53 crc kubenswrapper[4805]: I1203 14:09:53.871680 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 14:09:53 crc kubenswrapper[4805]: I1203 14:09:53.871721 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 14:09:53 crc kubenswrapper[4805]: I1203 14:09:53.871735 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T14:09:53Z","lastTransitionTime":"2025-12-03T14:09:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 14:09:53 crc kubenswrapper[4805]: I1203 14:09:53.874482 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:49Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:49Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://541ace7ccbb66dbd191c3ca8c1e4479e3edbf80dfc67326bea8aaee81fdcfc3c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://821e3e9a902aacecf6c41e8b8d8aa1ac951ce2fbb6505b7148b16348a3754e55\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T14:09:53Z is after 2025-08-24T17:21:41Z" Dec 03 14:09:53 crc kubenswrapper[4805]: I1203 14:09:53.890904 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-gskh4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6098937f-e3f6-45e8-a647-4994a79cd711\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://28d5653f26ec61647808b3a733ff7eac70aa69ac184660f24147e8cfaf0eb802\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nwrw9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9ad6b68a42fd33ee167e67873c6eed3f11cdf6dc71d22e952c021436eb967dd9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nwrw9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T14:09:50Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-gskh4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T14:09:53Z is after 2025-08-24T17:21:41Z" Dec 03 14:09:53 crc kubenswrapper[4805]: I1203 14:09:53.914496 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-psjpm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"deeec30d-cad2-4dfd-adb6-64f6646581be\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:50Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:50Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wdlzx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wdlzx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wdlzx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wdlzx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wdlzx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wdlzx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wdlzx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wdlzx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2f875252fbe6a673a72fc9c56bb96899f62732305ad77b3108e3b244bfa91ecb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2f875252fbe6a673a72fc9c56bb96899f62732305ad77b3108e3b244bfa91ecb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T14:09:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T14:09:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wdlzx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T14:09:50Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-psjpm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T14:09:53Z is after 2025-08-24T17:21:41Z" Dec 03 14:09:53 crc kubenswrapper[4805]: I1203 14:09:53.925112 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-pdzgk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"860769de-6a5b-4cb9-a752-264e7e2e785f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5e2f0a0cc21505d609f81918365d5e4f5d21959ac12ab147d6a51601c9de4c14\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ggn4n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T14:09:50Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-pdzgk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T14:09:53Z is after 2025-08-24T17:21:41Z" Dec 03 14:09:53 crc kubenswrapper[4805]: I1203 14:09:53.936397 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:48Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:48Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T14:09:53Z is after 2025-08-24T17:21:41Z" Dec 03 14:09:53 crc kubenswrapper[4805]: I1203 14:09:53.947191 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:51Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:51Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b3123d3fa879097db88ed2f3cfb05631f440c24c5405503b2de3546f12e2e899\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T14:09:53Z is after 2025-08-24T17:21:41Z" Dec 03 14:09:53 crc kubenswrapper[4805]: I1203 14:09:53.956861 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-bdtdg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"53a0e9e3-b6d8-44b7-a32b-aa3450442865\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f557e17ad105d6972d3c1480656c6073f67e28c0f5773aa365549951f4d161c7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9ww54\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T14:09:51Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-bdtdg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T14:09:53Z is after 2025-08-24T17:21:41Z" Dec 03 14:09:53 crc kubenswrapper[4805]: I1203 14:09:53.978318 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 14:09:53 crc kubenswrapper[4805]: I1203 14:09:53.978364 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 14:09:53 crc kubenswrapper[4805]: I1203 14:09:53.978380 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 14:09:53 crc kubenswrapper[4805]: I1203 14:09:53.978399 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 14:09:53 crc kubenswrapper[4805]: I1203 14:09:53.978414 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T14:09:53Z","lastTransitionTime":"2025-12-03T14:09:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 14:09:53 crc kubenswrapper[4805]: I1203 14:09:53.994548 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b6d36671-7173-4e9c-843f-005b4dcf576c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a837d7d491eaacbe4279b4014bb5fcbe9b514ba5cc82c708227bcd9b25a346e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6e5f04faf3b23197bb8ea11d3e574dc86f31e21d798734962001ec8c95de7c84\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2ff1a979387da41b00e95eb9cc8b2c9de966380a12fae687591df3c5b18a4497\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1e66e1cb18573a14e3c3ae16af4dbd9f8997a6ad38680c3ab4505646ebd48e7c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7d5b5c85e3de43e7e95111aa967248550607c26200c783251b68696f202f92a1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://474f72dac37a808fc4b9832306e76979b151d7221807ff3086013466ed1419c3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://474f72dac37a808fc4b9832306e76979b151d7221807ff3086013466ed1419c3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T14:09:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T14:09:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://74c8cea1a61c4775a0ad2f4ab4bb8a36437cf23d2ee953311d6ea967e297510a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://74c8cea1a61c4775a0ad2f4ab4bb8a36437cf23d2ee953311d6ea967e297510a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T14:09:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T14:09:31Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://a4f2f56a809adc01fba82eba0652d78f58770c132f77b1424f8b17068b1927e3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a4f2f56a809adc01fba82eba0652d78f58770c132f77b1424f8b17068b1927e3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T14:09:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T14:09:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T14:09:30Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T14:09:53Z is after 2025-08-24T17:21:41Z" Dec 03 14:09:54 crc kubenswrapper[4805]: I1203 14:09:54.020086 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ced8291d-0ce5-4e74-9eaf-2bf033774039\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:30Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:30Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3e5a34636f0a8c4d9179074596d28bc2eb7a504573398c596a637defd829907a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://29cd95be3f964415943c873e7914ffcea4abe50dfdf17f8a9cd03fc38df35981\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://951173b3d3d581275a6f6e43209b5e0611a701ca7a10f81504d723e24a2d21a3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e64730957bbbb2feb4d506e0ed9731a0594bf66fa1e8fcdb8ec944ded6308eed\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5da55f769e23ad52bac9c77da62c78e70fa34db8ed1a435e4eaf4dad3fd2b68a\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-03T14:09:48Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1203 14:09:42.992179 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1203 14:09:42.993233 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1830926909/tls.crt::/tmp/serving-cert-1830926909/tls.key\\\\\\\"\\\\nI1203 14:09:48.622403 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1203 14:09:48.626476 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1203 14:09:48.626527 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1203 14:09:48.626574 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1203 14:09:48.626582 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1203 14:09:48.633990 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1203 14:09:48.634024 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 14:09:48.634029 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 14:09:48.634034 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1203 14:09:48.634038 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1203 14:09:48.634042 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1203 14:09:48.634045 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1203 14:09:48.634344 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1203 14:09:48.644497 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T14:09:32Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bd92a160eb530c5687ebd7cd5b83b41820a2b808a0fd9514e5ace9d237e1bd26\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:32Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://af9c940135012c629447528c61743ed1e58708ab57ce76604c6da0479aa459cf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://af9c940135012c629447528c61743ed1e58708ab57ce76604c6da0479aa459cf\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T14:09:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T14:09:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T14:09:30Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T14:09:54Z is after 2025-08-24T17:21:41Z" Dec 03 14:09:54 crc kubenswrapper[4805]: I1203 14:09:54.032636 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:49Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:49Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://38339331a3209d6787d1def81b72f4a8514bf91080a139cad6d8ac5cc6bef10e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T14:09:54Z is after 2025-08-24T17:21:41Z" Dec 03 14:09:54 crc kubenswrapper[4805]: I1203 14:09:54.043338 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:48Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:48Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T14:09:54Z is after 2025-08-24T17:21:41Z" Dec 03 14:09:54 crc kubenswrapper[4805]: I1203 14:09:54.053762 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-k4mhs" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c09c1cdc-f461-45c9-8444-5b8764bccabc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3354c1a44f891182d04116fc76234a2b337d50567d032a3584f902a9516387cc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qdcd7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T14:09:50Z\\\"}}\" for pod \"openshift-multus\"/\"multus-k4mhs\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T14:09:54Z is after 2025-08-24T17:21:41Z" Dec 03 14:09:54 crc kubenswrapper[4805]: I1203 14:09:54.064662 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-bpk49" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"de00a81f-82e1-4223-9614-f98606b9a8ef\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:50Z\\\",\\\"message\\\":\\\"containers with incomplete status: [cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:50Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:50Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-25d7t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7741e0e70416ebe094bfabc6009e98d7d4c0fbcba1fc82df13b9ba40774b456a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7741e0e70416ebe094bfabc6009e98d7d4c0fbcba1fc82df13b9ba40774b456a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T14:09:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T14:09:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-25d7t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ee3499c02eb52bbf08c447d2433d435a74dfb70670dc4cbaac44f5fbff9e8a29\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-25d7t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-25d7t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-25d7t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-25d7t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-25d7t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T14:09:50Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-bpk49\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T14:09:54Z is after 2025-08-24T17:21:41Z" Dec 03 14:09:54 crc kubenswrapper[4805]: I1203 14:09:54.078861 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:48Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:48Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T14:09:54Z is after 2025-08-24T17:21:41Z" Dec 03 14:09:54 crc kubenswrapper[4805]: I1203 14:09:54.080304 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 14:09:54 crc kubenswrapper[4805]: I1203 14:09:54.080329 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 14:09:54 crc kubenswrapper[4805]: I1203 14:09:54.080338 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 14:09:54 crc kubenswrapper[4805]: I1203 14:09:54.080350 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 14:09:54 crc kubenswrapper[4805]: I1203 14:09:54.080359 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T14:09:54Z","lastTransitionTime":"2025-12-03T14:09:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 14:09:54 crc kubenswrapper[4805]: I1203 14:09:54.094234 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:49Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:49Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://541ace7ccbb66dbd191c3ca8c1e4479e3edbf80dfc67326bea8aaee81fdcfc3c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://821e3e9a902aacecf6c41e8b8d8aa1ac951ce2fbb6505b7148b16348a3754e55\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T14:09:54Z is after 2025-08-24T17:21:41Z" Dec 03 14:09:54 crc kubenswrapper[4805]: I1203 14:09:54.103156 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-gskh4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6098937f-e3f6-45e8-a647-4994a79cd711\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://28d5653f26ec61647808b3a733ff7eac70aa69ac184660f24147e8cfaf0eb802\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nwrw9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9ad6b68a42fd33ee167e67873c6eed3f11cdf6dc71d22e952c021436eb967dd9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nwrw9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T14:09:50Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-gskh4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T14:09:54Z is after 2025-08-24T17:21:41Z" Dec 03 14:09:54 crc kubenswrapper[4805]: I1203 14:09:54.119583 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-psjpm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"deeec30d-cad2-4dfd-adb6-64f6646581be\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:50Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:50Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wdlzx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wdlzx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wdlzx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wdlzx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wdlzx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wdlzx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wdlzx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wdlzx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2f875252fbe6a673a72fc9c56bb96899f62732305ad77b3108e3b244bfa91ecb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2f875252fbe6a673a72fc9c56bb96899f62732305ad77b3108e3b244bfa91ecb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T14:09:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T14:09:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wdlzx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T14:09:50Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-psjpm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T14:09:54Z is after 2025-08-24T17:21:41Z" Dec 03 14:09:54 crc kubenswrapper[4805]: I1203 14:09:54.128873 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-pdzgk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"860769de-6a5b-4cb9-a752-264e7e2e785f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5e2f0a0cc21505d609f81918365d5e4f5d21959ac12ab147d6a51601c9de4c14\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ggn4n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T14:09:50Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-pdzgk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T14:09:54Z is after 2025-08-24T17:21:41Z" Dec 03 14:09:54 crc kubenswrapper[4805]: I1203 14:09:54.146624 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b6d36671-7173-4e9c-843f-005b4dcf576c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a837d7d491eaacbe4279b4014bb5fcbe9b514ba5cc82c708227bcd9b25a346e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6e5f04faf3b23197bb8ea11d3e574dc86f31e21d798734962001ec8c95de7c84\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2ff1a979387da41b00e95eb9cc8b2c9de966380a12fae687591df3c5b18a4497\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1e66e1cb18573a14e3c3ae16af4dbd9f8997a6ad38680c3ab4505646ebd48e7c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7d5b5c85e3de43e7e95111aa967248550607c26200c783251b68696f202f92a1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://474f72dac37a808fc4b9832306e76979b151d7221807ff3086013466ed1419c3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://474f72dac37a808fc4b9832306e76979b151d7221807ff3086013466ed1419c3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T14:09:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T14:09:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://74c8cea1a61c4775a0ad2f4ab4bb8a36437cf23d2ee953311d6ea967e297510a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://74c8cea1a61c4775a0ad2f4ab4bb8a36437cf23d2ee953311d6ea967e297510a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T14:09:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T14:09:31Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://a4f2f56a809adc01fba82eba0652d78f58770c132f77b1424f8b17068b1927e3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a4f2f56a809adc01fba82eba0652d78f58770c132f77b1424f8b17068b1927e3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T14:09:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T14:09:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T14:09:30Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T14:09:54Z is after 2025-08-24T17:21:41Z" Dec 03 14:09:54 crc kubenswrapper[4805]: I1203 14:09:54.159878 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ced8291d-0ce5-4e74-9eaf-2bf033774039\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:30Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:30Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3e5a34636f0a8c4d9179074596d28bc2eb7a504573398c596a637defd829907a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://29cd95be3f964415943c873e7914ffcea4abe50dfdf17f8a9cd03fc38df35981\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://951173b3d3d581275a6f6e43209b5e0611a701ca7a10f81504d723e24a2d21a3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e64730957bbbb2feb4d506e0ed9731a0594bf66fa1e8fcdb8ec944ded6308eed\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5da55f769e23ad52bac9c77da62c78e70fa34db8ed1a435e4eaf4dad3fd2b68a\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-03T14:09:48Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1203 14:09:42.992179 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1203 14:09:42.993233 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1830926909/tls.crt::/tmp/serving-cert-1830926909/tls.key\\\\\\\"\\\\nI1203 14:09:48.622403 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1203 14:09:48.626476 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1203 14:09:48.626527 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1203 14:09:48.626574 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1203 14:09:48.626582 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1203 14:09:48.633990 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1203 14:09:48.634024 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 14:09:48.634029 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 14:09:48.634034 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1203 14:09:48.634038 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1203 14:09:48.634042 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1203 14:09:48.634045 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1203 14:09:48.634344 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1203 14:09:48.644497 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T14:09:32Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bd92a160eb530c5687ebd7cd5b83b41820a2b808a0fd9514e5ace9d237e1bd26\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:32Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://af9c940135012c629447528c61743ed1e58708ab57ce76604c6da0479aa459cf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://af9c940135012c629447528c61743ed1e58708ab57ce76604c6da0479aa459cf\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T14:09:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T14:09:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T14:09:30Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T14:09:54Z is after 2025-08-24T17:21:41Z" Dec 03 14:09:54 crc kubenswrapper[4805]: I1203 14:09:54.160763 4805 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 03 14:09:54 crc kubenswrapper[4805]: I1203 14:09:54.165881 4805 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 03 14:09:54 crc kubenswrapper[4805]: I1203 14:09:54.167861 4805 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-controller-manager/kube-controller-manager-crc"] Dec 03 14:09:54 crc kubenswrapper[4805]: I1203 14:09:54.170978 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:49Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:49Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://38339331a3209d6787d1def81b72f4a8514bf91080a139cad6d8ac5cc6bef10e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T14:09:54Z is after 2025-08-24T17:21:41Z" Dec 03 14:09:54 crc kubenswrapper[4805]: I1203 14:09:54.182319 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 14:09:54 crc kubenswrapper[4805]: I1203 14:09:54.182356 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 14:09:54 crc kubenswrapper[4805]: I1203 14:09:54.182364 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 14:09:54 crc kubenswrapper[4805]: I1203 14:09:54.182379 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 14:09:54 crc kubenswrapper[4805]: I1203 14:09:54.182386 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T14:09:54Z","lastTransitionTime":"2025-12-03T14:09:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 14:09:54 crc kubenswrapper[4805]: I1203 14:09:54.202464 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:48Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:48Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T14:09:54Z is after 2025-08-24T17:21:41Z" Dec 03 14:09:54 crc kubenswrapper[4805]: I1203 14:09:54.245685 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:51Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:51Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b3123d3fa879097db88ed2f3cfb05631f440c24c5405503b2de3546f12e2e899\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T14:09:54Z is after 2025-08-24T17:21:41Z" Dec 03 14:09:54 crc kubenswrapper[4805]: I1203 14:09:54.280339 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-bdtdg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"53a0e9e3-b6d8-44b7-a32b-aa3450442865\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f557e17ad105d6972d3c1480656c6073f67e28c0f5773aa365549951f4d161c7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9ww54\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T14:09:51Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-bdtdg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T14:09:54Z is after 2025-08-24T17:21:41Z" Dec 03 14:09:54 crc kubenswrapper[4805]: I1203 14:09:54.285364 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 14:09:54 crc kubenswrapper[4805]: I1203 14:09:54.285397 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 14:09:54 crc kubenswrapper[4805]: I1203 14:09:54.285406 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 14:09:54 crc kubenswrapper[4805]: I1203 14:09:54.285422 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 14:09:54 crc kubenswrapper[4805]: I1203 14:09:54.285432 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T14:09:54Z","lastTransitionTime":"2025-12-03T14:09:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 14:09:54 crc kubenswrapper[4805]: I1203 14:09:54.322513 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:48Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:48Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T14:09:54Z is after 2025-08-24T17:21:41Z" Dec 03 14:09:54 crc kubenswrapper[4805]: I1203 14:09:54.362799 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-k4mhs" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c09c1cdc-f461-45c9-8444-5b8764bccabc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3354c1a44f891182d04116fc76234a2b337d50567d032a3584f902a9516387cc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qdcd7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T14:09:50Z\\\"}}\" for pod \"openshift-multus\"/\"multus-k4mhs\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T14:09:54Z is after 2025-08-24T17:21:41Z" Dec 03 14:09:54 crc kubenswrapper[4805]: I1203 14:09:54.388227 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 14:09:54 crc kubenswrapper[4805]: I1203 14:09:54.388276 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 14:09:54 crc kubenswrapper[4805]: I1203 14:09:54.388288 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 14:09:54 crc kubenswrapper[4805]: I1203 14:09:54.388308 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 14:09:54 crc kubenswrapper[4805]: I1203 14:09:54.388319 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T14:09:54Z","lastTransitionTime":"2025-12-03T14:09:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 14:09:54 crc kubenswrapper[4805]: I1203 14:09:54.406090 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-bpk49" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"de00a81f-82e1-4223-9614-f98606b9a8ef\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:50Z\\\",\\\"message\\\":\\\"containers with incomplete status: [bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:50Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:50Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-25d7t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7741e0e70416ebe094bfabc6009e98d7d4c0fbcba1fc82df13b9ba40774b456a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7741e0e70416ebe094bfabc6009e98d7d4c0fbcba1fc82df13b9ba40774b456a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T14:09:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T14:09:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-25d7t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ee3499c02eb52bbf08c447d2433d435a74dfb70670dc4cbaac44f5fbff9e8a29\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ee3499c02eb52bbf08c447d2433d435a74dfb70670dc4cbaac44f5fbff9e8a29\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T14:09:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T14:09:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-25d7t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-25d7t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-25d7t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-25d7t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-25d7t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T14:09:50Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-bpk49\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T14:09:54Z is after 2025-08-24T17:21:41Z" Dec 03 14:09:54 crc kubenswrapper[4805]: I1203 14:09:54.441643 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-pdzgk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"860769de-6a5b-4cb9-a752-264e7e2e785f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5e2f0a0cc21505d609f81918365d5e4f5d21959ac12ab147d6a51601c9de4c14\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ggn4n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T14:09:50Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-pdzgk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T14:09:54Z is after 2025-08-24T17:21:41Z" Dec 03 14:09:54 crc kubenswrapper[4805]: I1203 14:09:54.481687 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ced8291d-0ce5-4e74-9eaf-2bf033774039\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:30Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:30Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3e5a34636f0a8c4d9179074596d28bc2eb7a504573398c596a637defd829907a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://29cd95be3f964415943c873e7914ffcea4abe50dfdf17f8a9cd03fc38df35981\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://951173b3d3d581275a6f6e43209b5e0611a701ca7a10f81504d723e24a2d21a3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e64730957bbbb2feb4d506e0ed9731a0594bf66fa1e8fcdb8ec944ded6308eed\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5da55f769e23ad52bac9c77da62c78e70fa34db8ed1a435e4eaf4dad3fd2b68a\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-03T14:09:48Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1203 14:09:42.992179 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1203 14:09:42.993233 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1830926909/tls.crt::/tmp/serving-cert-1830926909/tls.key\\\\\\\"\\\\nI1203 14:09:48.622403 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1203 14:09:48.626476 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1203 14:09:48.626527 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1203 14:09:48.626574 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1203 14:09:48.626582 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1203 14:09:48.633990 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1203 14:09:48.634024 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 14:09:48.634029 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 14:09:48.634034 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1203 14:09:48.634038 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1203 14:09:48.634042 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1203 14:09:48.634045 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1203 14:09:48.634344 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1203 14:09:48.644497 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T14:09:32Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bd92a160eb530c5687ebd7cd5b83b41820a2b808a0fd9514e5ace9d237e1bd26\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:32Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://af9c940135012c629447528c61743ed1e58708ab57ce76604c6da0479aa459cf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://af9c940135012c629447528c61743ed1e58708ab57ce76604c6da0479aa459cf\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T14:09:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T14:09:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T14:09:30Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T14:09:54Z is after 2025-08-24T17:21:41Z" Dec 03 14:09:54 crc kubenswrapper[4805]: I1203 14:09:54.490412 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 14:09:54 crc kubenswrapper[4805]: I1203 14:09:54.490449 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 14:09:54 crc kubenswrapper[4805]: I1203 14:09:54.490462 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 14:09:54 crc kubenswrapper[4805]: I1203 14:09:54.490478 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 14:09:54 crc kubenswrapper[4805]: I1203 14:09:54.490490 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T14:09:54Z","lastTransitionTime":"2025-12-03T14:09:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 14:09:54 crc kubenswrapper[4805]: I1203 14:09:54.522277 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:49Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:49Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://38339331a3209d6787d1def81b72f4a8514bf91080a139cad6d8ac5cc6bef10e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T14:09:54Z is after 2025-08-24T17:21:41Z" Dec 03 14:09:54 crc kubenswrapper[4805]: I1203 14:09:54.561384 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:48Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:48Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T14:09:54Z is after 2025-08-24T17:21:41Z" Dec 03 14:09:54 crc kubenswrapper[4805]: I1203 14:09:54.593856 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 14:09:54 crc kubenswrapper[4805]: I1203 14:09:54.593926 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 14:09:54 crc kubenswrapper[4805]: I1203 14:09:54.593936 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 14:09:54 crc kubenswrapper[4805]: I1203 14:09:54.593949 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 14:09:54 crc kubenswrapper[4805]: I1203 14:09:54.593959 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T14:09:54Z","lastTransitionTime":"2025-12-03T14:09:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 14:09:54 crc kubenswrapper[4805]: I1203 14:09:54.602151 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:51Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:51Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b3123d3fa879097db88ed2f3cfb05631f440c24c5405503b2de3546f12e2e899\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T14:09:54Z is after 2025-08-24T17:21:41Z" Dec 03 14:09:54 crc kubenswrapper[4805]: I1203 14:09:54.639307 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-bdtdg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"53a0e9e3-b6d8-44b7-a32b-aa3450442865\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f557e17ad105d6972d3c1480656c6073f67e28c0f5773aa365549951f4d161c7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9ww54\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T14:09:51Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-bdtdg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T14:09:54Z is after 2025-08-24T17:21:41Z" Dec 03 14:09:54 crc kubenswrapper[4805]: I1203 14:09:54.694018 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 14:09:54 crc kubenswrapper[4805]: I1203 14:09:54.694061 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 14:09:54 crc kubenswrapper[4805]: I1203 14:09:54.694103 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 14:09:54 crc kubenswrapper[4805]: I1203 14:09:54.694019 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b6d36671-7173-4e9c-843f-005b4dcf576c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a837d7d491eaacbe4279b4014bb5fcbe9b514ba5cc82c708227bcd9b25a346e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6e5f04faf3b23197bb8ea11d3e574dc86f31e21d798734962001ec8c95de7c84\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2ff1a979387da41b00e95eb9cc8b2c9de966380a12fae687591df3c5b18a4497\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1e66e1cb18573a14e3c3ae16af4dbd9f8997a6ad38680c3ab4505646ebd48e7c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7d5b5c85e3de43e7e95111aa967248550607c26200c783251b68696f202f92a1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://474f72dac37a808fc4b9832306e76979b151d7221807ff3086013466ed1419c3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://474f72dac37a808fc4b9832306e76979b151d7221807ff3086013466ed1419c3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T14:09:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T14:09:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://74c8cea1a61c4775a0ad2f4ab4bb8a36437cf23d2ee953311d6ea967e297510a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://74c8cea1a61c4775a0ad2f4ab4bb8a36437cf23d2ee953311d6ea967e297510a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T14:09:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T14:09:31Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://a4f2f56a809adc01fba82eba0652d78f58770c132f77b1424f8b17068b1927e3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a4f2f56a809adc01fba82eba0652d78f58770c132f77b1424f8b17068b1927e3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T14:09:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T14:09:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T14:09:30Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T14:09:54Z is after 2025-08-24T17:21:41Z" Dec 03 14:09:54 crc kubenswrapper[4805]: E1203 14:09:54.694185 4805 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 03 14:09:54 crc kubenswrapper[4805]: E1203 14:09:54.694353 4805 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 03 14:09:54 crc kubenswrapper[4805]: E1203 14:09:54.694427 4805 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 03 14:09:54 crc kubenswrapper[4805]: I1203 14:09:54.696247 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 14:09:54 crc kubenswrapper[4805]: I1203 14:09:54.696292 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 14:09:54 crc kubenswrapper[4805]: I1203 14:09:54.696304 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 14:09:54 crc kubenswrapper[4805]: I1203 14:09:54.696325 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 14:09:54 crc kubenswrapper[4805]: I1203 14:09:54.696338 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T14:09:54Z","lastTransitionTime":"2025-12-03T14:09:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 14:09:54 crc kubenswrapper[4805]: I1203 14:09:54.722985 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:48Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:48Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T14:09:54Z is after 2025-08-24T17:21:41Z" Dec 03 14:09:54 crc kubenswrapper[4805]: I1203 14:09:54.761077 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-k4mhs" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c09c1cdc-f461-45c9-8444-5b8764bccabc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3354c1a44f891182d04116fc76234a2b337d50567d032a3584f902a9516387cc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qdcd7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T14:09:50Z\\\"}}\" for pod \"openshift-multus\"/\"multus-k4mhs\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T14:09:54Z is after 2025-08-24T17:21:41Z" Dec 03 14:09:54 crc kubenswrapper[4805]: I1203 14:09:54.798731 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 14:09:54 crc kubenswrapper[4805]: I1203 14:09:54.798765 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 14:09:54 crc kubenswrapper[4805]: I1203 14:09:54.798774 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 14:09:54 crc kubenswrapper[4805]: I1203 14:09:54.798786 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 14:09:54 crc kubenswrapper[4805]: I1203 14:09:54.798796 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T14:09:54Z","lastTransitionTime":"2025-12-03T14:09:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 14:09:54 crc kubenswrapper[4805]: I1203 14:09:54.807440 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-bpk49" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"de00a81f-82e1-4223-9614-f98606b9a8ef\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:50Z\\\",\\\"message\\\":\\\"containers with incomplete status: [bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:50Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:50Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-25d7t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7741e0e70416ebe094bfabc6009e98d7d4c0fbcba1fc82df13b9ba40774b456a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7741e0e70416ebe094bfabc6009e98d7d4c0fbcba1fc82df13b9ba40774b456a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T14:09:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T14:09:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-25d7t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ee3499c02eb52bbf08c447d2433d435a74dfb70670dc4cbaac44f5fbff9e8a29\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ee3499c02eb52bbf08c447d2433d435a74dfb70670dc4cbaac44f5fbff9e8a29\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T14:09:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T14:09:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-25d7t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-25d7t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-25d7t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-25d7t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-25d7t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T14:09:50Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-bpk49\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T14:09:54Z is after 2025-08-24T17:21:41Z" Dec 03 14:09:54 crc kubenswrapper[4805]: I1203 14:09:54.840882 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"750d2b4b-738a-4992-8902-742528f6e875\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://abd2d0156113bb5182727a53aac594410ed051c19258beb8a07b808eff0b42a7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8673a29c0e16f4e707d79f0610722ccabb023ff2b4a18d7cf9d305183817d952\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://35766e7206b1be7a53d13d6890d759cf81dbd999643c12349268e4bf16fbbdb0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5b9844a11b2b903978d0d464b469e26c9211100da61d018dde9636dca95b684e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T14:09:30Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T14:09:54Z is after 2025-08-24T17:21:41Z" Dec 03 14:09:54 crc kubenswrapper[4805]: I1203 14:09:54.851536 4805 generic.go:334] "Generic (PLEG): container finished" podID="de00a81f-82e1-4223-9614-f98606b9a8ef" containerID="ec8488b1d69799951be7ef40e5e028732265c4ddcdea2d453923e567864771f7" exitCode=0 Dec 03 14:09:54 crc kubenswrapper[4805]: I1203 14:09:54.851573 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-bpk49" event={"ID":"de00a81f-82e1-4223-9614-f98606b9a8ef","Type":"ContainerDied","Data":"ec8488b1d69799951be7ef40e5e028732265c4ddcdea2d453923e567864771f7"} Dec 03 14:09:54 crc kubenswrapper[4805]: I1203 14:09:54.881166 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:49Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:49Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://541ace7ccbb66dbd191c3ca8c1e4479e3edbf80dfc67326bea8aaee81fdcfc3c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://821e3e9a902aacecf6c41e8b8d8aa1ac951ce2fbb6505b7148b16348a3754e55\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T14:09:54Z is after 2025-08-24T17:21:41Z" Dec 03 14:09:54 crc kubenswrapper[4805]: I1203 14:09:54.901737 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 14:09:54 crc kubenswrapper[4805]: I1203 14:09:54.901773 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 14:09:54 crc kubenswrapper[4805]: I1203 14:09:54.901781 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 14:09:54 crc kubenswrapper[4805]: I1203 14:09:54.901794 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 14:09:54 crc kubenswrapper[4805]: I1203 14:09:54.901803 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T14:09:54Z","lastTransitionTime":"2025-12-03T14:09:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 14:09:54 crc kubenswrapper[4805]: I1203 14:09:54.921153 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-gskh4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6098937f-e3f6-45e8-a647-4994a79cd711\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://28d5653f26ec61647808b3a733ff7eac70aa69ac184660f24147e8cfaf0eb802\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nwrw9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9ad6b68a42fd33ee167e67873c6eed3f11cdf6dc71d22e952c021436eb967dd9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nwrw9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T14:09:50Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-gskh4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T14:09:54Z is after 2025-08-24T17:21:41Z" Dec 03 14:09:54 crc kubenswrapper[4805]: I1203 14:09:54.965022 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-psjpm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"deeec30d-cad2-4dfd-adb6-64f6646581be\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:50Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:50Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wdlzx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wdlzx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wdlzx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wdlzx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wdlzx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wdlzx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wdlzx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wdlzx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2f875252fbe6a673a72fc9c56bb96899f62732305ad77b3108e3b244bfa91ecb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2f875252fbe6a673a72fc9c56bb96899f62732305ad77b3108e3b244bfa91ecb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T14:09:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T14:09:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wdlzx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T14:09:50Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-psjpm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T14:09:54Z is after 2025-08-24T17:21:41Z" Dec 03 14:09:55 crc kubenswrapper[4805]: I1203 14:09:55.002310 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:48Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:48Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T14:09:55Z is after 2025-08-24T17:21:41Z" Dec 03 14:09:55 crc kubenswrapper[4805]: I1203 14:09:55.004430 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 14:09:55 crc kubenswrapper[4805]: I1203 14:09:55.004494 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 14:09:55 crc kubenswrapper[4805]: I1203 14:09:55.004504 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 14:09:55 crc kubenswrapper[4805]: I1203 14:09:55.004520 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 14:09:55 crc kubenswrapper[4805]: I1203 14:09:55.004533 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T14:09:55Z","lastTransitionTime":"2025-12-03T14:09:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 14:09:55 crc kubenswrapper[4805]: I1203 14:09:55.039636 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-pdzgk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"860769de-6a5b-4cb9-a752-264e7e2e785f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5e2f0a0cc21505d609f81918365d5e4f5d21959ac12ab147d6a51601c9de4c14\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ggn4n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T14:09:50Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-pdzgk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T14:09:55Z is after 2025-08-24T17:21:41Z" Dec 03 14:09:55 crc kubenswrapper[4805]: I1203 14:09:55.082994 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:49Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:49Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://38339331a3209d6787d1def81b72f4a8514bf91080a139cad6d8ac5cc6bef10e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T14:09:55Z is after 2025-08-24T17:21:41Z" Dec 03 14:09:55 crc kubenswrapper[4805]: I1203 14:09:55.107215 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 14:09:55 crc kubenswrapper[4805]: I1203 14:09:55.107257 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 14:09:55 crc kubenswrapper[4805]: I1203 14:09:55.107273 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 14:09:55 crc kubenswrapper[4805]: I1203 14:09:55.107290 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 14:09:55 crc kubenswrapper[4805]: I1203 14:09:55.107301 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T14:09:55Z","lastTransitionTime":"2025-12-03T14:09:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 14:09:55 crc kubenswrapper[4805]: I1203 14:09:55.119630 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:48Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:48Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T14:09:55Z is after 2025-08-24T17:21:41Z" Dec 03 14:09:55 crc kubenswrapper[4805]: I1203 14:09:55.165442 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:51Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:51Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b3123d3fa879097db88ed2f3cfb05631f440c24c5405503b2de3546f12e2e899\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T14:09:55Z is after 2025-08-24T17:21:41Z" Dec 03 14:09:55 crc kubenswrapper[4805]: I1203 14:09:55.205596 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-bdtdg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"53a0e9e3-b6d8-44b7-a32b-aa3450442865\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f557e17ad105d6972d3c1480656c6073f67e28c0f5773aa365549951f4d161c7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9ww54\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T14:09:51Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-bdtdg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T14:09:55Z is after 2025-08-24T17:21:41Z" Dec 03 14:09:55 crc kubenswrapper[4805]: I1203 14:09:55.209951 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 14:09:55 crc kubenswrapper[4805]: I1203 14:09:55.210022 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 14:09:55 crc kubenswrapper[4805]: I1203 14:09:55.210047 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 14:09:55 crc kubenswrapper[4805]: I1203 14:09:55.210078 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 14:09:55 crc kubenswrapper[4805]: I1203 14:09:55.210101 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T14:09:55Z","lastTransitionTime":"2025-12-03T14:09:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 14:09:55 crc kubenswrapper[4805]: I1203 14:09:55.258914 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b6d36671-7173-4e9c-843f-005b4dcf576c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a837d7d491eaacbe4279b4014bb5fcbe9b514ba5cc82c708227bcd9b25a346e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6e5f04faf3b23197bb8ea11d3e574dc86f31e21d798734962001ec8c95de7c84\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2ff1a979387da41b00e95eb9cc8b2c9de966380a12fae687591df3c5b18a4497\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1e66e1cb18573a14e3c3ae16af4dbd9f8997a6ad38680c3ab4505646ebd48e7c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7d5b5c85e3de43e7e95111aa967248550607c26200c783251b68696f202f92a1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://474f72dac37a808fc4b9832306e76979b151d7221807ff3086013466ed1419c3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://474f72dac37a808fc4b9832306e76979b151d7221807ff3086013466ed1419c3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T14:09:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T14:09:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://74c8cea1a61c4775a0ad2f4ab4bb8a36437cf23d2ee953311d6ea967e297510a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://74c8cea1a61c4775a0ad2f4ab4bb8a36437cf23d2ee953311d6ea967e297510a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T14:09:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T14:09:31Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://a4f2f56a809adc01fba82eba0652d78f58770c132f77b1424f8b17068b1927e3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a4f2f56a809adc01fba82eba0652d78f58770c132f77b1424f8b17068b1927e3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T14:09:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T14:09:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T14:09:30Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T14:09:55Z is after 2025-08-24T17:21:41Z" Dec 03 14:09:55 crc kubenswrapper[4805]: I1203 14:09:55.289632 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ced8291d-0ce5-4e74-9eaf-2bf033774039\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:30Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:30Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3e5a34636f0a8c4d9179074596d28bc2eb7a504573398c596a637defd829907a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://29cd95be3f964415943c873e7914ffcea4abe50dfdf17f8a9cd03fc38df35981\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://951173b3d3d581275a6f6e43209b5e0611a701ca7a10f81504d723e24a2d21a3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e64730957bbbb2feb4d506e0ed9731a0594bf66fa1e8fcdb8ec944ded6308eed\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5da55f769e23ad52bac9c77da62c78e70fa34db8ed1a435e4eaf4dad3fd2b68a\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-03T14:09:48Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1203 14:09:42.992179 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1203 14:09:42.993233 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1830926909/tls.crt::/tmp/serving-cert-1830926909/tls.key\\\\\\\"\\\\nI1203 14:09:48.622403 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1203 14:09:48.626476 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1203 14:09:48.626527 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1203 14:09:48.626574 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1203 14:09:48.626582 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1203 14:09:48.633990 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1203 14:09:48.634024 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 14:09:48.634029 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 14:09:48.634034 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1203 14:09:48.634038 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1203 14:09:48.634042 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1203 14:09:48.634045 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1203 14:09:48.634344 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1203 14:09:48.644497 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T14:09:32Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bd92a160eb530c5687ebd7cd5b83b41820a2b808a0fd9514e5ace9d237e1bd26\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:32Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://af9c940135012c629447528c61743ed1e58708ab57ce76604c6da0479aa459cf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://af9c940135012c629447528c61743ed1e58708ab57ce76604c6da0479aa459cf\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T14:09:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T14:09:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T14:09:30Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T14:09:55Z is after 2025-08-24T17:21:41Z" Dec 03 14:09:55 crc kubenswrapper[4805]: I1203 14:09:55.312668 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 14:09:55 crc kubenswrapper[4805]: I1203 14:09:55.312913 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 14:09:55 crc kubenswrapper[4805]: I1203 14:09:55.313068 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 14:09:55 crc kubenswrapper[4805]: I1203 14:09:55.313211 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 14:09:55 crc kubenswrapper[4805]: I1203 14:09:55.313351 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T14:09:55Z","lastTransitionTime":"2025-12-03T14:09:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 14:09:55 crc kubenswrapper[4805]: I1203 14:09:55.330761 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-bpk49" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"de00a81f-82e1-4223-9614-f98606b9a8ef\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:50Z\\\",\\\"message\\\":\\\"containers with incomplete status: [routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:50Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:50Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-25d7t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7741e0e70416ebe094bfabc6009e98d7d4c0fbcba1fc82df13b9ba40774b456a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7741e0e70416ebe094bfabc6009e98d7d4c0fbcba1fc82df13b9ba40774b456a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T14:09:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T14:09:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-25d7t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ee3499c02eb52bbf08c447d2433d435a74dfb70670dc4cbaac44f5fbff9e8a29\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ee3499c02eb52bbf08c447d2433d435a74dfb70670dc4cbaac44f5fbff9e8a29\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T14:09:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T14:09:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-25d7t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ec8488b1d69799951be7ef40e5e028732265c4ddcdea2d453923e567864771f7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ec8488b1d69799951be7ef40e5e028732265c4ddcdea2d453923e567864771f7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T14:09:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T14:09:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-25d7t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-25d7t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-25d7t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-25d7t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T14:09:50Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-bpk49\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T14:09:55Z is after 2025-08-24T17:21:41Z" Dec 03 14:09:55 crc kubenswrapper[4805]: I1203 14:09:55.363586 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"750d2b4b-738a-4992-8902-742528f6e875\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://abd2d0156113bb5182727a53aac594410ed051c19258beb8a07b808eff0b42a7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8673a29c0e16f4e707d79f0610722ccabb023ff2b4a18d7cf9d305183817d952\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://35766e7206b1be7a53d13d6890d759cf81dbd999643c12349268e4bf16fbbdb0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5b9844a11b2b903978d0d464b469e26c9211100da61d018dde9636dca95b684e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T14:09:30Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T14:09:55Z is after 2025-08-24T17:21:41Z" Dec 03 14:09:55 crc kubenswrapper[4805]: I1203 14:09:55.406155 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:48Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:48Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T14:09:55Z is after 2025-08-24T17:21:41Z" Dec 03 14:09:55 crc kubenswrapper[4805]: I1203 14:09:55.416316 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 14:09:55 crc kubenswrapper[4805]: I1203 14:09:55.416350 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 14:09:55 crc kubenswrapper[4805]: I1203 14:09:55.416360 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 14:09:55 crc kubenswrapper[4805]: I1203 14:09:55.416374 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 14:09:55 crc kubenswrapper[4805]: I1203 14:09:55.416383 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T14:09:55Z","lastTransitionTime":"2025-12-03T14:09:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 14:09:55 crc kubenswrapper[4805]: I1203 14:09:55.449523 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-k4mhs" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c09c1cdc-f461-45c9-8444-5b8764bccabc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3354c1a44f891182d04116fc76234a2b337d50567d032a3584f902a9516387cc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qdcd7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T14:09:50Z\\\"}}\" for pod \"openshift-multus\"/\"multus-k4mhs\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T14:09:55Z is after 2025-08-24T17:21:41Z" Dec 03 14:09:55 crc kubenswrapper[4805]: I1203 14:09:55.499981 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-psjpm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"deeec30d-cad2-4dfd-adb6-64f6646581be\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:50Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:50Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wdlzx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wdlzx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wdlzx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wdlzx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wdlzx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wdlzx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wdlzx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wdlzx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2f875252fbe6a673a72fc9c56bb96899f62732305ad77b3108e3b244bfa91ecb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2f875252fbe6a673a72fc9c56bb96899f62732305ad77b3108e3b244bfa91ecb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T14:09:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T14:09:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wdlzx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T14:09:50Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-psjpm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T14:09:55Z is after 2025-08-24T17:21:41Z" Dec 03 14:09:55 crc kubenswrapper[4805]: I1203 14:09:55.518583 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 14:09:55 crc kubenswrapper[4805]: I1203 14:09:55.518633 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 14:09:55 crc kubenswrapper[4805]: I1203 14:09:55.518649 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 14:09:55 crc kubenswrapper[4805]: I1203 14:09:55.518667 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 14:09:55 crc kubenswrapper[4805]: I1203 14:09:55.518682 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T14:09:55Z","lastTransitionTime":"2025-12-03T14:09:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 14:09:55 crc kubenswrapper[4805]: I1203 14:09:55.526814 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:48Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:48Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T14:09:55Z is after 2025-08-24T17:21:41Z" Dec 03 14:09:55 crc kubenswrapper[4805]: I1203 14:09:55.565374 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:49Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:49Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://541ace7ccbb66dbd191c3ca8c1e4479e3edbf80dfc67326bea8aaee81fdcfc3c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://821e3e9a902aacecf6c41e8b8d8aa1ac951ce2fbb6505b7148b16348a3754e55\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T14:09:55Z is after 2025-08-24T17:21:41Z" Dec 03 14:09:55 crc kubenswrapper[4805]: I1203 14:09:55.602291 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-gskh4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6098937f-e3f6-45e8-a647-4994a79cd711\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://28d5653f26ec61647808b3a733ff7eac70aa69ac184660f24147e8cfaf0eb802\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nwrw9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9ad6b68a42fd33ee167e67873c6eed3f11cdf6dc71d22e952c021436eb967dd9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nwrw9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T14:09:50Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-gskh4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T14:09:55Z is after 2025-08-24T17:21:41Z" Dec 03 14:09:55 crc kubenswrapper[4805]: I1203 14:09:55.620426 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 14:09:55 crc kubenswrapper[4805]: I1203 14:09:55.620455 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 14:09:55 crc kubenswrapper[4805]: I1203 14:09:55.620466 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 14:09:55 crc kubenswrapper[4805]: I1203 14:09:55.620482 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 14:09:55 crc kubenswrapper[4805]: I1203 14:09:55.620493 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T14:09:55Z","lastTransitionTime":"2025-12-03T14:09:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 14:09:55 crc kubenswrapper[4805]: I1203 14:09:55.723126 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 14:09:55 crc kubenswrapper[4805]: I1203 14:09:55.723172 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 14:09:55 crc kubenswrapper[4805]: I1203 14:09:55.723184 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 14:09:55 crc kubenswrapper[4805]: I1203 14:09:55.723201 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 14:09:55 crc kubenswrapper[4805]: I1203 14:09:55.723214 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T14:09:55Z","lastTransitionTime":"2025-12-03T14:09:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 14:09:55 crc kubenswrapper[4805]: I1203 14:09:55.825575 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 14:09:55 crc kubenswrapper[4805]: I1203 14:09:55.825620 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 14:09:55 crc kubenswrapper[4805]: I1203 14:09:55.825631 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 14:09:55 crc kubenswrapper[4805]: I1203 14:09:55.825647 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 14:09:55 crc kubenswrapper[4805]: I1203 14:09:55.825658 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T14:09:55Z","lastTransitionTime":"2025-12-03T14:09:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 14:09:55 crc kubenswrapper[4805]: I1203 14:09:55.857403 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-psjpm" event={"ID":"deeec30d-cad2-4dfd-adb6-64f6646581be","Type":"ContainerStarted","Data":"ab98e365c1a4242a47b8cbe2dd85109d366d6002c4de1f18ddc950a46bea6c12"} Dec 03 14:09:55 crc kubenswrapper[4805]: I1203 14:09:55.860103 4805 generic.go:334] "Generic (PLEG): container finished" podID="de00a81f-82e1-4223-9614-f98606b9a8ef" containerID="66707bb4e1fdacb265e89786df8a04605830f1c156394cacfcec9b54abe5d41a" exitCode=0 Dec 03 14:09:55 crc kubenswrapper[4805]: I1203 14:09:55.860153 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-bpk49" event={"ID":"de00a81f-82e1-4223-9614-f98606b9a8ef","Type":"ContainerDied","Data":"66707bb4e1fdacb265e89786df8a04605830f1c156394cacfcec9b54abe5d41a"} Dec 03 14:09:55 crc kubenswrapper[4805]: I1203 14:09:55.872924 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-pdzgk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"860769de-6a5b-4cb9-a752-264e7e2e785f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5e2f0a0cc21505d609f81918365d5e4f5d21959ac12ab147d6a51601c9de4c14\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ggn4n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T14:09:50Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-pdzgk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T14:09:55Z is after 2025-08-24T17:21:41Z" Dec 03 14:09:55 crc kubenswrapper[4805]: I1203 14:09:55.884495 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-bdtdg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"53a0e9e3-b6d8-44b7-a32b-aa3450442865\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f557e17ad105d6972d3c1480656c6073f67e28c0f5773aa365549951f4d161c7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9ww54\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T14:09:51Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-bdtdg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T14:09:55Z is after 2025-08-24T17:21:41Z" Dec 03 14:09:55 crc kubenswrapper[4805]: I1203 14:09:55.914126 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b6d36671-7173-4e9c-843f-005b4dcf576c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a837d7d491eaacbe4279b4014bb5fcbe9b514ba5cc82c708227bcd9b25a346e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6e5f04faf3b23197bb8ea11d3e574dc86f31e21d798734962001ec8c95de7c84\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2ff1a979387da41b00e95eb9cc8b2c9de966380a12fae687591df3c5b18a4497\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1e66e1cb18573a14e3c3ae16af4dbd9f8997a6ad38680c3ab4505646ebd48e7c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7d5b5c85e3de43e7e95111aa967248550607c26200c783251b68696f202f92a1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://474f72dac37a808fc4b9832306e76979b151d7221807ff3086013466ed1419c3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://474f72dac37a808fc4b9832306e76979b151d7221807ff3086013466ed1419c3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T14:09:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T14:09:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://74c8cea1a61c4775a0ad2f4ab4bb8a36437cf23d2ee953311d6ea967e297510a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://74c8cea1a61c4775a0ad2f4ab4bb8a36437cf23d2ee953311d6ea967e297510a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T14:09:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T14:09:31Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://a4f2f56a809adc01fba82eba0652d78f58770c132f77b1424f8b17068b1927e3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a4f2f56a809adc01fba82eba0652d78f58770c132f77b1424f8b17068b1927e3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T14:09:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T14:09:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T14:09:30Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T14:09:55Z is after 2025-08-24T17:21:41Z" Dec 03 14:09:55 crc kubenswrapper[4805]: I1203 14:09:55.928026 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ced8291d-0ce5-4e74-9eaf-2bf033774039\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:30Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:30Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3e5a34636f0a8c4d9179074596d28bc2eb7a504573398c596a637defd829907a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://29cd95be3f964415943c873e7914ffcea4abe50dfdf17f8a9cd03fc38df35981\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://951173b3d3d581275a6f6e43209b5e0611a701ca7a10f81504d723e24a2d21a3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e64730957bbbb2feb4d506e0ed9731a0594bf66fa1e8fcdb8ec944ded6308eed\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5da55f769e23ad52bac9c77da62c78e70fa34db8ed1a435e4eaf4dad3fd2b68a\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-03T14:09:48Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1203 14:09:42.992179 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1203 14:09:42.993233 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1830926909/tls.crt::/tmp/serving-cert-1830926909/tls.key\\\\\\\"\\\\nI1203 14:09:48.622403 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1203 14:09:48.626476 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1203 14:09:48.626527 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1203 14:09:48.626574 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1203 14:09:48.626582 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1203 14:09:48.633990 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1203 14:09:48.634024 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 14:09:48.634029 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 14:09:48.634034 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1203 14:09:48.634038 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1203 14:09:48.634042 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1203 14:09:48.634045 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1203 14:09:48.634344 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1203 14:09:48.644497 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T14:09:32Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bd92a160eb530c5687ebd7cd5b83b41820a2b808a0fd9514e5ace9d237e1bd26\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:32Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://af9c940135012c629447528c61743ed1e58708ab57ce76604c6da0479aa459cf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://af9c940135012c629447528c61743ed1e58708ab57ce76604c6da0479aa459cf\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T14:09:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T14:09:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T14:09:30Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T14:09:55Z is after 2025-08-24T17:21:41Z" Dec 03 14:09:55 crc kubenswrapper[4805]: I1203 14:09:55.928385 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 14:09:55 crc kubenswrapper[4805]: I1203 14:09:55.928429 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 14:09:55 crc kubenswrapper[4805]: I1203 14:09:55.928440 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 14:09:55 crc kubenswrapper[4805]: I1203 14:09:55.928459 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 14:09:55 crc kubenswrapper[4805]: I1203 14:09:55.928472 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T14:09:55Z","lastTransitionTime":"2025-12-03T14:09:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 14:09:55 crc kubenswrapper[4805]: I1203 14:09:55.940449 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:49Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:49Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://38339331a3209d6787d1def81b72f4a8514bf91080a139cad6d8ac5cc6bef10e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T14:09:55Z is after 2025-08-24T17:21:41Z" Dec 03 14:09:55 crc kubenswrapper[4805]: I1203 14:09:55.951359 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:48Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:48Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T14:09:55Z is after 2025-08-24T17:21:41Z" Dec 03 14:09:55 crc kubenswrapper[4805]: I1203 14:09:55.961574 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:51Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:51Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b3123d3fa879097db88ed2f3cfb05631f440c24c5405503b2de3546f12e2e899\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T14:09:55Z is after 2025-08-24T17:21:41Z" Dec 03 14:09:55 crc kubenswrapper[4805]: I1203 14:09:55.974129 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"750d2b4b-738a-4992-8902-742528f6e875\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://abd2d0156113bb5182727a53aac594410ed051c19258beb8a07b808eff0b42a7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8673a29c0e16f4e707d79f0610722ccabb023ff2b4a18d7cf9d305183817d952\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://35766e7206b1be7a53d13d6890d759cf81dbd999643c12349268e4bf16fbbdb0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5b9844a11b2b903978d0d464b469e26c9211100da61d018dde9636dca95b684e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T14:09:30Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T14:09:55Z is after 2025-08-24T17:21:41Z" Dec 03 14:09:55 crc kubenswrapper[4805]: I1203 14:09:55.985116 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:48Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:48Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T14:09:55Z is after 2025-08-24T17:21:41Z" Dec 03 14:09:56 crc kubenswrapper[4805]: I1203 14:09:56.001469 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-k4mhs" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c09c1cdc-f461-45c9-8444-5b8764bccabc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3354c1a44f891182d04116fc76234a2b337d50567d032a3584f902a9516387cc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qdcd7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T14:09:50Z\\\"}}\" for pod \"openshift-multus\"/\"multus-k4mhs\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T14:09:55Z is after 2025-08-24T17:21:41Z" Dec 03 14:09:56 crc kubenswrapper[4805]: I1203 14:09:56.029916 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 14:09:56 crc kubenswrapper[4805]: I1203 14:09:56.029966 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 14:09:56 crc kubenswrapper[4805]: I1203 14:09:56.029978 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 14:09:56 crc kubenswrapper[4805]: I1203 14:09:56.029993 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 14:09:56 crc kubenswrapper[4805]: I1203 14:09:56.030003 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T14:09:56Z","lastTransitionTime":"2025-12-03T14:09:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 14:09:56 crc kubenswrapper[4805]: I1203 14:09:56.041928 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-bpk49" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"de00a81f-82e1-4223-9614-f98606b9a8ef\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:50Z\\\",\\\"message\\\":\\\"containers with incomplete status: [whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:50Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:50Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-25d7t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7741e0e70416ebe094bfabc6009e98d7d4c0fbcba1fc82df13b9ba40774b456a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7741e0e70416ebe094bfabc6009e98d7d4c0fbcba1fc82df13b9ba40774b456a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T14:09:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T14:09:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-25d7t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ee3499c02eb52bbf08c447d2433d435a74dfb70670dc4cbaac44f5fbff9e8a29\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ee3499c02eb52bbf08c447d2433d435a74dfb70670dc4cbaac44f5fbff9e8a29\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T14:09:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T14:09:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-25d7t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ec8488b1d69799951be7ef40e5e028732265c4ddcdea2d453923e567864771f7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ec8488b1d69799951be7ef40e5e028732265c4ddcdea2d453923e567864771f7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T14:09:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T14:09:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-25d7t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://66707bb4e1fdacb265e89786df8a04605830f1c156394cacfcec9b54abe5d41a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://66707bb4e1fdacb265e89786df8a04605830f1c156394cacfcec9b54abe5d41a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T14:09:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T14:09:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-25d7t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-25d7t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-25d7t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T14:09:50Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-bpk49\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T14:09:56Z is after 2025-08-24T17:21:41Z" Dec 03 14:09:56 crc kubenswrapper[4805]: I1203 14:09:56.083725 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:48Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:48Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T14:09:56Z is after 2025-08-24T17:21:41Z" Dec 03 14:09:56 crc kubenswrapper[4805]: I1203 14:09:56.122612 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:49Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:49Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://541ace7ccbb66dbd191c3ca8c1e4479e3edbf80dfc67326bea8aaee81fdcfc3c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://821e3e9a902aacecf6c41e8b8d8aa1ac951ce2fbb6505b7148b16348a3754e55\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T14:09:56Z is after 2025-08-24T17:21:41Z" Dec 03 14:09:56 crc kubenswrapper[4805]: I1203 14:09:56.132646 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 14:09:56 crc kubenswrapper[4805]: I1203 14:09:56.132710 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 14:09:56 crc kubenswrapper[4805]: I1203 14:09:56.132724 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 14:09:56 crc kubenswrapper[4805]: I1203 14:09:56.132742 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 14:09:56 crc kubenswrapper[4805]: I1203 14:09:56.132754 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T14:09:56Z","lastTransitionTime":"2025-12-03T14:09:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 14:09:56 crc kubenswrapper[4805]: I1203 14:09:56.162150 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-gskh4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6098937f-e3f6-45e8-a647-4994a79cd711\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://28d5653f26ec61647808b3a733ff7eac70aa69ac184660f24147e8cfaf0eb802\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nwrw9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9ad6b68a42fd33ee167e67873c6eed3f11cdf6dc71d22e952c021436eb967dd9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nwrw9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T14:09:50Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-gskh4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T14:09:56Z is after 2025-08-24T17:21:41Z" Dec 03 14:09:56 crc kubenswrapper[4805]: I1203 14:09:56.235566 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 14:09:56 crc kubenswrapper[4805]: I1203 14:09:56.235611 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 14:09:56 crc kubenswrapper[4805]: I1203 14:09:56.235628 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 14:09:56 crc kubenswrapper[4805]: I1203 14:09:56.235650 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 14:09:56 crc kubenswrapper[4805]: I1203 14:09:56.235714 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T14:09:56Z","lastTransitionTime":"2025-12-03T14:09:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 14:09:56 crc kubenswrapper[4805]: I1203 14:09:56.240017 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-psjpm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"deeec30d-cad2-4dfd-adb6-64f6646581be\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:50Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:50Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wdlzx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wdlzx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wdlzx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wdlzx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wdlzx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wdlzx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wdlzx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wdlzx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2f875252fbe6a673a72fc9c56bb96899f62732305ad77b3108e3b244bfa91ecb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2f875252fbe6a673a72fc9c56bb96899f62732305ad77b3108e3b244bfa91ecb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T14:09:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T14:09:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wdlzx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T14:09:50Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-psjpm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T14:09:56Z is after 2025-08-24T17:21:41Z" Dec 03 14:09:56 crc kubenswrapper[4805]: I1203 14:09:56.338052 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 14:09:56 crc kubenswrapper[4805]: I1203 14:09:56.338092 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 14:09:56 crc kubenswrapper[4805]: I1203 14:09:56.338104 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 14:09:56 crc kubenswrapper[4805]: I1203 14:09:56.338118 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 14:09:56 crc kubenswrapper[4805]: I1203 14:09:56.338127 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T14:09:56Z","lastTransitionTime":"2025-12-03T14:09:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 14:09:56 crc kubenswrapper[4805]: I1203 14:09:56.425790 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 14:09:56 crc kubenswrapper[4805]: E1203 14:09:56.426101 4805 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-03 14:10:04.426072757 +0000 UTC m=+34.088989720 (durationBeforeRetry 8s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 14:09:56 crc kubenswrapper[4805]: I1203 14:09:56.440353 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 14:09:56 crc kubenswrapper[4805]: I1203 14:09:56.440400 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 14:09:56 crc kubenswrapper[4805]: I1203 14:09:56.440409 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 14:09:56 crc kubenswrapper[4805]: I1203 14:09:56.440424 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 14:09:56 crc kubenswrapper[4805]: I1203 14:09:56.440436 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T14:09:56Z","lastTransitionTime":"2025-12-03T14:09:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 14:09:56 crc kubenswrapper[4805]: I1203 14:09:56.526792 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 14:09:56 crc kubenswrapper[4805]: I1203 14:09:56.526907 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 14:09:56 crc kubenswrapper[4805]: I1203 14:09:56.526948 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 14:09:56 crc kubenswrapper[4805]: I1203 14:09:56.526986 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 14:09:56 crc kubenswrapper[4805]: E1203 14:09:56.527011 4805 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 03 14:09:56 crc kubenswrapper[4805]: E1203 14:09:56.527052 4805 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 03 14:09:56 crc kubenswrapper[4805]: E1203 14:09:56.527064 4805 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 03 14:09:56 crc kubenswrapper[4805]: E1203 14:09:56.527082 4805 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Dec 03 14:09:56 crc kubenswrapper[4805]: E1203 14:09:56.527105 4805 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 03 14:09:56 crc kubenswrapper[4805]: E1203 14:09:56.527125 4805 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-12-03 14:10:04.527107949 +0000 UTC m=+34.190024912 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 03 14:09:56 crc kubenswrapper[4805]: E1203 14:09:56.527164 4805 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-03 14:10:04.52714128 +0000 UTC m=+34.190058263 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Dec 03 14:09:56 crc kubenswrapper[4805]: E1203 14:09:56.527091 4805 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 03 14:09:56 crc kubenswrapper[4805]: E1203 14:09:56.527188 4805 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-03 14:10:04.527176171 +0000 UTC m=+34.190093194 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 03 14:09:56 crc kubenswrapper[4805]: E1203 14:09:56.527199 4805 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 03 14:09:56 crc kubenswrapper[4805]: E1203 14:09:56.527212 4805 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 03 14:09:56 crc kubenswrapper[4805]: E1203 14:09:56.527250 4805 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-12-03 14:10:04.527238103 +0000 UTC m=+34.190155106 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 03 14:09:56 crc kubenswrapper[4805]: I1203 14:09:56.543141 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 14:09:56 crc kubenswrapper[4805]: I1203 14:09:56.543189 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 14:09:56 crc kubenswrapper[4805]: I1203 14:09:56.543202 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 14:09:56 crc kubenswrapper[4805]: I1203 14:09:56.543219 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 14:09:56 crc kubenswrapper[4805]: I1203 14:09:56.543230 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T14:09:56Z","lastTransitionTime":"2025-12-03T14:09:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 14:09:56 crc kubenswrapper[4805]: I1203 14:09:56.646020 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 14:09:56 crc kubenswrapper[4805]: I1203 14:09:56.646058 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 14:09:56 crc kubenswrapper[4805]: I1203 14:09:56.646069 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 14:09:56 crc kubenswrapper[4805]: I1203 14:09:56.646084 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 14:09:56 crc kubenswrapper[4805]: I1203 14:09:56.646096 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T14:09:56Z","lastTransitionTime":"2025-12-03T14:09:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 14:09:56 crc kubenswrapper[4805]: I1203 14:09:56.693975 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 14:09:56 crc kubenswrapper[4805]: I1203 14:09:56.694014 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 14:09:56 crc kubenswrapper[4805]: E1203 14:09:56.694434 4805 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 03 14:09:56 crc kubenswrapper[4805]: E1203 14:09:56.698172 4805 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 03 14:09:56 crc kubenswrapper[4805]: I1203 14:09:56.698254 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 14:09:56 crc kubenswrapper[4805]: E1203 14:09:56.698382 4805 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 03 14:09:56 crc kubenswrapper[4805]: I1203 14:09:56.748202 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 14:09:56 crc kubenswrapper[4805]: I1203 14:09:56.748247 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 14:09:56 crc kubenswrapper[4805]: I1203 14:09:56.748259 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 14:09:56 crc kubenswrapper[4805]: I1203 14:09:56.748278 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 14:09:56 crc kubenswrapper[4805]: I1203 14:09:56.748290 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T14:09:56Z","lastTransitionTime":"2025-12-03T14:09:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 14:09:56 crc kubenswrapper[4805]: I1203 14:09:56.850921 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 14:09:56 crc kubenswrapper[4805]: I1203 14:09:56.850984 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 14:09:56 crc kubenswrapper[4805]: I1203 14:09:56.851001 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 14:09:56 crc kubenswrapper[4805]: I1203 14:09:56.851026 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 14:09:56 crc kubenswrapper[4805]: I1203 14:09:56.851045 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T14:09:56Z","lastTransitionTime":"2025-12-03T14:09:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 14:09:56 crc kubenswrapper[4805]: I1203 14:09:56.868817 4805 generic.go:334] "Generic (PLEG): container finished" podID="de00a81f-82e1-4223-9614-f98606b9a8ef" containerID="6207ab6600970ff15c8713c547d4920b2c77eb9bea990bb5da497e24870dcc0f" exitCode=0 Dec 03 14:09:56 crc kubenswrapper[4805]: I1203 14:09:56.868886 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-bpk49" event={"ID":"de00a81f-82e1-4223-9614-f98606b9a8ef","Type":"ContainerDied","Data":"6207ab6600970ff15c8713c547d4920b2c77eb9bea990bb5da497e24870dcc0f"} Dec 03 14:09:56 crc kubenswrapper[4805]: I1203 14:09:56.889937 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-bpk49" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"de00a81f-82e1-4223-9614-f98606b9a8ef\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:50Z\\\",\\\"message\\\":\\\"containers with incomplete status: [whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:50Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:50Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-25d7t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7741e0e70416ebe094bfabc6009e98d7d4c0fbcba1fc82df13b9ba40774b456a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7741e0e70416ebe094bfabc6009e98d7d4c0fbcba1fc82df13b9ba40774b456a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T14:09:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T14:09:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-25d7t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ee3499c02eb52bbf08c447d2433d435a74dfb70670dc4cbaac44f5fbff9e8a29\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ee3499c02eb52bbf08c447d2433d435a74dfb70670dc4cbaac44f5fbff9e8a29\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T14:09:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T14:09:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-25d7t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ec8488b1d69799951be7ef40e5e028732265c4ddcdea2d453923e567864771f7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ec8488b1d69799951be7ef40e5e028732265c4ddcdea2d453923e567864771f7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T14:09:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T14:09:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-25d7t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://66707bb4e1fdacb265e89786df8a04605830f1c156394cacfcec9b54abe5d41a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://66707bb4e1fdacb265e89786df8a04605830f1c156394cacfcec9b54abe5d41a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T14:09:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T14:09:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-25d7t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6207ab6600970ff15c8713c547d4920b2c77eb9bea990bb5da497e24870dcc0f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6207ab6600970ff15c8713c547d4920b2c77eb9bea990bb5da497e24870dcc0f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T14:09:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T14:09:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-25d7t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-25d7t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T14:09:50Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-bpk49\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T14:09:56Z is after 2025-08-24T17:21:41Z" Dec 03 14:09:56 crc kubenswrapper[4805]: I1203 14:09:56.902064 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"750d2b4b-738a-4992-8902-742528f6e875\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://abd2d0156113bb5182727a53aac594410ed051c19258beb8a07b808eff0b42a7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8673a29c0e16f4e707d79f0610722ccabb023ff2b4a18d7cf9d305183817d952\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://35766e7206b1be7a53d13d6890d759cf81dbd999643c12349268e4bf16fbbdb0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5b9844a11b2b903978d0d464b469e26c9211100da61d018dde9636dca95b684e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T14:09:30Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T14:09:56Z is after 2025-08-24T17:21:41Z" Dec 03 14:09:56 crc kubenswrapper[4805]: I1203 14:09:56.914209 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:48Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:48Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T14:09:56Z is after 2025-08-24T17:21:41Z" Dec 03 14:09:56 crc kubenswrapper[4805]: I1203 14:09:56.927106 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-k4mhs" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c09c1cdc-f461-45c9-8444-5b8764bccabc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3354c1a44f891182d04116fc76234a2b337d50567d032a3584f902a9516387cc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qdcd7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T14:09:50Z\\\"}}\" for pod \"openshift-multus\"/\"multus-k4mhs\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T14:09:56Z is after 2025-08-24T17:21:41Z" Dec 03 14:09:56 crc kubenswrapper[4805]: I1203 14:09:56.947146 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-psjpm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"deeec30d-cad2-4dfd-adb6-64f6646581be\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:50Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:50Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wdlzx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wdlzx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wdlzx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wdlzx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wdlzx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wdlzx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wdlzx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wdlzx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2f875252fbe6a673a72fc9c56bb96899f62732305ad77b3108e3b244bfa91ecb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2f875252fbe6a673a72fc9c56bb96899f62732305ad77b3108e3b244bfa91ecb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T14:09:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T14:09:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wdlzx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T14:09:50Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-psjpm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T14:09:56Z is after 2025-08-24T17:21:41Z" Dec 03 14:09:56 crc kubenswrapper[4805]: I1203 14:09:56.953007 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 14:09:56 crc kubenswrapper[4805]: I1203 14:09:56.953047 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 14:09:56 crc kubenswrapper[4805]: I1203 14:09:56.953058 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 14:09:56 crc kubenswrapper[4805]: I1203 14:09:56.953075 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 14:09:56 crc kubenswrapper[4805]: I1203 14:09:56.953086 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T14:09:56Z","lastTransitionTime":"2025-12-03T14:09:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 14:09:56 crc kubenswrapper[4805]: I1203 14:09:56.959391 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:48Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:48Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T14:09:56Z is after 2025-08-24T17:21:41Z" Dec 03 14:09:56 crc kubenswrapper[4805]: I1203 14:09:56.970442 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:49Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:49Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://541ace7ccbb66dbd191c3ca8c1e4479e3edbf80dfc67326bea8aaee81fdcfc3c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://821e3e9a902aacecf6c41e8b8d8aa1ac951ce2fbb6505b7148b16348a3754e55\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T14:09:56Z is after 2025-08-24T17:21:41Z" Dec 03 14:09:56 crc kubenswrapper[4805]: I1203 14:09:56.982379 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-gskh4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6098937f-e3f6-45e8-a647-4994a79cd711\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://28d5653f26ec61647808b3a733ff7eac70aa69ac184660f24147e8cfaf0eb802\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nwrw9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9ad6b68a42fd33ee167e67873c6eed3f11cdf6dc71d22e952c021436eb967dd9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nwrw9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T14:09:50Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-gskh4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T14:09:56Z is after 2025-08-24T17:21:41Z" Dec 03 14:09:56 crc kubenswrapper[4805]: I1203 14:09:56.991405 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-pdzgk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"860769de-6a5b-4cb9-a752-264e7e2e785f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5e2f0a0cc21505d609f81918365d5e4f5d21959ac12ab147d6a51601c9de4c14\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ggn4n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T14:09:50Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-pdzgk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T14:09:56Z is after 2025-08-24T17:21:41Z" Dec 03 14:09:57 crc kubenswrapper[4805]: I1203 14:09:57.004174 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:49Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:49Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://38339331a3209d6787d1def81b72f4a8514bf91080a139cad6d8ac5cc6bef10e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T14:09:57Z is after 2025-08-24T17:21:41Z" Dec 03 14:09:57 crc kubenswrapper[4805]: I1203 14:09:57.014876 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:48Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:48Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T14:09:57Z is after 2025-08-24T17:21:41Z" Dec 03 14:09:57 crc kubenswrapper[4805]: I1203 14:09:57.026900 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:51Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:51Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b3123d3fa879097db88ed2f3cfb05631f440c24c5405503b2de3546f12e2e899\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T14:09:57Z is after 2025-08-24T17:21:41Z" Dec 03 14:09:57 crc kubenswrapper[4805]: I1203 14:09:57.036093 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-bdtdg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"53a0e9e3-b6d8-44b7-a32b-aa3450442865\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f557e17ad105d6972d3c1480656c6073f67e28c0f5773aa365549951f4d161c7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9ww54\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T14:09:51Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-bdtdg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T14:09:57Z is after 2025-08-24T17:21:41Z" Dec 03 14:09:57 crc kubenswrapper[4805]: I1203 14:09:57.060338 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 14:09:57 crc kubenswrapper[4805]: I1203 14:09:57.060373 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 14:09:57 crc kubenswrapper[4805]: I1203 14:09:57.060382 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 14:09:57 crc kubenswrapper[4805]: I1203 14:09:57.060398 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 14:09:57 crc kubenswrapper[4805]: I1203 14:09:57.060410 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T14:09:57Z","lastTransitionTime":"2025-12-03T14:09:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 14:09:57 crc kubenswrapper[4805]: I1203 14:09:57.084544 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b6d36671-7173-4e9c-843f-005b4dcf576c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a837d7d491eaacbe4279b4014bb5fcbe9b514ba5cc82c708227bcd9b25a346e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6e5f04faf3b23197bb8ea11d3e574dc86f31e21d798734962001ec8c95de7c84\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2ff1a979387da41b00e95eb9cc8b2c9de966380a12fae687591df3c5b18a4497\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1e66e1cb18573a14e3c3ae16af4dbd9f8997a6ad38680c3ab4505646ebd48e7c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7d5b5c85e3de43e7e95111aa967248550607c26200c783251b68696f202f92a1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://474f72dac37a808fc4b9832306e76979b151d7221807ff3086013466ed1419c3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://474f72dac37a808fc4b9832306e76979b151d7221807ff3086013466ed1419c3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T14:09:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T14:09:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://74c8cea1a61c4775a0ad2f4ab4bb8a36437cf23d2ee953311d6ea967e297510a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://74c8cea1a61c4775a0ad2f4ab4bb8a36437cf23d2ee953311d6ea967e297510a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T14:09:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T14:09:31Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://a4f2f56a809adc01fba82eba0652d78f58770c132f77b1424f8b17068b1927e3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a4f2f56a809adc01fba82eba0652d78f58770c132f77b1424f8b17068b1927e3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T14:09:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T14:09:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T14:09:30Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T14:09:57Z is after 2025-08-24T17:21:41Z" Dec 03 14:09:57 crc kubenswrapper[4805]: I1203 14:09:57.103118 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ced8291d-0ce5-4e74-9eaf-2bf033774039\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:30Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:30Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3e5a34636f0a8c4d9179074596d28bc2eb7a504573398c596a637defd829907a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://29cd95be3f964415943c873e7914ffcea4abe50dfdf17f8a9cd03fc38df35981\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://951173b3d3d581275a6f6e43209b5e0611a701ca7a10f81504d723e24a2d21a3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e64730957bbbb2feb4d506e0ed9731a0594bf66fa1e8fcdb8ec944ded6308eed\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5da55f769e23ad52bac9c77da62c78e70fa34db8ed1a435e4eaf4dad3fd2b68a\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-03T14:09:48Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1203 14:09:42.992179 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1203 14:09:42.993233 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1830926909/tls.crt::/tmp/serving-cert-1830926909/tls.key\\\\\\\"\\\\nI1203 14:09:48.622403 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1203 14:09:48.626476 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1203 14:09:48.626527 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1203 14:09:48.626574 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1203 14:09:48.626582 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1203 14:09:48.633990 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1203 14:09:48.634024 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 14:09:48.634029 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 14:09:48.634034 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1203 14:09:48.634038 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1203 14:09:48.634042 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1203 14:09:48.634045 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1203 14:09:48.634344 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1203 14:09:48.644497 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T14:09:32Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bd92a160eb530c5687ebd7cd5b83b41820a2b808a0fd9514e5ace9d237e1bd26\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:32Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://af9c940135012c629447528c61743ed1e58708ab57ce76604c6da0479aa459cf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://af9c940135012c629447528c61743ed1e58708ab57ce76604c6da0479aa459cf\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T14:09:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T14:09:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T14:09:30Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T14:09:57Z is after 2025-08-24T17:21:41Z" Dec 03 14:09:57 crc kubenswrapper[4805]: I1203 14:09:57.162301 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 14:09:57 crc kubenswrapper[4805]: I1203 14:09:57.162338 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 14:09:57 crc kubenswrapper[4805]: I1203 14:09:57.162346 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 14:09:57 crc kubenswrapper[4805]: I1203 14:09:57.162358 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 14:09:57 crc kubenswrapper[4805]: I1203 14:09:57.162367 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T14:09:57Z","lastTransitionTime":"2025-12-03T14:09:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 14:09:57 crc kubenswrapper[4805]: I1203 14:09:57.264312 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 14:09:57 crc kubenswrapper[4805]: I1203 14:09:57.264355 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 14:09:57 crc kubenswrapper[4805]: I1203 14:09:57.264366 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 14:09:57 crc kubenswrapper[4805]: I1203 14:09:57.264382 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 14:09:57 crc kubenswrapper[4805]: I1203 14:09:57.264393 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T14:09:57Z","lastTransitionTime":"2025-12-03T14:09:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 14:09:57 crc kubenswrapper[4805]: I1203 14:09:57.367744 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 14:09:57 crc kubenswrapper[4805]: I1203 14:09:57.367785 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 14:09:57 crc kubenswrapper[4805]: I1203 14:09:57.367795 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 14:09:57 crc kubenswrapper[4805]: I1203 14:09:57.367810 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 14:09:57 crc kubenswrapper[4805]: I1203 14:09:57.367820 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T14:09:57Z","lastTransitionTime":"2025-12-03T14:09:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 14:09:57 crc kubenswrapper[4805]: I1203 14:09:57.470526 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 14:09:57 crc kubenswrapper[4805]: I1203 14:09:57.470611 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 14:09:57 crc kubenswrapper[4805]: I1203 14:09:57.470636 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 14:09:57 crc kubenswrapper[4805]: I1203 14:09:57.470664 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 14:09:57 crc kubenswrapper[4805]: I1203 14:09:57.470689 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T14:09:57Z","lastTransitionTime":"2025-12-03T14:09:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 14:09:57 crc kubenswrapper[4805]: I1203 14:09:57.573989 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 14:09:57 crc kubenswrapper[4805]: I1203 14:09:57.574027 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 14:09:57 crc kubenswrapper[4805]: I1203 14:09:57.574034 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 14:09:57 crc kubenswrapper[4805]: I1203 14:09:57.574049 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 14:09:57 crc kubenswrapper[4805]: I1203 14:09:57.574059 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T14:09:57Z","lastTransitionTime":"2025-12-03T14:09:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 14:09:57 crc kubenswrapper[4805]: I1203 14:09:57.676458 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 14:09:57 crc kubenswrapper[4805]: I1203 14:09:57.676532 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 14:09:57 crc kubenswrapper[4805]: I1203 14:09:57.676554 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 14:09:57 crc kubenswrapper[4805]: I1203 14:09:57.676585 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 14:09:57 crc kubenswrapper[4805]: I1203 14:09:57.676607 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T14:09:57Z","lastTransitionTime":"2025-12-03T14:09:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 14:09:57 crc kubenswrapper[4805]: I1203 14:09:57.780644 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 14:09:57 crc kubenswrapper[4805]: I1203 14:09:57.781185 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 14:09:57 crc kubenswrapper[4805]: I1203 14:09:57.781204 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 14:09:57 crc kubenswrapper[4805]: I1203 14:09:57.781227 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 14:09:57 crc kubenswrapper[4805]: I1203 14:09:57.781244 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T14:09:57Z","lastTransitionTime":"2025-12-03T14:09:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 14:09:57 crc kubenswrapper[4805]: I1203 14:09:57.881147 4805 generic.go:334] "Generic (PLEG): container finished" podID="de00a81f-82e1-4223-9614-f98606b9a8ef" containerID="8de65d4a205e2deec0b433e53c0e37feb94aa0ede6a4c58d7414b80829804505" exitCode=0 Dec 03 14:09:57 crc kubenswrapper[4805]: I1203 14:09:57.881231 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-bpk49" event={"ID":"de00a81f-82e1-4223-9614-f98606b9a8ef","Type":"ContainerDied","Data":"8de65d4a205e2deec0b433e53c0e37feb94aa0ede6a4c58d7414b80829804505"} Dec 03 14:09:57 crc kubenswrapper[4805]: I1203 14:09:57.885106 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 14:09:57 crc kubenswrapper[4805]: I1203 14:09:57.885146 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 14:09:57 crc kubenswrapper[4805]: I1203 14:09:57.885161 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 14:09:57 crc kubenswrapper[4805]: I1203 14:09:57.885184 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 14:09:57 crc kubenswrapper[4805]: I1203 14:09:57.885199 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T14:09:57Z","lastTransitionTime":"2025-12-03T14:09:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 14:09:57 crc kubenswrapper[4805]: I1203 14:09:57.909402 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ced8291d-0ce5-4e74-9eaf-2bf033774039\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:30Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:30Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3e5a34636f0a8c4d9179074596d28bc2eb7a504573398c596a637defd829907a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://29cd95be3f964415943c873e7914ffcea4abe50dfdf17f8a9cd03fc38df35981\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://951173b3d3d581275a6f6e43209b5e0611a701ca7a10f81504d723e24a2d21a3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e64730957bbbb2feb4d506e0ed9731a0594bf66fa1e8fcdb8ec944ded6308eed\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5da55f769e23ad52bac9c77da62c78e70fa34db8ed1a435e4eaf4dad3fd2b68a\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-03T14:09:48Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1203 14:09:42.992179 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1203 14:09:42.993233 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1830926909/tls.crt::/tmp/serving-cert-1830926909/tls.key\\\\\\\"\\\\nI1203 14:09:48.622403 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1203 14:09:48.626476 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1203 14:09:48.626527 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1203 14:09:48.626574 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1203 14:09:48.626582 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1203 14:09:48.633990 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1203 14:09:48.634024 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 14:09:48.634029 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 14:09:48.634034 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1203 14:09:48.634038 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1203 14:09:48.634042 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1203 14:09:48.634045 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1203 14:09:48.634344 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1203 14:09:48.644497 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T14:09:32Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bd92a160eb530c5687ebd7cd5b83b41820a2b808a0fd9514e5ace9d237e1bd26\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:32Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://af9c940135012c629447528c61743ed1e58708ab57ce76604c6da0479aa459cf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://af9c940135012c629447528c61743ed1e58708ab57ce76604c6da0479aa459cf\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T14:09:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T14:09:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T14:09:30Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T14:09:57Z is after 2025-08-24T17:21:41Z" Dec 03 14:09:57 crc kubenswrapper[4805]: I1203 14:09:57.932501 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:49Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:49Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://38339331a3209d6787d1def81b72f4a8514bf91080a139cad6d8ac5cc6bef10e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T14:09:57Z is after 2025-08-24T17:21:41Z" Dec 03 14:09:57 crc kubenswrapper[4805]: I1203 14:09:57.955292 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:48Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:48Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T14:09:57Z is after 2025-08-24T17:21:41Z" Dec 03 14:09:57 crc kubenswrapper[4805]: I1203 14:09:57.977313 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:51Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:51Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b3123d3fa879097db88ed2f3cfb05631f440c24c5405503b2de3546f12e2e899\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T14:09:57Z is after 2025-08-24T17:21:41Z" Dec 03 14:09:57 crc kubenswrapper[4805]: I1203 14:09:57.989063 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 14:09:57 crc kubenswrapper[4805]: I1203 14:09:57.989105 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 14:09:57 crc kubenswrapper[4805]: I1203 14:09:57.989115 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 14:09:57 crc kubenswrapper[4805]: I1203 14:09:57.989130 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 14:09:57 crc kubenswrapper[4805]: I1203 14:09:57.989141 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T14:09:57Z","lastTransitionTime":"2025-12-03T14:09:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 14:09:57 crc kubenswrapper[4805]: I1203 14:09:57.992032 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-bdtdg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"53a0e9e3-b6d8-44b7-a32b-aa3450442865\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f557e17ad105d6972d3c1480656c6073f67e28c0f5773aa365549951f4d161c7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9ww54\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T14:09:51Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-bdtdg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T14:09:57Z is after 2025-08-24T17:21:41Z" Dec 03 14:09:58 crc kubenswrapper[4805]: I1203 14:09:58.017451 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b6d36671-7173-4e9c-843f-005b4dcf576c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a837d7d491eaacbe4279b4014bb5fcbe9b514ba5cc82c708227bcd9b25a346e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6e5f04faf3b23197bb8ea11d3e574dc86f31e21d798734962001ec8c95de7c84\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2ff1a979387da41b00e95eb9cc8b2c9de966380a12fae687591df3c5b18a4497\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1e66e1cb18573a14e3c3ae16af4dbd9f8997a6ad38680c3ab4505646ebd48e7c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7d5b5c85e3de43e7e95111aa967248550607c26200c783251b68696f202f92a1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://474f72dac37a808fc4b9832306e76979b151d7221807ff3086013466ed1419c3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://474f72dac37a808fc4b9832306e76979b151d7221807ff3086013466ed1419c3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T14:09:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T14:09:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://74c8cea1a61c4775a0ad2f4ab4bb8a36437cf23d2ee953311d6ea967e297510a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://74c8cea1a61c4775a0ad2f4ab4bb8a36437cf23d2ee953311d6ea967e297510a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T14:09:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T14:09:31Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://a4f2f56a809adc01fba82eba0652d78f58770c132f77b1424f8b17068b1927e3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a4f2f56a809adc01fba82eba0652d78f58770c132f77b1424f8b17068b1927e3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T14:09:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T14:09:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T14:09:30Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T14:09:58Z is after 2025-08-24T17:21:41Z" Dec 03 14:09:58 crc kubenswrapper[4805]: I1203 14:09:58.029651 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:48Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:48Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T14:09:58Z is after 2025-08-24T17:21:41Z" Dec 03 14:09:58 crc kubenswrapper[4805]: I1203 14:09:58.044328 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-k4mhs" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c09c1cdc-f461-45c9-8444-5b8764bccabc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3354c1a44f891182d04116fc76234a2b337d50567d032a3584f902a9516387cc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qdcd7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T14:09:50Z\\\"}}\" for pod \"openshift-multus\"/\"multus-k4mhs\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T14:09:58Z is after 2025-08-24T17:21:41Z" Dec 03 14:09:58 crc kubenswrapper[4805]: I1203 14:09:58.059218 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-bpk49" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"de00a81f-82e1-4223-9614-f98606b9a8ef\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:50Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:50Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-25d7t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7741e0e70416ebe094bfabc6009e98d7d4c0fbcba1fc82df13b9ba40774b456a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7741e0e70416ebe094bfabc6009e98d7d4c0fbcba1fc82df13b9ba40774b456a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T14:09:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T14:09:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-25d7t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ee3499c02eb52bbf08c447d2433d435a74dfb70670dc4cbaac44f5fbff9e8a29\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ee3499c02eb52bbf08c447d2433d435a74dfb70670dc4cbaac44f5fbff9e8a29\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T14:09:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T14:09:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-25d7t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ec8488b1d69799951be7ef40e5e028732265c4ddcdea2d453923e567864771f7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ec8488b1d69799951be7ef40e5e028732265c4ddcdea2d453923e567864771f7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T14:09:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T14:09:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-25d7t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://66707bb4e1fdacb265e89786df8a04605830f1c156394cacfcec9b54abe5d41a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://66707bb4e1fdacb265e89786df8a04605830f1c156394cacfcec9b54abe5d41a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T14:09:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T14:09:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-25d7t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6207ab6600970ff15c8713c547d4920b2c77eb9bea990bb5da497e24870dcc0f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6207ab6600970ff15c8713c547d4920b2c77eb9bea990bb5da497e24870dcc0f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T14:09:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T14:09:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-25d7t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8de65d4a205e2deec0b433e53c0e37feb94aa0ede6a4c58d7414b80829804505\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8de65d4a205e2deec0b433e53c0e37feb94aa0ede6a4c58d7414b80829804505\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T14:09:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T14:09:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-25d7t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T14:09:50Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-bpk49\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T14:09:58Z is after 2025-08-24T17:21:41Z" Dec 03 14:09:58 crc kubenswrapper[4805]: I1203 14:09:58.073497 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"750d2b4b-738a-4992-8902-742528f6e875\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://abd2d0156113bb5182727a53aac594410ed051c19258beb8a07b808eff0b42a7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8673a29c0e16f4e707d79f0610722ccabb023ff2b4a18d7cf9d305183817d952\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://35766e7206b1be7a53d13d6890d759cf81dbd999643c12349268e4bf16fbbdb0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5b9844a11b2b903978d0d464b469e26c9211100da61d018dde9636dca95b684e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T14:09:30Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T14:09:58Z is after 2025-08-24T17:21:41Z" Dec 03 14:09:58 crc kubenswrapper[4805]: I1203 14:09:58.085824 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:49Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:49Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://541ace7ccbb66dbd191c3ca8c1e4479e3edbf80dfc67326bea8aaee81fdcfc3c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://821e3e9a902aacecf6c41e8b8d8aa1ac951ce2fbb6505b7148b16348a3754e55\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T14:09:58Z is after 2025-08-24T17:21:41Z" Dec 03 14:09:58 crc kubenswrapper[4805]: I1203 14:09:58.092132 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 14:09:58 crc kubenswrapper[4805]: I1203 14:09:58.092171 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 14:09:58 crc kubenswrapper[4805]: I1203 14:09:58.092182 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 14:09:58 crc kubenswrapper[4805]: I1203 14:09:58.092197 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 14:09:58 crc kubenswrapper[4805]: I1203 14:09:58.092205 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T14:09:58Z","lastTransitionTime":"2025-12-03T14:09:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 14:09:58 crc kubenswrapper[4805]: I1203 14:09:58.097367 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-gskh4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6098937f-e3f6-45e8-a647-4994a79cd711\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://28d5653f26ec61647808b3a733ff7eac70aa69ac184660f24147e8cfaf0eb802\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nwrw9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9ad6b68a42fd33ee167e67873c6eed3f11cdf6dc71d22e952c021436eb967dd9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nwrw9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T14:09:50Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-gskh4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T14:09:58Z is after 2025-08-24T17:21:41Z" Dec 03 14:09:58 crc kubenswrapper[4805]: I1203 14:09:58.119375 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-psjpm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"deeec30d-cad2-4dfd-adb6-64f6646581be\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:50Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:50Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wdlzx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wdlzx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wdlzx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wdlzx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wdlzx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wdlzx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wdlzx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wdlzx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2f875252fbe6a673a72fc9c56bb96899f62732305ad77b3108e3b244bfa91ecb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2f875252fbe6a673a72fc9c56bb96899f62732305ad77b3108e3b244bfa91ecb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T14:09:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T14:09:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wdlzx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T14:09:50Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-psjpm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T14:09:58Z is after 2025-08-24T17:21:41Z" Dec 03 14:09:58 crc kubenswrapper[4805]: I1203 14:09:58.134569 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:48Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:48Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T14:09:58Z is after 2025-08-24T17:21:41Z" Dec 03 14:09:58 crc kubenswrapper[4805]: I1203 14:09:58.146411 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-pdzgk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"860769de-6a5b-4cb9-a752-264e7e2e785f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5e2f0a0cc21505d609f81918365d5e4f5d21959ac12ab147d6a51601c9de4c14\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ggn4n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T14:09:50Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-pdzgk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T14:09:58Z is after 2025-08-24T17:21:41Z" Dec 03 14:09:58 crc kubenswrapper[4805]: I1203 14:09:58.195129 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 14:09:58 crc kubenswrapper[4805]: I1203 14:09:58.195166 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 14:09:58 crc kubenswrapper[4805]: I1203 14:09:58.195187 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 14:09:58 crc kubenswrapper[4805]: I1203 14:09:58.195204 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 14:09:58 crc kubenswrapper[4805]: I1203 14:09:58.195218 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T14:09:58Z","lastTransitionTime":"2025-12-03T14:09:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 14:09:58 crc kubenswrapper[4805]: I1203 14:09:58.298341 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 14:09:58 crc kubenswrapper[4805]: I1203 14:09:58.298380 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 14:09:58 crc kubenswrapper[4805]: I1203 14:09:58.298389 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 14:09:58 crc kubenswrapper[4805]: I1203 14:09:58.298404 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 14:09:58 crc kubenswrapper[4805]: I1203 14:09:58.298413 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T14:09:58Z","lastTransitionTime":"2025-12-03T14:09:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 14:09:58 crc kubenswrapper[4805]: I1203 14:09:58.400336 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 14:09:58 crc kubenswrapper[4805]: I1203 14:09:58.400675 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 14:09:58 crc kubenswrapper[4805]: I1203 14:09:58.400686 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 14:09:58 crc kubenswrapper[4805]: I1203 14:09:58.400702 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 14:09:58 crc kubenswrapper[4805]: I1203 14:09:58.400714 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T14:09:58Z","lastTransitionTime":"2025-12-03T14:09:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 14:09:58 crc kubenswrapper[4805]: I1203 14:09:58.503675 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 14:09:58 crc kubenswrapper[4805]: I1203 14:09:58.503724 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 14:09:58 crc kubenswrapper[4805]: I1203 14:09:58.503735 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 14:09:58 crc kubenswrapper[4805]: I1203 14:09:58.503752 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 14:09:58 crc kubenswrapper[4805]: I1203 14:09:58.503763 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T14:09:58Z","lastTransitionTime":"2025-12-03T14:09:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 14:09:58 crc kubenswrapper[4805]: I1203 14:09:58.606200 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 14:09:58 crc kubenswrapper[4805]: I1203 14:09:58.606244 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 14:09:58 crc kubenswrapper[4805]: I1203 14:09:58.606256 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 14:09:58 crc kubenswrapper[4805]: I1203 14:09:58.606273 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 14:09:58 crc kubenswrapper[4805]: I1203 14:09:58.606288 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T14:09:58Z","lastTransitionTime":"2025-12-03T14:09:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 14:09:58 crc kubenswrapper[4805]: I1203 14:09:58.693865 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 14:09:58 crc kubenswrapper[4805]: I1203 14:09:58.693945 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 14:09:58 crc kubenswrapper[4805]: I1203 14:09:58.693965 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 14:09:58 crc kubenswrapper[4805]: E1203 14:09:58.694086 4805 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 03 14:09:58 crc kubenswrapper[4805]: E1203 14:09:58.694204 4805 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 03 14:09:58 crc kubenswrapper[4805]: E1203 14:09:58.694324 4805 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 03 14:09:58 crc kubenswrapper[4805]: I1203 14:09:58.709021 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 14:09:58 crc kubenswrapper[4805]: I1203 14:09:58.709070 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 14:09:58 crc kubenswrapper[4805]: I1203 14:09:58.709085 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 14:09:58 crc kubenswrapper[4805]: I1203 14:09:58.709106 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 14:09:58 crc kubenswrapper[4805]: I1203 14:09:58.709119 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T14:09:58Z","lastTransitionTime":"2025-12-03T14:09:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 14:09:58 crc kubenswrapper[4805]: I1203 14:09:58.812187 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 14:09:58 crc kubenswrapper[4805]: I1203 14:09:58.812245 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 14:09:58 crc kubenswrapper[4805]: I1203 14:09:58.812264 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 14:09:58 crc kubenswrapper[4805]: I1203 14:09:58.812287 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 14:09:58 crc kubenswrapper[4805]: I1203 14:09:58.812312 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T14:09:58Z","lastTransitionTime":"2025-12-03T14:09:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 14:09:58 crc kubenswrapper[4805]: I1203 14:09:58.894245 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-psjpm" event={"ID":"deeec30d-cad2-4dfd-adb6-64f6646581be","Type":"ContainerStarted","Data":"9e3aa4c113c750c2e2ead5d33f909a53247a221db998125f2494e7058d0c689a"} Dec 03 14:09:58 crc kubenswrapper[4805]: I1203 14:09:58.895081 4805 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-psjpm" Dec 03 14:09:58 crc kubenswrapper[4805]: I1203 14:09:58.899913 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-bpk49" event={"ID":"de00a81f-82e1-4223-9614-f98606b9a8ef","Type":"ContainerStarted","Data":"04f4ad27e02e2024ae9e7cd1b16116267e69cecda882dd723cdb274c41a0e7db"} Dec 03 14:09:58 crc kubenswrapper[4805]: I1203 14:09:58.910683 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-pdzgk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"860769de-6a5b-4cb9-a752-264e7e2e785f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5e2f0a0cc21505d609f81918365d5e4f5d21959ac12ab147d6a51601c9de4c14\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ggn4n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T14:09:50Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-pdzgk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T14:09:58Z is after 2025-08-24T17:21:41Z" Dec 03 14:09:58 crc kubenswrapper[4805]: I1203 14:09:58.915583 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 14:09:58 crc kubenswrapper[4805]: I1203 14:09:58.915625 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 14:09:58 crc kubenswrapper[4805]: I1203 14:09:58.915641 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 14:09:58 crc kubenswrapper[4805]: I1203 14:09:58.915663 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 14:09:58 crc kubenswrapper[4805]: I1203 14:09:58.915680 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T14:09:58Z","lastTransitionTime":"2025-12-03T14:09:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 14:09:58 crc kubenswrapper[4805]: I1203 14:09:58.929353 4805 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-psjpm" Dec 03 14:09:58 crc kubenswrapper[4805]: I1203 14:09:58.930997 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:49Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:49Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://38339331a3209d6787d1def81b72f4a8514bf91080a139cad6d8ac5cc6bef10e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T14:09:58Z is after 2025-08-24T17:21:41Z" Dec 03 14:09:58 crc kubenswrapper[4805]: I1203 14:09:58.944995 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:48Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:48Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T14:09:58Z is after 2025-08-24T17:21:41Z" Dec 03 14:09:58 crc kubenswrapper[4805]: I1203 14:09:58.956577 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:51Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:51Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b3123d3fa879097db88ed2f3cfb05631f440c24c5405503b2de3546f12e2e899\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T14:09:58Z is after 2025-08-24T17:21:41Z" Dec 03 14:09:58 crc kubenswrapper[4805]: I1203 14:09:58.966037 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-bdtdg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"53a0e9e3-b6d8-44b7-a32b-aa3450442865\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f557e17ad105d6972d3c1480656c6073f67e28c0f5773aa365549951f4d161c7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9ww54\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T14:09:51Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-bdtdg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T14:09:58Z is after 2025-08-24T17:21:41Z" Dec 03 14:09:58 crc kubenswrapper[4805]: I1203 14:09:58.993621 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b6d36671-7173-4e9c-843f-005b4dcf576c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a837d7d491eaacbe4279b4014bb5fcbe9b514ba5cc82c708227bcd9b25a346e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6e5f04faf3b23197bb8ea11d3e574dc86f31e21d798734962001ec8c95de7c84\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2ff1a979387da41b00e95eb9cc8b2c9de966380a12fae687591df3c5b18a4497\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1e66e1cb18573a14e3c3ae16af4dbd9f8997a6ad38680c3ab4505646ebd48e7c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7d5b5c85e3de43e7e95111aa967248550607c26200c783251b68696f202f92a1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://474f72dac37a808fc4b9832306e76979b151d7221807ff3086013466ed1419c3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://474f72dac37a808fc4b9832306e76979b151d7221807ff3086013466ed1419c3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T14:09:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T14:09:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://74c8cea1a61c4775a0ad2f4ab4bb8a36437cf23d2ee953311d6ea967e297510a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://74c8cea1a61c4775a0ad2f4ab4bb8a36437cf23d2ee953311d6ea967e297510a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T14:09:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T14:09:31Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://a4f2f56a809adc01fba82eba0652d78f58770c132f77b1424f8b17068b1927e3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a4f2f56a809adc01fba82eba0652d78f58770c132f77b1424f8b17068b1927e3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T14:09:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T14:09:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T14:09:30Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T14:09:58Z is after 2025-08-24T17:21:41Z" Dec 03 14:09:59 crc kubenswrapper[4805]: I1203 14:09:59.007542 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ced8291d-0ce5-4e74-9eaf-2bf033774039\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:30Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:30Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3e5a34636f0a8c4d9179074596d28bc2eb7a504573398c596a637defd829907a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://29cd95be3f964415943c873e7914ffcea4abe50dfdf17f8a9cd03fc38df35981\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://951173b3d3d581275a6f6e43209b5e0611a701ca7a10f81504d723e24a2d21a3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e64730957bbbb2feb4d506e0ed9731a0594bf66fa1e8fcdb8ec944ded6308eed\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5da55f769e23ad52bac9c77da62c78e70fa34db8ed1a435e4eaf4dad3fd2b68a\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-03T14:09:48Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1203 14:09:42.992179 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1203 14:09:42.993233 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1830926909/tls.crt::/tmp/serving-cert-1830926909/tls.key\\\\\\\"\\\\nI1203 14:09:48.622403 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1203 14:09:48.626476 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1203 14:09:48.626527 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1203 14:09:48.626574 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1203 14:09:48.626582 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1203 14:09:48.633990 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1203 14:09:48.634024 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 14:09:48.634029 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 14:09:48.634034 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1203 14:09:48.634038 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1203 14:09:48.634042 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1203 14:09:48.634045 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1203 14:09:48.634344 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1203 14:09:48.644497 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T14:09:32Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bd92a160eb530c5687ebd7cd5b83b41820a2b808a0fd9514e5ace9d237e1bd26\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:32Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://af9c940135012c629447528c61743ed1e58708ab57ce76604c6da0479aa459cf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://af9c940135012c629447528c61743ed1e58708ab57ce76604c6da0479aa459cf\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T14:09:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T14:09:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T14:09:30Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T14:09:59Z is after 2025-08-24T17:21:41Z" Dec 03 14:09:59 crc kubenswrapper[4805]: I1203 14:09:59.017932 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 14:09:59 crc kubenswrapper[4805]: I1203 14:09:59.017993 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 14:09:59 crc kubenswrapper[4805]: I1203 14:09:59.018017 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 14:09:59 crc kubenswrapper[4805]: I1203 14:09:59.018047 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 14:09:59 crc kubenswrapper[4805]: I1203 14:09:59.018068 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T14:09:59Z","lastTransitionTime":"2025-12-03T14:09:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 14:09:59 crc kubenswrapper[4805]: I1203 14:09:59.029110 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-k4mhs" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c09c1cdc-f461-45c9-8444-5b8764bccabc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3354c1a44f891182d04116fc76234a2b337d50567d032a3584f902a9516387cc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qdcd7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T14:09:50Z\\\"}}\" for pod \"openshift-multus\"/\"multus-k4mhs\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T14:09:59Z is after 2025-08-24T17:21:41Z" Dec 03 14:09:59 crc kubenswrapper[4805]: I1203 14:09:59.044751 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-bpk49" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"de00a81f-82e1-4223-9614-f98606b9a8ef\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:50Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:50Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-25d7t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7741e0e70416ebe094bfabc6009e98d7d4c0fbcba1fc82df13b9ba40774b456a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7741e0e70416ebe094bfabc6009e98d7d4c0fbcba1fc82df13b9ba40774b456a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T14:09:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T14:09:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-25d7t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ee3499c02eb52bbf08c447d2433d435a74dfb70670dc4cbaac44f5fbff9e8a29\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ee3499c02eb52bbf08c447d2433d435a74dfb70670dc4cbaac44f5fbff9e8a29\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T14:09:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T14:09:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-25d7t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ec8488b1d69799951be7ef40e5e028732265c4ddcdea2d453923e567864771f7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ec8488b1d69799951be7ef40e5e028732265c4ddcdea2d453923e567864771f7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T14:09:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T14:09:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-25d7t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://66707bb4e1fdacb265e89786df8a04605830f1c156394cacfcec9b54abe5d41a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://66707bb4e1fdacb265e89786df8a04605830f1c156394cacfcec9b54abe5d41a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T14:09:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T14:09:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-25d7t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6207ab6600970ff15c8713c547d4920b2c77eb9bea990bb5da497e24870dcc0f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6207ab6600970ff15c8713c547d4920b2c77eb9bea990bb5da497e24870dcc0f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T14:09:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T14:09:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-25d7t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8de65d4a205e2deec0b433e53c0e37feb94aa0ede6a4c58d7414b80829804505\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8de65d4a205e2deec0b433e53c0e37feb94aa0ede6a4c58d7414b80829804505\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T14:09:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T14:09:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-25d7t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T14:09:50Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-bpk49\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T14:09:59Z is after 2025-08-24T17:21:41Z" Dec 03 14:09:59 crc kubenswrapper[4805]: I1203 14:09:59.059214 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"750d2b4b-738a-4992-8902-742528f6e875\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://abd2d0156113bb5182727a53aac594410ed051c19258beb8a07b808eff0b42a7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8673a29c0e16f4e707d79f0610722ccabb023ff2b4a18d7cf9d305183817d952\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://35766e7206b1be7a53d13d6890d759cf81dbd999643c12349268e4bf16fbbdb0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5b9844a11b2b903978d0d464b469e26c9211100da61d018dde9636dca95b684e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T14:09:30Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T14:09:59Z is after 2025-08-24T17:21:41Z" Dec 03 14:09:59 crc kubenswrapper[4805]: I1203 14:09:59.076662 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:48Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:48Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T14:09:59Z is after 2025-08-24T17:21:41Z" Dec 03 14:09:59 crc kubenswrapper[4805]: I1203 14:09:59.090522 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-gskh4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6098937f-e3f6-45e8-a647-4994a79cd711\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://28d5653f26ec61647808b3a733ff7eac70aa69ac184660f24147e8cfaf0eb802\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nwrw9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9ad6b68a42fd33ee167e67873c6eed3f11cdf6dc71d22e952c021436eb967dd9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nwrw9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T14:09:50Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-gskh4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T14:09:59Z is after 2025-08-24T17:21:41Z" Dec 03 14:09:59 crc kubenswrapper[4805]: I1203 14:09:59.120736 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 14:09:59 crc kubenswrapper[4805]: I1203 14:09:59.120777 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 14:09:59 crc kubenswrapper[4805]: I1203 14:09:59.120790 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 14:09:59 crc kubenswrapper[4805]: I1203 14:09:59.120808 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 14:09:59 crc kubenswrapper[4805]: I1203 14:09:59.120819 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T14:09:59Z","lastTransitionTime":"2025-12-03T14:09:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 14:09:59 crc kubenswrapper[4805]: I1203 14:09:59.122021 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-psjpm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"deeec30d-cad2-4dfd-adb6-64f6646581be\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:50Z\\\",\\\"message\\\":\\\"containers with unready status: [nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:50Z\\\",\\\"message\\\":\\\"containers with unready status: [nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4b8d00cdf0f37f2e5cd4e952cf85d3de8d72c43fdf786088fd9f26ed17b1e21f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wdlzx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bd67c6de01a8d196a1f51d4d90640c9e69de6f05d5e201b25e6eaa5c33440ea8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wdlzx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cb8bf05ef093eb0727a066d4bbbdf9af9422867cd4dadacafc6eeb016867a61e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wdlzx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0304467102ed053628c0d3e06548c74b6492336b6fd044246508d1e8bdb17bc3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wdlzx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6ce80ccafb3262a11fbe9923317a9d38b61e5dc5b3e7e7ae136f143d6b643fd7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wdlzx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://47d76962beb8c0f196d50af96fb3ec611097823ca80a1c8bd718f2d18840128c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wdlzx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9e3aa4c113c750c2e2ead5d33f909a53247a221db998125f2494e7058d0c689a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wdlzx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ab98e365c1a4242a47b8cbe2dd85109d366d6002c4de1f18ddc950a46bea6c12\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wdlzx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2f875252fbe6a673a72fc9c56bb96899f62732305ad77b3108e3b244bfa91ecb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2f875252fbe6a673a72fc9c56bb96899f62732305ad77b3108e3b244bfa91ecb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T14:09:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T14:09:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wdlzx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T14:09:50Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-psjpm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T14:09:59Z is after 2025-08-24T17:21:41Z" Dec 03 14:09:59 crc kubenswrapper[4805]: I1203 14:09:59.139745 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:48Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:48Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T14:09:59Z is after 2025-08-24T17:21:41Z" Dec 03 14:09:59 crc kubenswrapper[4805]: I1203 14:09:59.157497 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:49Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:49Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://541ace7ccbb66dbd191c3ca8c1e4479e3edbf80dfc67326bea8aaee81fdcfc3c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://821e3e9a902aacecf6c41e8b8d8aa1ac951ce2fbb6505b7148b16348a3754e55\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T14:09:59Z is after 2025-08-24T17:21:41Z" Dec 03 14:09:59 crc kubenswrapper[4805]: I1203 14:09:59.171815 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-k4mhs" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c09c1cdc-f461-45c9-8444-5b8764bccabc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3354c1a44f891182d04116fc76234a2b337d50567d032a3584f902a9516387cc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qdcd7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T14:09:50Z\\\"}}\" for pod \"openshift-multus\"/\"multus-k4mhs\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T14:09:59Z is after 2025-08-24T17:21:41Z" Dec 03 14:09:59 crc kubenswrapper[4805]: I1203 14:09:59.186948 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-bpk49" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"de00a81f-82e1-4223-9614-f98606b9a8ef\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://04f4ad27e02e2024ae9e7cd1b16116267e69cecda882dd723cdb274c41a0e7db\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-25d7t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7741e0e70416ebe094bfabc6009e98d7d4c0fbcba1fc82df13b9ba40774b456a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7741e0e70416ebe094bfabc6009e98d7d4c0fbcba1fc82df13b9ba40774b456a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T14:09:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T14:09:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-25d7t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ee3499c02eb52bbf08c447d2433d435a74dfb70670dc4cbaac44f5fbff9e8a29\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ee3499c02eb52bbf08c447d2433d435a74dfb70670dc4cbaac44f5fbff9e8a29\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T14:09:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T14:09:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-25d7t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ec8488b1d69799951be7ef40e5e028732265c4ddcdea2d453923e567864771f7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ec8488b1d69799951be7ef40e5e028732265c4ddcdea2d453923e567864771f7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T14:09:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T14:09:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-25d7t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://66707bb4e1fdacb265e89786df8a04605830f1c156394cacfcec9b54abe5d41a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://66707bb4e1fdacb265e89786df8a04605830f1c156394cacfcec9b54abe5d41a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T14:09:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T14:09:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-25d7t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6207ab6600970ff15c8713c547d4920b2c77eb9bea990bb5da497e24870dcc0f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6207ab6600970ff15c8713c547d4920b2c77eb9bea990bb5da497e24870dcc0f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T14:09:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T14:09:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-25d7t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8de65d4a205e2deec0b433e53c0e37feb94aa0ede6a4c58d7414b80829804505\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8de65d4a205e2deec0b433e53c0e37feb94aa0ede6a4c58d7414b80829804505\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T14:09:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T14:09:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-25d7t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T14:09:50Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-bpk49\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T14:09:59Z is after 2025-08-24T17:21:41Z" Dec 03 14:09:59 crc kubenswrapper[4805]: I1203 14:09:59.201798 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"750d2b4b-738a-4992-8902-742528f6e875\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://abd2d0156113bb5182727a53aac594410ed051c19258beb8a07b808eff0b42a7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8673a29c0e16f4e707d79f0610722ccabb023ff2b4a18d7cf9d305183817d952\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://35766e7206b1be7a53d13d6890d759cf81dbd999643c12349268e4bf16fbbdb0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5b9844a11b2b903978d0d464b469e26c9211100da61d018dde9636dca95b684e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T14:09:30Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T14:09:59Z is after 2025-08-24T17:21:41Z" Dec 03 14:09:59 crc kubenswrapper[4805]: I1203 14:09:59.216029 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:48Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:48Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T14:09:59Z is after 2025-08-24T17:21:41Z" Dec 03 14:09:59 crc kubenswrapper[4805]: I1203 14:09:59.223717 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 14:09:59 crc kubenswrapper[4805]: I1203 14:09:59.223749 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 14:09:59 crc kubenswrapper[4805]: I1203 14:09:59.223762 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 14:09:59 crc kubenswrapper[4805]: I1203 14:09:59.223785 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 14:09:59 crc kubenswrapper[4805]: I1203 14:09:59.223797 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T14:09:59Z","lastTransitionTime":"2025-12-03T14:09:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 14:09:59 crc kubenswrapper[4805]: I1203 14:09:59.226647 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-gskh4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6098937f-e3f6-45e8-a647-4994a79cd711\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://28d5653f26ec61647808b3a733ff7eac70aa69ac184660f24147e8cfaf0eb802\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nwrw9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9ad6b68a42fd33ee167e67873c6eed3f11cdf6dc71d22e952c021436eb967dd9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nwrw9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T14:09:50Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-gskh4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T14:09:59Z is after 2025-08-24T17:21:41Z" Dec 03 14:09:59 crc kubenswrapper[4805]: I1203 14:09:59.243596 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-psjpm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"deeec30d-cad2-4dfd-adb6-64f6646581be\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:50Z\\\",\\\"message\\\":\\\"containers with unready status: [nbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:50Z\\\",\\\"message\\\":\\\"containers with unready status: [nbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4b8d00cdf0f37f2e5cd4e952cf85d3de8d72c43fdf786088fd9f26ed17b1e21f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wdlzx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bd67c6de01a8d196a1f51d4d90640c9e69de6f05d5e201b25e6eaa5c33440ea8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wdlzx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cb8bf05ef093eb0727a066d4bbbdf9af9422867cd4dadacafc6eeb016867a61e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wdlzx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0304467102ed053628c0d3e06548c74b6492336b6fd044246508d1e8bdb17bc3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wdlzx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6ce80ccafb3262a11fbe9923317a9d38b61e5dc5b3e7e7ae136f143d6b643fd7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wdlzx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://47d76962beb8c0f196d50af96fb3ec611097823ca80a1c8bd718f2d18840128c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wdlzx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9e3aa4c113c750c2e2ead5d33f909a53247a221db998125f2494e7058d0c689a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wdlzx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ab98e365c1a4242a47b8cbe2dd85109d366d6002c4de1f18ddc950a46bea6c12\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wdlzx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2f875252fbe6a673a72fc9c56bb96899f62732305ad77b3108e3b244bfa91ecb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2f875252fbe6a673a72fc9c56bb96899f62732305ad77b3108e3b244bfa91ecb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T14:09:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T14:09:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wdlzx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T14:09:50Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-psjpm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T14:09:59Z is after 2025-08-24T17:21:41Z" Dec 03 14:09:59 crc kubenswrapper[4805]: I1203 14:09:59.254669 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:48Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:48Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T14:09:59Z is after 2025-08-24T17:21:41Z" Dec 03 14:09:59 crc kubenswrapper[4805]: I1203 14:09:59.264756 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:49Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:49Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://541ace7ccbb66dbd191c3ca8c1e4479e3edbf80dfc67326bea8aaee81fdcfc3c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://821e3e9a902aacecf6c41e8b8d8aa1ac951ce2fbb6505b7148b16348a3754e55\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T14:09:59Z is after 2025-08-24T17:21:41Z" Dec 03 14:09:59 crc kubenswrapper[4805]: I1203 14:09:59.276943 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-pdzgk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"860769de-6a5b-4cb9-a752-264e7e2e785f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5e2f0a0cc21505d609f81918365d5e4f5d21959ac12ab147d6a51601c9de4c14\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ggn4n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T14:09:50Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-pdzgk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T14:09:59Z is after 2025-08-24T17:21:41Z" Dec 03 14:09:59 crc kubenswrapper[4805]: I1203 14:09:59.289460 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:49Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:49Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://38339331a3209d6787d1def81b72f4a8514bf91080a139cad6d8ac5cc6bef10e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T14:09:59Z is after 2025-08-24T17:21:41Z" Dec 03 14:09:59 crc kubenswrapper[4805]: I1203 14:09:59.305717 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:48Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:48Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T14:09:59Z is after 2025-08-24T17:21:41Z" Dec 03 14:09:59 crc kubenswrapper[4805]: I1203 14:09:59.318115 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:51Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:51Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b3123d3fa879097db88ed2f3cfb05631f440c24c5405503b2de3546f12e2e899\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T14:09:59Z is after 2025-08-24T17:21:41Z" Dec 03 14:09:59 crc kubenswrapper[4805]: I1203 14:09:59.325806 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 14:09:59 crc kubenswrapper[4805]: I1203 14:09:59.325830 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 14:09:59 crc kubenswrapper[4805]: I1203 14:09:59.325849 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 14:09:59 crc kubenswrapper[4805]: I1203 14:09:59.325862 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 14:09:59 crc kubenswrapper[4805]: I1203 14:09:59.325871 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T14:09:59Z","lastTransitionTime":"2025-12-03T14:09:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 14:09:59 crc kubenswrapper[4805]: I1203 14:09:59.328960 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-bdtdg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"53a0e9e3-b6d8-44b7-a32b-aa3450442865\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f557e17ad105d6972d3c1480656c6073f67e28c0f5773aa365549951f4d161c7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9ww54\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T14:09:51Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-bdtdg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T14:09:59Z is after 2025-08-24T17:21:41Z" Dec 03 14:09:59 crc kubenswrapper[4805]: I1203 14:09:59.350232 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b6d36671-7173-4e9c-843f-005b4dcf576c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a837d7d491eaacbe4279b4014bb5fcbe9b514ba5cc82c708227bcd9b25a346e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6e5f04faf3b23197bb8ea11d3e574dc86f31e21d798734962001ec8c95de7c84\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2ff1a979387da41b00e95eb9cc8b2c9de966380a12fae687591df3c5b18a4497\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1e66e1cb18573a14e3c3ae16af4dbd9f8997a6ad38680c3ab4505646ebd48e7c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7d5b5c85e3de43e7e95111aa967248550607c26200c783251b68696f202f92a1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://474f72dac37a808fc4b9832306e76979b151d7221807ff3086013466ed1419c3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://474f72dac37a808fc4b9832306e76979b151d7221807ff3086013466ed1419c3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T14:09:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T14:09:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://74c8cea1a61c4775a0ad2f4ab4bb8a36437cf23d2ee953311d6ea967e297510a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://74c8cea1a61c4775a0ad2f4ab4bb8a36437cf23d2ee953311d6ea967e297510a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T14:09:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T14:09:31Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://a4f2f56a809adc01fba82eba0652d78f58770c132f77b1424f8b17068b1927e3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a4f2f56a809adc01fba82eba0652d78f58770c132f77b1424f8b17068b1927e3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T14:09:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T14:09:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T14:09:30Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T14:09:59Z is after 2025-08-24T17:21:41Z" Dec 03 14:09:59 crc kubenswrapper[4805]: I1203 14:09:59.365101 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ced8291d-0ce5-4e74-9eaf-2bf033774039\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:30Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:30Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3e5a34636f0a8c4d9179074596d28bc2eb7a504573398c596a637defd829907a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://29cd95be3f964415943c873e7914ffcea4abe50dfdf17f8a9cd03fc38df35981\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://951173b3d3d581275a6f6e43209b5e0611a701ca7a10f81504d723e24a2d21a3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e64730957bbbb2feb4d506e0ed9731a0594bf66fa1e8fcdb8ec944ded6308eed\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5da55f769e23ad52bac9c77da62c78e70fa34db8ed1a435e4eaf4dad3fd2b68a\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-03T14:09:48Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1203 14:09:42.992179 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1203 14:09:42.993233 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1830926909/tls.crt::/tmp/serving-cert-1830926909/tls.key\\\\\\\"\\\\nI1203 14:09:48.622403 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1203 14:09:48.626476 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1203 14:09:48.626527 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1203 14:09:48.626574 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1203 14:09:48.626582 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1203 14:09:48.633990 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1203 14:09:48.634024 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 14:09:48.634029 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 14:09:48.634034 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1203 14:09:48.634038 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1203 14:09:48.634042 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1203 14:09:48.634045 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1203 14:09:48.634344 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1203 14:09:48.644497 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T14:09:32Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bd92a160eb530c5687ebd7cd5b83b41820a2b808a0fd9514e5ace9d237e1bd26\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:32Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://af9c940135012c629447528c61743ed1e58708ab57ce76604c6da0479aa459cf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://af9c940135012c629447528c61743ed1e58708ab57ce76604c6da0479aa459cf\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T14:09:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T14:09:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T14:09:30Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T14:09:59Z is after 2025-08-24T17:21:41Z" Dec 03 14:09:59 crc kubenswrapper[4805]: I1203 14:09:59.427805 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 14:09:59 crc kubenswrapper[4805]: I1203 14:09:59.427854 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 14:09:59 crc kubenswrapper[4805]: I1203 14:09:59.427867 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 14:09:59 crc kubenswrapper[4805]: I1203 14:09:59.427885 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 14:09:59 crc kubenswrapper[4805]: I1203 14:09:59.427895 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T14:09:59Z","lastTransitionTime":"2025-12-03T14:09:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 14:09:59 crc kubenswrapper[4805]: I1203 14:09:59.530613 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 14:09:59 crc kubenswrapper[4805]: I1203 14:09:59.530659 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 14:09:59 crc kubenswrapper[4805]: I1203 14:09:59.530668 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 14:09:59 crc kubenswrapper[4805]: I1203 14:09:59.530686 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 14:09:59 crc kubenswrapper[4805]: I1203 14:09:59.530703 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T14:09:59Z","lastTransitionTime":"2025-12-03T14:09:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 14:09:59 crc kubenswrapper[4805]: I1203 14:09:59.633554 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 14:09:59 crc kubenswrapper[4805]: I1203 14:09:59.633606 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 14:09:59 crc kubenswrapper[4805]: I1203 14:09:59.633617 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 14:09:59 crc kubenswrapper[4805]: I1203 14:09:59.633637 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 14:09:59 crc kubenswrapper[4805]: I1203 14:09:59.633649 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T14:09:59Z","lastTransitionTime":"2025-12-03T14:09:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 14:09:59 crc kubenswrapper[4805]: I1203 14:09:59.735803 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 14:09:59 crc kubenswrapper[4805]: I1203 14:09:59.735860 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 14:09:59 crc kubenswrapper[4805]: I1203 14:09:59.735877 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 14:09:59 crc kubenswrapper[4805]: I1203 14:09:59.735893 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 14:09:59 crc kubenswrapper[4805]: I1203 14:09:59.735903 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T14:09:59Z","lastTransitionTime":"2025-12-03T14:09:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 14:09:59 crc kubenswrapper[4805]: I1203 14:09:59.838717 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 14:09:59 crc kubenswrapper[4805]: I1203 14:09:59.838786 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 14:09:59 crc kubenswrapper[4805]: I1203 14:09:59.838795 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 14:09:59 crc kubenswrapper[4805]: I1203 14:09:59.838826 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 14:09:59 crc kubenswrapper[4805]: I1203 14:09:59.838869 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T14:09:59Z","lastTransitionTime":"2025-12-03T14:09:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 14:09:59 crc kubenswrapper[4805]: I1203 14:09:59.902564 4805 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Dec 03 14:09:59 crc kubenswrapper[4805]: I1203 14:09:59.902750 4805 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-psjpm" Dec 03 14:09:59 crc kubenswrapper[4805]: I1203 14:09:59.924706 4805 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-psjpm" Dec 03 14:09:59 crc kubenswrapper[4805]: I1203 14:09:59.936308 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-pdzgk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"860769de-6a5b-4cb9-a752-264e7e2e785f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5e2f0a0cc21505d609f81918365d5e4f5d21959ac12ab147d6a51601c9de4c14\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ggn4n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T14:09:50Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-pdzgk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T14:09:59Z is after 2025-08-24T17:21:41Z" Dec 03 14:09:59 crc kubenswrapper[4805]: I1203 14:09:59.940788 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 14:09:59 crc kubenswrapper[4805]: I1203 14:09:59.940863 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 14:09:59 crc kubenswrapper[4805]: I1203 14:09:59.940875 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 14:09:59 crc kubenswrapper[4805]: I1203 14:09:59.940889 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 14:09:59 crc kubenswrapper[4805]: I1203 14:09:59.940897 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T14:09:59Z","lastTransitionTime":"2025-12-03T14:09:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 14:09:59 crc kubenswrapper[4805]: I1203 14:09:59.950352 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:49Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:49Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://38339331a3209d6787d1def81b72f4a8514bf91080a139cad6d8ac5cc6bef10e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T14:09:59Z is after 2025-08-24T17:21:41Z" Dec 03 14:09:59 crc kubenswrapper[4805]: I1203 14:09:59.960398 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:48Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:48Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T14:09:59Z is after 2025-08-24T17:21:41Z" Dec 03 14:09:59 crc kubenswrapper[4805]: I1203 14:09:59.969719 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:51Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:51Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b3123d3fa879097db88ed2f3cfb05631f440c24c5405503b2de3546f12e2e899\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T14:09:59Z is after 2025-08-24T17:21:41Z" Dec 03 14:09:59 crc kubenswrapper[4805]: I1203 14:09:59.980714 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-bdtdg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"53a0e9e3-b6d8-44b7-a32b-aa3450442865\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f557e17ad105d6972d3c1480656c6073f67e28c0f5773aa365549951f4d161c7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9ww54\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T14:09:51Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-bdtdg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T14:09:59Z is after 2025-08-24T17:21:41Z" Dec 03 14:09:59 crc kubenswrapper[4805]: I1203 14:09:59.999607 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b6d36671-7173-4e9c-843f-005b4dcf576c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a837d7d491eaacbe4279b4014bb5fcbe9b514ba5cc82c708227bcd9b25a346e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6e5f04faf3b23197bb8ea11d3e574dc86f31e21d798734962001ec8c95de7c84\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2ff1a979387da41b00e95eb9cc8b2c9de966380a12fae687591df3c5b18a4497\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1e66e1cb18573a14e3c3ae16af4dbd9f8997a6ad38680c3ab4505646ebd48e7c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7d5b5c85e3de43e7e95111aa967248550607c26200c783251b68696f202f92a1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://474f72dac37a808fc4b9832306e76979b151d7221807ff3086013466ed1419c3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://474f72dac37a808fc4b9832306e76979b151d7221807ff3086013466ed1419c3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T14:09:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T14:09:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://74c8cea1a61c4775a0ad2f4ab4bb8a36437cf23d2ee953311d6ea967e297510a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://74c8cea1a61c4775a0ad2f4ab4bb8a36437cf23d2ee953311d6ea967e297510a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T14:09:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T14:09:31Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://a4f2f56a809adc01fba82eba0652d78f58770c132f77b1424f8b17068b1927e3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a4f2f56a809adc01fba82eba0652d78f58770c132f77b1424f8b17068b1927e3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T14:09:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T14:09:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T14:09:30Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T14:09:59Z is after 2025-08-24T17:21:41Z" Dec 03 14:10:00 crc kubenswrapper[4805]: I1203 14:10:00.012422 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ced8291d-0ce5-4e74-9eaf-2bf033774039\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:30Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:30Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3e5a34636f0a8c4d9179074596d28bc2eb7a504573398c596a637defd829907a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://29cd95be3f964415943c873e7914ffcea4abe50dfdf17f8a9cd03fc38df35981\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://951173b3d3d581275a6f6e43209b5e0611a701ca7a10f81504d723e24a2d21a3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e64730957bbbb2feb4d506e0ed9731a0594bf66fa1e8fcdb8ec944ded6308eed\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5da55f769e23ad52bac9c77da62c78e70fa34db8ed1a435e4eaf4dad3fd2b68a\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-03T14:09:48Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1203 14:09:42.992179 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1203 14:09:42.993233 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1830926909/tls.crt::/tmp/serving-cert-1830926909/tls.key\\\\\\\"\\\\nI1203 14:09:48.622403 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1203 14:09:48.626476 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1203 14:09:48.626527 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1203 14:09:48.626574 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1203 14:09:48.626582 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1203 14:09:48.633990 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1203 14:09:48.634024 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 14:09:48.634029 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 14:09:48.634034 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1203 14:09:48.634038 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1203 14:09:48.634042 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1203 14:09:48.634045 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1203 14:09:48.634344 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1203 14:09:48.644497 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T14:09:32Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bd92a160eb530c5687ebd7cd5b83b41820a2b808a0fd9514e5ace9d237e1bd26\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:32Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://af9c940135012c629447528c61743ed1e58708ab57ce76604c6da0479aa459cf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://af9c940135012c629447528c61743ed1e58708ab57ce76604c6da0479aa459cf\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T14:09:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T14:09:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T14:09:30Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T14:10:00Z is after 2025-08-24T17:21:41Z" Dec 03 14:10:00 crc kubenswrapper[4805]: I1203 14:10:00.023878 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-k4mhs" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c09c1cdc-f461-45c9-8444-5b8764bccabc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3354c1a44f891182d04116fc76234a2b337d50567d032a3584f902a9516387cc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qdcd7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T14:09:50Z\\\"}}\" for pod \"openshift-multus\"/\"multus-k4mhs\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T14:10:00Z is after 2025-08-24T17:21:41Z" Dec 03 14:10:00 crc kubenswrapper[4805]: I1203 14:10:00.039152 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-bpk49" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"de00a81f-82e1-4223-9614-f98606b9a8ef\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://04f4ad27e02e2024ae9e7cd1b16116267e69cecda882dd723cdb274c41a0e7db\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-25d7t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7741e0e70416ebe094bfabc6009e98d7d4c0fbcba1fc82df13b9ba40774b456a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7741e0e70416ebe094bfabc6009e98d7d4c0fbcba1fc82df13b9ba40774b456a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T14:09:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T14:09:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-25d7t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ee3499c02eb52bbf08c447d2433d435a74dfb70670dc4cbaac44f5fbff9e8a29\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ee3499c02eb52bbf08c447d2433d435a74dfb70670dc4cbaac44f5fbff9e8a29\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T14:09:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T14:09:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-25d7t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ec8488b1d69799951be7ef40e5e028732265c4ddcdea2d453923e567864771f7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ec8488b1d69799951be7ef40e5e028732265c4ddcdea2d453923e567864771f7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T14:09:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T14:09:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-25d7t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://66707bb4e1fdacb265e89786df8a04605830f1c156394cacfcec9b54abe5d41a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://66707bb4e1fdacb265e89786df8a04605830f1c156394cacfcec9b54abe5d41a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T14:09:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T14:09:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-25d7t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6207ab6600970ff15c8713c547d4920b2c77eb9bea990bb5da497e24870dcc0f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6207ab6600970ff15c8713c547d4920b2c77eb9bea990bb5da497e24870dcc0f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T14:09:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T14:09:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-25d7t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8de65d4a205e2deec0b433e53c0e37feb94aa0ede6a4c58d7414b80829804505\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8de65d4a205e2deec0b433e53c0e37feb94aa0ede6a4c58d7414b80829804505\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T14:09:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T14:09:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-25d7t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T14:09:50Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-bpk49\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T14:10:00Z is after 2025-08-24T17:21:41Z" Dec 03 14:10:00 crc kubenswrapper[4805]: I1203 14:10:00.042310 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 14:10:00 crc kubenswrapper[4805]: I1203 14:10:00.042339 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 14:10:00 crc kubenswrapper[4805]: I1203 14:10:00.042347 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 14:10:00 crc kubenswrapper[4805]: I1203 14:10:00.042360 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 14:10:00 crc kubenswrapper[4805]: I1203 14:10:00.042370 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T14:10:00Z","lastTransitionTime":"2025-12-03T14:10:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 14:10:00 crc kubenswrapper[4805]: I1203 14:10:00.050205 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"750d2b4b-738a-4992-8902-742528f6e875\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://abd2d0156113bb5182727a53aac594410ed051c19258beb8a07b808eff0b42a7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8673a29c0e16f4e707d79f0610722ccabb023ff2b4a18d7cf9d305183817d952\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://35766e7206b1be7a53d13d6890d759cf81dbd999643c12349268e4bf16fbbdb0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5b9844a11b2b903978d0d464b469e26c9211100da61d018dde9636dca95b684e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T14:09:30Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T14:10:00Z is after 2025-08-24T17:21:41Z" Dec 03 14:10:00 crc kubenswrapper[4805]: I1203 14:10:00.061304 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:48Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:48Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T14:10:00Z is after 2025-08-24T17:21:41Z" Dec 03 14:10:00 crc kubenswrapper[4805]: I1203 14:10:00.073174 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-gskh4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6098937f-e3f6-45e8-a647-4994a79cd711\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://28d5653f26ec61647808b3a733ff7eac70aa69ac184660f24147e8cfaf0eb802\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nwrw9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9ad6b68a42fd33ee167e67873c6eed3f11cdf6dc71d22e952c021436eb967dd9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nwrw9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T14:09:50Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-gskh4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T14:10:00Z is after 2025-08-24T17:21:41Z" Dec 03 14:10:00 crc kubenswrapper[4805]: I1203 14:10:00.091792 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-psjpm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"deeec30d-cad2-4dfd-adb6-64f6646581be\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:50Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:50Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4b8d00cdf0f37f2e5cd4e952cf85d3de8d72c43fdf786088fd9f26ed17b1e21f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wdlzx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bd67c6de01a8d196a1f51d4d90640c9e69de6f05d5e201b25e6eaa5c33440ea8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wdlzx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cb8bf05ef093eb0727a066d4bbbdf9af9422867cd4dadacafc6eeb016867a61e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wdlzx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0304467102ed053628c0d3e06548c74b6492336b6fd044246508d1e8bdb17bc3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wdlzx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6ce80ccafb3262a11fbe9923317a9d38b61e5dc5b3e7e7ae136f143d6b643fd7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wdlzx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://47d76962beb8c0f196d50af96fb3ec611097823ca80a1c8bd718f2d18840128c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wdlzx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9e3aa4c113c750c2e2ead5d33f909a53247a221db998125f2494e7058d0c689a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wdlzx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ab98e365c1a4242a47b8cbe2dd85109d366d6002c4de1f18ddc950a46bea6c12\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wdlzx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2f875252fbe6a673a72fc9c56bb96899f62732305ad77b3108e3b244bfa91ecb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2f875252fbe6a673a72fc9c56bb96899f62732305ad77b3108e3b244bfa91ecb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T14:09:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T14:09:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wdlzx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T14:09:50Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-psjpm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T14:10:00Z is after 2025-08-24T17:21:41Z" Dec 03 14:10:00 crc kubenswrapper[4805]: I1203 14:10:00.104185 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:48Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:48Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T14:10:00Z is after 2025-08-24T17:21:41Z" Dec 03 14:10:00 crc kubenswrapper[4805]: I1203 14:10:00.115719 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:49Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:49Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://541ace7ccbb66dbd191c3ca8c1e4479e3edbf80dfc67326bea8aaee81fdcfc3c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://821e3e9a902aacecf6c41e8b8d8aa1ac951ce2fbb6505b7148b16348a3754e55\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T14:10:00Z is after 2025-08-24T17:21:41Z" Dec 03 14:10:00 crc kubenswrapper[4805]: I1203 14:10:00.144257 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 14:10:00 crc kubenswrapper[4805]: I1203 14:10:00.144286 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 14:10:00 crc kubenswrapper[4805]: I1203 14:10:00.144294 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 14:10:00 crc kubenswrapper[4805]: I1203 14:10:00.144307 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 14:10:00 crc kubenswrapper[4805]: I1203 14:10:00.144316 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T14:10:00Z","lastTransitionTime":"2025-12-03T14:10:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 14:10:00 crc kubenswrapper[4805]: I1203 14:10:00.246340 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 14:10:00 crc kubenswrapper[4805]: I1203 14:10:00.246382 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 14:10:00 crc kubenswrapper[4805]: I1203 14:10:00.246393 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 14:10:00 crc kubenswrapper[4805]: I1203 14:10:00.246412 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 14:10:00 crc kubenswrapper[4805]: I1203 14:10:00.246425 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T14:10:00Z","lastTransitionTime":"2025-12-03T14:10:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 14:10:00 crc kubenswrapper[4805]: I1203 14:10:00.348544 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 14:10:00 crc kubenswrapper[4805]: I1203 14:10:00.348630 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 14:10:00 crc kubenswrapper[4805]: I1203 14:10:00.348645 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 14:10:00 crc kubenswrapper[4805]: I1203 14:10:00.348662 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 14:10:00 crc kubenswrapper[4805]: I1203 14:10:00.348675 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T14:10:00Z","lastTransitionTime":"2025-12-03T14:10:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 14:10:00 crc kubenswrapper[4805]: I1203 14:10:00.450933 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 14:10:00 crc kubenswrapper[4805]: I1203 14:10:00.450975 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 14:10:00 crc kubenswrapper[4805]: I1203 14:10:00.450987 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 14:10:00 crc kubenswrapper[4805]: I1203 14:10:00.451001 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 14:10:00 crc kubenswrapper[4805]: I1203 14:10:00.451012 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T14:10:00Z","lastTransitionTime":"2025-12-03T14:10:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 14:10:00 crc kubenswrapper[4805]: I1203 14:10:00.553571 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 14:10:00 crc kubenswrapper[4805]: I1203 14:10:00.553615 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 14:10:00 crc kubenswrapper[4805]: I1203 14:10:00.553627 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 14:10:00 crc kubenswrapper[4805]: I1203 14:10:00.553642 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 14:10:00 crc kubenswrapper[4805]: I1203 14:10:00.553652 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T14:10:00Z","lastTransitionTime":"2025-12-03T14:10:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 14:10:00 crc kubenswrapper[4805]: I1203 14:10:00.656648 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 14:10:00 crc kubenswrapper[4805]: I1203 14:10:00.656720 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 14:10:00 crc kubenswrapper[4805]: I1203 14:10:00.656742 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 14:10:00 crc kubenswrapper[4805]: I1203 14:10:00.656769 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 14:10:00 crc kubenswrapper[4805]: I1203 14:10:00.656791 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T14:10:00Z","lastTransitionTime":"2025-12-03T14:10:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 14:10:00 crc kubenswrapper[4805]: I1203 14:10:00.693962 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 14:10:00 crc kubenswrapper[4805]: I1203 14:10:00.693985 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 14:10:00 crc kubenswrapper[4805]: E1203 14:10:00.694088 4805 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 03 14:10:00 crc kubenswrapper[4805]: I1203 14:10:00.694132 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 14:10:00 crc kubenswrapper[4805]: E1203 14:10:00.694235 4805 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 03 14:10:00 crc kubenswrapper[4805]: E1203 14:10:00.694374 4805 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 03 14:10:00 crc kubenswrapper[4805]: I1203 14:10:00.706082 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-gskh4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6098937f-e3f6-45e8-a647-4994a79cd711\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://28d5653f26ec61647808b3a733ff7eac70aa69ac184660f24147e8cfaf0eb802\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nwrw9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9ad6b68a42fd33ee167e67873c6eed3f11cdf6dc71d22e952c021436eb967dd9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nwrw9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T14:09:50Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-gskh4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T14:10:00Z is after 2025-08-24T17:21:41Z" Dec 03 14:10:00 crc kubenswrapper[4805]: I1203 14:10:00.724817 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-psjpm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"deeec30d-cad2-4dfd-adb6-64f6646581be\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:50Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:50Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4b8d00cdf0f37f2e5cd4e952cf85d3de8d72c43fdf786088fd9f26ed17b1e21f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wdlzx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bd67c6de01a8d196a1f51d4d90640c9e69de6f05d5e201b25e6eaa5c33440ea8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wdlzx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cb8bf05ef093eb0727a066d4bbbdf9af9422867cd4dadacafc6eeb016867a61e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wdlzx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0304467102ed053628c0d3e06548c74b6492336b6fd044246508d1e8bdb17bc3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wdlzx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6ce80ccafb3262a11fbe9923317a9d38b61e5dc5b3e7e7ae136f143d6b643fd7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wdlzx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://47d76962beb8c0f196d50af96fb3ec611097823ca80a1c8bd718f2d18840128c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wdlzx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9e3aa4c113c750c2e2ead5d33f909a53247a221db998125f2494e7058d0c689a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wdlzx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ab98e365c1a4242a47b8cbe2dd85109d366d6002c4de1f18ddc950a46bea6c12\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wdlzx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2f875252fbe6a673a72fc9c56bb96899f62732305ad77b3108e3b244bfa91ecb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2f875252fbe6a673a72fc9c56bb96899f62732305ad77b3108e3b244bfa91ecb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T14:09:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T14:09:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wdlzx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T14:09:50Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-psjpm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T14:10:00Z is after 2025-08-24T17:21:41Z" Dec 03 14:10:00 crc kubenswrapper[4805]: I1203 14:10:00.737915 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:48Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:48Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T14:10:00Z is after 2025-08-24T17:21:41Z" Dec 03 14:10:00 crc kubenswrapper[4805]: I1203 14:10:00.755060 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:49Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:49Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://541ace7ccbb66dbd191c3ca8c1e4479e3edbf80dfc67326bea8aaee81fdcfc3c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://821e3e9a902aacecf6c41e8b8d8aa1ac951ce2fbb6505b7148b16348a3754e55\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T14:10:00Z is after 2025-08-24T17:21:41Z" Dec 03 14:10:00 crc kubenswrapper[4805]: I1203 14:10:00.758547 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 14:10:00 crc kubenswrapper[4805]: I1203 14:10:00.758617 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 14:10:00 crc kubenswrapper[4805]: I1203 14:10:00.758633 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 14:10:00 crc kubenswrapper[4805]: I1203 14:10:00.758658 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 14:10:00 crc kubenswrapper[4805]: I1203 14:10:00.758673 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T14:10:00Z","lastTransitionTime":"2025-12-03T14:10:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 14:10:00 crc kubenswrapper[4805]: I1203 14:10:00.768894 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-pdzgk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"860769de-6a5b-4cb9-a752-264e7e2e785f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5e2f0a0cc21505d609f81918365d5e4f5d21959ac12ab147d6a51601c9de4c14\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ggn4n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T14:09:50Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-pdzgk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T14:10:00Z is after 2025-08-24T17:21:41Z" Dec 03 14:10:00 crc kubenswrapper[4805]: I1203 14:10:00.782325 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:49Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:49Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://38339331a3209d6787d1def81b72f4a8514bf91080a139cad6d8ac5cc6bef10e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T14:10:00Z is after 2025-08-24T17:21:41Z" Dec 03 14:10:00 crc kubenswrapper[4805]: I1203 14:10:00.797600 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:48Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:48Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T14:10:00Z is after 2025-08-24T17:21:41Z" Dec 03 14:10:00 crc kubenswrapper[4805]: I1203 14:10:00.810861 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:51Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:51Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b3123d3fa879097db88ed2f3cfb05631f440c24c5405503b2de3546f12e2e899\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T14:10:00Z is after 2025-08-24T17:21:41Z" Dec 03 14:10:00 crc kubenswrapper[4805]: I1203 14:10:00.820800 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-bdtdg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"53a0e9e3-b6d8-44b7-a32b-aa3450442865\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f557e17ad105d6972d3c1480656c6073f67e28c0f5773aa365549951f4d161c7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9ww54\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T14:09:51Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-bdtdg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T14:10:00Z is after 2025-08-24T17:21:41Z" Dec 03 14:10:00 crc kubenswrapper[4805]: I1203 14:10:00.839086 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b6d36671-7173-4e9c-843f-005b4dcf576c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a837d7d491eaacbe4279b4014bb5fcbe9b514ba5cc82c708227bcd9b25a346e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6e5f04faf3b23197bb8ea11d3e574dc86f31e21d798734962001ec8c95de7c84\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2ff1a979387da41b00e95eb9cc8b2c9de966380a12fae687591df3c5b18a4497\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1e66e1cb18573a14e3c3ae16af4dbd9f8997a6ad38680c3ab4505646ebd48e7c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7d5b5c85e3de43e7e95111aa967248550607c26200c783251b68696f202f92a1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://474f72dac37a808fc4b9832306e76979b151d7221807ff3086013466ed1419c3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://474f72dac37a808fc4b9832306e76979b151d7221807ff3086013466ed1419c3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T14:09:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T14:09:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://74c8cea1a61c4775a0ad2f4ab4bb8a36437cf23d2ee953311d6ea967e297510a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://74c8cea1a61c4775a0ad2f4ab4bb8a36437cf23d2ee953311d6ea967e297510a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T14:09:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T14:09:31Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://a4f2f56a809adc01fba82eba0652d78f58770c132f77b1424f8b17068b1927e3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a4f2f56a809adc01fba82eba0652d78f58770c132f77b1424f8b17068b1927e3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T14:09:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T14:09:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T14:09:30Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T14:10:00Z is after 2025-08-24T17:21:41Z" Dec 03 14:10:00 crc kubenswrapper[4805]: I1203 14:10:00.850623 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ced8291d-0ce5-4e74-9eaf-2bf033774039\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:30Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:30Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3e5a34636f0a8c4d9179074596d28bc2eb7a504573398c596a637defd829907a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://29cd95be3f964415943c873e7914ffcea4abe50dfdf17f8a9cd03fc38df35981\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://951173b3d3d581275a6f6e43209b5e0611a701ca7a10f81504d723e24a2d21a3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e64730957bbbb2feb4d506e0ed9731a0594bf66fa1e8fcdb8ec944ded6308eed\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5da55f769e23ad52bac9c77da62c78e70fa34db8ed1a435e4eaf4dad3fd2b68a\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-03T14:09:48Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1203 14:09:42.992179 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1203 14:09:42.993233 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1830926909/tls.crt::/tmp/serving-cert-1830926909/tls.key\\\\\\\"\\\\nI1203 14:09:48.622403 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1203 14:09:48.626476 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1203 14:09:48.626527 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1203 14:09:48.626574 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1203 14:09:48.626582 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1203 14:09:48.633990 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1203 14:09:48.634024 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 14:09:48.634029 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 14:09:48.634034 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1203 14:09:48.634038 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1203 14:09:48.634042 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1203 14:09:48.634045 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1203 14:09:48.634344 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1203 14:09:48.644497 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T14:09:32Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bd92a160eb530c5687ebd7cd5b83b41820a2b808a0fd9514e5ace9d237e1bd26\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:32Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://af9c940135012c629447528c61743ed1e58708ab57ce76604c6da0479aa459cf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://af9c940135012c629447528c61743ed1e58708ab57ce76604c6da0479aa459cf\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T14:09:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T14:09:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T14:09:30Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T14:10:00Z is after 2025-08-24T17:21:41Z" Dec 03 14:10:00 crc kubenswrapper[4805]: I1203 14:10:00.861421 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 14:10:00 crc kubenswrapper[4805]: I1203 14:10:00.861481 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 14:10:00 crc kubenswrapper[4805]: I1203 14:10:00.861492 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 14:10:00 crc kubenswrapper[4805]: I1203 14:10:00.861510 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 14:10:00 crc kubenswrapper[4805]: I1203 14:10:00.861520 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T14:10:00Z","lastTransitionTime":"2025-12-03T14:10:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 14:10:00 crc kubenswrapper[4805]: I1203 14:10:00.864439 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-k4mhs" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c09c1cdc-f461-45c9-8444-5b8764bccabc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3354c1a44f891182d04116fc76234a2b337d50567d032a3584f902a9516387cc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qdcd7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T14:09:50Z\\\"}}\" for pod \"openshift-multus\"/\"multus-k4mhs\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T14:10:00Z is after 2025-08-24T17:21:41Z" Dec 03 14:10:00 crc kubenswrapper[4805]: I1203 14:10:00.876636 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-bpk49" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"de00a81f-82e1-4223-9614-f98606b9a8ef\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://04f4ad27e02e2024ae9e7cd1b16116267e69cecda882dd723cdb274c41a0e7db\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-25d7t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7741e0e70416ebe094bfabc6009e98d7d4c0fbcba1fc82df13b9ba40774b456a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7741e0e70416ebe094bfabc6009e98d7d4c0fbcba1fc82df13b9ba40774b456a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T14:09:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T14:09:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-25d7t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ee3499c02eb52bbf08c447d2433d435a74dfb70670dc4cbaac44f5fbff9e8a29\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ee3499c02eb52bbf08c447d2433d435a74dfb70670dc4cbaac44f5fbff9e8a29\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T14:09:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T14:09:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-25d7t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ec8488b1d69799951be7ef40e5e028732265c4ddcdea2d453923e567864771f7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ec8488b1d69799951be7ef40e5e028732265c4ddcdea2d453923e567864771f7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T14:09:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T14:09:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-25d7t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://66707bb4e1fdacb265e89786df8a04605830f1c156394cacfcec9b54abe5d41a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://66707bb4e1fdacb265e89786df8a04605830f1c156394cacfcec9b54abe5d41a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T14:09:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T14:09:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-25d7t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6207ab6600970ff15c8713c547d4920b2c77eb9bea990bb5da497e24870dcc0f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6207ab6600970ff15c8713c547d4920b2c77eb9bea990bb5da497e24870dcc0f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T14:09:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T14:09:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-25d7t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8de65d4a205e2deec0b433e53c0e37feb94aa0ede6a4c58d7414b80829804505\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8de65d4a205e2deec0b433e53c0e37feb94aa0ede6a4c58d7414b80829804505\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T14:09:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T14:09:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-25d7t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T14:09:50Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-bpk49\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T14:10:00Z is after 2025-08-24T17:21:41Z" Dec 03 14:10:00 crc kubenswrapper[4805]: I1203 14:10:00.888455 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"750d2b4b-738a-4992-8902-742528f6e875\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://abd2d0156113bb5182727a53aac594410ed051c19258beb8a07b808eff0b42a7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8673a29c0e16f4e707d79f0610722ccabb023ff2b4a18d7cf9d305183817d952\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://35766e7206b1be7a53d13d6890d759cf81dbd999643c12349268e4bf16fbbdb0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5b9844a11b2b903978d0d464b469e26c9211100da61d018dde9636dca95b684e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T14:09:30Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T14:10:00Z is after 2025-08-24T17:21:41Z" Dec 03 14:10:00 crc kubenswrapper[4805]: I1203 14:10:00.902611 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:48Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:48Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T14:10:00Z is after 2025-08-24T17:21:41Z" Dec 03 14:10:00 crc kubenswrapper[4805]: I1203 14:10:00.907175 4805 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-psjpm_deeec30d-cad2-4dfd-adb6-64f6646581be/ovnkube-controller/0.log" Dec 03 14:10:00 crc kubenswrapper[4805]: I1203 14:10:00.909601 4805 generic.go:334] "Generic (PLEG): container finished" podID="deeec30d-cad2-4dfd-adb6-64f6646581be" containerID="9e3aa4c113c750c2e2ead5d33f909a53247a221db998125f2494e7058d0c689a" exitCode=1 Dec 03 14:10:00 crc kubenswrapper[4805]: I1203 14:10:00.909644 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-psjpm" event={"ID":"deeec30d-cad2-4dfd-adb6-64f6646581be","Type":"ContainerDied","Data":"9e3aa4c113c750c2e2ead5d33f909a53247a221db998125f2494e7058d0c689a"} Dec 03 14:10:00 crc kubenswrapper[4805]: I1203 14:10:00.910191 4805 scope.go:117] "RemoveContainer" containerID="9e3aa4c113c750c2e2ead5d33f909a53247a221db998125f2494e7058d0c689a" Dec 03 14:10:00 crc kubenswrapper[4805]: I1203 14:10:00.928971 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:51Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:51Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b3123d3fa879097db88ed2f3cfb05631f440c24c5405503b2de3546f12e2e899\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T14:10:00Z is after 2025-08-24T17:21:41Z" Dec 03 14:10:00 crc kubenswrapper[4805]: I1203 14:10:00.940016 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-bdtdg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"53a0e9e3-b6d8-44b7-a32b-aa3450442865\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f557e17ad105d6972d3c1480656c6073f67e28c0f5773aa365549951f4d161c7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9ww54\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T14:09:51Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-bdtdg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T14:10:00Z is after 2025-08-24T17:21:41Z" Dec 03 14:10:00 crc kubenswrapper[4805]: I1203 14:10:00.961260 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b6d36671-7173-4e9c-843f-005b4dcf576c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a837d7d491eaacbe4279b4014bb5fcbe9b514ba5cc82c708227bcd9b25a346e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6e5f04faf3b23197bb8ea11d3e574dc86f31e21d798734962001ec8c95de7c84\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2ff1a979387da41b00e95eb9cc8b2c9de966380a12fae687591df3c5b18a4497\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1e66e1cb18573a14e3c3ae16af4dbd9f8997a6ad38680c3ab4505646ebd48e7c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7d5b5c85e3de43e7e95111aa967248550607c26200c783251b68696f202f92a1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://474f72dac37a808fc4b9832306e76979b151d7221807ff3086013466ed1419c3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://474f72dac37a808fc4b9832306e76979b151d7221807ff3086013466ed1419c3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T14:09:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T14:09:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://74c8cea1a61c4775a0ad2f4ab4bb8a36437cf23d2ee953311d6ea967e297510a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://74c8cea1a61c4775a0ad2f4ab4bb8a36437cf23d2ee953311d6ea967e297510a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T14:09:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T14:09:31Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://a4f2f56a809adc01fba82eba0652d78f58770c132f77b1424f8b17068b1927e3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a4f2f56a809adc01fba82eba0652d78f58770c132f77b1424f8b17068b1927e3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T14:09:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T14:09:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T14:09:30Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T14:10:00Z is after 2025-08-24T17:21:41Z" Dec 03 14:10:00 crc kubenswrapper[4805]: I1203 14:10:00.962785 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 14:10:00 crc kubenswrapper[4805]: I1203 14:10:00.962818 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 14:10:00 crc kubenswrapper[4805]: I1203 14:10:00.962831 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 14:10:00 crc kubenswrapper[4805]: I1203 14:10:00.962879 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 14:10:00 crc kubenswrapper[4805]: I1203 14:10:00.962890 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T14:10:00Z","lastTransitionTime":"2025-12-03T14:10:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 14:10:00 crc kubenswrapper[4805]: I1203 14:10:00.977940 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ced8291d-0ce5-4e74-9eaf-2bf033774039\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:30Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:30Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3e5a34636f0a8c4d9179074596d28bc2eb7a504573398c596a637defd829907a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://29cd95be3f964415943c873e7914ffcea4abe50dfdf17f8a9cd03fc38df35981\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://951173b3d3d581275a6f6e43209b5e0611a701ca7a10f81504d723e24a2d21a3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e64730957bbbb2feb4d506e0ed9731a0594bf66fa1e8fcdb8ec944ded6308eed\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5da55f769e23ad52bac9c77da62c78e70fa34db8ed1a435e4eaf4dad3fd2b68a\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-03T14:09:48Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1203 14:09:42.992179 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1203 14:09:42.993233 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1830926909/tls.crt::/tmp/serving-cert-1830926909/tls.key\\\\\\\"\\\\nI1203 14:09:48.622403 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1203 14:09:48.626476 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1203 14:09:48.626527 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1203 14:09:48.626574 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1203 14:09:48.626582 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1203 14:09:48.633990 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1203 14:09:48.634024 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 14:09:48.634029 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 14:09:48.634034 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1203 14:09:48.634038 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1203 14:09:48.634042 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1203 14:09:48.634045 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1203 14:09:48.634344 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1203 14:09:48.644497 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T14:09:32Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bd92a160eb530c5687ebd7cd5b83b41820a2b808a0fd9514e5ace9d237e1bd26\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:32Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://af9c940135012c629447528c61743ed1e58708ab57ce76604c6da0479aa459cf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://af9c940135012c629447528c61743ed1e58708ab57ce76604c6da0479aa459cf\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T14:09:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T14:09:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T14:09:30Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T14:10:00Z is after 2025-08-24T17:21:41Z" Dec 03 14:10:00 crc kubenswrapper[4805]: I1203 14:10:00.992093 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:49Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:49Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://38339331a3209d6787d1def81b72f4a8514bf91080a139cad6d8ac5cc6bef10e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T14:10:00Z is after 2025-08-24T17:21:41Z" Dec 03 14:10:01 crc kubenswrapper[4805]: I1203 14:10:01.003423 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:48Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:48Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T14:10:01Z is after 2025-08-24T17:21:41Z" Dec 03 14:10:01 crc kubenswrapper[4805]: I1203 14:10:01.016706 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"750d2b4b-738a-4992-8902-742528f6e875\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://abd2d0156113bb5182727a53aac594410ed051c19258beb8a07b808eff0b42a7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8673a29c0e16f4e707d79f0610722ccabb023ff2b4a18d7cf9d305183817d952\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://35766e7206b1be7a53d13d6890d759cf81dbd999643c12349268e4bf16fbbdb0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5b9844a11b2b903978d0d464b469e26c9211100da61d018dde9636dca95b684e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T14:09:30Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T14:10:01Z is after 2025-08-24T17:21:41Z" Dec 03 14:10:01 crc kubenswrapper[4805]: I1203 14:10:01.029679 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:48Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:48Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T14:10:01Z is after 2025-08-24T17:21:41Z" Dec 03 14:10:01 crc kubenswrapper[4805]: I1203 14:10:01.044433 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-k4mhs" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c09c1cdc-f461-45c9-8444-5b8764bccabc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3354c1a44f891182d04116fc76234a2b337d50567d032a3584f902a9516387cc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qdcd7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T14:09:50Z\\\"}}\" for pod \"openshift-multus\"/\"multus-k4mhs\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T14:10:01Z is after 2025-08-24T17:21:41Z" Dec 03 14:10:01 crc kubenswrapper[4805]: I1203 14:10:01.059687 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-bpk49" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"de00a81f-82e1-4223-9614-f98606b9a8ef\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://04f4ad27e02e2024ae9e7cd1b16116267e69cecda882dd723cdb274c41a0e7db\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-25d7t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7741e0e70416ebe094bfabc6009e98d7d4c0fbcba1fc82df13b9ba40774b456a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7741e0e70416ebe094bfabc6009e98d7d4c0fbcba1fc82df13b9ba40774b456a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T14:09:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T14:09:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-25d7t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ee3499c02eb52bbf08c447d2433d435a74dfb70670dc4cbaac44f5fbff9e8a29\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ee3499c02eb52bbf08c447d2433d435a74dfb70670dc4cbaac44f5fbff9e8a29\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T14:09:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T14:09:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-25d7t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ec8488b1d69799951be7ef40e5e028732265c4ddcdea2d453923e567864771f7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ec8488b1d69799951be7ef40e5e028732265c4ddcdea2d453923e567864771f7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T14:09:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T14:09:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-25d7t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://66707bb4e1fdacb265e89786df8a04605830f1c156394cacfcec9b54abe5d41a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://66707bb4e1fdacb265e89786df8a04605830f1c156394cacfcec9b54abe5d41a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T14:09:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T14:09:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-25d7t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6207ab6600970ff15c8713c547d4920b2c77eb9bea990bb5da497e24870dcc0f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6207ab6600970ff15c8713c547d4920b2c77eb9bea990bb5da497e24870dcc0f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T14:09:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T14:09:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-25d7t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8de65d4a205e2deec0b433e53c0e37feb94aa0ede6a4c58d7414b80829804505\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8de65d4a205e2deec0b433e53c0e37feb94aa0ede6a4c58d7414b80829804505\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T14:09:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T14:09:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-25d7t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T14:09:50Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-bpk49\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T14:10:01Z is after 2025-08-24T17:21:41Z" Dec 03 14:10:01 crc kubenswrapper[4805]: I1203 14:10:01.064850 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 14:10:01 crc kubenswrapper[4805]: I1203 14:10:01.064884 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 14:10:01 crc kubenswrapper[4805]: I1203 14:10:01.064898 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 14:10:01 crc kubenswrapper[4805]: I1203 14:10:01.064914 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 14:10:01 crc kubenswrapper[4805]: I1203 14:10:01.064927 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T14:10:01Z","lastTransitionTime":"2025-12-03T14:10:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 14:10:01 crc kubenswrapper[4805]: I1203 14:10:01.074580 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:48Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:48Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T14:10:01Z is after 2025-08-24T17:21:41Z" Dec 03 14:10:01 crc kubenswrapper[4805]: I1203 14:10:01.088209 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:49Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:49Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://541ace7ccbb66dbd191c3ca8c1e4479e3edbf80dfc67326bea8aaee81fdcfc3c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://821e3e9a902aacecf6c41e8b8d8aa1ac951ce2fbb6505b7148b16348a3754e55\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T14:10:01Z is after 2025-08-24T17:21:41Z" Dec 03 14:10:01 crc kubenswrapper[4805]: I1203 14:10:01.101319 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-gskh4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6098937f-e3f6-45e8-a647-4994a79cd711\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://28d5653f26ec61647808b3a733ff7eac70aa69ac184660f24147e8cfaf0eb802\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nwrw9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9ad6b68a42fd33ee167e67873c6eed3f11cdf6dc71d22e952c021436eb967dd9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nwrw9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T14:09:50Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-gskh4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T14:10:01Z is after 2025-08-24T17:21:41Z" Dec 03 14:10:01 crc kubenswrapper[4805]: I1203 14:10:01.122453 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-psjpm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"deeec30d-cad2-4dfd-adb6-64f6646581be\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:50Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:50Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4b8d00cdf0f37f2e5cd4e952cf85d3de8d72c43fdf786088fd9f26ed17b1e21f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wdlzx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bd67c6de01a8d196a1f51d4d90640c9e69de6f05d5e201b25e6eaa5c33440ea8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wdlzx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cb8bf05ef093eb0727a066d4bbbdf9af9422867cd4dadacafc6eeb016867a61e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wdlzx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0304467102ed053628c0d3e06548c74b6492336b6fd044246508d1e8bdb17bc3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wdlzx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6ce80ccafb3262a11fbe9923317a9d38b61e5dc5b3e7e7ae136f143d6b643fd7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wdlzx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://47d76962beb8c0f196d50af96fb3ec611097823ca80a1c8bd718f2d18840128c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wdlzx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9e3aa4c113c750c2e2ead5d33f909a53247a221db998125f2494e7058d0c689a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9e3aa4c113c750c2e2ead5d33f909a53247a221db998125f2494e7058d0c689a\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-03T14:10:00Z\\\",\\\"message\\\":\\\" 6125 reflector.go:311] Stopping reflector *v1.NetworkPolicy (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1203 14:10:00.108261 6125 reflector.go:311] Stopping reflector *v1.Pod (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1203 14:10:00.108416 6125 reflector.go:311] Stopping reflector *v1.Namespace (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1203 14:10:00.108932 6125 reflector.go:311] Stopping reflector *v1.EndpointSlice (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1203 14:10:00.108462 6125 reflector.go:311] Stopping reflector *v1.Node (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1203 14:10:00.109063 6125 reflector.go:311] Stopping reflector *v1.Service (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1203 14:10:00.109526 6125 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1203 14:10:00.109568 6125 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI1203 14:10:00.109575 6125 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI1203 14:10:00.109591 6125 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI1203 14:10:00.109612 6125 factory.go:656] Stopping watch factory\\\\nI1203 14:10:00.109628 6125 ovnkube.go:599] Stopped ovnkube\\\\nI1203 14:10:00.109650 6125 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1203 14\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T14:09:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wdlzx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ab98e365c1a4242a47b8cbe2dd85109d366d6002c4de1f18ddc950a46bea6c12\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wdlzx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2f875252fbe6a673a72fc9c56bb96899f62732305ad77b3108e3b244bfa91ecb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2f875252fbe6a673a72fc9c56bb96899f62732305ad77b3108e3b244bfa91ecb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T14:09:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T14:09:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wdlzx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T14:09:50Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-psjpm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T14:10:01Z is after 2025-08-24T17:21:41Z" Dec 03 14:10:01 crc kubenswrapper[4805]: I1203 14:10:01.134145 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-pdzgk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"860769de-6a5b-4cb9-a752-264e7e2e785f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5e2f0a0cc21505d609f81918365d5e4f5d21959ac12ab147d6a51601c9de4c14\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ggn4n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T14:09:50Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-pdzgk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T14:10:01Z is after 2025-08-24T17:21:41Z" Dec 03 14:10:01 crc kubenswrapper[4805]: I1203 14:10:01.167158 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 14:10:01 crc kubenswrapper[4805]: I1203 14:10:01.167194 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 14:10:01 crc kubenswrapper[4805]: I1203 14:10:01.167206 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 14:10:01 crc kubenswrapper[4805]: I1203 14:10:01.167222 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 14:10:01 crc kubenswrapper[4805]: I1203 14:10:01.167233 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T14:10:01Z","lastTransitionTime":"2025-12-03T14:10:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 14:10:01 crc kubenswrapper[4805]: I1203 14:10:01.269000 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 14:10:01 crc kubenswrapper[4805]: I1203 14:10:01.269035 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 14:10:01 crc kubenswrapper[4805]: I1203 14:10:01.269044 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 14:10:01 crc kubenswrapper[4805]: I1203 14:10:01.269057 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 14:10:01 crc kubenswrapper[4805]: I1203 14:10:01.269066 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T14:10:01Z","lastTransitionTime":"2025-12-03T14:10:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 14:10:01 crc kubenswrapper[4805]: I1203 14:10:01.370986 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 14:10:01 crc kubenswrapper[4805]: I1203 14:10:01.371024 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 14:10:01 crc kubenswrapper[4805]: I1203 14:10:01.371032 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 14:10:01 crc kubenswrapper[4805]: I1203 14:10:01.371045 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 14:10:01 crc kubenswrapper[4805]: I1203 14:10:01.371053 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T14:10:01Z","lastTransitionTime":"2025-12-03T14:10:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 14:10:01 crc kubenswrapper[4805]: I1203 14:10:01.473428 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 14:10:01 crc kubenswrapper[4805]: I1203 14:10:01.473472 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 14:10:01 crc kubenswrapper[4805]: I1203 14:10:01.473483 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 14:10:01 crc kubenswrapper[4805]: I1203 14:10:01.473498 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 14:10:01 crc kubenswrapper[4805]: I1203 14:10:01.473509 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T14:10:01Z","lastTransitionTime":"2025-12-03T14:10:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 14:10:01 crc kubenswrapper[4805]: I1203 14:10:01.576071 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 14:10:01 crc kubenswrapper[4805]: I1203 14:10:01.576111 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 14:10:01 crc kubenswrapper[4805]: I1203 14:10:01.576121 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 14:10:01 crc kubenswrapper[4805]: I1203 14:10:01.576134 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 14:10:01 crc kubenswrapper[4805]: I1203 14:10:01.576144 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T14:10:01Z","lastTransitionTime":"2025-12-03T14:10:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 14:10:01 crc kubenswrapper[4805]: I1203 14:10:01.679409 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 14:10:01 crc kubenswrapper[4805]: I1203 14:10:01.679439 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 14:10:01 crc kubenswrapper[4805]: I1203 14:10:01.679449 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 14:10:01 crc kubenswrapper[4805]: I1203 14:10:01.679465 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 14:10:01 crc kubenswrapper[4805]: I1203 14:10:01.679476 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T14:10:01Z","lastTransitionTime":"2025-12-03T14:10:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 14:10:01 crc kubenswrapper[4805]: I1203 14:10:01.782272 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 14:10:01 crc kubenswrapper[4805]: I1203 14:10:01.782335 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 14:10:01 crc kubenswrapper[4805]: I1203 14:10:01.782354 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 14:10:01 crc kubenswrapper[4805]: I1203 14:10:01.782389 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 14:10:01 crc kubenswrapper[4805]: I1203 14:10:01.782409 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T14:10:01Z","lastTransitionTime":"2025-12-03T14:10:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 14:10:01 crc kubenswrapper[4805]: I1203 14:10:01.884785 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 14:10:01 crc kubenswrapper[4805]: I1203 14:10:01.884870 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 14:10:01 crc kubenswrapper[4805]: I1203 14:10:01.884891 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 14:10:01 crc kubenswrapper[4805]: I1203 14:10:01.884916 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 14:10:01 crc kubenswrapper[4805]: I1203 14:10:01.884933 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T14:10:01Z","lastTransitionTime":"2025-12-03T14:10:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 14:10:01 crc kubenswrapper[4805]: I1203 14:10:01.923383 4805 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-psjpm_deeec30d-cad2-4dfd-adb6-64f6646581be/ovnkube-controller/0.log" Dec 03 14:10:01 crc kubenswrapper[4805]: I1203 14:10:01.927584 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-psjpm" event={"ID":"deeec30d-cad2-4dfd-adb6-64f6646581be","Type":"ContainerStarted","Data":"644e56c400524d739e2a208fb7eee1551ebbd28ba220df91c9dd84ba09d0b400"} Dec 03 14:10:01 crc kubenswrapper[4805]: I1203 14:10:01.927692 4805 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Dec 03 14:10:01 crc kubenswrapper[4805]: I1203 14:10:01.950714 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"750d2b4b-738a-4992-8902-742528f6e875\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://abd2d0156113bb5182727a53aac594410ed051c19258beb8a07b808eff0b42a7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8673a29c0e16f4e707d79f0610722ccabb023ff2b4a18d7cf9d305183817d952\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://35766e7206b1be7a53d13d6890d759cf81dbd999643c12349268e4bf16fbbdb0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5b9844a11b2b903978d0d464b469e26c9211100da61d018dde9636dca95b684e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T14:09:30Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T14:10:01Z is after 2025-08-24T17:21:41Z" Dec 03 14:10:01 crc kubenswrapper[4805]: I1203 14:10:01.968619 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:48Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:48Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T14:10:01Z is after 2025-08-24T17:21:41Z" Dec 03 14:10:01 crc kubenswrapper[4805]: I1203 14:10:01.981802 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 14:10:01 crc kubenswrapper[4805]: I1203 14:10:01.981884 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 14:10:01 crc kubenswrapper[4805]: I1203 14:10:01.981903 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 14:10:01 crc kubenswrapper[4805]: I1203 14:10:01.981928 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 14:10:01 crc kubenswrapper[4805]: I1203 14:10:01.981946 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T14:10:01Z","lastTransitionTime":"2025-12-03T14:10:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 14:10:01 crc kubenswrapper[4805]: I1203 14:10:01.988259 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-k4mhs" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c09c1cdc-f461-45c9-8444-5b8764bccabc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3354c1a44f891182d04116fc76234a2b337d50567d032a3584f902a9516387cc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qdcd7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T14:09:50Z\\\"}}\" for pod \"openshift-multus\"/\"multus-k4mhs\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T14:10:01Z is after 2025-08-24T17:21:41Z" Dec 03 14:10:02 crc kubenswrapper[4805]: E1203 14:10:02.002222 4805 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T14:10:01Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T14:10:01Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T14:10:01Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T14:10:01Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T14:10:01Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T14:10:01Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T14:10:01Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T14:10:01Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"e4edfbcd-1de1-486e-9c0b-65f9fbba52ad\\\",\\\"systemUUID\\\":\\\"7a729b24-19ef-490e-ab4c-d31c52ed5658\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T14:10:01Z is after 2025-08-24T17:21:41Z" Dec 03 14:10:02 crc kubenswrapper[4805]: I1203 14:10:02.005347 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-bpk49" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"de00a81f-82e1-4223-9614-f98606b9a8ef\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://04f4ad27e02e2024ae9e7cd1b16116267e69cecda882dd723cdb274c41a0e7db\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-25d7t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7741e0e70416ebe094bfabc6009e98d7d4c0fbcba1fc82df13b9ba40774b456a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7741e0e70416ebe094bfabc6009e98d7d4c0fbcba1fc82df13b9ba40774b456a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T14:09:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T14:09:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-25d7t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ee3499c02eb52bbf08c447d2433d435a74dfb70670dc4cbaac44f5fbff9e8a29\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ee3499c02eb52bbf08c447d2433d435a74dfb70670dc4cbaac44f5fbff9e8a29\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T14:09:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T14:09:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-25d7t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ec8488b1d69799951be7ef40e5e028732265c4ddcdea2d453923e567864771f7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ec8488b1d69799951be7ef40e5e028732265c4ddcdea2d453923e567864771f7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T14:09:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T14:09:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-25d7t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://66707bb4e1fdacb265e89786df8a04605830f1c156394cacfcec9b54abe5d41a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://66707bb4e1fdacb265e89786df8a04605830f1c156394cacfcec9b54abe5d41a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T14:09:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T14:09:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-25d7t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6207ab6600970ff15c8713c547d4920b2c77eb9bea990bb5da497e24870dcc0f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6207ab6600970ff15c8713c547d4920b2c77eb9bea990bb5da497e24870dcc0f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T14:09:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T14:09:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-25d7t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8de65d4a205e2deec0b433e53c0e37feb94aa0ede6a4c58d7414b80829804505\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8de65d4a205e2deec0b433e53c0e37feb94aa0ede6a4c58d7414b80829804505\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T14:09:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T14:09:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-25d7t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T14:09:50Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-bpk49\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T14:10:02Z is after 2025-08-24T17:21:41Z" Dec 03 14:10:02 crc kubenswrapper[4805]: I1203 14:10:02.007623 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 14:10:02 crc kubenswrapper[4805]: I1203 14:10:02.007695 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 14:10:02 crc kubenswrapper[4805]: I1203 14:10:02.007721 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 14:10:02 crc kubenswrapper[4805]: I1203 14:10:02.007753 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 14:10:02 crc kubenswrapper[4805]: I1203 14:10:02.007776 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T14:10:02Z","lastTransitionTime":"2025-12-03T14:10:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 14:10:02 crc kubenswrapper[4805]: I1203 14:10:02.023912 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:48Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:48Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T14:10:02Z is after 2025-08-24T17:21:41Z" Dec 03 14:10:02 crc kubenswrapper[4805]: E1203 14:10:02.024990 4805 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T14:10:02Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T14:10:02Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T14:10:02Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T14:10:02Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T14:10:02Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T14:10:02Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T14:10:02Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T14:10:02Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"e4edfbcd-1de1-486e-9c0b-65f9fbba52ad\\\",\\\"systemUUID\\\":\\\"7a729b24-19ef-490e-ab4c-d31c52ed5658\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T14:10:02Z is after 2025-08-24T17:21:41Z" Dec 03 14:10:02 crc kubenswrapper[4805]: I1203 14:10:02.030298 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 14:10:02 crc kubenswrapper[4805]: I1203 14:10:02.030392 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 14:10:02 crc kubenswrapper[4805]: I1203 14:10:02.030410 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 14:10:02 crc kubenswrapper[4805]: I1203 14:10:02.030434 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 14:10:02 crc kubenswrapper[4805]: I1203 14:10:02.030453 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T14:10:02Z","lastTransitionTime":"2025-12-03T14:10:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 14:10:02 crc kubenswrapper[4805]: I1203 14:10:02.043457 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:49Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:49Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://541ace7ccbb66dbd191c3ca8c1e4479e3edbf80dfc67326bea8aaee81fdcfc3c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://821e3e9a902aacecf6c41e8b8d8aa1ac951ce2fbb6505b7148b16348a3754e55\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T14:10:02Z is after 2025-08-24T17:21:41Z" Dec 03 14:10:02 crc kubenswrapper[4805]: E1203 14:10:02.050532 4805 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T14:10:02Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T14:10:02Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T14:10:02Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T14:10:02Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T14:10:02Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T14:10:02Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T14:10:02Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T14:10:02Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"e4edfbcd-1de1-486e-9c0b-65f9fbba52ad\\\",\\\"systemUUID\\\":\\\"7a729b24-19ef-490e-ab4c-d31c52ed5658\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T14:10:02Z is after 2025-08-24T17:21:41Z" Dec 03 14:10:02 crc kubenswrapper[4805]: I1203 14:10:02.055734 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 14:10:02 crc kubenswrapper[4805]: I1203 14:10:02.055788 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 14:10:02 crc kubenswrapper[4805]: I1203 14:10:02.055805 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 14:10:02 crc kubenswrapper[4805]: I1203 14:10:02.055829 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 14:10:02 crc kubenswrapper[4805]: I1203 14:10:02.055876 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T14:10:02Z","lastTransitionTime":"2025-12-03T14:10:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 14:10:02 crc kubenswrapper[4805]: I1203 14:10:02.064402 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-gskh4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6098937f-e3f6-45e8-a647-4994a79cd711\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://28d5653f26ec61647808b3a733ff7eac70aa69ac184660f24147e8cfaf0eb802\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nwrw9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9ad6b68a42fd33ee167e67873c6eed3f11cdf6dc71d22e952c021436eb967dd9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nwrw9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T14:09:50Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-gskh4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T14:10:02Z is after 2025-08-24T17:21:41Z" Dec 03 14:10:02 crc kubenswrapper[4805]: E1203 14:10:02.075080 4805 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T14:10:02Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T14:10:02Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T14:10:02Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T14:10:02Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T14:10:02Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T14:10:02Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T14:10:02Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T14:10:02Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"e4edfbcd-1de1-486e-9c0b-65f9fbba52ad\\\",\\\"systemUUID\\\":\\\"7a729b24-19ef-490e-ab4c-d31c52ed5658\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T14:10:02Z is after 2025-08-24T17:21:41Z" Dec 03 14:10:02 crc kubenswrapper[4805]: I1203 14:10:02.080640 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 14:10:02 crc kubenswrapper[4805]: I1203 14:10:02.080700 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 14:10:02 crc kubenswrapper[4805]: I1203 14:10:02.080723 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 14:10:02 crc kubenswrapper[4805]: I1203 14:10:02.080747 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 14:10:02 crc kubenswrapper[4805]: I1203 14:10:02.080766 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T14:10:02Z","lastTransitionTime":"2025-12-03T14:10:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 14:10:02 crc kubenswrapper[4805]: I1203 14:10:02.093979 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-psjpm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"deeec30d-cad2-4dfd-adb6-64f6646581be\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:50Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:50Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4b8d00cdf0f37f2e5cd4e952cf85d3de8d72c43fdf786088fd9f26ed17b1e21f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wdlzx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bd67c6de01a8d196a1f51d4d90640c9e69de6f05d5e201b25e6eaa5c33440ea8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wdlzx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cb8bf05ef093eb0727a066d4bbbdf9af9422867cd4dadacafc6eeb016867a61e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wdlzx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0304467102ed053628c0d3e06548c74b6492336b6fd044246508d1e8bdb17bc3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wdlzx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6ce80ccafb3262a11fbe9923317a9d38b61e5dc5b3e7e7ae136f143d6b643fd7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wdlzx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://47d76962beb8c0f196d50af96fb3ec611097823ca80a1c8bd718f2d18840128c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wdlzx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://644e56c400524d739e2a208fb7eee1551ebbd28ba220df91c9dd84ba09d0b400\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9e3aa4c113c750c2e2ead5d33f909a53247a221db998125f2494e7058d0c689a\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-03T14:10:00Z\\\",\\\"message\\\":\\\" 6125 reflector.go:311] Stopping reflector *v1.NetworkPolicy (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1203 14:10:00.108261 6125 reflector.go:311] Stopping reflector *v1.Pod (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1203 14:10:00.108416 6125 reflector.go:311] Stopping reflector *v1.Namespace (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1203 14:10:00.108932 6125 reflector.go:311] Stopping reflector *v1.EndpointSlice (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1203 14:10:00.108462 6125 reflector.go:311] Stopping reflector *v1.Node (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1203 14:10:00.109063 6125 reflector.go:311] Stopping reflector *v1.Service (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1203 14:10:00.109526 6125 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1203 14:10:00.109568 6125 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI1203 14:10:00.109575 6125 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI1203 14:10:00.109591 6125 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI1203 14:10:00.109612 6125 factory.go:656] Stopping watch factory\\\\nI1203 14:10:00.109628 6125 ovnkube.go:599] Stopped ovnkube\\\\nI1203 14:10:00.109650 6125 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1203 14\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T14:09:58Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:10:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wdlzx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ab98e365c1a4242a47b8cbe2dd85109d366d6002c4de1f18ddc950a46bea6c12\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wdlzx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2f875252fbe6a673a72fc9c56bb96899f62732305ad77b3108e3b244bfa91ecb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2f875252fbe6a673a72fc9c56bb96899f62732305ad77b3108e3b244bfa91ecb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T14:09:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T14:09:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wdlzx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T14:09:50Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-psjpm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T14:10:02Z is after 2025-08-24T17:21:41Z" Dec 03 14:10:02 crc kubenswrapper[4805]: E1203 14:10:02.101820 4805 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T14:10:02Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T14:10:02Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T14:10:02Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T14:10:02Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T14:10:02Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T14:10:02Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T14:10:02Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T14:10:02Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"e4edfbcd-1de1-486e-9c0b-65f9fbba52ad\\\",\\\"systemUUID\\\":\\\"7a729b24-19ef-490e-ab4c-d31c52ed5658\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T14:10:02Z is after 2025-08-24T17:21:41Z" Dec 03 14:10:02 crc kubenswrapper[4805]: E1203 14:10:02.102100 4805 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Dec 03 14:10:02 crc kubenswrapper[4805]: I1203 14:10:02.104194 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 14:10:02 crc kubenswrapper[4805]: I1203 14:10:02.104236 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 14:10:02 crc kubenswrapper[4805]: I1203 14:10:02.104252 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 14:10:02 crc kubenswrapper[4805]: I1203 14:10:02.104278 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 14:10:02 crc kubenswrapper[4805]: I1203 14:10:02.104296 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T14:10:02Z","lastTransitionTime":"2025-12-03T14:10:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 14:10:02 crc kubenswrapper[4805]: I1203 14:10:02.109278 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-pdzgk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"860769de-6a5b-4cb9-a752-264e7e2e785f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5e2f0a0cc21505d609f81918365d5e4f5d21959ac12ab147d6a51601c9de4c14\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ggn4n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T14:09:50Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-pdzgk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T14:10:02Z is after 2025-08-24T17:21:41Z" Dec 03 14:10:02 crc kubenswrapper[4805]: I1203 14:10:02.142006 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b6d36671-7173-4e9c-843f-005b4dcf576c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a837d7d491eaacbe4279b4014bb5fcbe9b514ba5cc82c708227bcd9b25a346e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6e5f04faf3b23197bb8ea11d3e574dc86f31e21d798734962001ec8c95de7c84\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2ff1a979387da41b00e95eb9cc8b2c9de966380a12fae687591df3c5b18a4497\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1e66e1cb18573a14e3c3ae16af4dbd9f8997a6ad38680c3ab4505646ebd48e7c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7d5b5c85e3de43e7e95111aa967248550607c26200c783251b68696f202f92a1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://474f72dac37a808fc4b9832306e76979b151d7221807ff3086013466ed1419c3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://474f72dac37a808fc4b9832306e76979b151d7221807ff3086013466ed1419c3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T14:09:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T14:09:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://74c8cea1a61c4775a0ad2f4ab4bb8a36437cf23d2ee953311d6ea967e297510a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://74c8cea1a61c4775a0ad2f4ab4bb8a36437cf23d2ee953311d6ea967e297510a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T14:09:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T14:09:31Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://a4f2f56a809adc01fba82eba0652d78f58770c132f77b1424f8b17068b1927e3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a4f2f56a809adc01fba82eba0652d78f58770c132f77b1424f8b17068b1927e3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T14:09:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T14:09:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T14:09:30Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T14:10:02Z is after 2025-08-24T17:21:41Z" Dec 03 14:10:02 crc kubenswrapper[4805]: I1203 14:10:02.162381 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ced8291d-0ce5-4e74-9eaf-2bf033774039\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:30Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:30Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3e5a34636f0a8c4d9179074596d28bc2eb7a504573398c596a637defd829907a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://29cd95be3f964415943c873e7914ffcea4abe50dfdf17f8a9cd03fc38df35981\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://951173b3d3d581275a6f6e43209b5e0611a701ca7a10f81504d723e24a2d21a3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e64730957bbbb2feb4d506e0ed9731a0594bf66fa1e8fcdb8ec944ded6308eed\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5da55f769e23ad52bac9c77da62c78e70fa34db8ed1a435e4eaf4dad3fd2b68a\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-03T14:09:48Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1203 14:09:42.992179 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1203 14:09:42.993233 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1830926909/tls.crt::/tmp/serving-cert-1830926909/tls.key\\\\\\\"\\\\nI1203 14:09:48.622403 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1203 14:09:48.626476 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1203 14:09:48.626527 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1203 14:09:48.626574 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1203 14:09:48.626582 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1203 14:09:48.633990 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1203 14:09:48.634024 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 14:09:48.634029 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 14:09:48.634034 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1203 14:09:48.634038 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1203 14:09:48.634042 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1203 14:09:48.634045 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1203 14:09:48.634344 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1203 14:09:48.644497 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T14:09:32Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bd92a160eb530c5687ebd7cd5b83b41820a2b808a0fd9514e5ace9d237e1bd26\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:32Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://af9c940135012c629447528c61743ed1e58708ab57ce76604c6da0479aa459cf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://af9c940135012c629447528c61743ed1e58708ab57ce76604c6da0479aa459cf\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T14:09:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T14:09:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T14:09:30Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T14:10:02Z is after 2025-08-24T17:21:41Z" Dec 03 14:10:02 crc kubenswrapper[4805]: I1203 14:10:02.178290 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:49Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:49Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://38339331a3209d6787d1def81b72f4a8514bf91080a139cad6d8ac5cc6bef10e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T14:10:02Z is after 2025-08-24T17:21:41Z" Dec 03 14:10:02 crc kubenswrapper[4805]: I1203 14:10:02.192647 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:48Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:48Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T14:10:02Z is after 2025-08-24T17:21:41Z" Dec 03 14:10:02 crc kubenswrapper[4805]: I1203 14:10:02.204200 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:51Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:51Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b3123d3fa879097db88ed2f3cfb05631f440c24c5405503b2de3546f12e2e899\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T14:10:02Z is after 2025-08-24T17:21:41Z" Dec 03 14:10:02 crc kubenswrapper[4805]: I1203 14:10:02.206424 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 14:10:02 crc kubenswrapper[4805]: I1203 14:10:02.206458 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 14:10:02 crc kubenswrapper[4805]: I1203 14:10:02.206469 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 14:10:02 crc kubenswrapper[4805]: I1203 14:10:02.206488 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 14:10:02 crc kubenswrapper[4805]: I1203 14:10:02.206501 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T14:10:02Z","lastTransitionTime":"2025-12-03T14:10:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 14:10:02 crc kubenswrapper[4805]: I1203 14:10:02.215136 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-bdtdg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"53a0e9e3-b6d8-44b7-a32b-aa3450442865\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f557e17ad105d6972d3c1480656c6073f67e28c0f5773aa365549951f4d161c7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9ww54\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T14:09:51Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-bdtdg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T14:10:02Z is after 2025-08-24T17:21:41Z" Dec 03 14:10:02 crc kubenswrapper[4805]: I1203 14:10:02.308670 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 14:10:02 crc kubenswrapper[4805]: I1203 14:10:02.308707 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 14:10:02 crc kubenswrapper[4805]: I1203 14:10:02.308715 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 14:10:02 crc kubenswrapper[4805]: I1203 14:10:02.308745 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 14:10:02 crc kubenswrapper[4805]: I1203 14:10:02.308764 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T14:10:02Z","lastTransitionTime":"2025-12-03T14:10:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 14:10:02 crc kubenswrapper[4805]: I1203 14:10:02.411327 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 14:10:02 crc kubenswrapper[4805]: I1203 14:10:02.411375 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 14:10:02 crc kubenswrapper[4805]: I1203 14:10:02.411384 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 14:10:02 crc kubenswrapper[4805]: I1203 14:10:02.411398 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 14:10:02 crc kubenswrapper[4805]: I1203 14:10:02.411408 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T14:10:02Z","lastTransitionTime":"2025-12-03T14:10:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 14:10:02 crc kubenswrapper[4805]: I1203 14:10:02.514760 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 14:10:02 crc kubenswrapper[4805]: I1203 14:10:02.514830 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 14:10:02 crc kubenswrapper[4805]: I1203 14:10:02.514874 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 14:10:02 crc kubenswrapper[4805]: I1203 14:10:02.514898 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 14:10:02 crc kubenswrapper[4805]: I1203 14:10:02.514914 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T14:10:02Z","lastTransitionTime":"2025-12-03T14:10:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 14:10:02 crc kubenswrapper[4805]: I1203 14:10:02.618524 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 14:10:02 crc kubenswrapper[4805]: I1203 14:10:02.618592 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 14:10:02 crc kubenswrapper[4805]: I1203 14:10:02.618614 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 14:10:02 crc kubenswrapper[4805]: I1203 14:10:02.618678 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 14:10:02 crc kubenswrapper[4805]: I1203 14:10:02.618698 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T14:10:02Z","lastTransitionTime":"2025-12-03T14:10:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 14:10:02 crc kubenswrapper[4805]: I1203 14:10:02.693489 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 14:10:02 crc kubenswrapper[4805]: I1203 14:10:02.693541 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 14:10:02 crc kubenswrapper[4805]: E1203 14:10:02.693608 4805 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 03 14:10:02 crc kubenswrapper[4805]: E1203 14:10:02.694207 4805 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 03 14:10:02 crc kubenswrapper[4805]: I1203 14:10:02.694247 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 14:10:02 crc kubenswrapper[4805]: E1203 14:10:02.694494 4805 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 03 14:10:02 crc kubenswrapper[4805]: I1203 14:10:02.721817 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 14:10:02 crc kubenswrapper[4805]: I1203 14:10:02.721915 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 14:10:02 crc kubenswrapper[4805]: I1203 14:10:02.721933 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 14:10:02 crc kubenswrapper[4805]: I1203 14:10:02.721959 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 14:10:02 crc kubenswrapper[4805]: I1203 14:10:02.721984 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T14:10:02Z","lastTransitionTime":"2025-12-03T14:10:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 14:10:02 crc kubenswrapper[4805]: I1203 14:10:02.825423 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 14:10:02 crc kubenswrapper[4805]: I1203 14:10:02.825521 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 14:10:02 crc kubenswrapper[4805]: I1203 14:10:02.825545 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 14:10:02 crc kubenswrapper[4805]: I1203 14:10:02.825569 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 14:10:02 crc kubenswrapper[4805]: I1203 14:10:02.825587 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T14:10:02Z","lastTransitionTime":"2025-12-03T14:10:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 14:10:02 crc kubenswrapper[4805]: I1203 14:10:02.928439 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 14:10:02 crc kubenswrapper[4805]: I1203 14:10:02.928512 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 14:10:02 crc kubenswrapper[4805]: I1203 14:10:02.928530 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 14:10:02 crc kubenswrapper[4805]: I1203 14:10:02.928553 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 14:10:02 crc kubenswrapper[4805]: I1203 14:10:02.928571 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T14:10:02Z","lastTransitionTime":"2025-12-03T14:10:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 14:10:02 crc kubenswrapper[4805]: I1203 14:10:02.932894 4805 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-psjpm_deeec30d-cad2-4dfd-adb6-64f6646581be/ovnkube-controller/1.log" Dec 03 14:10:02 crc kubenswrapper[4805]: I1203 14:10:02.933658 4805 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-psjpm_deeec30d-cad2-4dfd-adb6-64f6646581be/ovnkube-controller/0.log" Dec 03 14:10:02 crc kubenswrapper[4805]: I1203 14:10:02.938075 4805 generic.go:334] "Generic (PLEG): container finished" podID="deeec30d-cad2-4dfd-adb6-64f6646581be" containerID="644e56c400524d739e2a208fb7eee1551ebbd28ba220df91c9dd84ba09d0b400" exitCode=1 Dec 03 14:10:02 crc kubenswrapper[4805]: I1203 14:10:02.938132 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-psjpm" event={"ID":"deeec30d-cad2-4dfd-adb6-64f6646581be","Type":"ContainerDied","Data":"644e56c400524d739e2a208fb7eee1551ebbd28ba220df91c9dd84ba09d0b400"} Dec 03 14:10:02 crc kubenswrapper[4805]: I1203 14:10:02.938172 4805 scope.go:117] "RemoveContainer" containerID="9e3aa4c113c750c2e2ead5d33f909a53247a221db998125f2494e7058d0c689a" Dec 03 14:10:02 crc kubenswrapper[4805]: I1203 14:10:02.940414 4805 scope.go:117] "RemoveContainer" containerID="644e56c400524d739e2a208fb7eee1551ebbd28ba220df91c9dd84ba09d0b400" Dec 03 14:10:02 crc kubenswrapper[4805]: E1203 14:10:02.940755 4805 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-psjpm_openshift-ovn-kubernetes(deeec30d-cad2-4dfd-adb6-64f6646581be)\"" pod="openshift-ovn-kubernetes/ovnkube-node-psjpm" podUID="deeec30d-cad2-4dfd-adb6-64f6646581be" Dec 03 14:10:02 crc kubenswrapper[4805]: I1203 14:10:02.955054 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-pdzgk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"860769de-6a5b-4cb9-a752-264e7e2e785f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5e2f0a0cc21505d609f81918365d5e4f5d21959ac12ab147d6a51601c9de4c14\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ggn4n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T14:09:50Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-pdzgk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T14:10:02Z is after 2025-08-24T17:21:41Z" Dec 03 14:10:02 crc kubenswrapper[4805]: I1203 14:10:02.974525 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:49Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:49Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://38339331a3209d6787d1def81b72f4a8514bf91080a139cad6d8ac5cc6bef10e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T14:10:02Z is after 2025-08-24T17:21:41Z" Dec 03 14:10:02 crc kubenswrapper[4805]: I1203 14:10:02.992415 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:48Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:48Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T14:10:02Z is after 2025-08-24T17:21:41Z" Dec 03 14:10:03 crc kubenswrapper[4805]: I1203 14:10:03.010684 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:51Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:51Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b3123d3fa879097db88ed2f3cfb05631f440c24c5405503b2de3546f12e2e899\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T14:10:03Z is after 2025-08-24T17:21:41Z" Dec 03 14:10:03 crc kubenswrapper[4805]: I1203 14:10:03.027049 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-bdtdg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"53a0e9e3-b6d8-44b7-a32b-aa3450442865\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f557e17ad105d6972d3c1480656c6073f67e28c0f5773aa365549951f4d161c7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9ww54\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T14:09:51Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-bdtdg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T14:10:03Z is after 2025-08-24T17:21:41Z" Dec 03 14:10:03 crc kubenswrapper[4805]: I1203 14:10:03.031603 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 14:10:03 crc kubenswrapper[4805]: I1203 14:10:03.031661 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 14:10:03 crc kubenswrapper[4805]: I1203 14:10:03.031675 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 14:10:03 crc kubenswrapper[4805]: I1203 14:10:03.031696 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 14:10:03 crc kubenswrapper[4805]: I1203 14:10:03.031713 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T14:10:03Z","lastTransitionTime":"2025-12-03T14:10:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 14:10:03 crc kubenswrapper[4805]: I1203 14:10:03.067247 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b6d36671-7173-4e9c-843f-005b4dcf576c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a837d7d491eaacbe4279b4014bb5fcbe9b514ba5cc82c708227bcd9b25a346e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6e5f04faf3b23197bb8ea11d3e574dc86f31e21d798734962001ec8c95de7c84\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2ff1a979387da41b00e95eb9cc8b2c9de966380a12fae687591df3c5b18a4497\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1e66e1cb18573a14e3c3ae16af4dbd9f8997a6ad38680c3ab4505646ebd48e7c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7d5b5c85e3de43e7e95111aa967248550607c26200c783251b68696f202f92a1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://474f72dac37a808fc4b9832306e76979b151d7221807ff3086013466ed1419c3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://474f72dac37a808fc4b9832306e76979b151d7221807ff3086013466ed1419c3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T14:09:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T14:09:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://74c8cea1a61c4775a0ad2f4ab4bb8a36437cf23d2ee953311d6ea967e297510a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://74c8cea1a61c4775a0ad2f4ab4bb8a36437cf23d2ee953311d6ea967e297510a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T14:09:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T14:09:31Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://a4f2f56a809adc01fba82eba0652d78f58770c132f77b1424f8b17068b1927e3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a4f2f56a809adc01fba82eba0652d78f58770c132f77b1424f8b17068b1927e3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T14:09:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T14:09:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T14:09:30Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T14:10:03Z is after 2025-08-24T17:21:41Z" Dec 03 14:10:03 crc kubenswrapper[4805]: I1203 14:10:03.106204 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ced8291d-0ce5-4e74-9eaf-2bf033774039\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:30Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:30Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3e5a34636f0a8c4d9179074596d28bc2eb7a504573398c596a637defd829907a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://29cd95be3f964415943c873e7914ffcea4abe50dfdf17f8a9cd03fc38df35981\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://951173b3d3d581275a6f6e43209b5e0611a701ca7a10f81504d723e24a2d21a3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e64730957bbbb2feb4d506e0ed9731a0594bf66fa1e8fcdb8ec944ded6308eed\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5da55f769e23ad52bac9c77da62c78e70fa34db8ed1a435e4eaf4dad3fd2b68a\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-03T14:09:48Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1203 14:09:42.992179 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1203 14:09:42.993233 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1830926909/tls.crt::/tmp/serving-cert-1830926909/tls.key\\\\\\\"\\\\nI1203 14:09:48.622403 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1203 14:09:48.626476 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1203 14:09:48.626527 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1203 14:09:48.626574 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1203 14:09:48.626582 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1203 14:09:48.633990 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1203 14:09:48.634024 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 14:09:48.634029 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 14:09:48.634034 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1203 14:09:48.634038 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1203 14:09:48.634042 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1203 14:09:48.634045 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1203 14:09:48.634344 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1203 14:09:48.644497 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T14:09:32Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bd92a160eb530c5687ebd7cd5b83b41820a2b808a0fd9514e5ace9d237e1bd26\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:32Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://af9c940135012c629447528c61743ed1e58708ab57ce76604c6da0479aa459cf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://af9c940135012c629447528c61743ed1e58708ab57ce76604c6da0479aa459cf\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T14:09:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T14:09:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T14:09:30Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T14:10:03Z is after 2025-08-24T17:21:41Z" Dec 03 14:10:03 crc kubenswrapper[4805]: I1203 14:10:03.120701 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-k4mhs" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c09c1cdc-f461-45c9-8444-5b8764bccabc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3354c1a44f891182d04116fc76234a2b337d50567d032a3584f902a9516387cc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qdcd7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T14:09:50Z\\\"}}\" for pod \"openshift-multus\"/\"multus-k4mhs\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T14:10:03Z is after 2025-08-24T17:21:41Z" Dec 03 14:10:03 crc kubenswrapper[4805]: I1203 14:10:03.133954 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 14:10:03 crc kubenswrapper[4805]: I1203 14:10:03.133995 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 14:10:03 crc kubenswrapper[4805]: I1203 14:10:03.134004 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 14:10:03 crc kubenswrapper[4805]: I1203 14:10:03.134018 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 14:10:03 crc kubenswrapper[4805]: I1203 14:10:03.134027 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T14:10:03Z","lastTransitionTime":"2025-12-03T14:10:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 14:10:03 crc kubenswrapper[4805]: I1203 14:10:03.138122 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-bpk49" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"de00a81f-82e1-4223-9614-f98606b9a8ef\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://04f4ad27e02e2024ae9e7cd1b16116267e69cecda882dd723cdb274c41a0e7db\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-25d7t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7741e0e70416ebe094bfabc6009e98d7d4c0fbcba1fc82df13b9ba40774b456a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7741e0e70416ebe094bfabc6009e98d7d4c0fbcba1fc82df13b9ba40774b456a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T14:09:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T14:09:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-25d7t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ee3499c02eb52bbf08c447d2433d435a74dfb70670dc4cbaac44f5fbff9e8a29\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ee3499c02eb52bbf08c447d2433d435a74dfb70670dc4cbaac44f5fbff9e8a29\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T14:09:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T14:09:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-25d7t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ec8488b1d69799951be7ef40e5e028732265c4ddcdea2d453923e567864771f7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ec8488b1d69799951be7ef40e5e028732265c4ddcdea2d453923e567864771f7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T14:09:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T14:09:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-25d7t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://66707bb4e1fdacb265e89786df8a04605830f1c156394cacfcec9b54abe5d41a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://66707bb4e1fdacb265e89786df8a04605830f1c156394cacfcec9b54abe5d41a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T14:09:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T14:09:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-25d7t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6207ab6600970ff15c8713c547d4920b2c77eb9bea990bb5da497e24870dcc0f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6207ab6600970ff15c8713c547d4920b2c77eb9bea990bb5da497e24870dcc0f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T14:09:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T14:09:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-25d7t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8de65d4a205e2deec0b433e53c0e37feb94aa0ede6a4c58d7414b80829804505\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8de65d4a205e2deec0b433e53c0e37feb94aa0ede6a4c58d7414b80829804505\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T14:09:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T14:09:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-25d7t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T14:09:50Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-bpk49\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T14:10:03Z is after 2025-08-24T17:21:41Z" Dec 03 14:10:03 crc kubenswrapper[4805]: I1203 14:10:03.152133 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"750d2b4b-738a-4992-8902-742528f6e875\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://abd2d0156113bb5182727a53aac594410ed051c19258beb8a07b808eff0b42a7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8673a29c0e16f4e707d79f0610722ccabb023ff2b4a18d7cf9d305183817d952\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://35766e7206b1be7a53d13d6890d759cf81dbd999643c12349268e4bf16fbbdb0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5b9844a11b2b903978d0d464b469e26c9211100da61d018dde9636dca95b684e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T14:09:30Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T14:10:03Z is after 2025-08-24T17:21:41Z" Dec 03 14:10:03 crc kubenswrapper[4805]: I1203 14:10:03.166345 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:48Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:48Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T14:10:03Z is after 2025-08-24T17:21:41Z" Dec 03 14:10:03 crc kubenswrapper[4805]: I1203 14:10:03.180628 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-gskh4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6098937f-e3f6-45e8-a647-4994a79cd711\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://28d5653f26ec61647808b3a733ff7eac70aa69ac184660f24147e8cfaf0eb802\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nwrw9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9ad6b68a42fd33ee167e67873c6eed3f11cdf6dc71d22e952c021436eb967dd9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nwrw9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T14:09:50Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-gskh4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T14:10:03Z is after 2025-08-24T17:21:41Z" Dec 03 14:10:03 crc kubenswrapper[4805]: I1203 14:10:03.185448 4805 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-69h75"] Dec 03 14:10:03 crc kubenswrapper[4805]: I1203 14:10:03.185898 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-69h75" Dec 03 14:10:03 crc kubenswrapper[4805]: I1203 14:10:03.187661 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-control-plane-metrics-cert" Dec 03 14:10:03 crc kubenswrapper[4805]: I1203 14:10:03.187879 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-kubernetes-control-plane-dockercfg-gs7dd" Dec 03 14:10:03 crc kubenswrapper[4805]: I1203 14:10:03.204100 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-psjpm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"deeec30d-cad2-4dfd-adb6-64f6646581be\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:50Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:50Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4b8d00cdf0f37f2e5cd4e952cf85d3de8d72c43fdf786088fd9f26ed17b1e21f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wdlzx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bd67c6de01a8d196a1f51d4d90640c9e69de6f05d5e201b25e6eaa5c33440ea8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wdlzx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cb8bf05ef093eb0727a066d4bbbdf9af9422867cd4dadacafc6eeb016867a61e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wdlzx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0304467102ed053628c0d3e06548c74b6492336b6fd044246508d1e8bdb17bc3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wdlzx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6ce80ccafb3262a11fbe9923317a9d38b61e5dc5b3e7e7ae136f143d6b643fd7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wdlzx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://47d76962beb8c0f196d50af96fb3ec611097823ca80a1c8bd718f2d18840128c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wdlzx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://644e56c400524d739e2a208fb7eee1551ebbd28ba220df91c9dd84ba09d0b400\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9e3aa4c113c750c2e2ead5d33f909a53247a221db998125f2494e7058d0c689a\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-03T14:10:00Z\\\",\\\"message\\\":\\\" 6125 reflector.go:311] Stopping reflector *v1.NetworkPolicy (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1203 14:10:00.108261 6125 reflector.go:311] Stopping reflector *v1.Pod (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1203 14:10:00.108416 6125 reflector.go:311] Stopping reflector *v1.Namespace (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1203 14:10:00.108932 6125 reflector.go:311] Stopping reflector *v1.EndpointSlice (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1203 14:10:00.108462 6125 reflector.go:311] Stopping reflector *v1.Node (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1203 14:10:00.109063 6125 reflector.go:311] Stopping reflector *v1.Service (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1203 14:10:00.109526 6125 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1203 14:10:00.109568 6125 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI1203 14:10:00.109575 6125 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI1203 14:10:00.109591 6125 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI1203 14:10:00.109612 6125 factory.go:656] Stopping watch factory\\\\nI1203 14:10:00.109628 6125 ovnkube.go:599] Stopped ovnkube\\\\nI1203 14:10:00.109650 6125 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1203 14\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T14:09:58Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://644e56c400524d739e2a208fb7eee1551ebbd28ba220df91c9dd84ba09d0b400\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-03T14:10:02Z\\\",\\\"message\\\":\\\"tring(nil)}, nodeEndpoints:map[string]services.lbEndpoints{}, externalTrafficLocal:false, internalTrafficLocal:false, hasNodePort:false}}\\\\nF1203 14:10:01.661807 6255 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T14:10:01Z is after 2025-08-24T17:21:41Z]\\\\nI1203 14:10:01.660569 6255 services_controller.go:451] Built service openshift-operator-lifecycle-manager/package-server-manager-metrics cluster-wide LB for network=default: []services.LB{services.LB{Name:\\\\\\\"Service_openshift-operator-lifecycle-manager/package-server-manager-metrics_TCP_cluster\\\\\\\", UUID:\\\\\\\"\\\\\\\", Protocol:\\\\\\\"TCP\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Serv\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T14:10:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wdlzx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ab98e365c1a4242a47b8cbe2dd85109d366d6002c4de1f18ddc950a46bea6c12\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wdlzx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2f875252fbe6a673a72fc9c56bb96899f62732305ad77b3108e3b244bfa91ecb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2f875252fbe6a673a72fc9c56bb96899f62732305ad77b3108e3b244bfa91ecb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T14:09:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T14:09:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wdlzx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T14:09:50Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-psjpm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T14:10:03Z is after 2025-08-24T17:21:41Z" Dec 03 14:10:03 crc kubenswrapper[4805]: I1203 14:10:03.217552 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:48Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:48Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T14:10:03Z is after 2025-08-24T17:21:41Z" Dec 03 14:10:03 crc kubenswrapper[4805]: I1203 14:10:03.232966 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:49Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:49Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://541ace7ccbb66dbd191c3ca8c1e4479e3edbf80dfc67326bea8aaee81fdcfc3c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://821e3e9a902aacecf6c41e8b8d8aa1ac951ce2fbb6505b7148b16348a3754e55\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T14:10:03Z is after 2025-08-24T17:21:41Z" Dec 03 14:10:03 crc kubenswrapper[4805]: I1203 14:10:03.236718 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 14:10:03 crc kubenswrapper[4805]: I1203 14:10:03.236746 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 14:10:03 crc kubenswrapper[4805]: I1203 14:10:03.236756 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 14:10:03 crc kubenswrapper[4805]: I1203 14:10:03.236772 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 14:10:03 crc kubenswrapper[4805]: I1203 14:10:03.236786 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T14:10:03Z","lastTransitionTime":"2025-12-03T14:10:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 14:10:03 crc kubenswrapper[4805]: I1203 14:10:03.248106 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"750d2b4b-738a-4992-8902-742528f6e875\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://abd2d0156113bb5182727a53aac594410ed051c19258beb8a07b808eff0b42a7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8673a29c0e16f4e707d79f0610722ccabb023ff2b4a18d7cf9d305183817d952\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://35766e7206b1be7a53d13d6890d759cf81dbd999643c12349268e4bf16fbbdb0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5b9844a11b2b903978d0d464b469e26c9211100da61d018dde9636dca95b684e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T14:09:30Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T14:10:03Z is after 2025-08-24T17:21:41Z" Dec 03 14:10:03 crc kubenswrapper[4805]: I1203 14:10:03.262456 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:48Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:48Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T14:10:03Z is after 2025-08-24T17:21:41Z" Dec 03 14:10:03 crc kubenswrapper[4805]: I1203 14:10:03.277363 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-k4mhs" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c09c1cdc-f461-45c9-8444-5b8764bccabc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3354c1a44f891182d04116fc76234a2b337d50567d032a3584f902a9516387cc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qdcd7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T14:09:50Z\\\"}}\" for pod \"openshift-multus\"/\"multus-k4mhs\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T14:10:03Z is after 2025-08-24T17:21:41Z" Dec 03 14:10:03 crc kubenswrapper[4805]: I1203 14:10:03.293953 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-bpk49" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"de00a81f-82e1-4223-9614-f98606b9a8ef\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://04f4ad27e02e2024ae9e7cd1b16116267e69cecda882dd723cdb274c41a0e7db\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-25d7t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7741e0e70416ebe094bfabc6009e98d7d4c0fbcba1fc82df13b9ba40774b456a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7741e0e70416ebe094bfabc6009e98d7d4c0fbcba1fc82df13b9ba40774b456a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T14:09:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T14:09:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-25d7t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ee3499c02eb52bbf08c447d2433d435a74dfb70670dc4cbaac44f5fbff9e8a29\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ee3499c02eb52bbf08c447d2433d435a74dfb70670dc4cbaac44f5fbff9e8a29\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T14:09:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T14:09:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-25d7t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ec8488b1d69799951be7ef40e5e028732265c4ddcdea2d453923e567864771f7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ec8488b1d69799951be7ef40e5e028732265c4ddcdea2d453923e567864771f7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T14:09:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T14:09:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-25d7t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://66707bb4e1fdacb265e89786df8a04605830f1c156394cacfcec9b54abe5d41a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://66707bb4e1fdacb265e89786df8a04605830f1c156394cacfcec9b54abe5d41a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T14:09:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T14:09:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-25d7t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6207ab6600970ff15c8713c547d4920b2c77eb9bea990bb5da497e24870dcc0f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6207ab6600970ff15c8713c547d4920b2c77eb9bea990bb5da497e24870dcc0f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T14:09:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T14:09:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-25d7t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8de65d4a205e2deec0b433e53c0e37feb94aa0ede6a4c58d7414b80829804505\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8de65d4a205e2deec0b433e53c0e37feb94aa0ede6a4c58d7414b80829804505\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T14:09:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T14:09:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-25d7t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T14:09:50Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-bpk49\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T14:10:03Z is after 2025-08-24T17:21:41Z" Dec 03 14:10:03 crc kubenswrapper[4805]: I1203 14:10:03.296380 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/5d1c694f-9847-4086-8b37-3573ddeb3ace-ovnkube-config\") pod \"ovnkube-control-plane-749d76644c-69h75\" (UID: \"5d1c694f-9847-4086-8b37-3573ddeb3ace\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-69h75" Dec 03 14:10:03 crc kubenswrapper[4805]: I1203 14:10:03.296410 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-s578b\" (UniqueName: \"kubernetes.io/projected/5d1c694f-9847-4086-8b37-3573ddeb3ace-kube-api-access-s578b\") pod \"ovnkube-control-plane-749d76644c-69h75\" (UID: \"5d1c694f-9847-4086-8b37-3573ddeb3ace\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-69h75" Dec 03 14:10:03 crc kubenswrapper[4805]: I1203 14:10:03.296436 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/5d1c694f-9847-4086-8b37-3573ddeb3ace-ovn-control-plane-metrics-cert\") pod \"ovnkube-control-plane-749d76644c-69h75\" (UID: \"5d1c694f-9847-4086-8b37-3573ddeb3ace\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-69h75" Dec 03 14:10:03 crc kubenswrapper[4805]: I1203 14:10:03.296476 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/5d1c694f-9847-4086-8b37-3573ddeb3ace-env-overrides\") pod \"ovnkube-control-plane-749d76644c-69h75\" (UID: \"5d1c694f-9847-4086-8b37-3573ddeb3ace\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-69h75" Dec 03 14:10:03 crc kubenswrapper[4805]: I1203 14:10:03.307277 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:48Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:48Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T14:10:03Z is after 2025-08-24T17:21:41Z" Dec 03 14:10:03 crc kubenswrapper[4805]: I1203 14:10:03.320989 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:49Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:49Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://541ace7ccbb66dbd191c3ca8c1e4479e3edbf80dfc67326bea8aaee81fdcfc3c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://821e3e9a902aacecf6c41e8b8d8aa1ac951ce2fbb6505b7148b16348a3754e55\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T14:10:03Z is after 2025-08-24T17:21:41Z" Dec 03 14:10:03 crc kubenswrapper[4805]: I1203 14:10:03.334900 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-gskh4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6098937f-e3f6-45e8-a647-4994a79cd711\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://28d5653f26ec61647808b3a733ff7eac70aa69ac184660f24147e8cfaf0eb802\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nwrw9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9ad6b68a42fd33ee167e67873c6eed3f11cdf6dc71d22e952c021436eb967dd9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nwrw9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T14:09:50Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-gskh4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T14:10:03Z is after 2025-08-24T17:21:41Z" Dec 03 14:10:03 crc kubenswrapper[4805]: I1203 14:10:03.339661 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 14:10:03 crc kubenswrapper[4805]: I1203 14:10:03.339685 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 14:10:03 crc kubenswrapper[4805]: I1203 14:10:03.339692 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 14:10:03 crc kubenswrapper[4805]: I1203 14:10:03.339705 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 14:10:03 crc kubenswrapper[4805]: I1203 14:10:03.339714 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T14:10:03Z","lastTransitionTime":"2025-12-03T14:10:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 14:10:03 crc kubenswrapper[4805]: I1203 14:10:03.357372 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-psjpm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"deeec30d-cad2-4dfd-adb6-64f6646581be\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:50Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:50Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4b8d00cdf0f37f2e5cd4e952cf85d3de8d72c43fdf786088fd9f26ed17b1e21f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wdlzx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bd67c6de01a8d196a1f51d4d90640c9e69de6f05d5e201b25e6eaa5c33440ea8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wdlzx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cb8bf05ef093eb0727a066d4bbbdf9af9422867cd4dadacafc6eeb016867a61e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wdlzx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0304467102ed053628c0d3e06548c74b6492336b6fd044246508d1e8bdb17bc3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wdlzx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6ce80ccafb3262a11fbe9923317a9d38b61e5dc5b3e7e7ae136f143d6b643fd7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wdlzx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://47d76962beb8c0f196d50af96fb3ec611097823ca80a1c8bd718f2d18840128c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wdlzx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://644e56c400524d739e2a208fb7eee1551ebbd28ba220df91c9dd84ba09d0b400\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9e3aa4c113c750c2e2ead5d33f909a53247a221db998125f2494e7058d0c689a\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-03T14:10:00Z\\\",\\\"message\\\":\\\" 6125 reflector.go:311] Stopping reflector *v1.NetworkPolicy (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1203 14:10:00.108261 6125 reflector.go:311] Stopping reflector *v1.Pod (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1203 14:10:00.108416 6125 reflector.go:311] Stopping reflector *v1.Namespace (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1203 14:10:00.108932 6125 reflector.go:311] Stopping reflector *v1.EndpointSlice (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1203 14:10:00.108462 6125 reflector.go:311] Stopping reflector *v1.Node (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1203 14:10:00.109063 6125 reflector.go:311] Stopping reflector *v1.Service (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1203 14:10:00.109526 6125 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1203 14:10:00.109568 6125 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI1203 14:10:00.109575 6125 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI1203 14:10:00.109591 6125 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI1203 14:10:00.109612 6125 factory.go:656] Stopping watch factory\\\\nI1203 14:10:00.109628 6125 ovnkube.go:599] Stopped ovnkube\\\\nI1203 14:10:00.109650 6125 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1203 14\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T14:09:58Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://644e56c400524d739e2a208fb7eee1551ebbd28ba220df91c9dd84ba09d0b400\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-03T14:10:02Z\\\",\\\"message\\\":\\\"tring(nil)}, nodeEndpoints:map[string]services.lbEndpoints{}, externalTrafficLocal:false, internalTrafficLocal:false, hasNodePort:false}}\\\\nF1203 14:10:01.661807 6255 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T14:10:01Z is after 2025-08-24T17:21:41Z]\\\\nI1203 14:10:01.660569 6255 services_controller.go:451] Built service openshift-operator-lifecycle-manager/package-server-manager-metrics cluster-wide LB for network=default: []services.LB{services.LB{Name:\\\\\\\"Service_openshift-operator-lifecycle-manager/package-server-manager-metrics_TCP_cluster\\\\\\\", UUID:\\\\\\\"\\\\\\\", Protocol:\\\\\\\"TCP\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Serv\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T14:10:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wdlzx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ab98e365c1a4242a47b8cbe2dd85109d366d6002c4de1f18ddc950a46bea6c12\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wdlzx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2f875252fbe6a673a72fc9c56bb96899f62732305ad77b3108e3b244bfa91ecb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2f875252fbe6a673a72fc9c56bb96899f62732305ad77b3108e3b244bfa91ecb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T14:09:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T14:09:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wdlzx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T14:09:50Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-psjpm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T14:10:03Z is after 2025-08-24T17:21:41Z" Dec 03 14:10:03 crc kubenswrapper[4805]: I1203 14:10:03.369043 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-pdzgk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"860769de-6a5b-4cb9-a752-264e7e2e785f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5e2f0a0cc21505d609f81918365d5e4f5d21959ac12ab147d6a51601c9de4c14\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ggn4n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T14:09:50Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-pdzgk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T14:10:03Z is after 2025-08-24T17:21:41Z" Dec 03 14:10:03 crc kubenswrapper[4805]: I1203 14:10:03.379588 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-69h75" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5d1c694f-9847-4086-8b37-3573ddeb3ace\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:10:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:10:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:10:03Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:10:03Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s578b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s578b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T14:10:03Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-69h75\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T14:10:03Z is after 2025-08-24T17:21:41Z" Dec 03 14:10:03 crc kubenswrapper[4805]: I1203 14:10:03.397099 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/5d1c694f-9847-4086-8b37-3573ddeb3ace-ovn-control-plane-metrics-cert\") pod \"ovnkube-control-plane-749d76644c-69h75\" (UID: \"5d1c694f-9847-4086-8b37-3573ddeb3ace\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-69h75" Dec 03 14:10:03 crc kubenswrapper[4805]: I1203 14:10:03.397154 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/5d1c694f-9847-4086-8b37-3573ddeb3ace-env-overrides\") pod \"ovnkube-control-plane-749d76644c-69h75\" (UID: \"5d1c694f-9847-4086-8b37-3573ddeb3ace\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-69h75" Dec 03 14:10:03 crc kubenswrapper[4805]: I1203 14:10:03.397188 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/5d1c694f-9847-4086-8b37-3573ddeb3ace-ovnkube-config\") pod \"ovnkube-control-plane-749d76644c-69h75\" (UID: \"5d1c694f-9847-4086-8b37-3573ddeb3ace\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-69h75" Dec 03 14:10:03 crc kubenswrapper[4805]: I1203 14:10:03.397206 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s578b\" (UniqueName: \"kubernetes.io/projected/5d1c694f-9847-4086-8b37-3573ddeb3ace-kube-api-access-s578b\") pod \"ovnkube-control-plane-749d76644c-69h75\" (UID: \"5d1c694f-9847-4086-8b37-3573ddeb3ace\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-69h75" Dec 03 14:10:03 crc kubenswrapper[4805]: I1203 14:10:03.398070 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/5d1c694f-9847-4086-8b37-3573ddeb3ace-env-overrides\") pod \"ovnkube-control-plane-749d76644c-69h75\" (UID: \"5d1c694f-9847-4086-8b37-3573ddeb3ace\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-69h75" Dec 03 14:10:03 crc kubenswrapper[4805]: I1203 14:10:03.398118 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/5d1c694f-9847-4086-8b37-3573ddeb3ace-ovnkube-config\") pod \"ovnkube-control-plane-749d76644c-69h75\" (UID: \"5d1c694f-9847-4086-8b37-3573ddeb3ace\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-69h75" Dec 03 14:10:03 crc kubenswrapper[4805]: I1203 14:10:03.402384 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b6d36671-7173-4e9c-843f-005b4dcf576c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a837d7d491eaacbe4279b4014bb5fcbe9b514ba5cc82c708227bcd9b25a346e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6e5f04faf3b23197bb8ea11d3e574dc86f31e21d798734962001ec8c95de7c84\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2ff1a979387da41b00e95eb9cc8b2c9de966380a12fae687591df3c5b18a4497\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1e66e1cb18573a14e3c3ae16af4dbd9f8997a6ad38680c3ab4505646ebd48e7c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7d5b5c85e3de43e7e95111aa967248550607c26200c783251b68696f202f92a1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://474f72dac37a808fc4b9832306e76979b151d7221807ff3086013466ed1419c3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://474f72dac37a808fc4b9832306e76979b151d7221807ff3086013466ed1419c3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T14:09:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T14:09:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://74c8cea1a61c4775a0ad2f4ab4bb8a36437cf23d2ee953311d6ea967e297510a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://74c8cea1a61c4775a0ad2f4ab4bb8a36437cf23d2ee953311d6ea967e297510a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T14:09:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T14:09:31Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://a4f2f56a809adc01fba82eba0652d78f58770c132f77b1424f8b17068b1927e3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a4f2f56a809adc01fba82eba0652d78f58770c132f77b1424f8b17068b1927e3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T14:09:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T14:09:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T14:09:30Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T14:10:03Z is after 2025-08-24T17:21:41Z" Dec 03 14:10:03 crc kubenswrapper[4805]: I1203 14:10:03.403599 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/5d1c694f-9847-4086-8b37-3573ddeb3ace-ovn-control-plane-metrics-cert\") pod \"ovnkube-control-plane-749d76644c-69h75\" (UID: \"5d1c694f-9847-4086-8b37-3573ddeb3ace\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-69h75" Dec 03 14:10:03 crc kubenswrapper[4805]: I1203 14:10:03.414628 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ced8291d-0ce5-4e74-9eaf-2bf033774039\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:30Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:30Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3e5a34636f0a8c4d9179074596d28bc2eb7a504573398c596a637defd829907a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://29cd95be3f964415943c873e7914ffcea4abe50dfdf17f8a9cd03fc38df35981\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://951173b3d3d581275a6f6e43209b5e0611a701ca7a10f81504d723e24a2d21a3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e64730957bbbb2feb4d506e0ed9731a0594bf66fa1e8fcdb8ec944ded6308eed\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5da55f769e23ad52bac9c77da62c78e70fa34db8ed1a435e4eaf4dad3fd2b68a\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-03T14:09:48Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1203 14:09:42.992179 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1203 14:09:42.993233 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1830926909/tls.crt::/tmp/serving-cert-1830926909/tls.key\\\\\\\"\\\\nI1203 14:09:48.622403 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1203 14:09:48.626476 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1203 14:09:48.626527 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1203 14:09:48.626574 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1203 14:09:48.626582 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1203 14:09:48.633990 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1203 14:09:48.634024 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 14:09:48.634029 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 14:09:48.634034 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1203 14:09:48.634038 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1203 14:09:48.634042 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1203 14:09:48.634045 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1203 14:09:48.634344 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1203 14:09:48.644497 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T14:09:32Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bd92a160eb530c5687ebd7cd5b83b41820a2b808a0fd9514e5ace9d237e1bd26\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:32Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://af9c940135012c629447528c61743ed1e58708ab57ce76604c6da0479aa459cf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://af9c940135012c629447528c61743ed1e58708ab57ce76604c6da0479aa459cf\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T14:09:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T14:09:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T14:09:30Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T14:10:03Z is after 2025-08-24T17:21:41Z" Dec 03 14:10:03 crc kubenswrapper[4805]: I1203 14:10:03.417870 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s578b\" (UniqueName: \"kubernetes.io/projected/5d1c694f-9847-4086-8b37-3573ddeb3ace-kube-api-access-s578b\") pod \"ovnkube-control-plane-749d76644c-69h75\" (UID: \"5d1c694f-9847-4086-8b37-3573ddeb3ace\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-69h75" Dec 03 14:10:03 crc kubenswrapper[4805]: I1203 14:10:03.429765 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:49Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:49Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://38339331a3209d6787d1def81b72f4a8514bf91080a139cad6d8ac5cc6bef10e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T14:10:03Z is after 2025-08-24T17:21:41Z" Dec 03 14:10:03 crc kubenswrapper[4805]: I1203 14:10:03.441856 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 14:10:03 crc kubenswrapper[4805]: I1203 14:10:03.442083 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 14:10:03 crc kubenswrapper[4805]: I1203 14:10:03.442149 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 14:10:03 crc kubenswrapper[4805]: I1203 14:10:03.442223 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 14:10:03 crc kubenswrapper[4805]: I1203 14:10:03.442288 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T14:10:03Z","lastTransitionTime":"2025-12-03T14:10:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 14:10:03 crc kubenswrapper[4805]: I1203 14:10:03.442444 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:48Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:48Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T14:10:03Z is after 2025-08-24T17:21:41Z" Dec 03 14:10:03 crc kubenswrapper[4805]: I1203 14:10:03.453615 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:51Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:51Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b3123d3fa879097db88ed2f3cfb05631f440c24c5405503b2de3546f12e2e899\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T14:10:03Z is after 2025-08-24T17:21:41Z" Dec 03 14:10:03 crc kubenswrapper[4805]: I1203 14:10:03.463135 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-bdtdg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"53a0e9e3-b6d8-44b7-a32b-aa3450442865\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f557e17ad105d6972d3c1480656c6073f67e28c0f5773aa365549951f4d161c7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9ww54\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T14:09:51Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-bdtdg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T14:10:03Z is after 2025-08-24T17:21:41Z" Dec 03 14:10:03 crc kubenswrapper[4805]: I1203 14:10:03.498620 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-69h75" Dec 03 14:10:03 crc kubenswrapper[4805]: W1203 14:10:03.516332 4805 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod5d1c694f_9847_4086_8b37_3573ddeb3ace.slice/crio-92602eff1b651070aab3e95cdc767ed1faf6617b4c53f2806be2f13e0cb42171 WatchSource:0}: Error finding container 92602eff1b651070aab3e95cdc767ed1faf6617b4c53f2806be2f13e0cb42171: Status 404 returned error can't find the container with id 92602eff1b651070aab3e95cdc767ed1faf6617b4c53f2806be2f13e0cb42171 Dec 03 14:10:03 crc kubenswrapper[4805]: I1203 14:10:03.544420 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 14:10:03 crc kubenswrapper[4805]: I1203 14:10:03.544459 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 14:10:03 crc kubenswrapper[4805]: I1203 14:10:03.544470 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 14:10:03 crc kubenswrapper[4805]: I1203 14:10:03.544485 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 14:10:03 crc kubenswrapper[4805]: I1203 14:10:03.544497 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T14:10:03Z","lastTransitionTime":"2025-12-03T14:10:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 14:10:03 crc kubenswrapper[4805]: I1203 14:10:03.647355 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 14:10:03 crc kubenswrapper[4805]: I1203 14:10:03.647403 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 14:10:03 crc kubenswrapper[4805]: I1203 14:10:03.647416 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 14:10:03 crc kubenswrapper[4805]: I1203 14:10:03.647435 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 14:10:03 crc kubenswrapper[4805]: I1203 14:10:03.647448 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T14:10:03Z","lastTransitionTime":"2025-12-03T14:10:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 14:10:03 crc kubenswrapper[4805]: I1203 14:10:03.749999 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 14:10:03 crc kubenswrapper[4805]: I1203 14:10:03.750031 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 14:10:03 crc kubenswrapper[4805]: I1203 14:10:03.750040 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 14:10:03 crc kubenswrapper[4805]: I1203 14:10:03.750056 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 14:10:03 crc kubenswrapper[4805]: I1203 14:10:03.750067 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T14:10:03Z","lastTransitionTime":"2025-12-03T14:10:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 14:10:03 crc kubenswrapper[4805]: I1203 14:10:03.852982 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 14:10:03 crc kubenswrapper[4805]: I1203 14:10:03.853017 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 14:10:03 crc kubenswrapper[4805]: I1203 14:10:03.853025 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 14:10:03 crc kubenswrapper[4805]: I1203 14:10:03.853039 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 14:10:03 crc kubenswrapper[4805]: I1203 14:10:03.853049 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T14:10:03Z","lastTransitionTime":"2025-12-03T14:10:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 14:10:03 crc kubenswrapper[4805]: I1203 14:10:03.947049 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-69h75" event={"ID":"5d1c694f-9847-4086-8b37-3573ddeb3ace","Type":"ContainerStarted","Data":"c7c4d434b035cac2a91a4f0846a2f4fb5bd68f881c28487860adb6db85e8ce75"} Dec 03 14:10:03 crc kubenswrapper[4805]: I1203 14:10:03.947608 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-69h75" event={"ID":"5d1c694f-9847-4086-8b37-3573ddeb3ace","Type":"ContainerStarted","Data":"abc1d9b8582ca9fe4a51fb608fb39129a179ddcdfde0797a6369a55b2e0124cf"} Dec 03 14:10:03 crc kubenswrapper[4805]: I1203 14:10:03.947682 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-69h75" event={"ID":"5d1c694f-9847-4086-8b37-3573ddeb3ace","Type":"ContainerStarted","Data":"92602eff1b651070aab3e95cdc767ed1faf6617b4c53f2806be2f13e0cb42171"} Dec 03 14:10:03 crc kubenswrapper[4805]: I1203 14:10:03.951519 4805 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-psjpm_deeec30d-cad2-4dfd-adb6-64f6646581be/ovnkube-controller/1.log" Dec 03 14:10:03 crc kubenswrapper[4805]: I1203 14:10:03.955660 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 14:10:03 crc kubenswrapper[4805]: I1203 14:10:03.955756 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 14:10:03 crc kubenswrapper[4805]: I1203 14:10:03.955823 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 14:10:03 crc kubenswrapper[4805]: I1203 14:10:03.955923 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 14:10:03 crc kubenswrapper[4805]: I1203 14:10:03.956014 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T14:10:03Z","lastTransitionTime":"2025-12-03T14:10:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 14:10:03 crc kubenswrapper[4805]: I1203 14:10:03.957811 4805 scope.go:117] "RemoveContainer" containerID="644e56c400524d739e2a208fb7eee1551ebbd28ba220df91c9dd84ba09d0b400" Dec 03 14:10:03 crc kubenswrapper[4805]: E1203 14:10:03.958023 4805 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-psjpm_openshift-ovn-kubernetes(deeec30d-cad2-4dfd-adb6-64f6646581be)\"" pod="openshift-ovn-kubernetes/ovnkube-node-psjpm" podUID="deeec30d-cad2-4dfd-adb6-64f6646581be" Dec 03 14:10:03 crc kubenswrapper[4805]: I1203 14:10:03.959999 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:48Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:48Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T14:10:03Z is after 2025-08-24T17:21:41Z" Dec 03 14:10:03 crc kubenswrapper[4805]: I1203 14:10:03.972143 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:51Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:51Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b3123d3fa879097db88ed2f3cfb05631f440c24c5405503b2de3546f12e2e899\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T14:10:03Z is after 2025-08-24T17:21:41Z" Dec 03 14:10:03 crc kubenswrapper[4805]: I1203 14:10:03.982037 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-bdtdg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"53a0e9e3-b6d8-44b7-a32b-aa3450442865\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f557e17ad105d6972d3c1480656c6073f67e28c0f5773aa365549951f4d161c7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9ww54\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T14:09:51Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-bdtdg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T14:10:03Z is after 2025-08-24T17:21:41Z" Dec 03 14:10:04 crc kubenswrapper[4805]: I1203 14:10:04.002925 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b6d36671-7173-4e9c-843f-005b4dcf576c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a837d7d491eaacbe4279b4014bb5fcbe9b514ba5cc82c708227bcd9b25a346e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6e5f04faf3b23197bb8ea11d3e574dc86f31e21d798734962001ec8c95de7c84\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2ff1a979387da41b00e95eb9cc8b2c9de966380a12fae687591df3c5b18a4497\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1e66e1cb18573a14e3c3ae16af4dbd9f8997a6ad38680c3ab4505646ebd48e7c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7d5b5c85e3de43e7e95111aa967248550607c26200c783251b68696f202f92a1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://474f72dac37a808fc4b9832306e76979b151d7221807ff3086013466ed1419c3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://474f72dac37a808fc4b9832306e76979b151d7221807ff3086013466ed1419c3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T14:09:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T14:09:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://74c8cea1a61c4775a0ad2f4ab4bb8a36437cf23d2ee953311d6ea967e297510a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://74c8cea1a61c4775a0ad2f4ab4bb8a36437cf23d2ee953311d6ea967e297510a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T14:09:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T14:09:31Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://a4f2f56a809adc01fba82eba0652d78f58770c132f77b1424f8b17068b1927e3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a4f2f56a809adc01fba82eba0652d78f58770c132f77b1424f8b17068b1927e3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T14:09:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T14:09:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T14:09:30Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T14:10:04Z is after 2025-08-24T17:21:41Z" Dec 03 14:10:04 crc kubenswrapper[4805]: I1203 14:10:04.022483 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ced8291d-0ce5-4e74-9eaf-2bf033774039\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:30Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:30Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3e5a34636f0a8c4d9179074596d28bc2eb7a504573398c596a637defd829907a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://29cd95be3f964415943c873e7914ffcea4abe50dfdf17f8a9cd03fc38df35981\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://951173b3d3d581275a6f6e43209b5e0611a701ca7a10f81504d723e24a2d21a3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e64730957bbbb2feb4d506e0ed9731a0594bf66fa1e8fcdb8ec944ded6308eed\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5da55f769e23ad52bac9c77da62c78e70fa34db8ed1a435e4eaf4dad3fd2b68a\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-03T14:09:48Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1203 14:09:42.992179 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1203 14:09:42.993233 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1830926909/tls.crt::/tmp/serving-cert-1830926909/tls.key\\\\\\\"\\\\nI1203 14:09:48.622403 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1203 14:09:48.626476 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1203 14:09:48.626527 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1203 14:09:48.626574 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1203 14:09:48.626582 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1203 14:09:48.633990 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1203 14:09:48.634024 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 14:09:48.634029 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 14:09:48.634034 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1203 14:09:48.634038 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1203 14:09:48.634042 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1203 14:09:48.634045 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1203 14:09:48.634344 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1203 14:09:48.644497 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T14:09:32Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bd92a160eb530c5687ebd7cd5b83b41820a2b808a0fd9514e5ace9d237e1bd26\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:32Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://af9c940135012c629447528c61743ed1e58708ab57ce76604c6da0479aa459cf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://af9c940135012c629447528c61743ed1e58708ab57ce76604c6da0479aa459cf\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T14:09:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T14:09:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T14:09:30Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T14:10:04Z is after 2025-08-24T17:21:41Z" Dec 03 14:10:04 crc kubenswrapper[4805]: I1203 14:10:04.040279 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:49Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:49Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://38339331a3209d6787d1def81b72f4a8514bf91080a139cad6d8ac5cc6bef10e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T14:10:04Z is after 2025-08-24T17:21:41Z" Dec 03 14:10:04 crc kubenswrapper[4805]: I1203 14:10:04.055265 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"750d2b4b-738a-4992-8902-742528f6e875\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://abd2d0156113bb5182727a53aac594410ed051c19258beb8a07b808eff0b42a7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8673a29c0e16f4e707d79f0610722ccabb023ff2b4a18d7cf9d305183817d952\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://35766e7206b1be7a53d13d6890d759cf81dbd999643c12349268e4bf16fbbdb0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5b9844a11b2b903978d0d464b469e26c9211100da61d018dde9636dca95b684e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T14:09:30Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T14:10:04Z is after 2025-08-24T17:21:41Z" Dec 03 14:10:04 crc kubenswrapper[4805]: I1203 14:10:04.058373 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 14:10:04 crc kubenswrapper[4805]: I1203 14:10:04.058470 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 14:10:04 crc kubenswrapper[4805]: I1203 14:10:04.058611 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 14:10:04 crc kubenswrapper[4805]: I1203 14:10:04.058697 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 14:10:04 crc kubenswrapper[4805]: I1203 14:10:04.058763 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T14:10:04Z","lastTransitionTime":"2025-12-03T14:10:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 14:10:04 crc kubenswrapper[4805]: I1203 14:10:04.075259 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:48Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:48Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T14:10:04Z is after 2025-08-24T17:21:41Z" Dec 03 14:10:04 crc kubenswrapper[4805]: I1203 14:10:04.088900 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-k4mhs" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c09c1cdc-f461-45c9-8444-5b8764bccabc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3354c1a44f891182d04116fc76234a2b337d50567d032a3584f902a9516387cc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qdcd7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T14:09:50Z\\\"}}\" for pod \"openshift-multus\"/\"multus-k4mhs\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T14:10:04Z is after 2025-08-24T17:21:41Z" Dec 03 14:10:04 crc kubenswrapper[4805]: I1203 14:10:04.102033 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-bpk49" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"de00a81f-82e1-4223-9614-f98606b9a8ef\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://04f4ad27e02e2024ae9e7cd1b16116267e69cecda882dd723cdb274c41a0e7db\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-25d7t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7741e0e70416ebe094bfabc6009e98d7d4c0fbcba1fc82df13b9ba40774b456a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7741e0e70416ebe094bfabc6009e98d7d4c0fbcba1fc82df13b9ba40774b456a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T14:09:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T14:09:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-25d7t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ee3499c02eb52bbf08c447d2433d435a74dfb70670dc4cbaac44f5fbff9e8a29\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ee3499c02eb52bbf08c447d2433d435a74dfb70670dc4cbaac44f5fbff9e8a29\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T14:09:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T14:09:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-25d7t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ec8488b1d69799951be7ef40e5e028732265c4ddcdea2d453923e567864771f7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ec8488b1d69799951be7ef40e5e028732265c4ddcdea2d453923e567864771f7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T14:09:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T14:09:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-25d7t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://66707bb4e1fdacb265e89786df8a04605830f1c156394cacfcec9b54abe5d41a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://66707bb4e1fdacb265e89786df8a04605830f1c156394cacfcec9b54abe5d41a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T14:09:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T14:09:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-25d7t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6207ab6600970ff15c8713c547d4920b2c77eb9bea990bb5da497e24870dcc0f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6207ab6600970ff15c8713c547d4920b2c77eb9bea990bb5da497e24870dcc0f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T14:09:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T14:09:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-25d7t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8de65d4a205e2deec0b433e53c0e37feb94aa0ede6a4c58d7414b80829804505\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8de65d4a205e2deec0b433e53c0e37feb94aa0ede6a4c58d7414b80829804505\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T14:09:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T14:09:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-25d7t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T14:09:50Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-bpk49\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T14:10:04Z is after 2025-08-24T17:21:41Z" Dec 03 14:10:04 crc kubenswrapper[4805]: I1203 14:10:04.115132 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:48Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:48Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T14:10:04Z is after 2025-08-24T17:21:41Z" Dec 03 14:10:04 crc kubenswrapper[4805]: I1203 14:10:04.128909 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:49Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:49Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://541ace7ccbb66dbd191c3ca8c1e4479e3edbf80dfc67326bea8aaee81fdcfc3c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://821e3e9a902aacecf6c41e8b8d8aa1ac951ce2fbb6505b7148b16348a3754e55\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T14:10:04Z is after 2025-08-24T17:21:41Z" Dec 03 14:10:04 crc kubenswrapper[4805]: I1203 14:10:04.140169 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-gskh4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6098937f-e3f6-45e8-a647-4994a79cd711\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://28d5653f26ec61647808b3a733ff7eac70aa69ac184660f24147e8cfaf0eb802\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nwrw9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9ad6b68a42fd33ee167e67873c6eed3f11cdf6dc71d22e952c021436eb967dd9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nwrw9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T14:09:50Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-gskh4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T14:10:04Z is after 2025-08-24T17:21:41Z" Dec 03 14:10:04 crc kubenswrapper[4805]: I1203 14:10:04.156617 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-psjpm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"deeec30d-cad2-4dfd-adb6-64f6646581be\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:50Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:50Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4b8d00cdf0f37f2e5cd4e952cf85d3de8d72c43fdf786088fd9f26ed17b1e21f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wdlzx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bd67c6de01a8d196a1f51d4d90640c9e69de6f05d5e201b25e6eaa5c33440ea8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wdlzx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cb8bf05ef093eb0727a066d4bbbdf9af9422867cd4dadacafc6eeb016867a61e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wdlzx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0304467102ed053628c0d3e06548c74b6492336b6fd044246508d1e8bdb17bc3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wdlzx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6ce80ccafb3262a11fbe9923317a9d38b61e5dc5b3e7e7ae136f143d6b643fd7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wdlzx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://47d76962beb8c0f196d50af96fb3ec611097823ca80a1c8bd718f2d18840128c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wdlzx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://644e56c400524d739e2a208fb7eee1551ebbd28ba220df91c9dd84ba09d0b400\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9e3aa4c113c750c2e2ead5d33f909a53247a221db998125f2494e7058d0c689a\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-03T14:10:00Z\\\",\\\"message\\\":\\\" 6125 reflector.go:311] Stopping reflector *v1.NetworkPolicy (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1203 14:10:00.108261 6125 reflector.go:311] Stopping reflector *v1.Pod (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1203 14:10:00.108416 6125 reflector.go:311] Stopping reflector *v1.Namespace (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1203 14:10:00.108932 6125 reflector.go:311] Stopping reflector *v1.EndpointSlice (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1203 14:10:00.108462 6125 reflector.go:311] Stopping reflector *v1.Node (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1203 14:10:00.109063 6125 reflector.go:311] Stopping reflector *v1.Service (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1203 14:10:00.109526 6125 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1203 14:10:00.109568 6125 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI1203 14:10:00.109575 6125 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI1203 14:10:00.109591 6125 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI1203 14:10:00.109612 6125 factory.go:656] Stopping watch factory\\\\nI1203 14:10:00.109628 6125 ovnkube.go:599] Stopped ovnkube\\\\nI1203 14:10:00.109650 6125 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1203 14\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T14:09:58Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://644e56c400524d739e2a208fb7eee1551ebbd28ba220df91c9dd84ba09d0b400\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-03T14:10:02Z\\\",\\\"message\\\":\\\"tring(nil)}, nodeEndpoints:map[string]services.lbEndpoints{}, externalTrafficLocal:false, internalTrafficLocal:false, hasNodePort:false}}\\\\nF1203 14:10:01.661807 6255 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T14:10:01Z is after 2025-08-24T17:21:41Z]\\\\nI1203 14:10:01.660569 6255 services_controller.go:451] Built service openshift-operator-lifecycle-manager/package-server-manager-metrics cluster-wide LB for network=default: []services.LB{services.LB{Name:\\\\\\\"Service_openshift-operator-lifecycle-manager/package-server-manager-metrics_TCP_cluster\\\\\\\", UUID:\\\\\\\"\\\\\\\", Protocol:\\\\\\\"TCP\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Serv\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T14:10:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wdlzx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ab98e365c1a4242a47b8cbe2dd85109d366d6002c4de1f18ddc950a46bea6c12\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wdlzx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2f875252fbe6a673a72fc9c56bb96899f62732305ad77b3108e3b244bfa91ecb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2f875252fbe6a673a72fc9c56bb96899f62732305ad77b3108e3b244bfa91ecb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T14:09:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T14:09:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wdlzx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T14:09:50Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-psjpm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T14:10:04Z is after 2025-08-24T17:21:41Z" Dec 03 14:10:04 crc kubenswrapper[4805]: I1203 14:10:04.160763 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 14:10:04 crc kubenswrapper[4805]: I1203 14:10:04.160892 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 14:10:04 crc kubenswrapper[4805]: I1203 14:10:04.160961 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 14:10:04 crc kubenswrapper[4805]: I1203 14:10:04.161031 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 14:10:04 crc kubenswrapper[4805]: I1203 14:10:04.161115 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T14:10:04Z","lastTransitionTime":"2025-12-03T14:10:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 14:10:04 crc kubenswrapper[4805]: I1203 14:10:04.166986 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-pdzgk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"860769de-6a5b-4cb9-a752-264e7e2e785f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5e2f0a0cc21505d609f81918365d5e4f5d21959ac12ab147d6a51601c9de4c14\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ggn4n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T14:09:50Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-pdzgk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T14:10:04Z is after 2025-08-24T17:21:41Z" Dec 03 14:10:04 crc kubenswrapper[4805]: I1203 14:10:04.178024 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-69h75" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5d1c694f-9847-4086-8b37-3573ddeb3ace\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:10:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:10:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:10:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:10:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://abc1d9b8582ca9fe4a51fb608fb39129a179ddcdfde0797a6369a55b2e0124cf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:10:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s578b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c7c4d434b035cac2a91a4f0846a2f4fb5bd68f881c28487860adb6db85e8ce75\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:10:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s578b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T14:10:03Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-69h75\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T14:10:04Z is after 2025-08-24T17:21:41Z" Dec 03 14:10:04 crc kubenswrapper[4805]: I1203 14:10:04.190440 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:48Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:48Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T14:10:04Z is after 2025-08-24T17:21:41Z" Dec 03 14:10:04 crc kubenswrapper[4805]: I1203 14:10:04.202285 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:49Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:49Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://541ace7ccbb66dbd191c3ca8c1e4479e3edbf80dfc67326bea8aaee81fdcfc3c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://821e3e9a902aacecf6c41e8b8d8aa1ac951ce2fbb6505b7148b16348a3754e55\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T14:10:04Z is after 2025-08-24T17:21:41Z" Dec 03 14:10:04 crc kubenswrapper[4805]: I1203 14:10:04.214388 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-gskh4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6098937f-e3f6-45e8-a647-4994a79cd711\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://28d5653f26ec61647808b3a733ff7eac70aa69ac184660f24147e8cfaf0eb802\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nwrw9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9ad6b68a42fd33ee167e67873c6eed3f11cdf6dc71d22e952c021436eb967dd9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nwrw9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T14:09:50Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-gskh4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T14:10:04Z is after 2025-08-24T17:21:41Z" Dec 03 14:10:04 crc kubenswrapper[4805]: I1203 14:10:04.243559 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-psjpm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"deeec30d-cad2-4dfd-adb6-64f6646581be\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:50Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:50Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4b8d00cdf0f37f2e5cd4e952cf85d3de8d72c43fdf786088fd9f26ed17b1e21f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wdlzx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bd67c6de01a8d196a1f51d4d90640c9e69de6f05d5e201b25e6eaa5c33440ea8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wdlzx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cb8bf05ef093eb0727a066d4bbbdf9af9422867cd4dadacafc6eeb016867a61e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wdlzx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0304467102ed053628c0d3e06548c74b6492336b6fd044246508d1e8bdb17bc3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wdlzx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6ce80ccafb3262a11fbe9923317a9d38b61e5dc5b3e7e7ae136f143d6b643fd7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wdlzx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://47d76962beb8c0f196d50af96fb3ec611097823ca80a1c8bd718f2d18840128c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wdlzx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://644e56c400524d739e2a208fb7eee1551ebbd28ba220df91c9dd84ba09d0b400\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://644e56c400524d739e2a208fb7eee1551ebbd28ba220df91c9dd84ba09d0b400\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-03T14:10:02Z\\\",\\\"message\\\":\\\"tring(nil)}, nodeEndpoints:map[string]services.lbEndpoints{}, externalTrafficLocal:false, internalTrafficLocal:false, hasNodePort:false}}\\\\nF1203 14:10:01.661807 6255 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T14:10:01Z is after 2025-08-24T17:21:41Z]\\\\nI1203 14:10:01.660569 6255 services_controller.go:451] Built service openshift-operator-lifecycle-manager/package-server-manager-metrics cluster-wide LB for network=default: []services.LB{services.LB{Name:\\\\\\\"Service_openshift-operator-lifecycle-manager/package-server-manager-metrics_TCP_cluster\\\\\\\", UUID:\\\\\\\"\\\\\\\", Protocol:\\\\\\\"TCP\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Serv\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T14:10:01Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-psjpm_openshift-ovn-kubernetes(deeec30d-cad2-4dfd-adb6-64f6646581be)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wdlzx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ab98e365c1a4242a47b8cbe2dd85109d366d6002c4de1f18ddc950a46bea6c12\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wdlzx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2f875252fbe6a673a72fc9c56bb96899f62732305ad77b3108e3b244bfa91ecb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2f875252fbe6a673a72fc9c56bb96899f62732305ad77b3108e3b244bfa91ecb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T14:09:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T14:09:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wdlzx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T14:09:50Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-psjpm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T14:10:04Z is after 2025-08-24T17:21:41Z" Dec 03 14:10:04 crc kubenswrapper[4805]: I1203 14:10:04.255824 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-pdzgk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"860769de-6a5b-4cb9-a752-264e7e2e785f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5e2f0a0cc21505d609f81918365d5e4f5d21959ac12ab147d6a51601c9de4c14\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ggn4n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T14:09:50Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-pdzgk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T14:10:04Z is after 2025-08-24T17:21:41Z" Dec 03 14:10:04 crc kubenswrapper[4805]: I1203 14:10:04.263156 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 14:10:04 crc kubenswrapper[4805]: I1203 14:10:04.263300 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 14:10:04 crc kubenswrapper[4805]: I1203 14:10:04.263374 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 14:10:04 crc kubenswrapper[4805]: I1203 14:10:04.263445 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 14:10:04 crc kubenswrapper[4805]: I1203 14:10:04.263510 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T14:10:04Z","lastTransitionTime":"2025-12-03T14:10:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 14:10:04 crc kubenswrapper[4805]: I1203 14:10:04.271322 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-69h75" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5d1c694f-9847-4086-8b37-3573ddeb3ace\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:10:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:10:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:10:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:10:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://abc1d9b8582ca9fe4a51fb608fb39129a179ddcdfde0797a6369a55b2e0124cf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:10:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s578b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c7c4d434b035cac2a91a4f0846a2f4fb5bd68f881c28487860adb6db85e8ce75\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:10:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s578b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T14:10:03Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-69h75\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T14:10:04Z is after 2025-08-24T17:21:41Z" Dec 03 14:10:04 crc kubenswrapper[4805]: I1203 14:10:04.283060 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-bdtdg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"53a0e9e3-b6d8-44b7-a32b-aa3450442865\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f557e17ad105d6972d3c1480656c6073f67e28c0f5773aa365549951f4d161c7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9ww54\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T14:09:51Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-bdtdg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T14:10:04Z is after 2025-08-24T17:21:41Z" Dec 03 14:10:04 crc kubenswrapper[4805]: I1203 14:10:04.305064 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b6d36671-7173-4e9c-843f-005b4dcf576c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a837d7d491eaacbe4279b4014bb5fcbe9b514ba5cc82c708227bcd9b25a346e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6e5f04faf3b23197bb8ea11d3e574dc86f31e21d798734962001ec8c95de7c84\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2ff1a979387da41b00e95eb9cc8b2c9de966380a12fae687591df3c5b18a4497\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1e66e1cb18573a14e3c3ae16af4dbd9f8997a6ad38680c3ab4505646ebd48e7c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7d5b5c85e3de43e7e95111aa967248550607c26200c783251b68696f202f92a1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://474f72dac37a808fc4b9832306e76979b151d7221807ff3086013466ed1419c3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://474f72dac37a808fc4b9832306e76979b151d7221807ff3086013466ed1419c3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T14:09:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T14:09:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://74c8cea1a61c4775a0ad2f4ab4bb8a36437cf23d2ee953311d6ea967e297510a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://74c8cea1a61c4775a0ad2f4ab4bb8a36437cf23d2ee953311d6ea967e297510a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T14:09:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T14:09:31Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://a4f2f56a809adc01fba82eba0652d78f58770c132f77b1424f8b17068b1927e3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a4f2f56a809adc01fba82eba0652d78f58770c132f77b1424f8b17068b1927e3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T14:09:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T14:09:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T14:09:30Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T14:10:04Z is after 2025-08-24T17:21:41Z" Dec 03 14:10:04 crc kubenswrapper[4805]: I1203 14:10:04.325155 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ced8291d-0ce5-4e74-9eaf-2bf033774039\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:30Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:30Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3e5a34636f0a8c4d9179074596d28bc2eb7a504573398c596a637defd829907a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://29cd95be3f964415943c873e7914ffcea4abe50dfdf17f8a9cd03fc38df35981\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://951173b3d3d581275a6f6e43209b5e0611a701ca7a10f81504d723e24a2d21a3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e64730957bbbb2feb4d506e0ed9731a0594bf66fa1e8fcdb8ec944ded6308eed\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5da55f769e23ad52bac9c77da62c78e70fa34db8ed1a435e4eaf4dad3fd2b68a\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-03T14:09:48Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1203 14:09:42.992179 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1203 14:09:42.993233 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1830926909/tls.crt::/tmp/serving-cert-1830926909/tls.key\\\\\\\"\\\\nI1203 14:09:48.622403 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1203 14:09:48.626476 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1203 14:09:48.626527 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1203 14:09:48.626574 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1203 14:09:48.626582 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1203 14:09:48.633990 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1203 14:09:48.634024 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 14:09:48.634029 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 14:09:48.634034 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1203 14:09:48.634038 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1203 14:09:48.634042 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1203 14:09:48.634045 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1203 14:09:48.634344 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1203 14:09:48.644497 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T14:09:32Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bd92a160eb530c5687ebd7cd5b83b41820a2b808a0fd9514e5ace9d237e1bd26\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:32Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://af9c940135012c629447528c61743ed1e58708ab57ce76604c6da0479aa459cf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://af9c940135012c629447528c61743ed1e58708ab57ce76604c6da0479aa459cf\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T14:09:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T14:09:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T14:09:30Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T14:10:04Z is after 2025-08-24T17:21:41Z" Dec 03 14:10:04 crc kubenswrapper[4805]: I1203 14:10:04.344706 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:49Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:49Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://38339331a3209d6787d1def81b72f4a8514bf91080a139cad6d8ac5cc6bef10e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T14:10:04Z is after 2025-08-24T17:21:41Z" Dec 03 14:10:04 crc kubenswrapper[4805]: I1203 14:10:04.362587 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:48Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:48Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T14:10:04Z is after 2025-08-24T17:21:41Z" Dec 03 14:10:04 crc kubenswrapper[4805]: I1203 14:10:04.366906 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 14:10:04 crc kubenswrapper[4805]: I1203 14:10:04.367076 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 14:10:04 crc kubenswrapper[4805]: I1203 14:10:04.367167 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 14:10:04 crc kubenswrapper[4805]: I1203 14:10:04.367256 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 14:10:04 crc kubenswrapper[4805]: I1203 14:10:04.367351 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T14:10:04Z","lastTransitionTime":"2025-12-03T14:10:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 14:10:04 crc kubenswrapper[4805]: I1203 14:10:04.381411 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:51Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:51Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b3123d3fa879097db88ed2f3cfb05631f440c24c5405503b2de3546f12e2e899\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T14:10:04Z is after 2025-08-24T17:21:41Z" Dec 03 14:10:04 crc kubenswrapper[4805]: I1203 14:10:04.397043 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"750d2b4b-738a-4992-8902-742528f6e875\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://abd2d0156113bb5182727a53aac594410ed051c19258beb8a07b808eff0b42a7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8673a29c0e16f4e707d79f0610722ccabb023ff2b4a18d7cf9d305183817d952\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://35766e7206b1be7a53d13d6890d759cf81dbd999643c12349268e4bf16fbbdb0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5b9844a11b2b903978d0d464b469e26c9211100da61d018dde9636dca95b684e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T14:09:30Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T14:10:04Z is after 2025-08-24T17:21:41Z" Dec 03 14:10:04 crc kubenswrapper[4805]: I1203 14:10:04.411322 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:48Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:48Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T14:10:04Z is after 2025-08-24T17:21:41Z" Dec 03 14:10:04 crc kubenswrapper[4805]: I1203 14:10:04.429908 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-k4mhs" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c09c1cdc-f461-45c9-8444-5b8764bccabc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3354c1a44f891182d04116fc76234a2b337d50567d032a3584f902a9516387cc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qdcd7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T14:09:50Z\\\"}}\" for pod \"openshift-multus\"/\"multus-k4mhs\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T14:10:04Z is after 2025-08-24T17:21:41Z" Dec 03 14:10:04 crc kubenswrapper[4805]: I1203 14:10:04.447516 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-bpk49" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"de00a81f-82e1-4223-9614-f98606b9a8ef\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://04f4ad27e02e2024ae9e7cd1b16116267e69cecda882dd723cdb274c41a0e7db\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-25d7t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7741e0e70416ebe094bfabc6009e98d7d4c0fbcba1fc82df13b9ba40774b456a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7741e0e70416ebe094bfabc6009e98d7d4c0fbcba1fc82df13b9ba40774b456a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T14:09:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T14:09:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-25d7t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ee3499c02eb52bbf08c447d2433d435a74dfb70670dc4cbaac44f5fbff9e8a29\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ee3499c02eb52bbf08c447d2433d435a74dfb70670dc4cbaac44f5fbff9e8a29\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T14:09:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T14:09:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-25d7t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ec8488b1d69799951be7ef40e5e028732265c4ddcdea2d453923e567864771f7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ec8488b1d69799951be7ef40e5e028732265c4ddcdea2d453923e567864771f7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T14:09:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T14:09:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-25d7t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://66707bb4e1fdacb265e89786df8a04605830f1c156394cacfcec9b54abe5d41a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://66707bb4e1fdacb265e89786df8a04605830f1c156394cacfcec9b54abe5d41a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T14:09:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T14:09:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-25d7t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6207ab6600970ff15c8713c547d4920b2c77eb9bea990bb5da497e24870dcc0f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6207ab6600970ff15c8713c547d4920b2c77eb9bea990bb5da497e24870dcc0f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T14:09:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T14:09:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-25d7t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8de65d4a205e2deec0b433e53c0e37feb94aa0ede6a4c58d7414b80829804505\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8de65d4a205e2deec0b433e53c0e37feb94aa0ede6a4c58d7414b80829804505\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T14:09:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T14:09:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-25d7t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T14:09:50Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-bpk49\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T14:10:04Z is after 2025-08-24T17:21:41Z" Dec 03 14:10:04 crc kubenswrapper[4805]: I1203 14:10:04.470111 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 14:10:04 crc kubenswrapper[4805]: I1203 14:10:04.470387 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 14:10:04 crc kubenswrapper[4805]: I1203 14:10:04.470494 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 14:10:04 crc kubenswrapper[4805]: I1203 14:10:04.470607 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 14:10:04 crc kubenswrapper[4805]: I1203 14:10:04.470700 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T14:10:04Z","lastTransitionTime":"2025-12-03T14:10:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 14:10:04 crc kubenswrapper[4805]: I1203 14:10:04.506950 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 14:10:04 crc kubenswrapper[4805]: E1203 14:10:04.507221 4805 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-03 14:10:20.507182778 +0000 UTC m=+50.170099741 (durationBeforeRetry 16s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 14:10:04 crc kubenswrapper[4805]: I1203 14:10:04.574060 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 14:10:04 crc kubenswrapper[4805]: I1203 14:10:04.574133 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 14:10:04 crc kubenswrapper[4805]: I1203 14:10:04.574158 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 14:10:04 crc kubenswrapper[4805]: I1203 14:10:04.574188 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 14:10:04 crc kubenswrapper[4805]: I1203 14:10:04.574211 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T14:10:04Z","lastTransitionTime":"2025-12-03T14:10:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 14:10:04 crc kubenswrapper[4805]: I1203 14:10:04.608982 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 14:10:04 crc kubenswrapper[4805]: I1203 14:10:04.609048 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 14:10:04 crc kubenswrapper[4805]: I1203 14:10:04.609105 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 14:10:04 crc kubenswrapper[4805]: I1203 14:10:04.609157 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 14:10:04 crc kubenswrapper[4805]: E1203 14:10:04.609245 4805 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 03 14:10:04 crc kubenswrapper[4805]: E1203 14:10:04.609284 4805 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 03 14:10:04 crc kubenswrapper[4805]: E1203 14:10:04.609298 4805 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 03 14:10:04 crc kubenswrapper[4805]: E1203 14:10:04.609327 4805 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 03 14:10:04 crc kubenswrapper[4805]: E1203 14:10:04.609355 4805 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 03 14:10:04 crc kubenswrapper[4805]: E1203 14:10:04.609381 4805 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 03 14:10:04 crc kubenswrapper[4805]: E1203 14:10:04.609471 4805 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Dec 03 14:10:04 crc kubenswrapper[4805]: E1203 14:10:04.609567 4805 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 03 14:10:04 crc kubenswrapper[4805]: E1203 14:10:04.609361 4805 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-12-03 14:10:20.609340141 +0000 UTC m=+50.272257074 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 03 14:10:04 crc kubenswrapper[4805]: E1203 14:10:04.609630 4805 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-12-03 14:10:20.609608139 +0000 UTC m=+50.272525112 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 03 14:10:04 crc kubenswrapper[4805]: E1203 14:10:04.609653 4805 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-03 14:10:20.60964142 +0000 UTC m=+50.272558383 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Dec 03 14:10:04 crc kubenswrapper[4805]: E1203 14:10:04.609675 4805 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-03 14:10:20.60966477 +0000 UTC m=+50.272581733 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 03 14:10:04 crc kubenswrapper[4805]: I1203 14:10:04.618261 4805 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/network-metrics-daemon-4gd9r"] Dec 03 14:10:04 crc kubenswrapper[4805]: I1203 14:10:04.619058 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-4gd9r" Dec 03 14:10:04 crc kubenswrapper[4805]: E1203 14:10:04.619162 4805 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-4gd9r" podUID="6b36a3db-1047-487a-9d6c-4f51ffdd8cc1" Dec 03 14:10:04 crc kubenswrapper[4805]: I1203 14:10:04.635336 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"750d2b4b-738a-4992-8902-742528f6e875\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://abd2d0156113bb5182727a53aac594410ed051c19258beb8a07b808eff0b42a7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8673a29c0e16f4e707d79f0610722ccabb023ff2b4a18d7cf9d305183817d952\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://35766e7206b1be7a53d13d6890d759cf81dbd999643c12349268e4bf16fbbdb0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5b9844a11b2b903978d0d464b469e26c9211100da61d018dde9636dca95b684e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T14:09:30Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T14:10:04Z is after 2025-08-24T17:21:41Z" Dec 03 14:10:04 crc kubenswrapper[4805]: I1203 14:10:04.654907 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:48Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:48Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T14:10:04Z is after 2025-08-24T17:21:41Z" Dec 03 14:10:04 crc kubenswrapper[4805]: I1203 14:10:04.671329 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-k4mhs" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c09c1cdc-f461-45c9-8444-5b8764bccabc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3354c1a44f891182d04116fc76234a2b337d50567d032a3584f902a9516387cc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qdcd7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T14:09:50Z\\\"}}\" for pod \"openshift-multus\"/\"multus-k4mhs\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T14:10:04Z is after 2025-08-24T17:21:41Z" Dec 03 14:10:04 crc kubenswrapper[4805]: I1203 14:10:04.676772 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 14:10:04 crc kubenswrapper[4805]: I1203 14:10:04.676873 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 14:10:04 crc kubenswrapper[4805]: I1203 14:10:04.676928 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 14:10:04 crc kubenswrapper[4805]: I1203 14:10:04.676960 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 14:10:04 crc kubenswrapper[4805]: I1203 14:10:04.676983 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T14:10:04Z","lastTransitionTime":"2025-12-03T14:10:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 14:10:04 crc kubenswrapper[4805]: I1203 14:10:04.690656 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-bpk49" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"de00a81f-82e1-4223-9614-f98606b9a8ef\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://04f4ad27e02e2024ae9e7cd1b16116267e69cecda882dd723cdb274c41a0e7db\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-25d7t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7741e0e70416ebe094bfabc6009e98d7d4c0fbcba1fc82df13b9ba40774b456a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7741e0e70416ebe094bfabc6009e98d7d4c0fbcba1fc82df13b9ba40774b456a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T14:09:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T14:09:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-25d7t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ee3499c02eb52bbf08c447d2433d435a74dfb70670dc4cbaac44f5fbff9e8a29\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ee3499c02eb52bbf08c447d2433d435a74dfb70670dc4cbaac44f5fbff9e8a29\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T14:09:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T14:09:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-25d7t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ec8488b1d69799951be7ef40e5e028732265c4ddcdea2d453923e567864771f7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ec8488b1d69799951be7ef40e5e028732265c4ddcdea2d453923e567864771f7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T14:09:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T14:09:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-25d7t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://66707bb4e1fdacb265e89786df8a04605830f1c156394cacfcec9b54abe5d41a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://66707bb4e1fdacb265e89786df8a04605830f1c156394cacfcec9b54abe5d41a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T14:09:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T14:09:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-25d7t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6207ab6600970ff15c8713c547d4920b2c77eb9bea990bb5da497e24870dcc0f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6207ab6600970ff15c8713c547d4920b2c77eb9bea990bb5da497e24870dcc0f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T14:09:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T14:09:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-25d7t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8de65d4a205e2deec0b433e53c0e37feb94aa0ede6a4c58d7414b80829804505\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8de65d4a205e2deec0b433e53c0e37feb94aa0ede6a4c58d7414b80829804505\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T14:09:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T14:09:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-25d7t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T14:09:50Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-bpk49\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T14:10:04Z is after 2025-08-24T17:21:41Z" Dec 03 14:10:04 crc kubenswrapper[4805]: I1203 14:10:04.693461 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 14:10:04 crc kubenswrapper[4805]: I1203 14:10:04.693492 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 14:10:04 crc kubenswrapper[4805]: I1203 14:10:04.693470 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 14:10:04 crc kubenswrapper[4805]: E1203 14:10:04.693642 4805 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 03 14:10:04 crc kubenswrapper[4805]: E1203 14:10:04.693709 4805 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 03 14:10:04 crc kubenswrapper[4805]: E1203 14:10:04.693878 4805 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 03 14:10:04 crc kubenswrapper[4805]: I1203 14:10:04.707953 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:48Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:48Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T14:10:04Z is after 2025-08-24T17:21:41Z" Dec 03 14:10:04 crc kubenswrapper[4805]: I1203 14:10:04.710050 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4t6zl\" (UniqueName: \"kubernetes.io/projected/6b36a3db-1047-487a-9d6c-4f51ffdd8cc1-kube-api-access-4t6zl\") pod \"network-metrics-daemon-4gd9r\" (UID: \"6b36a3db-1047-487a-9d6c-4f51ffdd8cc1\") " pod="openshift-multus/network-metrics-daemon-4gd9r" Dec 03 14:10:04 crc kubenswrapper[4805]: I1203 14:10:04.710245 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/6b36a3db-1047-487a-9d6c-4f51ffdd8cc1-metrics-certs\") pod \"network-metrics-daemon-4gd9r\" (UID: \"6b36a3db-1047-487a-9d6c-4f51ffdd8cc1\") " pod="openshift-multus/network-metrics-daemon-4gd9r" Dec 03 14:10:04 crc kubenswrapper[4805]: I1203 14:10:04.724334 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:49Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:49Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://541ace7ccbb66dbd191c3ca8c1e4479e3edbf80dfc67326bea8aaee81fdcfc3c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://821e3e9a902aacecf6c41e8b8d8aa1ac951ce2fbb6505b7148b16348a3754e55\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T14:10:04Z is after 2025-08-24T17:21:41Z" Dec 03 14:10:04 crc kubenswrapper[4805]: I1203 14:10:04.737449 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-gskh4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6098937f-e3f6-45e8-a647-4994a79cd711\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://28d5653f26ec61647808b3a733ff7eac70aa69ac184660f24147e8cfaf0eb802\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nwrw9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9ad6b68a42fd33ee167e67873c6eed3f11cdf6dc71d22e952c021436eb967dd9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nwrw9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T14:09:50Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-gskh4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T14:10:04Z is after 2025-08-24T17:21:41Z" Dec 03 14:10:04 crc kubenswrapper[4805]: I1203 14:10:04.761173 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-psjpm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"deeec30d-cad2-4dfd-adb6-64f6646581be\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:50Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:50Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4b8d00cdf0f37f2e5cd4e952cf85d3de8d72c43fdf786088fd9f26ed17b1e21f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wdlzx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bd67c6de01a8d196a1f51d4d90640c9e69de6f05d5e201b25e6eaa5c33440ea8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wdlzx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cb8bf05ef093eb0727a066d4bbbdf9af9422867cd4dadacafc6eeb016867a61e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wdlzx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0304467102ed053628c0d3e06548c74b6492336b6fd044246508d1e8bdb17bc3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wdlzx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6ce80ccafb3262a11fbe9923317a9d38b61e5dc5b3e7e7ae136f143d6b643fd7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wdlzx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://47d76962beb8c0f196d50af96fb3ec611097823ca80a1c8bd718f2d18840128c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wdlzx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://644e56c400524d739e2a208fb7eee1551ebbd28ba220df91c9dd84ba09d0b400\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://644e56c400524d739e2a208fb7eee1551ebbd28ba220df91c9dd84ba09d0b400\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-03T14:10:02Z\\\",\\\"message\\\":\\\"tring(nil)}, nodeEndpoints:map[string]services.lbEndpoints{}, externalTrafficLocal:false, internalTrafficLocal:false, hasNodePort:false}}\\\\nF1203 14:10:01.661807 6255 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T14:10:01Z is after 2025-08-24T17:21:41Z]\\\\nI1203 14:10:01.660569 6255 services_controller.go:451] Built service openshift-operator-lifecycle-manager/package-server-manager-metrics cluster-wide LB for network=default: []services.LB{services.LB{Name:\\\\\\\"Service_openshift-operator-lifecycle-manager/package-server-manager-metrics_TCP_cluster\\\\\\\", UUID:\\\\\\\"\\\\\\\", Protocol:\\\\\\\"TCP\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Serv\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T14:10:01Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-psjpm_openshift-ovn-kubernetes(deeec30d-cad2-4dfd-adb6-64f6646581be)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wdlzx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ab98e365c1a4242a47b8cbe2dd85109d366d6002c4de1f18ddc950a46bea6c12\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wdlzx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2f875252fbe6a673a72fc9c56bb96899f62732305ad77b3108e3b244bfa91ecb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2f875252fbe6a673a72fc9c56bb96899f62732305ad77b3108e3b244bfa91ecb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T14:09:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T14:09:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wdlzx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T14:09:50Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-psjpm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T14:10:04Z is after 2025-08-24T17:21:41Z" Dec 03 14:10:04 crc kubenswrapper[4805]: I1203 14:10:04.774626 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-4gd9r" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6b36a3db-1047-487a-9d6c-4f51ffdd8cc1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:10:04Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:10:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:10:04Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:10:04Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4t6zl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4t6zl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T14:10:04Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-4gd9r\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T14:10:04Z is after 2025-08-24T17:21:41Z" Dec 03 14:10:04 crc kubenswrapper[4805]: I1203 14:10:04.779359 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 14:10:04 crc kubenswrapper[4805]: I1203 14:10:04.779564 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 14:10:04 crc kubenswrapper[4805]: I1203 14:10:04.779647 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 14:10:04 crc kubenswrapper[4805]: I1203 14:10:04.779718 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 14:10:04 crc kubenswrapper[4805]: I1203 14:10:04.779783 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T14:10:04Z","lastTransitionTime":"2025-12-03T14:10:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 14:10:04 crc kubenswrapper[4805]: I1203 14:10:04.791519 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-pdzgk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"860769de-6a5b-4cb9-a752-264e7e2e785f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5e2f0a0cc21505d609f81918365d5e4f5d21959ac12ab147d6a51601c9de4c14\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ggn4n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T14:09:50Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-pdzgk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T14:10:04Z is after 2025-08-24T17:21:41Z" Dec 03 14:10:04 crc kubenswrapper[4805]: I1203 14:10:04.806995 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-69h75" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5d1c694f-9847-4086-8b37-3573ddeb3ace\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:10:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:10:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:10:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:10:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://abc1d9b8582ca9fe4a51fb608fb39129a179ddcdfde0797a6369a55b2e0124cf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:10:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s578b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c7c4d434b035cac2a91a4f0846a2f4fb5bd68f881c28487860adb6db85e8ce75\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:10:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s578b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T14:10:03Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-69h75\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T14:10:04Z is after 2025-08-24T17:21:41Z" Dec 03 14:10:04 crc kubenswrapper[4805]: I1203 14:10:04.810882 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4t6zl\" (UniqueName: \"kubernetes.io/projected/6b36a3db-1047-487a-9d6c-4f51ffdd8cc1-kube-api-access-4t6zl\") pod \"network-metrics-daemon-4gd9r\" (UID: \"6b36a3db-1047-487a-9d6c-4f51ffdd8cc1\") " pod="openshift-multus/network-metrics-daemon-4gd9r" Dec 03 14:10:04 crc kubenswrapper[4805]: I1203 14:10:04.810928 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/6b36a3db-1047-487a-9d6c-4f51ffdd8cc1-metrics-certs\") pod \"network-metrics-daemon-4gd9r\" (UID: \"6b36a3db-1047-487a-9d6c-4f51ffdd8cc1\") " pod="openshift-multus/network-metrics-daemon-4gd9r" Dec 03 14:10:04 crc kubenswrapper[4805]: E1203 14:10:04.811069 4805 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Dec 03 14:10:04 crc kubenswrapper[4805]: E1203 14:10:04.811127 4805 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/6b36a3db-1047-487a-9d6c-4f51ffdd8cc1-metrics-certs podName:6b36a3db-1047-487a-9d6c-4f51ffdd8cc1 nodeName:}" failed. No retries permitted until 2025-12-03 14:10:05.311111238 +0000 UTC m=+34.974028181 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/6b36a3db-1047-487a-9d6c-4f51ffdd8cc1-metrics-certs") pod "network-metrics-daemon-4gd9r" (UID: "6b36a3db-1047-487a-9d6c-4f51ffdd8cc1") : object "openshift-multus"/"metrics-daemon-secret" not registered Dec 03 14:10:04 crc kubenswrapper[4805]: I1203 14:10:04.819813 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-bdtdg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"53a0e9e3-b6d8-44b7-a32b-aa3450442865\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f557e17ad105d6972d3c1480656c6073f67e28c0f5773aa365549951f4d161c7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9ww54\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T14:09:51Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-bdtdg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T14:10:04Z is after 2025-08-24T17:21:41Z" Dec 03 14:10:04 crc kubenswrapper[4805]: I1203 14:10:04.828877 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4t6zl\" (UniqueName: \"kubernetes.io/projected/6b36a3db-1047-487a-9d6c-4f51ffdd8cc1-kube-api-access-4t6zl\") pod \"network-metrics-daemon-4gd9r\" (UID: \"6b36a3db-1047-487a-9d6c-4f51ffdd8cc1\") " pod="openshift-multus/network-metrics-daemon-4gd9r" Dec 03 14:10:04 crc kubenswrapper[4805]: I1203 14:10:04.840085 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b6d36671-7173-4e9c-843f-005b4dcf576c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a837d7d491eaacbe4279b4014bb5fcbe9b514ba5cc82c708227bcd9b25a346e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6e5f04faf3b23197bb8ea11d3e574dc86f31e21d798734962001ec8c95de7c84\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2ff1a979387da41b00e95eb9cc8b2c9de966380a12fae687591df3c5b18a4497\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1e66e1cb18573a14e3c3ae16af4dbd9f8997a6ad38680c3ab4505646ebd48e7c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7d5b5c85e3de43e7e95111aa967248550607c26200c783251b68696f202f92a1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://474f72dac37a808fc4b9832306e76979b151d7221807ff3086013466ed1419c3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://474f72dac37a808fc4b9832306e76979b151d7221807ff3086013466ed1419c3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T14:09:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T14:09:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://74c8cea1a61c4775a0ad2f4ab4bb8a36437cf23d2ee953311d6ea967e297510a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://74c8cea1a61c4775a0ad2f4ab4bb8a36437cf23d2ee953311d6ea967e297510a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T14:09:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T14:09:31Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://a4f2f56a809adc01fba82eba0652d78f58770c132f77b1424f8b17068b1927e3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a4f2f56a809adc01fba82eba0652d78f58770c132f77b1424f8b17068b1927e3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T14:09:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T14:09:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T14:09:30Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T14:10:04Z is after 2025-08-24T17:21:41Z" Dec 03 14:10:04 crc kubenswrapper[4805]: I1203 14:10:04.859483 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ced8291d-0ce5-4e74-9eaf-2bf033774039\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:30Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:30Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3e5a34636f0a8c4d9179074596d28bc2eb7a504573398c596a637defd829907a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://29cd95be3f964415943c873e7914ffcea4abe50dfdf17f8a9cd03fc38df35981\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://951173b3d3d581275a6f6e43209b5e0611a701ca7a10f81504d723e24a2d21a3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e64730957bbbb2feb4d506e0ed9731a0594bf66fa1e8fcdb8ec944ded6308eed\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5da55f769e23ad52bac9c77da62c78e70fa34db8ed1a435e4eaf4dad3fd2b68a\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-03T14:09:48Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1203 14:09:42.992179 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1203 14:09:42.993233 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1830926909/tls.crt::/tmp/serving-cert-1830926909/tls.key\\\\\\\"\\\\nI1203 14:09:48.622403 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1203 14:09:48.626476 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1203 14:09:48.626527 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1203 14:09:48.626574 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1203 14:09:48.626582 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1203 14:09:48.633990 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1203 14:09:48.634024 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 14:09:48.634029 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 14:09:48.634034 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1203 14:09:48.634038 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1203 14:09:48.634042 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1203 14:09:48.634045 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1203 14:09:48.634344 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1203 14:09:48.644497 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T14:09:32Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bd92a160eb530c5687ebd7cd5b83b41820a2b808a0fd9514e5ace9d237e1bd26\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:32Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://af9c940135012c629447528c61743ed1e58708ab57ce76604c6da0479aa459cf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://af9c940135012c629447528c61743ed1e58708ab57ce76604c6da0479aa459cf\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T14:09:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T14:09:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T14:09:30Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T14:10:04Z is after 2025-08-24T17:21:41Z" Dec 03 14:10:04 crc kubenswrapper[4805]: I1203 14:10:04.875954 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:49Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:49Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://38339331a3209d6787d1def81b72f4a8514bf91080a139cad6d8ac5cc6bef10e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T14:10:04Z is after 2025-08-24T17:21:41Z" Dec 03 14:10:04 crc kubenswrapper[4805]: I1203 14:10:04.883458 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 14:10:04 crc kubenswrapper[4805]: I1203 14:10:04.883521 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 14:10:04 crc kubenswrapper[4805]: I1203 14:10:04.883533 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 14:10:04 crc kubenswrapper[4805]: I1203 14:10:04.883577 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 14:10:04 crc kubenswrapper[4805]: I1203 14:10:04.883595 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T14:10:04Z","lastTransitionTime":"2025-12-03T14:10:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 14:10:04 crc kubenswrapper[4805]: I1203 14:10:04.892214 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:48Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:48Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T14:10:04Z is after 2025-08-24T17:21:41Z" Dec 03 14:10:04 crc kubenswrapper[4805]: I1203 14:10:04.911096 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:51Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:51Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b3123d3fa879097db88ed2f3cfb05631f440c24c5405503b2de3546f12e2e899\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T14:10:04Z is after 2025-08-24T17:21:41Z" Dec 03 14:10:04 crc kubenswrapper[4805]: I1203 14:10:04.986527 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 14:10:04 crc kubenswrapper[4805]: I1203 14:10:04.986586 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 14:10:04 crc kubenswrapper[4805]: I1203 14:10:04.986602 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 14:10:04 crc kubenswrapper[4805]: I1203 14:10:04.986625 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 14:10:04 crc kubenswrapper[4805]: I1203 14:10:04.986643 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T14:10:04Z","lastTransitionTime":"2025-12-03T14:10:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 14:10:05 crc kubenswrapper[4805]: I1203 14:10:05.089315 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 14:10:05 crc kubenswrapper[4805]: I1203 14:10:05.089368 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 14:10:05 crc kubenswrapper[4805]: I1203 14:10:05.089388 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 14:10:05 crc kubenswrapper[4805]: I1203 14:10:05.089410 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 14:10:05 crc kubenswrapper[4805]: I1203 14:10:05.089427 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T14:10:05Z","lastTransitionTime":"2025-12-03T14:10:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 14:10:05 crc kubenswrapper[4805]: I1203 14:10:05.193262 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 14:10:05 crc kubenswrapper[4805]: I1203 14:10:05.193336 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 14:10:05 crc kubenswrapper[4805]: I1203 14:10:05.193359 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 14:10:05 crc kubenswrapper[4805]: I1203 14:10:05.193388 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 14:10:05 crc kubenswrapper[4805]: I1203 14:10:05.193427 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T14:10:05Z","lastTransitionTime":"2025-12-03T14:10:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 14:10:05 crc kubenswrapper[4805]: I1203 14:10:05.295927 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 14:10:05 crc kubenswrapper[4805]: I1203 14:10:05.295967 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 14:10:05 crc kubenswrapper[4805]: I1203 14:10:05.295979 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 14:10:05 crc kubenswrapper[4805]: I1203 14:10:05.295996 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 14:10:05 crc kubenswrapper[4805]: I1203 14:10:05.296008 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T14:10:05Z","lastTransitionTime":"2025-12-03T14:10:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 14:10:05 crc kubenswrapper[4805]: I1203 14:10:05.316023 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/6b36a3db-1047-487a-9d6c-4f51ffdd8cc1-metrics-certs\") pod \"network-metrics-daemon-4gd9r\" (UID: \"6b36a3db-1047-487a-9d6c-4f51ffdd8cc1\") " pod="openshift-multus/network-metrics-daemon-4gd9r" Dec 03 14:10:05 crc kubenswrapper[4805]: E1203 14:10:05.316215 4805 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Dec 03 14:10:05 crc kubenswrapper[4805]: E1203 14:10:05.316322 4805 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/6b36a3db-1047-487a-9d6c-4f51ffdd8cc1-metrics-certs podName:6b36a3db-1047-487a-9d6c-4f51ffdd8cc1 nodeName:}" failed. No retries permitted until 2025-12-03 14:10:06.316295229 +0000 UTC m=+35.979212192 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/6b36a3db-1047-487a-9d6c-4f51ffdd8cc1-metrics-certs") pod "network-metrics-daemon-4gd9r" (UID: "6b36a3db-1047-487a-9d6c-4f51ffdd8cc1") : object "openshift-multus"/"metrics-daemon-secret" not registered Dec 03 14:10:05 crc kubenswrapper[4805]: I1203 14:10:05.398113 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 14:10:05 crc kubenswrapper[4805]: I1203 14:10:05.398169 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 14:10:05 crc kubenswrapper[4805]: I1203 14:10:05.398184 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 14:10:05 crc kubenswrapper[4805]: I1203 14:10:05.398205 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 14:10:05 crc kubenswrapper[4805]: I1203 14:10:05.398220 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T14:10:05Z","lastTransitionTime":"2025-12-03T14:10:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 14:10:05 crc kubenswrapper[4805]: I1203 14:10:05.501769 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 14:10:05 crc kubenswrapper[4805]: I1203 14:10:05.501874 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 14:10:05 crc kubenswrapper[4805]: I1203 14:10:05.501894 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 14:10:05 crc kubenswrapper[4805]: I1203 14:10:05.501920 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 14:10:05 crc kubenswrapper[4805]: I1203 14:10:05.501935 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T14:10:05Z","lastTransitionTime":"2025-12-03T14:10:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 14:10:05 crc kubenswrapper[4805]: I1203 14:10:05.604663 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 14:10:05 crc kubenswrapper[4805]: I1203 14:10:05.604749 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 14:10:05 crc kubenswrapper[4805]: I1203 14:10:05.604766 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 14:10:05 crc kubenswrapper[4805]: I1203 14:10:05.604788 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 14:10:05 crc kubenswrapper[4805]: I1203 14:10:05.604804 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T14:10:05Z","lastTransitionTime":"2025-12-03T14:10:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 14:10:05 crc kubenswrapper[4805]: I1203 14:10:05.707935 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 14:10:05 crc kubenswrapper[4805]: I1203 14:10:05.707989 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 14:10:05 crc kubenswrapper[4805]: I1203 14:10:05.708012 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 14:10:05 crc kubenswrapper[4805]: I1203 14:10:05.708040 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 14:10:05 crc kubenswrapper[4805]: I1203 14:10:05.708062 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T14:10:05Z","lastTransitionTime":"2025-12-03T14:10:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 14:10:05 crc kubenswrapper[4805]: I1203 14:10:05.810770 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 14:10:05 crc kubenswrapper[4805]: I1203 14:10:05.810806 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 14:10:05 crc kubenswrapper[4805]: I1203 14:10:05.810815 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 14:10:05 crc kubenswrapper[4805]: I1203 14:10:05.810828 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 14:10:05 crc kubenswrapper[4805]: I1203 14:10:05.810855 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T14:10:05Z","lastTransitionTime":"2025-12-03T14:10:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 14:10:05 crc kubenswrapper[4805]: I1203 14:10:05.913677 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 14:10:05 crc kubenswrapper[4805]: I1203 14:10:05.913746 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 14:10:05 crc kubenswrapper[4805]: I1203 14:10:05.913771 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 14:10:05 crc kubenswrapper[4805]: I1203 14:10:05.913799 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 14:10:05 crc kubenswrapper[4805]: I1203 14:10:05.913821 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T14:10:05Z","lastTransitionTime":"2025-12-03T14:10:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 14:10:06 crc kubenswrapper[4805]: I1203 14:10:06.017242 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 14:10:06 crc kubenswrapper[4805]: I1203 14:10:06.017295 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 14:10:06 crc kubenswrapper[4805]: I1203 14:10:06.017308 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 14:10:06 crc kubenswrapper[4805]: I1203 14:10:06.017326 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 14:10:06 crc kubenswrapper[4805]: I1203 14:10:06.017340 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T14:10:06Z","lastTransitionTime":"2025-12-03T14:10:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 14:10:06 crc kubenswrapper[4805]: I1203 14:10:06.119742 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 14:10:06 crc kubenswrapper[4805]: I1203 14:10:06.119791 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 14:10:06 crc kubenswrapper[4805]: I1203 14:10:06.119808 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 14:10:06 crc kubenswrapper[4805]: I1203 14:10:06.119833 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 14:10:06 crc kubenswrapper[4805]: I1203 14:10:06.119894 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T14:10:06Z","lastTransitionTime":"2025-12-03T14:10:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 14:10:06 crc kubenswrapper[4805]: I1203 14:10:06.222707 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 14:10:06 crc kubenswrapper[4805]: I1203 14:10:06.222764 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 14:10:06 crc kubenswrapper[4805]: I1203 14:10:06.222784 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 14:10:06 crc kubenswrapper[4805]: I1203 14:10:06.222806 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 14:10:06 crc kubenswrapper[4805]: I1203 14:10:06.222822 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T14:10:06Z","lastTransitionTime":"2025-12-03T14:10:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 14:10:06 crc kubenswrapper[4805]: I1203 14:10:06.325665 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 14:10:06 crc kubenswrapper[4805]: I1203 14:10:06.326224 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 14:10:06 crc kubenswrapper[4805]: I1203 14:10:06.326365 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 14:10:06 crc kubenswrapper[4805]: I1203 14:10:06.326493 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 14:10:06 crc kubenswrapper[4805]: I1203 14:10:06.326605 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T14:10:06Z","lastTransitionTime":"2025-12-03T14:10:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 14:10:06 crc kubenswrapper[4805]: I1203 14:10:06.328306 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/6b36a3db-1047-487a-9d6c-4f51ffdd8cc1-metrics-certs\") pod \"network-metrics-daemon-4gd9r\" (UID: \"6b36a3db-1047-487a-9d6c-4f51ffdd8cc1\") " pod="openshift-multus/network-metrics-daemon-4gd9r" Dec 03 14:10:06 crc kubenswrapper[4805]: E1203 14:10:06.328550 4805 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Dec 03 14:10:06 crc kubenswrapper[4805]: E1203 14:10:06.328648 4805 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/6b36a3db-1047-487a-9d6c-4f51ffdd8cc1-metrics-certs podName:6b36a3db-1047-487a-9d6c-4f51ffdd8cc1 nodeName:}" failed. No retries permitted until 2025-12-03 14:10:08.328620436 +0000 UTC m=+37.991537399 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/6b36a3db-1047-487a-9d6c-4f51ffdd8cc1-metrics-certs") pod "network-metrics-daemon-4gd9r" (UID: "6b36a3db-1047-487a-9d6c-4f51ffdd8cc1") : object "openshift-multus"/"metrics-daemon-secret" not registered Dec 03 14:10:06 crc kubenswrapper[4805]: I1203 14:10:06.428864 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 14:10:06 crc kubenswrapper[4805]: I1203 14:10:06.428921 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 14:10:06 crc kubenswrapper[4805]: I1203 14:10:06.428939 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 14:10:06 crc kubenswrapper[4805]: I1203 14:10:06.428961 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 14:10:06 crc kubenswrapper[4805]: I1203 14:10:06.428979 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T14:10:06Z","lastTransitionTime":"2025-12-03T14:10:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 14:10:06 crc kubenswrapper[4805]: I1203 14:10:06.531624 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 14:10:06 crc kubenswrapper[4805]: I1203 14:10:06.531695 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 14:10:06 crc kubenswrapper[4805]: I1203 14:10:06.531719 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 14:10:06 crc kubenswrapper[4805]: I1203 14:10:06.531750 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 14:10:06 crc kubenswrapper[4805]: I1203 14:10:06.531773 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T14:10:06Z","lastTransitionTime":"2025-12-03T14:10:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 14:10:06 crc kubenswrapper[4805]: I1203 14:10:06.634574 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 14:10:06 crc kubenswrapper[4805]: I1203 14:10:06.634627 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 14:10:06 crc kubenswrapper[4805]: I1203 14:10:06.634643 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 14:10:06 crc kubenswrapper[4805]: I1203 14:10:06.634665 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 14:10:06 crc kubenswrapper[4805]: I1203 14:10:06.634682 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T14:10:06Z","lastTransitionTime":"2025-12-03T14:10:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 14:10:06 crc kubenswrapper[4805]: I1203 14:10:06.693525 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 14:10:06 crc kubenswrapper[4805]: I1203 14:10:06.693589 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 14:10:06 crc kubenswrapper[4805]: I1203 14:10:06.693548 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-4gd9r" Dec 03 14:10:06 crc kubenswrapper[4805]: I1203 14:10:06.693525 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 14:10:06 crc kubenswrapper[4805]: E1203 14:10:06.693762 4805 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 03 14:10:06 crc kubenswrapper[4805]: E1203 14:10:06.694000 4805 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 03 14:10:06 crc kubenswrapper[4805]: E1203 14:10:06.694157 4805 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 03 14:10:06 crc kubenswrapper[4805]: E1203 14:10:06.694308 4805 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-4gd9r" podUID="6b36a3db-1047-487a-9d6c-4f51ffdd8cc1" Dec 03 14:10:06 crc kubenswrapper[4805]: I1203 14:10:06.736653 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 14:10:06 crc kubenswrapper[4805]: I1203 14:10:06.736712 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 14:10:06 crc kubenswrapper[4805]: I1203 14:10:06.736728 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 14:10:06 crc kubenswrapper[4805]: I1203 14:10:06.736750 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 14:10:06 crc kubenswrapper[4805]: I1203 14:10:06.736766 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T14:10:06Z","lastTransitionTime":"2025-12-03T14:10:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 14:10:06 crc kubenswrapper[4805]: I1203 14:10:06.840293 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 14:10:06 crc kubenswrapper[4805]: I1203 14:10:06.840357 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 14:10:06 crc kubenswrapper[4805]: I1203 14:10:06.840375 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 14:10:06 crc kubenswrapper[4805]: I1203 14:10:06.840403 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 14:10:06 crc kubenswrapper[4805]: I1203 14:10:06.840423 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T14:10:06Z","lastTransitionTime":"2025-12-03T14:10:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 14:10:06 crc kubenswrapper[4805]: I1203 14:10:06.943300 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 14:10:06 crc kubenswrapper[4805]: I1203 14:10:06.943372 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 14:10:06 crc kubenswrapper[4805]: I1203 14:10:06.943393 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 14:10:06 crc kubenswrapper[4805]: I1203 14:10:06.943422 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 14:10:06 crc kubenswrapper[4805]: I1203 14:10:06.943442 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T14:10:06Z","lastTransitionTime":"2025-12-03T14:10:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 14:10:07 crc kubenswrapper[4805]: I1203 14:10:07.045683 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 14:10:07 crc kubenswrapper[4805]: I1203 14:10:07.045755 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 14:10:07 crc kubenswrapper[4805]: I1203 14:10:07.045780 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 14:10:07 crc kubenswrapper[4805]: I1203 14:10:07.045807 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 14:10:07 crc kubenswrapper[4805]: I1203 14:10:07.045828 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T14:10:07Z","lastTransitionTime":"2025-12-03T14:10:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 14:10:07 crc kubenswrapper[4805]: I1203 14:10:07.148780 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 14:10:07 crc kubenswrapper[4805]: I1203 14:10:07.148858 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 14:10:07 crc kubenswrapper[4805]: I1203 14:10:07.148874 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 14:10:07 crc kubenswrapper[4805]: I1203 14:10:07.148898 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 14:10:07 crc kubenswrapper[4805]: I1203 14:10:07.148911 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T14:10:07Z","lastTransitionTime":"2025-12-03T14:10:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 14:10:07 crc kubenswrapper[4805]: I1203 14:10:07.251810 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 14:10:07 crc kubenswrapper[4805]: I1203 14:10:07.251923 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 14:10:07 crc kubenswrapper[4805]: I1203 14:10:07.251945 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 14:10:07 crc kubenswrapper[4805]: I1203 14:10:07.251973 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 14:10:07 crc kubenswrapper[4805]: I1203 14:10:07.251994 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T14:10:07Z","lastTransitionTime":"2025-12-03T14:10:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 14:10:07 crc kubenswrapper[4805]: I1203 14:10:07.354145 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 14:10:07 crc kubenswrapper[4805]: I1203 14:10:07.354192 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 14:10:07 crc kubenswrapper[4805]: I1203 14:10:07.354203 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 14:10:07 crc kubenswrapper[4805]: I1203 14:10:07.354219 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 14:10:07 crc kubenswrapper[4805]: I1203 14:10:07.354231 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T14:10:07Z","lastTransitionTime":"2025-12-03T14:10:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 14:10:07 crc kubenswrapper[4805]: I1203 14:10:07.457127 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 14:10:07 crc kubenswrapper[4805]: I1203 14:10:07.457204 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 14:10:07 crc kubenswrapper[4805]: I1203 14:10:07.457227 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 14:10:07 crc kubenswrapper[4805]: I1203 14:10:07.457254 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 14:10:07 crc kubenswrapper[4805]: I1203 14:10:07.457308 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T14:10:07Z","lastTransitionTime":"2025-12-03T14:10:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 14:10:07 crc kubenswrapper[4805]: I1203 14:10:07.560439 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 14:10:07 crc kubenswrapper[4805]: I1203 14:10:07.560808 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 14:10:07 crc kubenswrapper[4805]: I1203 14:10:07.561065 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 14:10:07 crc kubenswrapper[4805]: I1203 14:10:07.561315 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 14:10:07 crc kubenswrapper[4805]: I1203 14:10:07.561592 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T14:10:07Z","lastTransitionTime":"2025-12-03T14:10:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 14:10:07 crc kubenswrapper[4805]: I1203 14:10:07.664052 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 14:10:07 crc kubenswrapper[4805]: I1203 14:10:07.664112 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 14:10:07 crc kubenswrapper[4805]: I1203 14:10:07.664129 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 14:10:07 crc kubenswrapper[4805]: I1203 14:10:07.664152 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 14:10:07 crc kubenswrapper[4805]: I1203 14:10:07.664169 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T14:10:07Z","lastTransitionTime":"2025-12-03T14:10:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 14:10:07 crc kubenswrapper[4805]: I1203 14:10:07.766993 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 14:10:07 crc kubenswrapper[4805]: I1203 14:10:07.767204 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 14:10:07 crc kubenswrapper[4805]: I1203 14:10:07.767322 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 14:10:07 crc kubenswrapper[4805]: I1203 14:10:07.767446 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 14:10:07 crc kubenswrapper[4805]: I1203 14:10:07.767559 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T14:10:07Z","lastTransitionTime":"2025-12-03T14:10:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 14:10:07 crc kubenswrapper[4805]: I1203 14:10:07.870372 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 14:10:07 crc kubenswrapper[4805]: I1203 14:10:07.870408 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 14:10:07 crc kubenswrapper[4805]: I1203 14:10:07.870416 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 14:10:07 crc kubenswrapper[4805]: I1203 14:10:07.870452 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 14:10:07 crc kubenswrapper[4805]: I1203 14:10:07.870461 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T14:10:07Z","lastTransitionTime":"2025-12-03T14:10:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 14:10:07 crc kubenswrapper[4805]: I1203 14:10:07.972529 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 14:10:07 crc kubenswrapper[4805]: I1203 14:10:07.972638 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 14:10:07 crc kubenswrapper[4805]: I1203 14:10:07.972650 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 14:10:07 crc kubenswrapper[4805]: I1203 14:10:07.972666 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 14:10:07 crc kubenswrapper[4805]: I1203 14:10:07.972679 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T14:10:07Z","lastTransitionTime":"2025-12-03T14:10:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 14:10:08 crc kubenswrapper[4805]: I1203 14:10:08.075656 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 14:10:08 crc kubenswrapper[4805]: I1203 14:10:08.075705 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 14:10:08 crc kubenswrapper[4805]: I1203 14:10:08.075724 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 14:10:08 crc kubenswrapper[4805]: I1203 14:10:08.075745 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 14:10:08 crc kubenswrapper[4805]: I1203 14:10:08.075761 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T14:10:08Z","lastTransitionTime":"2025-12-03T14:10:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 14:10:08 crc kubenswrapper[4805]: I1203 14:10:08.177998 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 14:10:08 crc kubenswrapper[4805]: I1203 14:10:08.178049 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 14:10:08 crc kubenswrapper[4805]: I1203 14:10:08.178065 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 14:10:08 crc kubenswrapper[4805]: I1203 14:10:08.178129 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 14:10:08 crc kubenswrapper[4805]: I1203 14:10:08.178152 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T14:10:08Z","lastTransitionTime":"2025-12-03T14:10:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 14:10:08 crc kubenswrapper[4805]: I1203 14:10:08.281550 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 14:10:08 crc kubenswrapper[4805]: I1203 14:10:08.281606 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 14:10:08 crc kubenswrapper[4805]: I1203 14:10:08.281621 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 14:10:08 crc kubenswrapper[4805]: I1203 14:10:08.281643 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 14:10:08 crc kubenswrapper[4805]: I1203 14:10:08.281662 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T14:10:08Z","lastTransitionTime":"2025-12-03T14:10:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 14:10:08 crc kubenswrapper[4805]: I1203 14:10:08.350116 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/6b36a3db-1047-487a-9d6c-4f51ffdd8cc1-metrics-certs\") pod \"network-metrics-daemon-4gd9r\" (UID: \"6b36a3db-1047-487a-9d6c-4f51ffdd8cc1\") " pod="openshift-multus/network-metrics-daemon-4gd9r" Dec 03 14:10:08 crc kubenswrapper[4805]: E1203 14:10:08.350457 4805 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Dec 03 14:10:08 crc kubenswrapper[4805]: E1203 14:10:08.350616 4805 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/6b36a3db-1047-487a-9d6c-4f51ffdd8cc1-metrics-certs podName:6b36a3db-1047-487a-9d6c-4f51ffdd8cc1 nodeName:}" failed. No retries permitted until 2025-12-03 14:10:12.350582225 +0000 UTC m=+42.013499178 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/6b36a3db-1047-487a-9d6c-4f51ffdd8cc1-metrics-certs") pod "network-metrics-daemon-4gd9r" (UID: "6b36a3db-1047-487a-9d6c-4f51ffdd8cc1") : object "openshift-multus"/"metrics-daemon-secret" not registered Dec 03 14:10:08 crc kubenswrapper[4805]: I1203 14:10:08.384739 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 14:10:08 crc kubenswrapper[4805]: I1203 14:10:08.384802 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 14:10:08 crc kubenswrapper[4805]: I1203 14:10:08.384819 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 14:10:08 crc kubenswrapper[4805]: I1203 14:10:08.384871 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 14:10:08 crc kubenswrapper[4805]: I1203 14:10:08.384890 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T14:10:08Z","lastTransitionTime":"2025-12-03T14:10:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 14:10:08 crc kubenswrapper[4805]: I1203 14:10:08.487327 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 14:10:08 crc kubenswrapper[4805]: I1203 14:10:08.487374 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 14:10:08 crc kubenswrapper[4805]: I1203 14:10:08.487384 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 14:10:08 crc kubenswrapper[4805]: I1203 14:10:08.487401 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 14:10:08 crc kubenswrapper[4805]: I1203 14:10:08.487413 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T14:10:08Z","lastTransitionTime":"2025-12-03T14:10:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 14:10:08 crc kubenswrapper[4805]: I1203 14:10:08.589680 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 14:10:08 crc kubenswrapper[4805]: I1203 14:10:08.589720 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 14:10:08 crc kubenswrapper[4805]: I1203 14:10:08.589730 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 14:10:08 crc kubenswrapper[4805]: I1203 14:10:08.589745 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 14:10:08 crc kubenswrapper[4805]: I1203 14:10:08.589753 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T14:10:08Z","lastTransitionTime":"2025-12-03T14:10:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 14:10:08 crc kubenswrapper[4805]: I1203 14:10:08.692045 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 14:10:08 crc kubenswrapper[4805]: I1203 14:10:08.692113 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 14:10:08 crc kubenswrapper[4805]: I1203 14:10:08.692125 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 14:10:08 crc kubenswrapper[4805]: I1203 14:10:08.692140 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 14:10:08 crc kubenswrapper[4805]: I1203 14:10:08.692151 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T14:10:08Z","lastTransitionTime":"2025-12-03T14:10:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 14:10:08 crc kubenswrapper[4805]: I1203 14:10:08.694399 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-4gd9r" Dec 03 14:10:08 crc kubenswrapper[4805]: I1203 14:10:08.694436 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 14:10:08 crc kubenswrapper[4805]: I1203 14:10:08.694407 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 14:10:08 crc kubenswrapper[4805]: E1203 14:10:08.694551 4805 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-4gd9r" podUID="6b36a3db-1047-487a-9d6c-4f51ffdd8cc1" Dec 03 14:10:08 crc kubenswrapper[4805]: I1203 14:10:08.694727 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 14:10:08 crc kubenswrapper[4805]: E1203 14:10:08.695054 4805 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 03 14:10:08 crc kubenswrapper[4805]: E1203 14:10:08.695048 4805 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 03 14:10:08 crc kubenswrapper[4805]: E1203 14:10:08.695122 4805 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 03 14:10:08 crc kubenswrapper[4805]: I1203 14:10:08.795402 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 14:10:08 crc kubenswrapper[4805]: I1203 14:10:08.795442 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 14:10:08 crc kubenswrapper[4805]: I1203 14:10:08.795451 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 14:10:08 crc kubenswrapper[4805]: I1203 14:10:08.795465 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 14:10:08 crc kubenswrapper[4805]: I1203 14:10:08.795474 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T14:10:08Z","lastTransitionTime":"2025-12-03T14:10:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 14:10:08 crc kubenswrapper[4805]: I1203 14:10:08.845818 4805 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 03 14:10:08 crc kubenswrapper[4805]: I1203 14:10:08.860204 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-69h75" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5d1c694f-9847-4086-8b37-3573ddeb3ace\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:10:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:10:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:10:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:10:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://abc1d9b8582ca9fe4a51fb608fb39129a179ddcdfde0797a6369a55b2e0124cf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:10:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s578b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c7c4d434b035cac2a91a4f0846a2f4fb5bd68f881c28487860adb6db85e8ce75\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:10:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s578b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T14:10:03Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-69h75\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T14:10:08Z is after 2025-08-24T17:21:41Z" Dec 03 14:10:08 crc kubenswrapper[4805]: I1203 14:10:08.869890 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-pdzgk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"860769de-6a5b-4cb9-a752-264e7e2e785f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5e2f0a0cc21505d609f81918365d5e4f5d21959ac12ab147d6a51601c9de4c14\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ggn4n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T14:09:50Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-pdzgk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T14:10:08Z is after 2025-08-24T17:21:41Z" Dec 03 14:10:08 crc kubenswrapper[4805]: I1203 14:10:08.883466 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:49Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:49Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://38339331a3209d6787d1def81b72f4a8514bf91080a139cad6d8ac5cc6bef10e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T14:10:08Z is after 2025-08-24T17:21:41Z" Dec 03 14:10:08 crc kubenswrapper[4805]: I1203 14:10:08.897931 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 14:10:08 crc kubenswrapper[4805]: I1203 14:10:08.898159 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 14:10:08 crc kubenswrapper[4805]: I1203 14:10:08.898242 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 14:10:08 crc kubenswrapper[4805]: I1203 14:10:08.898341 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 14:10:08 crc kubenswrapper[4805]: I1203 14:10:08.898433 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T14:10:08Z","lastTransitionTime":"2025-12-03T14:10:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 14:10:08 crc kubenswrapper[4805]: I1203 14:10:08.899376 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:48Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:48Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T14:10:08Z is after 2025-08-24T17:21:41Z" Dec 03 14:10:08 crc kubenswrapper[4805]: I1203 14:10:08.914413 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:51Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:51Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b3123d3fa879097db88ed2f3cfb05631f440c24c5405503b2de3546f12e2e899\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T14:10:08Z is after 2025-08-24T17:21:41Z" Dec 03 14:10:08 crc kubenswrapper[4805]: I1203 14:10:08.925753 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-bdtdg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"53a0e9e3-b6d8-44b7-a32b-aa3450442865\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f557e17ad105d6972d3c1480656c6073f67e28c0f5773aa365549951f4d161c7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9ww54\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T14:09:51Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-bdtdg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T14:10:08Z is after 2025-08-24T17:21:41Z" Dec 03 14:10:08 crc kubenswrapper[4805]: I1203 14:10:08.955006 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b6d36671-7173-4e9c-843f-005b4dcf576c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a837d7d491eaacbe4279b4014bb5fcbe9b514ba5cc82c708227bcd9b25a346e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6e5f04faf3b23197bb8ea11d3e574dc86f31e21d798734962001ec8c95de7c84\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2ff1a979387da41b00e95eb9cc8b2c9de966380a12fae687591df3c5b18a4497\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1e66e1cb18573a14e3c3ae16af4dbd9f8997a6ad38680c3ab4505646ebd48e7c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7d5b5c85e3de43e7e95111aa967248550607c26200c783251b68696f202f92a1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://474f72dac37a808fc4b9832306e76979b151d7221807ff3086013466ed1419c3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://474f72dac37a808fc4b9832306e76979b151d7221807ff3086013466ed1419c3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T14:09:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T14:09:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://74c8cea1a61c4775a0ad2f4ab4bb8a36437cf23d2ee953311d6ea967e297510a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://74c8cea1a61c4775a0ad2f4ab4bb8a36437cf23d2ee953311d6ea967e297510a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T14:09:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T14:09:31Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://a4f2f56a809adc01fba82eba0652d78f58770c132f77b1424f8b17068b1927e3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a4f2f56a809adc01fba82eba0652d78f58770c132f77b1424f8b17068b1927e3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T14:09:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T14:09:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T14:09:30Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T14:10:08Z is after 2025-08-24T17:21:41Z" Dec 03 14:10:08 crc kubenswrapper[4805]: I1203 14:10:08.973891 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ced8291d-0ce5-4e74-9eaf-2bf033774039\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:10:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:10:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3e5a34636f0a8c4d9179074596d28bc2eb7a504573398c596a637defd829907a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://29cd95be3f964415943c873e7914ffcea4abe50dfdf17f8a9cd03fc38df35981\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://951173b3d3d581275a6f6e43209b5e0611a701ca7a10f81504d723e24a2d21a3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e64730957bbbb2feb4d506e0ed9731a0594bf66fa1e8fcdb8ec944ded6308eed\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5da55f769e23ad52bac9c77da62c78e70fa34db8ed1a435e4eaf4dad3fd2b68a\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-03T14:09:48Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1203 14:09:42.992179 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1203 14:09:42.993233 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1830926909/tls.crt::/tmp/serving-cert-1830926909/tls.key\\\\\\\"\\\\nI1203 14:09:48.622403 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1203 14:09:48.626476 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1203 14:09:48.626527 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1203 14:09:48.626574 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1203 14:09:48.626582 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1203 14:09:48.633990 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1203 14:09:48.634024 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 14:09:48.634029 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 14:09:48.634034 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1203 14:09:48.634038 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1203 14:09:48.634042 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1203 14:09:48.634045 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1203 14:09:48.634344 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1203 14:09:48.644497 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T14:09:32Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bd92a160eb530c5687ebd7cd5b83b41820a2b808a0fd9514e5ace9d237e1bd26\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:32Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://af9c940135012c629447528c61743ed1e58708ab57ce76604c6da0479aa459cf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://af9c940135012c629447528c61743ed1e58708ab57ce76604c6da0479aa459cf\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T14:09:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T14:09:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T14:09:30Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T14:10:08Z is after 2025-08-24T17:21:41Z" Dec 03 14:10:08 crc kubenswrapper[4805]: I1203 14:10:08.989597 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-k4mhs" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c09c1cdc-f461-45c9-8444-5b8764bccabc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3354c1a44f891182d04116fc76234a2b337d50567d032a3584f902a9516387cc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qdcd7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T14:09:50Z\\\"}}\" for pod \"openshift-multus\"/\"multus-k4mhs\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T14:10:08Z is after 2025-08-24T17:21:41Z" Dec 03 14:10:09 crc kubenswrapper[4805]: I1203 14:10:09.001525 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 14:10:09 crc kubenswrapper[4805]: I1203 14:10:09.001565 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 14:10:09 crc kubenswrapper[4805]: I1203 14:10:09.001578 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 14:10:09 crc kubenswrapper[4805]: I1203 14:10:09.001599 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 14:10:09 crc kubenswrapper[4805]: I1203 14:10:09.001612 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T14:10:09Z","lastTransitionTime":"2025-12-03T14:10:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 14:10:09 crc kubenswrapper[4805]: I1203 14:10:09.005391 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-bpk49" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"de00a81f-82e1-4223-9614-f98606b9a8ef\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://04f4ad27e02e2024ae9e7cd1b16116267e69cecda882dd723cdb274c41a0e7db\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-25d7t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7741e0e70416ebe094bfabc6009e98d7d4c0fbcba1fc82df13b9ba40774b456a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7741e0e70416ebe094bfabc6009e98d7d4c0fbcba1fc82df13b9ba40774b456a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T14:09:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T14:09:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-25d7t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ee3499c02eb52bbf08c447d2433d435a74dfb70670dc4cbaac44f5fbff9e8a29\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ee3499c02eb52bbf08c447d2433d435a74dfb70670dc4cbaac44f5fbff9e8a29\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T14:09:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T14:09:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-25d7t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ec8488b1d69799951be7ef40e5e028732265c4ddcdea2d453923e567864771f7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ec8488b1d69799951be7ef40e5e028732265c4ddcdea2d453923e567864771f7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T14:09:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T14:09:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-25d7t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://66707bb4e1fdacb265e89786df8a04605830f1c156394cacfcec9b54abe5d41a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://66707bb4e1fdacb265e89786df8a04605830f1c156394cacfcec9b54abe5d41a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T14:09:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T14:09:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-25d7t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6207ab6600970ff15c8713c547d4920b2c77eb9bea990bb5da497e24870dcc0f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6207ab6600970ff15c8713c547d4920b2c77eb9bea990bb5da497e24870dcc0f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T14:09:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T14:09:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-25d7t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8de65d4a205e2deec0b433e53c0e37feb94aa0ede6a4c58d7414b80829804505\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8de65d4a205e2deec0b433e53c0e37feb94aa0ede6a4c58d7414b80829804505\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T14:09:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T14:09:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-25d7t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T14:09:50Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-bpk49\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T14:10:09Z is after 2025-08-24T17:21:41Z" Dec 03 14:10:09 crc kubenswrapper[4805]: I1203 14:10:09.018198 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"750d2b4b-738a-4992-8902-742528f6e875\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://abd2d0156113bb5182727a53aac594410ed051c19258beb8a07b808eff0b42a7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8673a29c0e16f4e707d79f0610722ccabb023ff2b4a18d7cf9d305183817d952\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://35766e7206b1be7a53d13d6890d759cf81dbd999643c12349268e4bf16fbbdb0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5b9844a11b2b903978d0d464b469e26c9211100da61d018dde9636dca95b684e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T14:09:30Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T14:10:09Z is after 2025-08-24T17:21:41Z" Dec 03 14:10:09 crc kubenswrapper[4805]: I1203 14:10:09.034998 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:48Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:48Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T14:10:09Z is after 2025-08-24T17:21:41Z" Dec 03 14:10:09 crc kubenswrapper[4805]: I1203 14:10:09.049321 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-gskh4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6098937f-e3f6-45e8-a647-4994a79cd711\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://28d5653f26ec61647808b3a733ff7eac70aa69ac184660f24147e8cfaf0eb802\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nwrw9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9ad6b68a42fd33ee167e67873c6eed3f11cdf6dc71d22e952c021436eb967dd9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nwrw9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T14:09:50Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-gskh4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T14:10:09Z is after 2025-08-24T17:21:41Z" Dec 03 14:10:09 crc kubenswrapper[4805]: I1203 14:10:09.075004 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-psjpm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"deeec30d-cad2-4dfd-adb6-64f6646581be\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:50Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:50Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4b8d00cdf0f37f2e5cd4e952cf85d3de8d72c43fdf786088fd9f26ed17b1e21f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wdlzx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bd67c6de01a8d196a1f51d4d90640c9e69de6f05d5e201b25e6eaa5c33440ea8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wdlzx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cb8bf05ef093eb0727a066d4bbbdf9af9422867cd4dadacafc6eeb016867a61e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wdlzx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0304467102ed053628c0d3e06548c74b6492336b6fd044246508d1e8bdb17bc3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wdlzx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6ce80ccafb3262a11fbe9923317a9d38b61e5dc5b3e7e7ae136f143d6b643fd7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wdlzx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://47d76962beb8c0f196d50af96fb3ec611097823ca80a1c8bd718f2d18840128c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wdlzx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://644e56c400524d739e2a208fb7eee1551ebbd28ba220df91c9dd84ba09d0b400\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://644e56c400524d739e2a208fb7eee1551ebbd28ba220df91c9dd84ba09d0b400\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-03T14:10:02Z\\\",\\\"message\\\":\\\"tring(nil)}, nodeEndpoints:map[string]services.lbEndpoints{}, externalTrafficLocal:false, internalTrafficLocal:false, hasNodePort:false}}\\\\nF1203 14:10:01.661807 6255 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T14:10:01Z is after 2025-08-24T17:21:41Z]\\\\nI1203 14:10:01.660569 6255 services_controller.go:451] Built service openshift-operator-lifecycle-manager/package-server-manager-metrics cluster-wide LB for network=default: []services.LB{services.LB{Name:\\\\\\\"Service_openshift-operator-lifecycle-manager/package-server-manager-metrics_TCP_cluster\\\\\\\", UUID:\\\\\\\"\\\\\\\", Protocol:\\\\\\\"TCP\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Serv\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T14:10:01Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-psjpm_openshift-ovn-kubernetes(deeec30d-cad2-4dfd-adb6-64f6646581be)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wdlzx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ab98e365c1a4242a47b8cbe2dd85109d366d6002c4de1f18ddc950a46bea6c12\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wdlzx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2f875252fbe6a673a72fc9c56bb96899f62732305ad77b3108e3b244bfa91ecb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2f875252fbe6a673a72fc9c56bb96899f62732305ad77b3108e3b244bfa91ecb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T14:09:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T14:09:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wdlzx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T14:09:50Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-psjpm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T14:10:09Z is after 2025-08-24T17:21:41Z" Dec 03 14:10:09 crc kubenswrapper[4805]: I1203 14:10:09.088485 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-4gd9r" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6b36a3db-1047-487a-9d6c-4f51ffdd8cc1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:10:04Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:10:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:10:04Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:10:04Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4t6zl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4t6zl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T14:10:04Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-4gd9r\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T14:10:09Z is after 2025-08-24T17:21:41Z" Dec 03 14:10:09 crc kubenswrapper[4805]: I1203 14:10:09.103564 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:48Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:48Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T14:10:09Z is after 2025-08-24T17:21:41Z" Dec 03 14:10:09 crc kubenswrapper[4805]: I1203 14:10:09.104654 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 14:10:09 crc kubenswrapper[4805]: I1203 14:10:09.104723 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 14:10:09 crc kubenswrapper[4805]: I1203 14:10:09.104737 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 14:10:09 crc kubenswrapper[4805]: I1203 14:10:09.104755 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 14:10:09 crc kubenswrapper[4805]: I1203 14:10:09.104767 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T14:10:09Z","lastTransitionTime":"2025-12-03T14:10:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 14:10:09 crc kubenswrapper[4805]: I1203 14:10:09.120936 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:49Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:49Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://541ace7ccbb66dbd191c3ca8c1e4479e3edbf80dfc67326bea8aaee81fdcfc3c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://821e3e9a902aacecf6c41e8b8d8aa1ac951ce2fbb6505b7148b16348a3754e55\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T14:10:09Z is after 2025-08-24T17:21:41Z" Dec 03 14:10:09 crc kubenswrapper[4805]: I1203 14:10:09.207464 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 14:10:09 crc kubenswrapper[4805]: I1203 14:10:09.207509 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 14:10:09 crc kubenswrapper[4805]: I1203 14:10:09.207526 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 14:10:09 crc kubenswrapper[4805]: I1203 14:10:09.207548 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 14:10:09 crc kubenswrapper[4805]: I1203 14:10:09.207564 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T14:10:09Z","lastTransitionTime":"2025-12-03T14:10:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 14:10:09 crc kubenswrapper[4805]: I1203 14:10:09.310526 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 14:10:09 crc kubenswrapper[4805]: I1203 14:10:09.310589 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 14:10:09 crc kubenswrapper[4805]: I1203 14:10:09.310606 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 14:10:09 crc kubenswrapper[4805]: I1203 14:10:09.310629 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 14:10:09 crc kubenswrapper[4805]: I1203 14:10:09.310647 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T14:10:09Z","lastTransitionTime":"2025-12-03T14:10:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 14:10:09 crc kubenswrapper[4805]: I1203 14:10:09.413719 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 14:10:09 crc kubenswrapper[4805]: I1203 14:10:09.413937 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 14:10:09 crc kubenswrapper[4805]: I1203 14:10:09.413964 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 14:10:09 crc kubenswrapper[4805]: I1203 14:10:09.413994 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 14:10:09 crc kubenswrapper[4805]: I1203 14:10:09.414014 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T14:10:09Z","lastTransitionTime":"2025-12-03T14:10:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 14:10:09 crc kubenswrapper[4805]: I1203 14:10:09.517225 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 14:10:09 crc kubenswrapper[4805]: I1203 14:10:09.517438 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 14:10:09 crc kubenswrapper[4805]: I1203 14:10:09.517533 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 14:10:09 crc kubenswrapper[4805]: I1203 14:10:09.517620 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 14:10:09 crc kubenswrapper[4805]: I1203 14:10:09.517727 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T14:10:09Z","lastTransitionTime":"2025-12-03T14:10:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 14:10:09 crc kubenswrapper[4805]: I1203 14:10:09.620713 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 14:10:09 crc kubenswrapper[4805]: I1203 14:10:09.620758 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 14:10:09 crc kubenswrapper[4805]: I1203 14:10:09.620770 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 14:10:09 crc kubenswrapper[4805]: I1203 14:10:09.620788 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 14:10:09 crc kubenswrapper[4805]: I1203 14:10:09.620799 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T14:10:09Z","lastTransitionTime":"2025-12-03T14:10:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 14:10:09 crc kubenswrapper[4805]: I1203 14:10:09.724125 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 14:10:09 crc kubenswrapper[4805]: I1203 14:10:09.724177 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 14:10:09 crc kubenswrapper[4805]: I1203 14:10:09.724194 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 14:10:09 crc kubenswrapper[4805]: I1203 14:10:09.724218 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 14:10:09 crc kubenswrapper[4805]: I1203 14:10:09.724235 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T14:10:09Z","lastTransitionTime":"2025-12-03T14:10:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 14:10:09 crc kubenswrapper[4805]: I1203 14:10:09.826476 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 14:10:09 crc kubenswrapper[4805]: I1203 14:10:09.826522 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 14:10:09 crc kubenswrapper[4805]: I1203 14:10:09.826537 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 14:10:09 crc kubenswrapper[4805]: I1203 14:10:09.826562 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 14:10:09 crc kubenswrapper[4805]: I1203 14:10:09.826580 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T14:10:09Z","lastTransitionTime":"2025-12-03T14:10:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 14:10:09 crc kubenswrapper[4805]: I1203 14:10:09.929729 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 14:10:09 crc kubenswrapper[4805]: I1203 14:10:09.929768 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 14:10:09 crc kubenswrapper[4805]: I1203 14:10:09.929778 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 14:10:09 crc kubenswrapper[4805]: I1203 14:10:09.929794 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 14:10:09 crc kubenswrapper[4805]: I1203 14:10:09.929803 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T14:10:09Z","lastTransitionTime":"2025-12-03T14:10:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 14:10:10 crc kubenswrapper[4805]: I1203 14:10:10.032643 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 14:10:10 crc kubenswrapper[4805]: I1203 14:10:10.033018 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 14:10:10 crc kubenswrapper[4805]: I1203 14:10:10.033155 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 14:10:10 crc kubenswrapper[4805]: I1203 14:10:10.033276 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 14:10:10 crc kubenswrapper[4805]: I1203 14:10:10.033397 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T14:10:10Z","lastTransitionTime":"2025-12-03T14:10:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 14:10:10 crc kubenswrapper[4805]: I1203 14:10:10.136395 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 14:10:10 crc kubenswrapper[4805]: I1203 14:10:10.136480 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 14:10:10 crc kubenswrapper[4805]: I1203 14:10:10.136510 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 14:10:10 crc kubenswrapper[4805]: I1203 14:10:10.136537 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 14:10:10 crc kubenswrapper[4805]: I1203 14:10:10.136555 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T14:10:10Z","lastTransitionTime":"2025-12-03T14:10:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 14:10:10 crc kubenswrapper[4805]: I1203 14:10:10.239414 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 14:10:10 crc kubenswrapper[4805]: I1203 14:10:10.239459 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 14:10:10 crc kubenswrapper[4805]: I1203 14:10:10.239471 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 14:10:10 crc kubenswrapper[4805]: I1203 14:10:10.239486 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 14:10:10 crc kubenswrapper[4805]: I1203 14:10:10.239497 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T14:10:10Z","lastTransitionTime":"2025-12-03T14:10:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 14:10:10 crc kubenswrapper[4805]: I1203 14:10:10.341878 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 14:10:10 crc kubenswrapper[4805]: I1203 14:10:10.341938 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 14:10:10 crc kubenswrapper[4805]: I1203 14:10:10.341950 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 14:10:10 crc kubenswrapper[4805]: I1203 14:10:10.341967 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 14:10:10 crc kubenswrapper[4805]: I1203 14:10:10.341979 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T14:10:10Z","lastTransitionTime":"2025-12-03T14:10:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 14:10:10 crc kubenswrapper[4805]: I1203 14:10:10.444281 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 14:10:10 crc kubenswrapper[4805]: I1203 14:10:10.444320 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 14:10:10 crc kubenswrapper[4805]: I1203 14:10:10.444331 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 14:10:10 crc kubenswrapper[4805]: I1203 14:10:10.444363 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 14:10:10 crc kubenswrapper[4805]: I1203 14:10:10.444373 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T14:10:10Z","lastTransitionTime":"2025-12-03T14:10:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 14:10:10 crc kubenswrapper[4805]: I1203 14:10:10.547357 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 14:10:10 crc kubenswrapper[4805]: I1203 14:10:10.547443 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 14:10:10 crc kubenswrapper[4805]: I1203 14:10:10.547458 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 14:10:10 crc kubenswrapper[4805]: I1203 14:10:10.547475 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 14:10:10 crc kubenswrapper[4805]: I1203 14:10:10.547488 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T14:10:10Z","lastTransitionTime":"2025-12-03T14:10:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 14:10:10 crc kubenswrapper[4805]: I1203 14:10:10.649608 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 14:10:10 crc kubenswrapper[4805]: I1203 14:10:10.649659 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 14:10:10 crc kubenswrapper[4805]: I1203 14:10:10.649675 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 14:10:10 crc kubenswrapper[4805]: I1203 14:10:10.649698 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 14:10:10 crc kubenswrapper[4805]: I1203 14:10:10.649716 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T14:10:10Z","lastTransitionTime":"2025-12-03T14:10:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 14:10:10 crc kubenswrapper[4805]: I1203 14:10:10.694260 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-4gd9r" Dec 03 14:10:10 crc kubenswrapper[4805]: I1203 14:10:10.694373 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 14:10:10 crc kubenswrapper[4805]: E1203 14:10:10.694464 4805 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-4gd9r" podUID="6b36a3db-1047-487a-9d6c-4f51ffdd8cc1" Dec 03 14:10:10 crc kubenswrapper[4805]: E1203 14:10:10.694559 4805 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 03 14:10:10 crc kubenswrapper[4805]: I1203 14:10:10.694280 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 14:10:10 crc kubenswrapper[4805]: E1203 14:10:10.694688 4805 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 03 14:10:10 crc kubenswrapper[4805]: I1203 14:10:10.694260 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 14:10:10 crc kubenswrapper[4805]: E1203 14:10:10.694779 4805 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 03 14:10:10 crc kubenswrapper[4805]: I1203 14:10:10.708798 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:51Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:51Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b3123d3fa879097db88ed2f3cfb05631f440c24c5405503b2de3546f12e2e899\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T14:10:10Z is after 2025-08-24T17:21:41Z" Dec 03 14:10:10 crc kubenswrapper[4805]: I1203 14:10:10.722168 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-bdtdg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"53a0e9e3-b6d8-44b7-a32b-aa3450442865\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f557e17ad105d6972d3c1480656c6073f67e28c0f5773aa365549951f4d161c7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9ww54\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T14:09:51Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-bdtdg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T14:10:10Z is after 2025-08-24T17:21:41Z" Dec 03 14:10:10 crc kubenswrapper[4805]: I1203 14:10:10.742206 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b6d36671-7173-4e9c-843f-005b4dcf576c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a837d7d491eaacbe4279b4014bb5fcbe9b514ba5cc82c708227bcd9b25a346e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6e5f04faf3b23197bb8ea11d3e574dc86f31e21d798734962001ec8c95de7c84\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2ff1a979387da41b00e95eb9cc8b2c9de966380a12fae687591df3c5b18a4497\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1e66e1cb18573a14e3c3ae16af4dbd9f8997a6ad38680c3ab4505646ebd48e7c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7d5b5c85e3de43e7e95111aa967248550607c26200c783251b68696f202f92a1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://474f72dac37a808fc4b9832306e76979b151d7221807ff3086013466ed1419c3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://474f72dac37a808fc4b9832306e76979b151d7221807ff3086013466ed1419c3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T14:09:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T14:09:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://74c8cea1a61c4775a0ad2f4ab4bb8a36437cf23d2ee953311d6ea967e297510a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://74c8cea1a61c4775a0ad2f4ab4bb8a36437cf23d2ee953311d6ea967e297510a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T14:09:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T14:09:31Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://a4f2f56a809adc01fba82eba0652d78f58770c132f77b1424f8b17068b1927e3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a4f2f56a809adc01fba82eba0652d78f58770c132f77b1424f8b17068b1927e3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T14:09:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T14:09:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T14:09:30Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T14:10:10Z is after 2025-08-24T17:21:41Z" Dec 03 14:10:10 crc kubenswrapper[4805]: I1203 14:10:10.751722 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 14:10:10 crc kubenswrapper[4805]: I1203 14:10:10.751749 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 14:10:10 crc kubenswrapper[4805]: I1203 14:10:10.751757 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 14:10:10 crc kubenswrapper[4805]: I1203 14:10:10.751770 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 14:10:10 crc kubenswrapper[4805]: I1203 14:10:10.751780 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T14:10:10Z","lastTransitionTime":"2025-12-03T14:10:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 14:10:10 crc kubenswrapper[4805]: I1203 14:10:10.758042 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ced8291d-0ce5-4e74-9eaf-2bf033774039\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:10:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:10:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3e5a34636f0a8c4d9179074596d28bc2eb7a504573398c596a637defd829907a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://29cd95be3f964415943c873e7914ffcea4abe50dfdf17f8a9cd03fc38df35981\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://951173b3d3d581275a6f6e43209b5e0611a701ca7a10f81504d723e24a2d21a3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e64730957bbbb2feb4d506e0ed9731a0594bf66fa1e8fcdb8ec944ded6308eed\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5da55f769e23ad52bac9c77da62c78e70fa34db8ed1a435e4eaf4dad3fd2b68a\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-03T14:09:48Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1203 14:09:42.992179 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1203 14:09:42.993233 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1830926909/tls.crt::/tmp/serving-cert-1830926909/tls.key\\\\\\\"\\\\nI1203 14:09:48.622403 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1203 14:09:48.626476 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1203 14:09:48.626527 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1203 14:09:48.626574 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1203 14:09:48.626582 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1203 14:09:48.633990 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1203 14:09:48.634024 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 14:09:48.634029 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 14:09:48.634034 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1203 14:09:48.634038 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1203 14:09:48.634042 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1203 14:09:48.634045 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1203 14:09:48.634344 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1203 14:09:48.644497 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T14:09:32Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bd92a160eb530c5687ebd7cd5b83b41820a2b808a0fd9514e5ace9d237e1bd26\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:32Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://af9c940135012c629447528c61743ed1e58708ab57ce76604c6da0479aa459cf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://af9c940135012c629447528c61743ed1e58708ab57ce76604c6da0479aa459cf\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T14:09:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T14:09:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T14:09:30Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T14:10:10Z is after 2025-08-24T17:21:41Z" Dec 03 14:10:10 crc kubenswrapper[4805]: I1203 14:10:10.770333 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:49Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:49Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://38339331a3209d6787d1def81b72f4a8514bf91080a139cad6d8ac5cc6bef10e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T14:10:10Z is after 2025-08-24T17:21:41Z" Dec 03 14:10:10 crc kubenswrapper[4805]: I1203 14:10:10.784299 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:48Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:48Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T14:10:10Z is after 2025-08-24T17:21:41Z" Dec 03 14:10:10 crc kubenswrapper[4805]: I1203 14:10:10.798605 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"750d2b4b-738a-4992-8902-742528f6e875\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://abd2d0156113bb5182727a53aac594410ed051c19258beb8a07b808eff0b42a7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8673a29c0e16f4e707d79f0610722ccabb023ff2b4a18d7cf9d305183817d952\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://35766e7206b1be7a53d13d6890d759cf81dbd999643c12349268e4bf16fbbdb0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5b9844a11b2b903978d0d464b469e26c9211100da61d018dde9636dca95b684e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T14:09:30Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T14:10:10Z is after 2025-08-24T17:21:41Z" Dec 03 14:10:10 crc kubenswrapper[4805]: I1203 14:10:10.810124 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:48Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:48Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T14:10:10Z is after 2025-08-24T17:21:41Z" Dec 03 14:10:10 crc kubenswrapper[4805]: I1203 14:10:10.822420 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-k4mhs" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c09c1cdc-f461-45c9-8444-5b8764bccabc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3354c1a44f891182d04116fc76234a2b337d50567d032a3584f902a9516387cc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qdcd7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T14:09:50Z\\\"}}\" for pod \"openshift-multus\"/\"multus-k4mhs\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T14:10:10Z is after 2025-08-24T17:21:41Z" Dec 03 14:10:10 crc kubenswrapper[4805]: I1203 14:10:10.839887 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-bpk49" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"de00a81f-82e1-4223-9614-f98606b9a8ef\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://04f4ad27e02e2024ae9e7cd1b16116267e69cecda882dd723cdb274c41a0e7db\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-25d7t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7741e0e70416ebe094bfabc6009e98d7d4c0fbcba1fc82df13b9ba40774b456a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7741e0e70416ebe094bfabc6009e98d7d4c0fbcba1fc82df13b9ba40774b456a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T14:09:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T14:09:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-25d7t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ee3499c02eb52bbf08c447d2433d435a74dfb70670dc4cbaac44f5fbff9e8a29\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ee3499c02eb52bbf08c447d2433d435a74dfb70670dc4cbaac44f5fbff9e8a29\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T14:09:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T14:09:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-25d7t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ec8488b1d69799951be7ef40e5e028732265c4ddcdea2d453923e567864771f7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ec8488b1d69799951be7ef40e5e028732265c4ddcdea2d453923e567864771f7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T14:09:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T14:09:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-25d7t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://66707bb4e1fdacb265e89786df8a04605830f1c156394cacfcec9b54abe5d41a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://66707bb4e1fdacb265e89786df8a04605830f1c156394cacfcec9b54abe5d41a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T14:09:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T14:09:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-25d7t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6207ab6600970ff15c8713c547d4920b2c77eb9bea990bb5da497e24870dcc0f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6207ab6600970ff15c8713c547d4920b2c77eb9bea990bb5da497e24870dcc0f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T14:09:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T14:09:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-25d7t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8de65d4a205e2deec0b433e53c0e37feb94aa0ede6a4c58d7414b80829804505\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8de65d4a205e2deec0b433e53c0e37feb94aa0ede6a4c58d7414b80829804505\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T14:09:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T14:09:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-25d7t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T14:09:50Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-bpk49\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T14:10:10Z is after 2025-08-24T17:21:41Z" Dec 03 14:10:10 crc kubenswrapper[4805]: I1203 14:10:10.852858 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:48Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:48Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T14:10:10Z is after 2025-08-24T17:21:41Z" Dec 03 14:10:10 crc kubenswrapper[4805]: I1203 14:10:10.854249 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 14:10:10 crc kubenswrapper[4805]: I1203 14:10:10.854303 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 14:10:10 crc kubenswrapper[4805]: I1203 14:10:10.854320 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 14:10:10 crc kubenswrapper[4805]: I1203 14:10:10.854335 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 14:10:10 crc kubenswrapper[4805]: I1203 14:10:10.854351 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T14:10:10Z","lastTransitionTime":"2025-12-03T14:10:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 14:10:10 crc kubenswrapper[4805]: I1203 14:10:10.868049 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:49Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:49Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://541ace7ccbb66dbd191c3ca8c1e4479e3edbf80dfc67326bea8aaee81fdcfc3c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://821e3e9a902aacecf6c41e8b8d8aa1ac951ce2fbb6505b7148b16348a3754e55\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T14:10:10Z is after 2025-08-24T17:21:41Z" Dec 03 14:10:10 crc kubenswrapper[4805]: I1203 14:10:10.883509 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-gskh4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6098937f-e3f6-45e8-a647-4994a79cd711\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://28d5653f26ec61647808b3a733ff7eac70aa69ac184660f24147e8cfaf0eb802\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nwrw9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9ad6b68a42fd33ee167e67873c6eed3f11cdf6dc71d22e952c021436eb967dd9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nwrw9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T14:09:50Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-gskh4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T14:10:10Z is after 2025-08-24T17:21:41Z" Dec 03 14:10:10 crc kubenswrapper[4805]: I1203 14:10:10.903106 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-psjpm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"deeec30d-cad2-4dfd-adb6-64f6646581be\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:50Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:50Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4b8d00cdf0f37f2e5cd4e952cf85d3de8d72c43fdf786088fd9f26ed17b1e21f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wdlzx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bd67c6de01a8d196a1f51d4d90640c9e69de6f05d5e201b25e6eaa5c33440ea8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wdlzx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cb8bf05ef093eb0727a066d4bbbdf9af9422867cd4dadacafc6eeb016867a61e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wdlzx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0304467102ed053628c0d3e06548c74b6492336b6fd044246508d1e8bdb17bc3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wdlzx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6ce80ccafb3262a11fbe9923317a9d38b61e5dc5b3e7e7ae136f143d6b643fd7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wdlzx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://47d76962beb8c0f196d50af96fb3ec611097823ca80a1c8bd718f2d18840128c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wdlzx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://644e56c400524d739e2a208fb7eee1551ebbd28ba220df91c9dd84ba09d0b400\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://644e56c400524d739e2a208fb7eee1551ebbd28ba220df91c9dd84ba09d0b400\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-03T14:10:02Z\\\",\\\"message\\\":\\\"tring(nil)}, nodeEndpoints:map[string]services.lbEndpoints{}, externalTrafficLocal:false, internalTrafficLocal:false, hasNodePort:false}}\\\\nF1203 14:10:01.661807 6255 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T14:10:01Z is after 2025-08-24T17:21:41Z]\\\\nI1203 14:10:01.660569 6255 services_controller.go:451] Built service openshift-operator-lifecycle-manager/package-server-manager-metrics cluster-wide LB for network=default: []services.LB{services.LB{Name:\\\\\\\"Service_openshift-operator-lifecycle-manager/package-server-manager-metrics_TCP_cluster\\\\\\\", UUID:\\\\\\\"\\\\\\\", Protocol:\\\\\\\"TCP\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Serv\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T14:10:01Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-psjpm_openshift-ovn-kubernetes(deeec30d-cad2-4dfd-adb6-64f6646581be)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wdlzx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ab98e365c1a4242a47b8cbe2dd85109d366d6002c4de1f18ddc950a46bea6c12\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wdlzx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2f875252fbe6a673a72fc9c56bb96899f62732305ad77b3108e3b244bfa91ecb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2f875252fbe6a673a72fc9c56bb96899f62732305ad77b3108e3b244bfa91ecb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T14:09:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T14:09:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wdlzx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T14:09:50Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-psjpm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T14:10:10Z is after 2025-08-24T17:21:41Z" Dec 03 14:10:10 crc kubenswrapper[4805]: I1203 14:10:10.914473 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-4gd9r" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6b36a3db-1047-487a-9d6c-4f51ffdd8cc1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:10:04Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:10:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:10:04Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:10:04Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4t6zl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4t6zl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T14:10:04Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-4gd9r\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T14:10:10Z is after 2025-08-24T17:21:41Z" Dec 03 14:10:10 crc kubenswrapper[4805]: I1203 14:10:10.925208 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-pdzgk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"860769de-6a5b-4cb9-a752-264e7e2e785f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5e2f0a0cc21505d609f81918365d5e4f5d21959ac12ab147d6a51601c9de4c14\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ggn4n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T14:09:50Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-pdzgk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T14:10:10Z is after 2025-08-24T17:21:41Z" Dec 03 14:10:10 crc kubenswrapper[4805]: I1203 14:10:10.936566 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-69h75" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5d1c694f-9847-4086-8b37-3573ddeb3ace\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:10:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:10:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:10:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:10:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://abc1d9b8582ca9fe4a51fb608fb39129a179ddcdfde0797a6369a55b2e0124cf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:10:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s578b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c7c4d434b035cac2a91a4f0846a2f4fb5bd68f881c28487860adb6db85e8ce75\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:10:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s578b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T14:10:03Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-69h75\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T14:10:10Z is after 2025-08-24T17:21:41Z" Dec 03 14:10:10 crc kubenswrapper[4805]: I1203 14:10:10.956595 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 14:10:10 crc kubenswrapper[4805]: I1203 14:10:10.956889 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 14:10:10 crc kubenswrapper[4805]: I1203 14:10:10.956973 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 14:10:10 crc kubenswrapper[4805]: I1203 14:10:10.957048 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 14:10:10 crc kubenswrapper[4805]: I1203 14:10:10.957106 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T14:10:10Z","lastTransitionTime":"2025-12-03T14:10:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 14:10:11 crc kubenswrapper[4805]: I1203 14:10:11.059594 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 14:10:11 crc kubenswrapper[4805]: I1203 14:10:11.059671 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 14:10:11 crc kubenswrapper[4805]: I1203 14:10:11.059687 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 14:10:11 crc kubenswrapper[4805]: I1203 14:10:11.059704 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 14:10:11 crc kubenswrapper[4805]: I1203 14:10:11.059716 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T14:10:11Z","lastTransitionTime":"2025-12-03T14:10:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 14:10:11 crc kubenswrapper[4805]: I1203 14:10:11.162976 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 14:10:11 crc kubenswrapper[4805]: I1203 14:10:11.163042 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 14:10:11 crc kubenswrapper[4805]: I1203 14:10:11.163059 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 14:10:11 crc kubenswrapper[4805]: I1203 14:10:11.163082 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 14:10:11 crc kubenswrapper[4805]: I1203 14:10:11.163099 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T14:10:11Z","lastTransitionTime":"2025-12-03T14:10:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 14:10:11 crc kubenswrapper[4805]: I1203 14:10:11.265722 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 14:10:11 crc kubenswrapper[4805]: I1203 14:10:11.265763 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 14:10:11 crc kubenswrapper[4805]: I1203 14:10:11.265773 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 14:10:11 crc kubenswrapper[4805]: I1203 14:10:11.265789 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 14:10:11 crc kubenswrapper[4805]: I1203 14:10:11.265800 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T14:10:11Z","lastTransitionTime":"2025-12-03T14:10:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 14:10:11 crc kubenswrapper[4805]: I1203 14:10:11.369299 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 14:10:11 crc kubenswrapper[4805]: I1203 14:10:11.369335 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 14:10:11 crc kubenswrapper[4805]: I1203 14:10:11.369344 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 14:10:11 crc kubenswrapper[4805]: I1203 14:10:11.369358 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 14:10:11 crc kubenswrapper[4805]: I1203 14:10:11.369367 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T14:10:11Z","lastTransitionTime":"2025-12-03T14:10:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 14:10:11 crc kubenswrapper[4805]: I1203 14:10:11.472745 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 14:10:11 crc kubenswrapper[4805]: I1203 14:10:11.472869 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 14:10:11 crc kubenswrapper[4805]: I1203 14:10:11.472911 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 14:10:11 crc kubenswrapper[4805]: I1203 14:10:11.472943 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 14:10:11 crc kubenswrapper[4805]: I1203 14:10:11.472970 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T14:10:11Z","lastTransitionTime":"2025-12-03T14:10:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 14:10:11 crc kubenswrapper[4805]: I1203 14:10:11.574881 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 14:10:11 crc kubenswrapper[4805]: I1203 14:10:11.574935 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 14:10:11 crc kubenswrapper[4805]: I1203 14:10:11.574950 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 14:10:11 crc kubenswrapper[4805]: I1203 14:10:11.574971 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 14:10:11 crc kubenswrapper[4805]: I1203 14:10:11.574985 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T14:10:11Z","lastTransitionTime":"2025-12-03T14:10:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 14:10:11 crc kubenswrapper[4805]: I1203 14:10:11.677275 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 14:10:11 crc kubenswrapper[4805]: I1203 14:10:11.677321 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 14:10:11 crc kubenswrapper[4805]: I1203 14:10:11.677331 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 14:10:11 crc kubenswrapper[4805]: I1203 14:10:11.677345 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 14:10:11 crc kubenswrapper[4805]: I1203 14:10:11.677354 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T14:10:11Z","lastTransitionTime":"2025-12-03T14:10:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 14:10:11 crc kubenswrapper[4805]: I1203 14:10:11.779709 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 14:10:11 crc kubenswrapper[4805]: I1203 14:10:11.779771 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 14:10:11 crc kubenswrapper[4805]: I1203 14:10:11.779792 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 14:10:11 crc kubenswrapper[4805]: I1203 14:10:11.779817 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 14:10:11 crc kubenswrapper[4805]: I1203 14:10:11.779876 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T14:10:11Z","lastTransitionTime":"2025-12-03T14:10:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 14:10:11 crc kubenswrapper[4805]: I1203 14:10:11.882546 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 14:10:11 crc kubenswrapper[4805]: I1203 14:10:11.882585 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 14:10:11 crc kubenswrapper[4805]: I1203 14:10:11.882594 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 14:10:11 crc kubenswrapper[4805]: I1203 14:10:11.882608 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 14:10:11 crc kubenswrapper[4805]: I1203 14:10:11.882621 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T14:10:11Z","lastTransitionTime":"2025-12-03T14:10:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 14:10:11 crc kubenswrapper[4805]: I1203 14:10:11.984465 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 14:10:11 crc kubenswrapper[4805]: I1203 14:10:11.984526 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 14:10:11 crc kubenswrapper[4805]: I1203 14:10:11.984547 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 14:10:11 crc kubenswrapper[4805]: I1203 14:10:11.984574 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 14:10:11 crc kubenswrapper[4805]: I1203 14:10:11.984594 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T14:10:11Z","lastTransitionTime":"2025-12-03T14:10:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 14:10:12 crc kubenswrapper[4805]: I1203 14:10:12.088230 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 14:10:12 crc kubenswrapper[4805]: I1203 14:10:12.088310 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 14:10:12 crc kubenswrapper[4805]: I1203 14:10:12.088339 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 14:10:12 crc kubenswrapper[4805]: I1203 14:10:12.088375 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 14:10:12 crc kubenswrapper[4805]: I1203 14:10:12.088400 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T14:10:12Z","lastTransitionTime":"2025-12-03T14:10:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 14:10:12 crc kubenswrapper[4805]: I1203 14:10:12.191079 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 14:10:12 crc kubenswrapper[4805]: I1203 14:10:12.191237 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 14:10:12 crc kubenswrapper[4805]: I1203 14:10:12.191261 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 14:10:12 crc kubenswrapper[4805]: I1203 14:10:12.191285 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 14:10:12 crc kubenswrapper[4805]: I1203 14:10:12.191301 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T14:10:12Z","lastTransitionTime":"2025-12-03T14:10:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 14:10:12 crc kubenswrapper[4805]: I1203 14:10:12.294442 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 14:10:12 crc kubenswrapper[4805]: I1203 14:10:12.294500 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 14:10:12 crc kubenswrapper[4805]: I1203 14:10:12.294519 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 14:10:12 crc kubenswrapper[4805]: I1203 14:10:12.294543 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 14:10:12 crc kubenswrapper[4805]: I1203 14:10:12.294561 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T14:10:12Z","lastTransitionTime":"2025-12-03T14:10:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 14:10:12 crc kubenswrapper[4805]: I1203 14:10:12.318666 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 14:10:12 crc kubenswrapper[4805]: I1203 14:10:12.318749 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 14:10:12 crc kubenswrapper[4805]: I1203 14:10:12.318772 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 14:10:12 crc kubenswrapper[4805]: I1203 14:10:12.318904 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 14:10:12 crc kubenswrapper[4805]: I1203 14:10:12.318932 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T14:10:12Z","lastTransitionTime":"2025-12-03T14:10:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 14:10:12 crc kubenswrapper[4805]: E1203 14:10:12.337139 4805 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T14:10:12Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T14:10:12Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T14:10:12Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T14:10:12Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T14:10:12Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T14:10:12Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T14:10:12Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T14:10:12Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"e4edfbcd-1de1-486e-9c0b-65f9fbba52ad\\\",\\\"systemUUID\\\":\\\"7a729b24-19ef-490e-ab4c-d31c52ed5658\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T14:10:12Z is after 2025-08-24T17:21:41Z" Dec 03 14:10:12 crc kubenswrapper[4805]: I1203 14:10:12.342541 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 14:10:12 crc kubenswrapper[4805]: I1203 14:10:12.342603 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 14:10:12 crc kubenswrapper[4805]: I1203 14:10:12.342619 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 14:10:12 crc kubenswrapper[4805]: I1203 14:10:12.342644 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 14:10:12 crc kubenswrapper[4805]: I1203 14:10:12.342663 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T14:10:12Z","lastTransitionTime":"2025-12-03T14:10:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 14:10:12 crc kubenswrapper[4805]: E1203 14:10:12.357606 4805 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T14:10:12Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T14:10:12Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T14:10:12Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T14:10:12Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T14:10:12Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T14:10:12Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T14:10:12Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T14:10:12Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"e4edfbcd-1de1-486e-9c0b-65f9fbba52ad\\\",\\\"systemUUID\\\":\\\"7a729b24-19ef-490e-ab4c-d31c52ed5658\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T14:10:12Z is after 2025-08-24T17:21:41Z" Dec 03 14:10:12 crc kubenswrapper[4805]: I1203 14:10:12.362390 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 14:10:12 crc kubenswrapper[4805]: I1203 14:10:12.362453 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 14:10:12 crc kubenswrapper[4805]: I1203 14:10:12.362477 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 14:10:12 crc kubenswrapper[4805]: I1203 14:10:12.362502 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 14:10:12 crc kubenswrapper[4805]: I1203 14:10:12.362522 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T14:10:12Z","lastTransitionTime":"2025-12-03T14:10:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 14:10:12 crc kubenswrapper[4805]: E1203 14:10:12.382961 4805 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T14:10:12Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T14:10:12Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T14:10:12Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T14:10:12Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T14:10:12Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T14:10:12Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T14:10:12Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T14:10:12Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"e4edfbcd-1de1-486e-9c0b-65f9fbba52ad\\\",\\\"systemUUID\\\":\\\"7a729b24-19ef-490e-ab4c-d31c52ed5658\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T14:10:12Z is after 2025-08-24T17:21:41Z" Dec 03 14:10:12 crc kubenswrapper[4805]: I1203 14:10:12.387878 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 14:10:12 crc kubenswrapper[4805]: I1203 14:10:12.387926 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 14:10:12 crc kubenswrapper[4805]: I1203 14:10:12.387943 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 14:10:12 crc kubenswrapper[4805]: I1203 14:10:12.387965 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 14:10:12 crc kubenswrapper[4805]: I1203 14:10:12.387981 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T14:10:12Z","lastTransitionTime":"2025-12-03T14:10:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 14:10:12 crc kubenswrapper[4805]: I1203 14:10:12.394482 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/6b36a3db-1047-487a-9d6c-4f51ffdd8cc1-metrics-certs\") pod \"network-metrics-daemon-4gd9r\" (UID: \"6b36a3db-1047-487a-9d6c-4f51ffdd8cc1\") " pod="openshift-multus/network-metrics-daemon-4gd9r" Dec 03 14:10:12 crc kubenswrapper[4805]: E1203 14:10:12.394729 4805 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Dec 03 14:10:12 crc kubenswrapper[4805]: E1203 14:10:12.394899 4805 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/6b36a3db-1047-487a-9d6c-4f51ffdd8cc1-metrics-certs podName:6b36a3db-1047-487a-9d6c-4f51ffdd8cc1 nodeName:}" failed. No retries permitted until 2025-12-03 14:10:20.394877085 +0000 UTC m=+50.057794018 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/6b36a3db-1047-487a-9d6c-4f51ffdd8cc1-metrics-certs") pod "network-metrics-daemon-4gd9r" (UID: "6b36a3db-1047-487a-9d6c-4f51ffdd8cc1") : object "openshift-multus"/"metrics-daemon-secret" not registered Dec 03 14:10:12 crc kubenswrapper[4805]: E1203 14:10:12.408916 4805 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T14:10:12Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T14:10:12Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T14:10:12Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T14:10:12Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T14:10:12Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T14:10:12Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T14:10:12Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T14:10:12Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"e4edfbcd-1de1-486e-9c0b-65f9fbba52ad\\\",\\\"systemUUID\\\":\\\"7a729b24-19ef-490e-ab4c-d31c52ed5658\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T14:10:12Z is after 2025-08-24T17:21:41Z" Dec 03 14:10:12 crc kubenswrapper[4805]: I1203 14:10:12.413693 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 14:10:12 crc kubenswrapper[4805]: I1203 14:10:12.414004 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 14:10:12 crc kubenswrapper[4805]: I1203 14:10:12.414233 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 14:10:12 crc kubenswrapper[4805]: I1203 14:10:12.414456 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 14:10:12 crc kubenswrapper[4805]: I1203 14:10:12.414676 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T14:10:12Z","lastTransitionTime":"2025-12-03T14:10:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 14:10:12 crc kubenswrapper[4805]: E1203 14:10:12.434340 4805 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T14:10:12Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T14:10:12Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T14:10:12Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T14:10:12Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T14:10:12Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T14:10:12Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T14:10:12Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T14:10:12Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"e4edfbcd-1de1-486e-9c0b-65f9fbba52ad\\\",\\\"systemUUID\\\":\\\"7a729b24-19ef-490e-ab4c-d31c52ed5658\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T14:10:12Z is after 2025-08-24T17:21:41Z" Dec 03 14:10:12 crc kubenswrapper[4805]: E1203 14:10:12.434623 4805 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Dec 03 14:10:12 crc kubenswrapper[4805]: I1203 14:10:12.436432 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 14:10:12 crc kubenswrapper[4805]: I1203 14:10:12.436484 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 14:10:12 crc kubenswrapper[4805]: I1203 14:10:12.436499 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 14:10:12 crc kubenswrapper[4805]: I1203 14:10:12.436518 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 14:10:12 crc kubenswrapper[4805]: I1203 14:10:12.436530 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T14:10:12Z","lastTransitionTime":"2025-12-03T14:10:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 14:10:12 crc kubenswrapper[4805]: I1203 14:10:12.539369 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 14:10:12 crc kubenswrapper[4805]: I1203 14:10:12.539426 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 14:10:12 crc kubenswrapper[4805]: I1203 14:10:12.539444 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 14:10:12 crc kubenswrapper[4805]: I1203 14:10:12.539468 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 14:10:12 crc kubenswrapper[4805]: I1203 14:10:12.539485 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T14:10:12Z","lastTransitionTime":"2025-12-03T14:10:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 14:10:12 crc kubenswrapper[4805]: I1203 14:10:12.641915 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 14:10:12 crc kubenswrapper[4805]: I1203 14:10:12.641973 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 14:10:12 crc kubenswrapper[4805]: I1203 14:10:12.641989 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 14:10:12 crc kubenswrapper[4805]: I1203 14:10:12.642010 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 14:10:12 crc kubenswrapper[4805]: I1203 14:10:12.642025 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T14:10:12Z","lastTransitionTime":"2025-12-03T14:10:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 14:10:12 crc kubenswrapper[4805]: I1203 14:10:12.693904 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 14:10:12 crc kubenswrapper[4805]: I1203 14:10:12.693960 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-4gd9r" Dec 03 14:10:12 crc kubenswrapper[4805]: E1203 14:10:12.694114 4805 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 03 14:10:12 crc kubenswrapper[4805]: E1203 14:10:12.694301 4805 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-4gd9r" podUID="6b36a3db-1047-487a-9d6c-4f51ffdd8cc1" Dec 03 14:10:12 crc kubenswrapper[4805]: I1203 14:10:12.694536 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 14:10:12 crc kubenswrapper[4805]: E1203 14:10:12.694622 4805 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 03 14:10:12 crc kubenswrapper[4805]: I1203 14:10:12.694803 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 14:10:12 crc kubenswrapper[4805]: E1203 14:10:12.695178 4805 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 03 14:10:12 crc kubenswrapper[4805]: I1203 14:10:12.744756 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 14:10:12 crc kubenswrapper[4805]: I1203 14:10:12.744807 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 14:10:12 crc kubenswrapper[4805]: I1203 14:10:12.744824 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 14:10:12 crc kubenswrapper[4805]: I1203 14:10:12.744882 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 14:10:12 crc kubenswrapper[4805]: I1203 14:10:12.744901 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T14:10:12Z","lastTransitionTime":"2025-12-03T14:10:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 14:10:12 crc kubenswrapper[4805]: I1203 14:10:12.847465 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 14:10:12 crc kubenswrapper[4805]: I1203 14:10:12.847519 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 14:10:12 crc kubenswrapper[4805]: I1203 14:10:12.847536 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 14:10:12 crc kubenswrapper[4805]: I1203 14:10:12.847560 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 14:10:12 crc kubenswrapper[4805]: I1203 14:10:12.847583 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T14:10:12Z","lastTransitionTime":"2025-12-03T14:10:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 14:10:12 crc kubenswrapper[4805]: I1203 14:10:12.950677 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 14:10:12 crc kubenswrapper[4805]: I1203 14:10:12.950734 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 14:10:12 crc kubenswrapper[4805]: I1203 14:10:12.950751 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 14:10:12 crc kubenswrapper[4805]: I1203 14:10:12.950772 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 14:10:12 crc kubenswrapper[4805]: I1203 14:10:12.950790 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T14:10:12Z","lastTransitionTime":"2025-12-03T14:10:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 14:10:13 crc kubenswrapper[4805]: I1203 14:10:13.053680 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 14:10:13 crc kubenswrapper[4805]: I1203 14:10:13.053759 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 14:10:13 crc kubenswrapper[4805]: I1203 14:10:13.053786 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 14:10:13 crc kubenswrapper[4805]: I1203 14:10:13.053811 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 14:10:13 crc kubenswrapper[4805]: I1203 14:10:13.053828 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T14:10:13Z","lastTransitionTime":"2025-12-03T14:10:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 14:10:13 crc kubenswrapper[4805]: I1203 14:10:13.156446 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 14:10:13 crc kubenswrapper[4805]: I1203 14:10:13.156509 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 14:10:13 crc kubenswrapper[4805]: I1203 14:10:13.156524 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 14:10:13 crc kubenswrapper[4805]: I1203 14:10:13.156546 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 14:10:13 crc kubenswrapper[4805]: I1203 14:10:13.156567 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T14:10:13Z","lastTransitionTime":"2025-12-03T14:10:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 14:10:13 crc kubenswrapper[4805]: I1203 14:10:13.259644 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 14:10:13 crc kubenswrapper[4805]: I1203 14:10:13.259709 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 14:10:13 crc kubenswrapper[4805]: I1203 14:10:13.259725 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 14:10:13 crc kubenswrapper[4805]: I1203 14:10:13.259749 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 14:10:13 crc kubenswrapper[4805]: I1203 14:10:13.259766 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T14:10:13Z","lastTransitionTime":"2025-12-03T14:10:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 14:10:13 crc kubenswrapper[4805]: I1203 14:10:13.362685 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 14:10:13 crc kubenswrapper[4805]: I1203 14:10:13.362719 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 14:10:13 crc kubenswrapper[4805]: I1203 14:10:13.362728 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 14:10:13 crc kubenswrapper[4805]: I1203 14:10:13.362741 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 14:10:13 crc kubenswrapper[4805]: I1203 14:10:13.362751 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T14:10:13Z","lastTransitionTime":"2025-12-03T14:10:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 14:10:13 crc kubenswrapper[4805]: I1203 14:10:13.465117 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 14:10:13 crc kubenswrapper[4805]: I1203 14:10:13.465144 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 14:10:13 crc kubenswrapper[4805]: I1203 14:10:13.465154 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 14:10:13 crc kubenswrapper[4805]: I1203 14:10:13.465169 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 14:10:13 crc kubenswrapper[4805]: I1203 14:10:13.465179 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T14:10:13Z","lastTransitionTime":"2025-12-03T14:10:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 14:10:13 crc kubenswrapper[4805]: I1203 14:10:13.567610 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 14:10:13 crc kubenswrapper[4805]: I1203 14:10:13.567644 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 14:10:13 crc kubenswrapper[4805]: I1203 14:10:13.567655 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 14:10:13 crc kubenswrapper[4805]: I1203 14:10:13.567669 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 14:10:13 crc kubenswrapper[4805]: I1203 14:10:13.567679 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T14:10:13Z","lastTransitionTime":"2025-12-03T14:10:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 14:10:13 crc kubenswrapper[4805]: I1203 14:10:13.670506 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 14:10:13 crc kubenswrapper[4805]: I1203 14:10:13.670545 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 14:10:13 crc kubenswrapper[4805]: I1203 14:10:13.670556 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 14:10:13 crc kubenswrapper[4805]: I1203 14:10:13.670574 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 14:10:13 crc kubenswrapper[4805]: I1203 14:10:13.670584 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T14:10:13Z","lastTransitionTime":"2025-12-03T14:10:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 14:10:13 crc kubenswrapper[4805]: I1203 14:10:13.773455 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 14:10:13 crc kubenswrapper[4805]: I1203 14:10:13.773511 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 14:10:13 crc kubenswrapper[4805]: I1203 14:10:13.773530 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 14:10:13 crc kubenswrapper[4805]: I1203 14:10:13.773552 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 14:10:13 crc kubenswrapper[4805]: I1203 14:10:13.773569 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T14:10:13Z","lastTransitionTime":"2025-12-03T14:10:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 14:10:13 crc kubenswrapper[4805]: I1203 14:10:13.876172 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 14:10:13 crc kubenswrapper[4805]: I1203 14:10:13.876224 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 14:10:13 crc kubenswrapper[4805]: I1203 14:10:13.876240 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 14:10:13 crc kubenswrapper[4805]: I1203 14:10:13.876262 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 14:10:13 crc kubenswrapper[4805]: I1203 14:10:13.876280 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T14:10:13Z","lastTransitionTime":"2025-12-03T14:10:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 14:10:13 crc kubenswrapper[4805]: I1203 14:10:13.978525 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 14:10:13 crc kubenswrapper[4805]: I1203 14:10:13.978599 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 14:10:13 crc kubenswrapper[4805]: I1203 14:10:13.978624 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 14:10:13 crc kubenswrapper[4805]: I1203 14:10:13.978655 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 14:10:13 crc kubenswrapper[4805]: I1203 14:10:13.978678 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T14:10:13Z","lastTransitionTime":"2025-12-03T14:10:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 14:10:14 crc kubenswrapper[4805]: I1203 14:10:14.081945 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 14:10:14 crc kubenswrapper[4805]: I1203 14:10:14.082008 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 14:10:14 crc kubenswrapper[4805]: I1203 14:10:14.082033 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 14:10:14 crc kubenswrapper[4805]: I1203 14:10:14.082060 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 14:10:14 crc kubenswrapper[4805]: I1203 14:10:14.082081 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T14:10:14Z","lastTransitionTime":"2025-12-03T14:10:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 14:10:14 crc kubenswrapper[4805]: I1203 14:10:14.185025 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 14:10:14 crc kubenswrapper[4805]: I1203 14:10:14.185091 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 14:10:14 crc kubenswrapper[4805]: I1203 14:10:14.185106 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 14:10:14 crc kubenswrapper[4805]: I1203 14:10:14.185126 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 14:10:14 crc kubenswrapper[4805]: I1203 14:10:14.185139 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T14:10:14Z","lastTransitionTime":"2025-12-03T14:10:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 14:10:14 crc kubenswrapper[4805]: I1203 14:10:14.287364 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 14:10:14 crc kubenswrapper[4805]: I1203 14:10:14.287416 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 14:10:14 crc kubenswrapper[4805]: I1203 14:10:14.287428 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 14:10:14 crc kubenswrapper[4805]: I1203 14:10:14.287450 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 14:10:14 crc kubenswrapper[4805]: I1203 14:10:14.287466 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T14:10:14Z","lastTransitionTime":"2025-12-03T14:10:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 14:10:14 crc kubenswrapper[4805]: I1203 14:10:14.390138 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 14:10:14 crc kubenswrapper[4805]: I1203 14:10:14.390198 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 14:10:14 crc kubenswrapper[4805]: I1203 14:10:14.390212 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 14:10:14 crc kubenswrapper[4805]: I1203 14:10:14.390228 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 14:10:14 crc kubenswrapper[4805]: I1203 14:10:14.390237 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T14:10:14Z","lastTransitionTime":"2025-12-03T14:10:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 14:10:14 crc kubenswrapper[4805]: I1203 14:10:14.493311 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 14:10:14 crc kubenswrapper[4805]: I1203 14:10:14.493379 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 14:10:14 crc kubenswrapper[4805]: I1203 14:10:14.493398 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 14:10:14 crc kubenswrapper[4805]: I1203 14:10:14.493454 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 14:10:14 crc kubenswrapper[4805]: I1203 14:10:14.493472 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T14:10:14Z","lastTransitionTime":"2025-12-03T14:10:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 14:10:14 crc kubenswrapper[4805]: I1203 14:10:14.595670 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 14:10:14 crc kubenswrapper[4805]: I1203 14:10:14.595718 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 14:10:14 crc kubenswrapper[4805]: I1203 14:10:14.595729 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 14:10:14 crc kubenswrapper[4805]: I1203 14:10:14.595745 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 14:10:14 crc kubenswrapper[4805]: I1203 14:10:14.595757 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T14:10:14Z","lastTransitionTime":"2025-12-03T14:10:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 14:10:14 crc kubenswrapper[4805]: I1203 14:10:14.694200 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 14:10:14 crc kubenswrapper[4805]: I1203 14:10:14.694355 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-4gd9r" Dec 03 14:10:14 crc kubenswrapper[4805]: I1203 14:10:14.694516 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 14:10:14 crc kubenswrapper[4805]: I1203 14:10:14.694549 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 14:10:14 crc kubenswrapper[4805]: E1203 14:10:14.694491 4805 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 03 14:10:14 crc kubenswrapper[4805]: E1203 14:10:14.694689 4805 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 03 14:10:14 crc kubenswrapper[4805]: E1203 14:10:14.694745 4805 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 03 14:10:14 crc kubenswrapper[4805]: E1203 14:10:14.694830 4805 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-4gd9r" podUID="6b36a3db-1047-487a-9d6c-4f51ffdd8cc1" Dec 03 14:10:14 crc kubenswrapper[4805]: I1203 14:10:14.697863 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 14:10:14 crc kubenswrapper[4805]: I1203 14:10:14.697887 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 14:10:14 crc kubenswrapper[4805]: I1203 14:10:14.697917 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 14:10:14 crc kubenswrapper[4805]: I1203 14:10:14.697932 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 14:10:14 crc kubenswrapper[4805]: I1203 14:10:14.697941 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T14:10:14Z","lastTransitionTime":"2025-12-03T14:10:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 14:10:14 crc kubenswrapper[4805]: I1203 14:10:14.799735 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 14:10:14 crc kubenswrapper[4805]: I1203 14:10:14.800202 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 14:10:14 crc kubenswrapper[4805]: I1203 14:10:14.800352 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 14:10:14 crc kubenswrapper[4805]: I1203 14:10:14.800486 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 14:10:14 crc kubenswrapper[4805]: I1203 14:10:14.800620 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T14:10:14Z","lastTransitionTime":"2025-12-03T14:10:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 14:10:14 crc kubenswrapper[4805]: I1203 14:10:14.903254 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 14:10:14 crc kubenswrapper[4805]: I1203 14:10:14.903313 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 14:10:14 crc kubenswrapper[4805]: I1203 14:10:14.903328 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 14:10:14 crc kubenswrapper[4805]: I1203 14:10:14.903348 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 14:10:14 crc kubenswrapper[4805]: I1203 14:10:14.903516 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T14:10:14Z","lastTransitionTime":"2025-12-03T14:10:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 14:10:15 crc kubenswrapper[4805]: I1203 14:10:15.006478 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 14:10:15 crc kubenswrapper[4805]: I1203 14:10:15.006532 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 14:10:15 crc kubenswrapper[4805]: I1203 14:10:15.006548 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 14:10:15 crc kubenswrapper[4805]: I1203 14:10:15.006573 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 14:10:15 crc kubenswrapper[4805]: I1203 14:10:15.006594 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T14:10:15Z","lastTransitionTime":"2025-12-03T14:10:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 14:10:15 crc kubenswrapper[4805]: I1203 14:10:15.109229 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 14:10:15 crc kubenswrapper[4805]: I1203 14:10:15.109311 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 14:10:15 crc kubenswrapper[4805]: I1203 14:10:15.109334 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 14:10:15 crc kubenswrapper[4805]: I1203 14:10:15.109363 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 14:10:15 crc kubenswrapper[4805]: I1203 14:10:15.109387 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T14:10:15Z","lastTransitionTime":"2025-12-03T14:10:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 14:10:15 crc kubenswrapper[4805]: I1203 14:10:15.212643 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 14:10:15 crc kubenswrapper[4805]: I1203 14:10:15.212692 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 14:10:15 crc kubenswrapper[4805]: I1203 14:10:15.212709 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 14:10:15 crc kubenswrapper[4805]: I1203 14:10:15.212731 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 14:10:15 crc kubenswrapper[4805]: I1203 14:10:15.212747 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T14:10:15Z","lastTransitionTime":"2025-12-03T14:10:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 14:10:15 crc kubenswrapper[4805]: I1203 14:10:15.316358 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 14:10:15 crc kubenswrapper[4805]: I1203 14:10:15.316424 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 14:10:15 crc kubenswrapper[4805]: I1203 14:10:15.316441 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 14:10:15 crc kubenswrapper[4805]: I1203 14:10:15.316465 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 14:10:15 crc kubenswrapper[4805]: I1203 14:10:15.316482 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T14:10:15Z","lastTransitionTime":"2025-12-03T14:10:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 14:10:15 crc kubenswrapper[4805]: I1203 14:10:15.419261 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 14:10:15 crc kubenswrapper[4805]: I1203 14:10:15.419320 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 14:10:15 crc kubenswrapper[4805]: I1203 14:10:15.419338 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 14:10:15 crc kubenswrapper[4805]: I1203 14:10:15.419363 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 14:10:15 crc kubenswrapper[4805]: I1203 14:10:15.419381 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T14:10:15Z","lastTransitionTime":"2025-12-03T14:10:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 14:10:15 crc kubenswrapper[4805]: I1203 14:10:15.522456 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 14:10:15 crc kubenswrapper[4805]: I1203 14:10:15.522510 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 14:10:15 crc kubenswrapper[4805]: I1203 14:10:15.522523 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 14:10:15 crc kubenswrapper[4805]: I1203 14:10:15.522541 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 14:10:15 crc kubenswrapper[4805]: I1203 14:10:15.522553 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T14:10:15Z","lastTransitionTime":"2025-12-03T14:10:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 14:10:15 crc kubenswrapper[4805]: I1203 14:10:15.625167 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 14:10:15 crc kubenswrapper[4805]: I1203 14:10:15.625212 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 14:10:15 crc kubenswrapper[4805]: I1203 14:10:15.625223 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 14:10:15 crc kubenswrapper[4805]: I1203 14:10:15.625239 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 14:10:15 crc kubenswrapper[4805]: I1203 14:10:15.625251 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T14:10:15Z","lastTransitionTime":"2025-12-03T14:10:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 14:10:15 crc kubenswrapper[4805]: I1203 14:10:15.694770 4805 scope.go:117] "RemoveContainer" containerID="644e56c400524d739e2a208fb7eee1551ebbd28ba220df91c9dd84ba09d0b400" Dec 03 14:10:15 crc kubenswrapper[4805]: I1203 14:10:15.727612 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 14:10:15 crc kubenswrapper[4805]: I1203 14:10:15.727659 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 14:10:15 crc kubenswrapper[4805]: I1203 14:10:15.727668 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 14:10:15 crc kubenswrapper[4805]: I1203 14:10:15.727686 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 14:10:15 crc kubenswrapper[4805]: I1203 14:10:15.727699 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T14:10:15Z","lastTransitionTime":"2025-12-03T14:10:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 14:10:15 crc kubenswrapper[4805]: I1203 14:10:15.830318 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 14:10:15 crc kubenswrapper[4805]: I1203 14:10:15.830721 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 14:10:15 crc kubenswrapper[4805]: I1203 14:10:15.830734 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 14:10:15 crc kubenswrapper[4805]: I1203 14:10:15.830753 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 14:10:15 crc kubenswrapper[4805]: I1203 14:10:15.830766 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T14:10:15Z","lastTransitionTime":"2025-12-03T14:10:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 14:10:15 crc kubenswrapper[4805]: I1203 14:10:15.934047 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 14:10:15 crc kubenswrapper[4805]: I1203 14:10:15.934098 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 14:10:15 crc kubenswrapper[4805]: I1203 14:10:15.934110 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 14:10:15 crc kubenswrapper[4805]: I1203 14:10:15.934126 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 14:10:15 crc kubenswrapper[4805]: I1203 14:10:15.934149 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T14:10:15Z","lastTransitionTime":"2025-12-03T14:10:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 14:10:15 crc kubenswrapper[4805]: I1203 14:10:15.999583 4805 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-psjpm_deeec30d-cad2-4dfd-adb6-64f6646581be/ovnkube-controller/1.log" Dec 03 14:10:16 crc kubenswrapper[4805]: I1203 14:10:16.003388 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-psjpm" event={"ID":"deeec30d-cad2-4dfd-adb6-64f6646581be","Type":"ContainerStarted","Data":"a0272667e733f2c5eec0fce0eb9c31a16c048babdaed90635895d689ae9303f4"} Dec 03 14:10:16 crc kubenswrapper[4805]: I1203 14:10:16.003582 4805 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Dec 03 14:10:16 crc kubenswrapper[4805]: I1203 14:10:16.029296 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:48Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:48Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T14:10:16Z is after 2025-08-24T17:21:41Z" Dec 03 14:10:16 crc kubenswrapper[4805]: I1203 14:10:16.036403 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 14:10:16 crc kubenswrapper[4805]: I1203 14:10:16.036451 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 14:10:16 crc kubenswrapper[4805]: I1203 14:10:16.036468 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 14:10:16 crc kubenswrapper[4805]: I1203 14:10:16.036492 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 14:10:16 crc kubenswrapper[4805]: I1203 14:10:16.036509 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T14:10:16Z","lastTransitionTime":"2025-12-03T14:10:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 14:10:16 crc kubenswrapper[4805]: I1203 14:10:16.051877 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:49Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:49Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://541ace7ccbb66dbd191c3ca8c1e4479e3edbf80dfc67326bea8aaee81fdcfc3c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://821e3e9a902aacecf6c41e8b8d8aa1ac951ce2fbb6505b7148b16348a3754e55\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T14:10:16Z is after 2025-08-24T17:21:41Z" Dec 03 14:10:16 crc kubenswrapper[4805]: I1203 14:10:16.068467 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-gskh4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6098937f-e3f6-45e8-a647-4994a79cd711\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://28d5653f26ec61647808b3a733ff7eac70aa69ac184660f24147e8cfaf0eb802\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nwrw9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9ad6b68a42fd33ee167e67873c6eed3f11cdf6dc71d22e952c021436eb967dd9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nwrw9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T14:09:50Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-gskh4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T14:10:16Z is after 2025-08-24T17:21:41Z" Dec 03 14:10:16 crc kubenswrapper[4805]: I1203 14:10:16.086677 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-psjpm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"deeec30d-cad2-4dfd-adb6-64f6646581be\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:50Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:50Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4b8d00cdf0f37f2e5cd4e952cf85d3de8d72c43fdf786088fd9f26ed17b1e21f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wdlzx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bd67c6de01a8d196a1f51d4d90640c9e69de6f05d5e201b25e6eaa5c33440ea8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wdlzx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cb8bf05ef093eb0727a066d4bbbdf9af9422867cd4dadacafc6eeb016867a61e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wdlzx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0304467102ed053628c0d3e06548c74b6492336b6fd044246508d1e8bdb17bc3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wdlzx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6ce80ccafb3262a11fbe9923317a9d38b61e5dc5b3e7e7ae136f143d6b643fd7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wdlzx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://47d76962beb8c0f196d50af96fb3ec611097823ca80a1c8bd718f2d18840128c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wdlzx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a0272667e733f2c5eec0fce0eb9c31a16c048babdaed90635895d689ae9303f4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://644e56c400524d739e2a208fb7eee1551ebbd28ba220df91c9dd84ba09d0b400\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-03T14:10:02Z\\\",\\\"message\\\":\\\"tring(nil)}, nodeEndpoints:map[string]services.lbEndpoints{}, externalTrafficLocal:false, internalTrafficLocal:false, hasNodePort:false}}\\\\nF1203 14:10:01.661807 6255 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T14:10:01Z is after 2025-08-24T17:21:41Z]\\\\nI1203 14:10:01.660569 6255 services_controller.go:451] Built service openshift-operator-lifecycle-manager/package-server-manager-metrics cluster-wide LB for network=default: []services.LB{services.LB{Name:\\\\\\\"Service_openshift-operator-lifecycle-manager/package-server-manager-metrics_TCP_cluster\\\\\\\", UUID:\\\\\\\"\\\\\\\", Protocol:\\\\\\\"TCP\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Serv\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T14:10:01Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:10:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wdlzx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ab98e365c1a4242a47b8cbe2dd85109d366d6002c4de1f18ddc950a46bea6c12\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wdlzx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2f875252fbe6a673a72fc9c56bb96899f62732305ad77b3108e3b244bfa91ecb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2f875252fbe6a673a72fc9c56bb96899f62732305ad77b3108e3b244bfa91ecb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T14:09:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T14:09:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wdlzx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T14:09:50Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-psjpm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T14:10:16Z is after 2025-08-24T17:21:41Z" Dec 03 14:10:16 crc kubenswrapper[4805]: I1203 14:10:16.103232 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-4gd9r" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6b36a3db-1047-487a-9d6c-4f51ffdd8cc1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:10:04Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:10:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:10:04Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:10:04Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4t6zl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4t6zl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T14:10:04Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-4gd9r\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T14:10:16Z is after 2025-08-24T17:21:41Z" Dec 03 14:10:16 crc kubenswrapper[4805]: I1203 14:10:16.121121 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-pdzgk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"860769de-6a5b-4cb9-a752-264e7e2e785f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5e2f0a0cc21505d609f81918365d5e4f5d21959ac12ab147d6a51601c9de4c14\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ggn4n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T14:09:50Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-pdzgk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T14:10:16Z is after 2025-08-24T17:21:41Z" Dec 03 14:10:16 crc kubenswrapper[4805]: I1203 14:10:16.135529 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-69h75" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5d1c694f-9847-4086-8b37-3573ddeb3ace\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:10:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:10:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:10:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:10:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://abc1d9b8582ca9fe4a51fb608fb39129a179ddcdfde0797a6369a55b2e0124cf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:10:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s578b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c7c4d434b035cac2a91a4f0846a2f4fb5bd68f881c28487860adb6db85e8ce75\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:10:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s578b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T14:10:03Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-69h75\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T14:10:16Z is after 2025-08-24T17:21:41Z" Dec 03 14:10:16 crc kubenswrapper[4805]: I1203 14:10:16.138436 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 14:10:16 crc kubenswrapper[4805]: I1203 14:10:16.138467 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 14:10:16 crc kubenswrapper[4805]: I1203 14:10:16.138478 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 14:10:16 crc kubenswrapper[4805]: I1203 14:10:16.138493 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 14:10:16 crc kubenswrapper[4805]: I1203 14:10:16.138504 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T14:10:16Z","lastTransitionTime":"2025-12-03T14:10:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 14:10:16 crc kubenswrapper[4805]: I1203 14:10:16.155187 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b6d36671-7173-4e9c-843f-005b4dcf576c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a837d7d491eaacbe4279b4014bb5fcbe9b514ba5cc82c708227bcd9b25a346e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6e5f04faf3b23197bb8ea11d3e574dc86f31e21d798734962001ec8c95de7c84\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2ff1a979387da41b00e95eb9cc8b2c9de966380a12fae687591df3c5b18a4497\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1e66e1cb18573a14e3c3ae16af4dbd9f8997a6ad38680c3ab4505646ebd48e7c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7d5b5c85e3de43e7e95111aa967248550607c26200c783251b68696f202f92a1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://474f72dac37a808fc4b9832306e76979b151d7221807ff3086013466ed1419c3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://474f72dac37a808fc4b9832306e76979b151d7221807ff3086013466ed1419c3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T14:09:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T14:09:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://74c8cea1a61c4775a0ad2f4ab4bb8a36437cf23d2ee953311d6ea967e297510a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://74c8cea1a61c4775a0ad2f4ab4bb8a36437cf23d2ee953311d6ea967e297510a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T14:09:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T14:09:31Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://a4f2f56a809adc01fba82eba0652d78f58770c132f77b1424f8b17068b1927e3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a4f2f56a809adc01fba82eba0652d78f58770c132f77b1424f8b17068b1927e3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T14:09:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T14:09:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T14:09:30Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T14:10:16Z is after 2025-08-24T17:21:41Z" Dec 03 14:10:16 crc kubenswrapper[4805]: I1203 14:10:16.172554 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ced8291d-0ce5-4e74-9eaf-2bf033774039\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:10:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:10:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3e5a34636f0a8c4d9179074596d28bc2eb7a504573398c596a637defd829907a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://29cd95be3f964415943c873e7914ffcea4abe50dfdf17f8a9cd03fc38df35981\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://951173b3d3d581275a6f6e43209b5e0611a701ca7a10f81504d723e24a2d21a3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e64730957bbbb2feb4d506e0ed9731a0594bf66fa1e8fcdb8ec944ded6308eed\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5da55f769e23ad52bac9c77da62c78e70fa34db8ed1a435e4eaf4dad3fd2b68a\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-03T14:09:48Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1203 14:09:42.992179 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1203 14:09:42.993233 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1830926909/tls.crt::/tmp/serving-cert-1830926909/tls.key\\\\\\\"\\\\nI1203 14:09:48.622403 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1203 14:09:48.626476 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1203 14:09:48.626527 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1203 14:09:48.626574 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1203 14:09:48.626582 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1203 14:09:48.633990 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1203 14:09:48.634024 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 14:09:48.634029 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 14:09:48.634034 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1203 14:09:48.634038 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1203 14:09:48.634042 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1203 14:09:48.634045 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1203 14:09:48.634344 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1203 14:09:48.644497 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T14:09:32Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bd92a160eb530c5687ebd7cd5b83b41820a2b808a0fd9514e5ace9d237e1bd26\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:32Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://af9c940135012c629447528c61743ed1e58708ab57ce76604c6da0479aa459cf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://af9c940135012c629447528c61743ed1e58708ab57ce76604c6da0479aa459cf\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T14:09:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T14:09:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T14:09:30Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T14:10:16Z is after 2025-08-24T17:21:41Z" Dec 03 14:10:16 crc kubenswrapper[4805]: I1203 14:10:16.188397 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:49Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:49Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://38339331a3209d6787d1def81b72f4a8514bf91080a139cad6d8ac5cc6bef10e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T14:10:16Z is after 2025-08-24T17:21:41Z" Dec 03 14:10:16 crc kubenswrapper[4805]: I1203 14:10:16.204997 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:48Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:48Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T14:10:16Z is after 2025-08-24T17:21:41Z" Dec 03 14:10:16 crc kubenswrapper[4805]: I1203 14:10:16.225615 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:51Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:51Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b3123d3fa879097db88ed2f3cfb05631f440c24c5405503b2de3546f12e2e899\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T14:10:16Z is after 2025-08-24T17:21:41Z" Dec 03 14:10:16 crc kubenswrapper[4805]: I1203 14:10:16.240875 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 14:10:16 crc kubenswrapper[4805]: I1203 14:10:16.240937 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 14:10:16 crc kubenswrapper[4805]: I1203 14:10:16.240953 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 14:10:16 crc kubenswrapper[4805]: I1203 14:10:16.240980 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 14:10:16 crc kubenswrapper[4805]: I1203 14:10:16.240997 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T14:10:16Z","lastTransitionTime":"2025-12-03T14:10:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 14:10:16 crc kubenswrapper[4805]: I1203 14:10:16.250417 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-bdtdg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"53a0e9e3-b6d8-44b7-a32b-aa3450442865\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f557e17ad105d6972d3c1480656c6073f67e28c0f5773aa365549951f4d161c7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9ww54\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T14:09:51Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-bdtdg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T14:10:16Z is after 2025-08-24T17:21:41Z" Dec 03 14:10:16 crc kubenswrapper[4805]: I1203 14:10:16.275493 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"750d2b4b-738a-4992-8902-742528f6e875\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://abd2d0156113bb5182727a53aac594410ed051c19258beb8a07b808eff0b42a7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8673a29c0e16f4e707d79f0610722ccabb023ff2b4a18d7cf9d305183817d952\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://35766e7206b1be7a53d13d6890d759cf81dbd999643c12349268e4bf16fbbdb0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5b9844a11b2b903978d0d464b469e26c9211100da61d018dde9636dca95b684e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T14:09:30Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T14:10:16Z is after 2025-08-24T17:21:41Z" Dec 03 14:10:16 crc kubenswrapper[4805]: I1203 14:10:16.315334 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:48Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:48Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T14:10:16Z is after 2025-08-24T17:21:41Z" Dec 03 14:10:16 crc kubenswrapper[4805]: I1203 14:10:16.335835 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-k4mhs" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c09c1cdc-f461-45c9-8444-5b8764bccabc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3354c1a44f891182d04116fc76234a2b337d50567d032a3584f902a9516387cc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qdcd7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T14:09:50Z\\\"}}\" for pod \"openshift-multus\"/\"multus-k4mhs\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T14:10:16Z is after 2025-08-24T17:21:41Z" Dec 03 14:10:16 crc kubenswrapper[4805]: I1203 14:10:16.342923 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 14:10:16 crc kubenswrapper[4805]: I1203 14:10:16.342995 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 14:10:16 crc kubenswrapper[4805]: I1203 14:10:16.343006 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 14:10:16 crc kubenswrapper[4805]: I1203 14:10:16.343051 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 14:10:16 crc kubenswrapper[4805]: I1203 14:10:16.343069 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T14:10:16Z","lastTransitionTime":"2025-12-03T14:10:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 14:10:16 crc kubenswrapper[4805]: I1203 14:10:16.351138 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-bpk49" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"de00a81f-82e1-4223-9614-f98606b9a8ef\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://04f4ad27e02e2024ae9e7cd1b16116267e69cecda882dd723cdb274c41a0e7db\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-25d7t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7741e0e70416ebe094bfabc6009e98d7d4c0fbcba1fc82df13b9ba40774b456a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7741e0e70416ebe094bfabc6009e98d7d4c0fbcba1fc82df13b9ba40774b456a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T14:09:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T14:09:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-25d7t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ee3499c02eb52bbf08c447d2433d435a74dfb70670dc4cbaac44f5fbff9e8a29\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ee3499c02eb52bbf08c447d2433d435a74dfb70670dc4cbaac44f5fbff9e8a29\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T14:09:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T14:09:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-25d7t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ec8488b1d69799951be7ef40e5e028732265c4ddcdea2d453923e567864771f7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ec8488b1d69799951be7ef40e5e028732265c4ddcdea2d453923e567864771f7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T14:09:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T14:09:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-25d7t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://66707bb4e1fdacb265e89786df8a04605830f1c156394cacfcec9b54abe5d41a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://66707bb4e1fdacb265e89786df8a04605830f1c156394cacfcec9b54abe5d41a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T14:09:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T14:09:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-25d7t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6207ab6600970ff15c8713c547d4920b2c77eb9bea990bb5da497e24870dcc0f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6207ab6600970ff15c8713c547d4920b2c77eb9bea990bb5da497e24870dcc0f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T14:09:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T14:09:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-25d7t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8de65d4a205e2deec0b433e53c0e37feb94aa0ede6a4c58d7414b80829804505\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8de65d4a205e2deec0b433e53c0e37feb94aa0ede6a4c58d7414b80829804505\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T14:09:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T14:09:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-25d7t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T14:09:50Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-bpk49\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T14:10:16Z is after 2025-08-24T17:21:41Z" Dec 03 14:10:16 crc kubenswrapper[4805]: I1203 14:10:16.445780 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 14:10:16 crc kubenswrapper[4805]: I1203 14:10:16.445812 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 14:10:16 crc kubenswrapper[4805]: I1203 14:10:16.445823 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 14:10:16 crc kubenswrapper[4805]: I1203 14:10:16.445836 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 14:10:16 crc kubenswrapper[4805]: I1203 14:10:16.445847 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T14:10:16Z","lastTransitionTime":"2025-12-03T14:10:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 14:10:16 crc kubenswrapper[4805]: I1203 14:10:16.548568 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 14:10:16 crc kubenswrapper[4805]: I1203 14:10:16.548605 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 14:10:16 crc kubenswrapper[4805]: I1203 14:10:16.548616 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 14:10:16 crc kubenswrapper[4805]: I1203 14:10:16.548633 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 14:10:16 crc kubenswrapper[4805]: I1203 14:10:16.548645 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T14:10:16Z","lastTransitionTime":"2025-12-03T14:10:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 14:10:16 crc kubenswrapper[4805]: I1203 14:10:16.650978 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 14:10:16 crc kubenswrapper[4805]: I1203 14:10:16.651018 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 14:10:16 crc kubenswrapper[4805]: I1203 14:10:16.651029 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 14:10:16 crc kubenswrapper[4805]: I1203 14:10:16.651044 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 14:10:16 crc kubenswrapper[4805]: I1203 14:10:16.651055 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T14:10:16Z","lastTransitionTime":"2025-12-03T14:10:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 14:10:16 crc kubenswrapper[4805]: I1203 14:10:16.714938 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 14:10:16 crc kubenswrapper[4805]: I1203 14:10:16.714995 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 14:10:16 crc kubenswrapper[4805]: I1203 14:10:16.715027 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 14:10:16 crc kubenswrapper[4805]: I1203 14:10:16.714956 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-4gd9r" Dec 03 14:10:16 crc kubenswrapper[4805]: E1203 14:10:16.715086 4805 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 03 14:10:16 crc kubenswrapper[4805]: E1203 14:10:16.715212 4805 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 03 14:10:16 crc kubenswrapper[4805]: E1203 14:10:16.715274 4805 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 03 14:10:16 crc kubenswrapper[4805]: E1203 14:10:16.715355 4805 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-4gd9r" podUID="6b36a3db-1047-487a-9d6c-4f51ffdd8cc1" Dec 03 14:10:16 crc kubenswrapper[4805]: I1203 14:10:16.753421 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 14:10:16 crc kubenswrapper[4805]: I1203 14:10:16.753453 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 14:10:16 crc kubenswrapper[4805]: I1203 14:10:16.753462 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 14:10:16 crc kubenswrapper[4805]: I1203 14:10:16.753478 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 14:10:16 crc kubenswrapper[4805]: I1203 14:10:16.753488 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T14:10:16Z","lastTransitionTime":"2025-12-03T14:10:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 14:10:16 crc kubenswrapper[4805]: I1203 14:10:16.855414 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 14:10:16 crc kubenswrapper[4805]: I1203 14:10:16.855446 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 14:10:16 crc kubenswrapper[4805]: I1203 14:10:16.855456 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 14:10:16 crc kubenswrapper[4805]: I1203 14:10:16.855472 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 14:10:16 crc kubenswrapper[4805]: I1203 14:10:16.855483 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T14:10:16Z","lastTransitionTime":"2025-12-03T14:10:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 14:10:16 crc kubenswrapper[4805]: I1203 14:10:16.959138 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 14:10:16 crc kubenswrapper[4805]: I1203 14:10:16.959199 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 14:10:16 crc kubenswrapper[4805]: I1203 14:10:16.959216 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 14:10:16 crc kubenswrapper[4805]: I1203 14:10:16.959248 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 14:10:16 crc kubenswrapper[4805]: I1203 14:10:16.959263 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T14:10:16Z","lastTransitionTime":"2025-12-03T14:10:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 14:10:17 crc kubenswrapper[4805]: I1203 14:10:17.062494 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 14:10:17 crc kubenswrapper[4805]: I1203 14:10:17.062568 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 14:10:17 crc kubenswrapper[4805]: I1203 14:10:17.062604 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 14:10:17 crc kubenswrapper[4805]: I1203 14:10:17.062630 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 14:10:17 crc kubenswrapper[4805]: I1203 14:10:17.062650 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T14:10:17Z","lastTransitionTime":"2025-12-03T14:10:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 14:10:17 crc kubenswrapper[4805]: I1203 14:10:17.165705 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 14:10:17 crc kubenswrapper[4805]: I1203 14:10:17.165795 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 14:10:17 crc kubenswrapper[4805]: I1203 14:10:17.165813 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 14:10:17 crc kubenswrapper[4805]: I1203 14:10:17.165864 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 14:10:17 crc kubenswrapper[4805]: I1203 14:10:17.165880 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T14:10:17Z","lastTransitionTime":"2025-12-03T14:10:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 14:10:17 crc kubenswrapper[4805]: I1203 14:10:17.269039 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 14:10:17 crc kubenswrapper[4805]: I1203 14:10:17.269195 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 14:10:17 crc kubenswrapper[4805]: I1203 14:10:17.269214 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 14:10:17 crc kubenswrapper[4805]: I1203 14:10:17.269274 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 14:10:17 crc kubenswrapper[4805]: I1203 14:10:17.269293 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T14:10:17Z","lastTransitionTime":"2025-12-03T14:10:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 14:10:17 crc kubenswrapper[4805]: I1203 14:10:17.372537 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 14:10:17 crc kubenswrapper[4805]: I1203 14:10:17.373211 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 14:10:17 crc kubenswrapper[4805]: I1203 14:10:17.373253 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 14:10:17 crc kubenswrapper[4805]: I1203 14:10:17.373284 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 14:10:17 crc kubenswrapper[4805]: I1203 14:10:17.373302 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T14:10:17Z","lastTransitionTime":"2025-12-03T14:10:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 14:10:17 crc kubenswrapper[4805]: I1203 14:10:17.477461 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 14:10:17 crc kubenswrapper[4805]: I1203 14:10:17.477572 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 14:10:17 crc kubenswrapper[4805]: I1203 14:10:17.477593 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 14:10:17 crc kubenswrapper[4805]: I1203 14:10:17.477618 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 14:10:17 crc kubenswrapper[4805]: I1203 14:10:17.477713 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T14:10:17Z","lastTransitionTime":"2025-12-03T14:10:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 14:10:17 crc kubenswrapper[4805]: I1203 14:10:17.581009 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 14:10:17 crc kubenswrapper[4805]: I1203 14:10:17.581095 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 14:10:17 crc kubenswrapper[4805]: I1203 14:10:17.581107 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 14:10:17 crc kubenswrapper[4805]: I1203 14:10:17.581340 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 14:10:17 crc kubenswrapper[4805]: I1203 14:10:17.581359 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T14:10:17Z","lastTransitionTime":"2025-12-03T14:10:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 14:10:17 crc kubenswrapper[4805]: I1203 14:10:17.684307 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 14:10:17 crc kubenswrapper[4805]: I1203 14:10:17.684355 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 14:10:17 crc kubenswrapper[4805]: I1203 14:10:17.684364 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 14:10:17 crc kubenswrapper[4805]: I1203 14:10:17.684381 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 14:10:17 crc kubenswrapper[4805]: I1203 14:10:17.684391 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T14:10:17Z","lastTransitionTime":"2025-12-03T14:10:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 14:10:17 crc kubenswrapper[4805]: I1203 14:10:17.786256 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 14:10:17 crc kubenswrapper[4805]: I1203 14:10:17.786351 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 14:10:17 crc kubenswrapper[4805]: I1203 14:10:17.786365 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 14:10:17 crc kubenswrapper[4805]: I1203 14:10:17.786381 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 14:10:17 crc kubenswrapper[4805]: I1203 14:10:17.786392 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T14:10:17Z","lastTransitionTime":"2025-12-03T14:10:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 14:10:17 crc kubenswrapper[4805]: I1203 14:10:17.888998 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 14:10:17 crc kubenswrapper[4805]: I1203 14:10:17.889070 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 14:10:17 crc kubenswrapper[4805]: I1203 14:10:17.889112 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 14:10:17 crc kubenswrapper[4805]: I1203 14:10:17.889145 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 14:10:17 crc kubenswrapper[4805]: I1203 14:10:17.889169 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T14:10:17Z","lastTransitionTime":"2025-12-03T14:10:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 14:10:17 crc kubenswrapper[4805]: I1203 14:10:17.992577 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 14:10:17 crc kubenswrapper[4805]: I1203 14:10:17.992628 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 14:10:17 crc kubenswrapper[4805]: I1203 14:10:17.992639 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 14:10:17 crc kubenswrapper[4805]: I1203 14:10:17.992656 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 14:10:17 crc kubenswrapper[4805]: I1203 14:10:17.992667 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T14:10:17Z","lastTransitionTime":"2025-12-03T14:10:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 14:10:18 crc kubenswrapper[4805]: I1203 14:10:18.013428 4805 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-psjpm_deeec30d-cad2-4dfd-adb6-64f6646581be/ovnkube-controller/2.log" Dec 03 14:10:18 crc kubenswrapper[4805]: I1203 14:10:18.014163 4805 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-psjpm_deeec30d-cad2-4dfd-adb6-64f6646581be/ovnkube-controller/1.log" Dec 03 14:10:18 crc kubenswrapper[4805]: I1203 14:10:18.017581 4805 generic.go:334] "Generic (PLEG): container finished" podID="deeec30d-cad2-4dfd-adb6-64f6646581be" containerID="a0272667e733f2c5eec0fce0eb9c31a16c048babdaed90635895d689ae9303f4" exitCode=1 Dec 03 14:10:18 crc kubenswrapper[4805]: I1203 14:10:18.017622 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-psjpm" event={"ID":"deeec30d-cad2-4dfd-adb6-64f6646581be","Type":"ContainerDied","Data":"a0272667e733f2c5eec0fce0eb9c31a16c048babdaed90635895d689ae9303f4"} Dec 03 14:10:18 crc kubenswrapper[4805]: I1203 14:10:18.017711 4805 scope.go:117] "RemoveContainer" containerID="644e56c400524d739e2a208fb7eee1551ebbd28ba220df91c9dd84ba09d0b400" Dec 03 14:10:18 crc kubenswrapper[4805]: I1203 14:10:18.018701 4805 scope.go:117] "RemoveContainer" containerID="a0272667e733f2c5eec0fce0eb9c31a16c048babdaed90635895d689ae9303f4" Dec 03 14:10:18 crc kubenswrapper[4805]: E1203 14:10:18.018945 4805 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-psjpm_openshift-ovn-kubernetes(deeec30d-cad2-4dfd-adb6-64f6646581be)\"" pod="openshift-ovn-kubernetes/ovnkube-node-psjpm" podUID="deeec30d-cad2-4dfd-adb6-64f6646581be" Dec 03 14:10:18 crc kubenswrapper[4805]: I1203 14:10:18.038386 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"750d2b4b-738a-4992-8902-742528f6e875\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://abd2d0156113bb5182727a53aac594410ed051c19258beb8a07b808eff0b42a7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8673a29c0e16f4e707d79f0610722ccabb023ff2b4a18d7cf9d305183817d952\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://35766e7206b1be7a53d13d6890d759cf81dbd999643c12349268e4bf16fbbdb0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5b9844a11b2b903978d0d464b469e26c9211100da61d018dde9636dca95b684e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T14:09:30Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T14:10:18Z is after 2025-08-24T17:21:41Z" Dec 03 14:10:18 crc kubenswrapper[4805]: I1203 14:10:18.050762 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:48Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:48Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T14:10:18Z is after 2025-08-24T17:21:41Z" Dec 03 14:10:18 crc kubenswrapper[4805]: I1203 14:10:18.070925 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-k4mhs" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c09c1cdc-f461-45c9-8444-5b8764bccabc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3354c1a44f891182d04116fc76234a2b337d50567d032a3584f902a9516387cc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qdcd7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T14:09:50Z\\\"}}\" for pod \"openshift-multus\"/\"multus-k4mhs\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T14:10:18Z is after 2025-08-24T17:21:41Z" Dec 03 14:10:18 crc kubenswrapper[4805]: I1203 14:10:18.091577 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-bpk49" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"de00a81f-82e1-4223-9614-f98606b9a8ef\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://04f4ad27e02e2024ae9e7cd1b16116267e69cecda882dd723cdb274c41a0e7db\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-25d7t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7741e0e70416ebe094bfabc6009e98d7d4c0fbcba1fc82df13b9ba40774b456a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7741e0e70416ebe094bfabc6009e98d7d4c0fbcba1fc82df13b9ba40774b456a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T14:09:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T14:09:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-25d7t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ee3499c02eb52bbf08c447d2433d435a74dfb70670dc4cbaac44f5fbff9e8a29\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ee3499c02eb52bbf08c447d2433d435a74dfb70670dc4cbaac44f5fbff9e8a29\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T14:09:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T14:09:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-25d7t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ec8488b1d69799951be7ef40e5e028732265c4ddcdea2d453923e567864771f7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ec8488b1d69799951be7ef40e5e028732265c4ddcdea2d453923e567864771f7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T14:09:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T14:09:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-25d7t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://66707bb4e1fdacb265e89786df8a04605830f1c156394cacfcec9b54abe5d41a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://66707bb4e1fdacb265e89786df8a04605830f1c156394cacfcec9b54abe5d41a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T14:09:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T14:09:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-25d7t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6207ab6600970ff15c8713c547d4920b2c77eb9bea990bb5da497e24870dcc0f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6207ab6600970ff15c8713c547d4920b2c77eb9bea990bb5da497e24870dcc0f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T14:09:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T14:09:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-25d7t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8de65d4a205e2deec0b433e53c0e37feb94aa0ede6a4c58d7414b80829804505\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8de65d4a205e2deec0b433e53c0e37feb94aa0ede6a4c58d7414b80829804505\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T14:09:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T14:09:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-25d7t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T14:09:50Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-bpk49\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T14:10:18Z is after 2025-08-24T17:21:41Z" Dec 03 14:10:18 crc kubenswrapper[4805]: I1203 14:10:18.095119 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 14:10:18 crc kubenswrapper[4805]: I1203 14:10:18.095149 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 14:10:18 crc kubenswrapper[4805]: I1203 14:10:18.095158 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 14:10:18 crc kubenswrapper[4805]: I1203 14:10:18.095171 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 14:10:18 crc kubenswrapper[4805]: I1203 14:10:18.095179 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T14:10:18Z","lastTransitionTime":"2025-12-03T14:10:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 14:10:18 crc kubenswrapper[4805]: I1203 14:10:18.105557 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:48Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:48Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T14:10:18Z is after 2025-08-24T17:21:41Z" Dec 03 14:10:18 crc kubenswrapper[4805]: I1203 14:10:18.120531 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:49Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:49Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://541ace7ccbb66dbd191c3ca8c1e4479e3edbf80dfc67326bea8aaee81fdcfc3c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://821e3e9a902aacecf6c41e8b8d8aa1ac951ce2fbb6505b7148b16348a3754e55\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T14:10:18Z is after 2025-08-24T17:21:41Z" Dec 03 14:10:18 crc kubenswrapper[4805]: I1203 14:10:18.136027 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-gskh4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6098937f-e3f6-45e8-a647-4994a79cd711\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://28d5653f26ec61647808b3a733ff7eac70aa69ac184660f24147e8cfaf0eb802\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nwrw9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9ad6b68a42fd33ee167e67873c6eed3f11cdf6dc71d22e952c021436eb967dd9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nwrw9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T14:09:50Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-gskh4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T14:10:18Z is after 2025-08-24T17:21:41Z" Dec 03 14:10:18 crc kubenswrapper[4805]: I1203 14:10:18.166293 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-psjpm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"deeec30d-cad2-4dfd-adb6-64f6646581be\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:50Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:50Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4b8d00cdf0f37f2e5cd4e952cf85d3de8d72c43fdf786088fd9f26ed17b1e21f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wdlzx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bd67c6de01a8d196a1f51d4d90640c9e69de6f05d5e201b25e6eaa5c33440ea8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wdlzx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cb8bf05ef093eb0727a066d4bbbdf9af9422867cd4dadacafc6eeb016867a61e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wdlzx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0304467102ed053628c0d3e06548c74b6492336b6fd044246508d1e8bdb17bc3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wdlzx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6ce80ccafb3262a11fbe9923317a9d38b61e5dc5b3e7e7ae136f143d6b643fd7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wdlzx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://47d76962beb8c0f196d50af96fb3ec611097823ca80a1c8bd718f2d18840128c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wdlzx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a0272667e733f2c5eec0fce0eb9c31a16c048babdaed90635895d689ae9303f4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://644e56c400524d739e2a208fb7eee1551ebbd28ba220df91c9dd84ba09d0b400\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-03T14:10:02Z\\\",\\\"message\\\":\\\"tring(nil)}, nodeEndpoints:map[string]services.lbEndpoints{}, externalTrafficLocal:false, internalTrafficLocal:false, hasNodePort:false}}\\\\nF1203 14:10:01.661807 6255 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T14:10:01Z is after 2025-08-24T17:21:41Z]\\\\nI1203 14:10:01.660569 6255 services_controller.go:451] Built service openshift-operator-lifecycle-manager/package-server-manager-metrics cluster-wide LB for network=default: []services.LB{services.LB{Name:\\\\\\\"Service_openshift-operator-lifecycle-manager/package-server-manager-metrics_TCP_cluster\\\\\\\", UUID:\\\\\\\"\\\\\\\", Protocol:\\\\\\\"TCP\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Serv\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T14:10:01Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a0272667e733f2c5eec0fce0eb9c31a16c048babdaed90635895d689ae9303f4\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-03T14:10:16Z\\\",\\\"message\\\":\\\"o create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T14:10:16Z is after 2025-08-24T17:21:41Z]\\\\nI1203 14:10:16.717659 6478 services_controller.go:473] Services do not match for network=default, existing lbs: []services.LB{services.LB{Name:\\\\\\\"Service_openshift-network-console/networking-console-plugin_TCP_cluster\\\\\\\", UUID:\\\\\\\"ab0b1d51-5ec6-479b-8881-93dfa8d30337\\\\\\\", Protocol:\\\\\\\"tcp\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-network-console/networking-console-plugin\\\\\\\"}, Opts:services.LBOpts{Reject:false, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{}, Templates:services.TemplateMap{}, Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGr\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T14:10:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wdlzx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ab98e365c1a4242a47b8cbe2dd85109d366d6002c4de1f18ddc950a46bea6c12\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wdlzx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2f875252fbe6a673a72fc9c56bb96899f62732305ad77b3108e3b244bfa91ecb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2f875252fbe6a673a72fc9c56bb96899f62732305ad77b3108e3b244bfa91ecb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T14:09:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T14:09:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wdlzx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T14:09:50Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-psjpm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T14:10:18Z is after 2025-08-24T17:21:41Z" Dec 03 14:10:18 crc kubenswrapper[4805]: I1203 14:10:18.182637 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-4gd9r" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6b36a3db-1047-487a-9d6c-4f51ffdd8cc1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:10:04Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:10:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:10:04Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:10:04Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4t6zl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4t6zl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T14:10:04Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-4gd9r\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T14:10:18Z is after 2025-08-24T17:21:41Z" Dec 03 14:10:18 crc kubenswrapper[4805]: I1203 14:10:18.197079 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-pdzgk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"860769de-6a5b-4cb9-a752-264e7e2e785f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5e2f0a0cc21505d609f81918365d5e4f5d21959ac12ab147d6a51601c9de4c14\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ggn4n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T14:09:50Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-pdzgk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T14:10:18Z is after 2025-08-24T17:21:41Z" Dec 03 14:10:18 crc kubenswrapper[4805]: I1203 14:10:18.197860 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 14:10:18 crc kubenswrapper[4805]: I1203 14:10:18.197892 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 14:10:18 crc kubenswrapper[4805]: I1203 14:10:18.197901 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 14:10:18 crc kubenswrapper[4805]: I1203 14:10:18.197915 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 14:10:18 crc kubenswrapper[4805]: I1203 14:10:18.197925 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T14:10:18Z","lastTransitionTime":"2025-12-03T14:10:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 14:10:18 crc kubenswrapper[4805]: I1203 14:10:18.209744 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-69h75" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5d1c694f-9847-4086-8b37-3573ddeb3ace\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:10:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:10:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:10:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:10:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://abc1d9b8582ca9fe4a51fb608fb39129a179ddcdfde0797a6369a55b2e0124cf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:10:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s578b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c7c4d434b035cac2a91a4f0846a2f4fb5bd68f881c28487860adb6db85e8ce75\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:10:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s578b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T14:10:03Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-69h75\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T14:10:18Z is after 2025-08-24T17:21:41Z" Dec 03 14:10:18 crc kubenswrapper[4805]: I1203 14:10:18.237259 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b6d36671-7173-4e9c-843f-005b4dcf576c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a837d7d491eaacbe4279b4014bb5fcbe9b514ba5cc82c708227bcd9b25a346e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6e5f04faf3b23197bb8ea11d3e574dc86f31e21d798734962001ec8c95de7c84\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2ff1a979387da41b00e95eb9cc8b2c9de966380a12fae687591df3c5b18a4497\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1e66e1cb18573a14e3c3ae16af4dbd9f8997a6ad38680c3ab4505646ebd48e7c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7d5b5c85e3de43e7e95111aa967248550607c26200c783251b68696f202f92a1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://474f72dac37a808fc4b9832306e76979b151d7221807ff3086013466ed1419c3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://474f72dac37a808fc4b9832306e76979b151d7221807ff3086013466ed1419c3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T14:09:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T14:09:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://74c8cea1a61c4775a0ad2f4ab4bb8a36437cf23d2ee953311d6ea967e297510a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://74c8cea1a61c4775a0ad2f4ab4bb8a36437cf23d2ee953311d6ea967e297510a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T14:09:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T14:09:31Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://a4f2f56a809adc01fba82eba0652d78f58770c132f77b1424f8b17068b1927e3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a4f2f56a809adc01fba82eba0652d78f58770c132f77b1424f8b17068b1927e3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T14:09:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T14:09:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T14:09:30Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T14:10:18Z is after 2025-08-24T17:21:41Z" Dec 03 14:10:18 crc kubenswrapper[4805]: I1203 14:10:18.260625 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ced8291d-0ce5-4e74-9eaf-2bf033774039\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:10:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:10:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3e5a34636f0a8c4d9179074596d28bc2eb7a504573398c596a637defd829907a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://29cd95be3f964415943c873e7914ffcea4abe50dfdf17f8a9cd03fc38df35981\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://951173b3d3d581275a6f6e43209b5e0611a701ca7a10f81504d723e24a2d21a3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e64730957bbbb2feb4d506e0ed9731a0594bf66fa1e8fcdb8ec944ded6308eed\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5da55f769e23ad52bac9c77da62c78e70fa34db8ed1a435e4eaf4dad3fd2b68a\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-03T14:09:48Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1203 14:09:42.992179 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1203 14:09:42.993233 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1830926909/tls.crt::/tmp/serving-cert-1830926909/tls.key\\\\\\\"\\\\nI1203 14:09:48.622403 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1203 14:09:48.626476 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1203 14:09:48.626527 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1203 14:09:48.626574 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1203 14:09:48.626582 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1203 14:09:48.633990 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1203 14:09:48.634024 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 14:09:48.634029 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 14:09:48.634034 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1203 14:09:48.634038 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1203 14:09:48.634042 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1203 14:09:48.634045 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1203 14:09:48.634344 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1203 14:09:48.644497 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T14:09:32Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bd92a160eb530c5687ebd7cd5b83b41820a2b808a0fd9514e5ace9d237e1bd26\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:32Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://af9c940135012c629447528c61743ed1e58708ab57ce76604c6da0479aa459cf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://af9c940135012c629447528c61743ed1e58708ab57ce76604c6da0479aa459cf\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T14:09:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T14:09:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T14:09:30Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T14:10:18Z is after 2025-08-24T17:21:41Z" Dec 03 14:10:18 crc kubenswrapper[4805]: I1203 14:10:18.280120 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:49Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:49Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://38339331a3209d6787d1def81b72f4a8514bf91080a139cad6d8ac5cc6bef10e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T14:10:18Z is after 2025-08-24T17:21:41Z" Dec 03 14:10:18 crc kubenswrapper[4805]: I1203 14:10:18.298485 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:48Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:48Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T14:10:18Z is after 2025-08-24T17:21:41Z" Dec 03 14:10:18 crc kubenswrapper[4805]: I1203 14:10:18.301200 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 14:10:18 crc kubenswrapper[4805]: I1203 14:10:18.301545 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 14:10:18 crc kubenswrapper[4805]: I1203 14:10:18.301816 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 14:10:18 crc kubenswrapper[4805]: I1203 14:10:18.302043 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 14:10:18 crc kubenswrapper[4805]: I1203 14:10:18.302251 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T14:10:18Z","lastTransitionTime":"2025-12-03T14:10:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 14:10:18 crc kubenswrapper[4805]: I1203 14:10:18.314663 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:51Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:51Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b3123d3fa879097db88ed2f3cfb05631f440c24c5405503b2de3546f12e2e899\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T14:10:18Z is after 2025-08-24T17:21:41Z" Dec 03 14:10:18 crc kubenswrapper[4805]: I1203 14:10:18.330947 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-bdtdg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"53a0e9e3-b6d8-44b7-a32b-aa3450442865\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f557e17ad105d6972d3c1480656c6073f67e28c0f5773aa365549951f4d161c7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9ww54\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T14:09:51Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-bdtdg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T14:10:18Z is after 2025-08-24T17:21:41Z" Dec 03 14:10:18 crc kubenswrapper[4805]: I1203 14:10:18.405561 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 14:10:18 crc kubenswrapper[4805]: I1203 14:10:18.405613 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 14:10:18 crc kubenswrapper[4805]: I1203 14:10:18.405625 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 14:10:18 crc kubenswrapper[4805]: I1203 14:10:18.405642 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 14:10:18 crc kubenswrapper[4805]: I1203 14:10:18.405654 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T14:10:18Z","lastTransitionTime":"2025-12-03T14:10:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 14:10:18 crc kubenswrapper[4805]: I1203 14:10:18.508610 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 14:10:18 crc kubenswrapper[4805]: I1203 14:10:18.508661 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 14:10:18 crc kubenswrapper[4805]: I1203 14:10:18.508678 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 14:10:18 crc kubenswrapper[4805]: I1203 14:10:18.508702 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 14:10:18 crc kubenswrapper[4805]: I1203 14:10:18.508719 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T14:10:18Z","lastTransitionTime":"2025-12-03T14:10:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 14:10:18 crc kubenswrapper[4805]: I1203 14:10:18.610928 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 14:10:18 crc kubenswrapper[4805]: I1203 14:10:18.610974 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 14:10:18 crc kubenswrapper[4805]: I1203 14:10:18.610988 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 14:10:18 crc kubenswrapper[4805]: I1203 14:10:18.611002 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 14:10:18 crc kubenswrapper[4805]: I1203 14:10:18.611011 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T14:10:18Z","lastTransitionTime":"2025-12-03T14:10:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 14:10:18 crc kubenswrapper[4805]: I1203 14:10:18.694224 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-4gd9r" Dec 03 14:10:18 crc kubenswrapper[4805]: I1203 14:10:18.694329 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 14:10:18 crc kubenswrapper[4805]: I1203 14:10:18.694241 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 14:10:18 crc kubenswrapper[4805]: E1203 14:10:18.694407 4805 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-4gd9r" podUID="6b36a3db-1047-487a-9d6c-4f51ffdd8cc1" Dec 03 14:10:18 crc kubenswrapper[4805]: E1203 14:10:18.694478 4805 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 03 14:10:18 crc kubenswrapper[4805]: I1203 14:10:18.694493 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 14:10:18 crc kubenswrapper[4805]: E1203 14:10:18.694560 4805 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 03 14:10:18 crc kubenswrapper[4805]: E1203 14:10:18.694702 4805 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 03 14:10:18 crc kubenswrapper[4805]: I1203 14:10:18.713226 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 14:10:18 crc kubenswrapper[4805]: I1203 14:10:18.713264 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 14:10:18 crc kubenswrapper[4805]: I1203 14:10:18.713273 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 14:10:18 crc kubenswrapper[4805]: I1203 14:10:18.713286 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 14:10:18 crc kubenswrapper[4805]: I1203 14:10:18.713295 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T14:10:18Z","lastTransitionTime":"2025-12-03T14:10:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 14:10:18 crc kubenswrapper[4805]: I1203 14:10:18.884202 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 14:10:18 crc kubenswrapper[4805]: I1203 14:10:18.884252 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 14:10:18 crc kubenswrapper[4805]: I1203 14:10:18.884263 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 14:10:18 crc kubenswrapper[4805]: I1203 14:10:18.884282 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 14:10:18 crc kubenswrapper[4805]: I1203 14:10:18.884294 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T14:10:18Z","lastTransitionTime":"2025-12-03T14:10:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 14:10:18 crc kubenswrapper[4805]: I1203 14:10:18.986439 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 14:10:18 crc kubenswrapper[4805]: I1203 14:10:18.986497 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 14:10:18 crc kubenswrapper[4805]: I1203 14:10:18.986507 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 14:10:18 crc kubenswrapper[4805]: I1203 14:10:18.986528 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 14:10:18 crc kubenswrapper[4805]: I1203 14:10:18.986542 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T14:10:18Z","lastTransitionTime":"2025-12-03T14:10:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 14:10:19 crc kubenswrapper[4805]: I1203 14:10:19.022417 4805 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-psjpm_deeec30d-cad2-4dfd-adb6-64f6646581be/ovnkube-controller/2.log" Dec 03 14:10:19 crc kubenswrapper[4805]: I1203 14:10:19.089493 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 14:10:19 crc kubenswrapper[4805]: I1203 14:10:19.089552 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 14:10:19 crc kubenswrapper[4805]: I1203 14:10:19.089567 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 14:10:19 crc kubenswrapper[4805]: I1203 14:10:19.089585 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 14:10:19 crc kubenswrapper[4805]: I1203 14:10:19.089596 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T14:10:19Z","lastTransitionTime":"2025-12-03T14:10:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 14:10:19 crc kubenswrapper[4805]: I1203 14:10:19.192274 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 14:10:19 crc kubenswrapper[4805]: I1203 14:10:19.192349 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 14:10:19 crc kubenswrapper[4805]: I1203 14:10:19.192378 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 14:10:19 crc kubenswrapper[4805]: I1203 14:10:19.192409 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 14:10:19 crc kubenswrapper[4805]: I1203 14:10:19.192430 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T14:10:19Z","lastTransitionTime":"2025-12-03T14:10:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 14:10:19 crc kubenswrapper[4805]: I1203 14:10:19.260185 4805 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-psjpm" Dec 03 14:10:19 crc kubenswrapper[4805]: I1203 14:10:19.260893 4805 scope.go:117] "RemoveContainer" containerID="a0272667e733f2c5eec0fce0eb9c31a16c048babdaed90635895d689ae9303f4" Dec 03 14:10:19 crc kubenswrapper[4805]: E1203 14:10:19.261043 4805 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-psjpm_openshift-ovn-kubernetes(deeec30d-cad2-4dfd-adb6-64f6646581be)\"" pod="openshift-ovn-kubernetes/ovnkube-node-psjpm" podUID="deeec30d-cad2-4dfd-adb6-64f6646581be" Dec 03 14:10:19 crc kubenswrapper[4805]: I1203 14:10:19.272936 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-pdzgk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"860769de-6a5b-4cb9-a752-264e7e2e785f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5e2f0a0cc21505d609f81918365d5e4f5d21959ac12ab147d6a51601c9de4c14\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ggn4n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T14:09:50Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-pdzgk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T14:10:19Z is after 2025-08-24T17:21:41Z" Dec 03 14:10:19 crc kubenswrapper[4805]: I1203 14:10:19.285272 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-69h75" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5d1c694f-9847-4086-8b37-3573ddeb3ace\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:10:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:10:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:10:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:10:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://abc1d9b8582ca9fe4a51fb608fb39129a179ddcdfde0797a6369a55b2e0124cf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:10:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s578b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c7c4d434b035cac2a91a4f0846a2f4fb5bd68f881c28487860adb6db85e8ce75\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:10:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s578b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T14:10:03Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-69h75\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T14:10:19Z is after 2025-08-24T17:21:41Z" Dec 03 14:10:19 crc kubenswrapper[4805]: I1203 14:10:19.295513 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 14:10:19 crc kubenswrapper[4805]: I1203 14:10:19.295587 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 14:10:19 crc kubenswrapper[4805]: I1203 14:10:19.295603 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 14:10:19 crc kubenswrapper[4805]: I1203 14:10:19.295631 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 14:10:19 crc kubenswrapper[4805]: I1203 14:10:19.295648 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T14:10:19Z","lastTransitionTime":"2025-12-03T14:10:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 14:10:19 crc kubenswrapper[4805]: I1203 14:10:19.296612 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:51Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:51Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b3123d3fa879097db88ed2f3cfb05631f440c24c5405503b2de3546f12e2e899\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T14:10:19Z is after 2025-08-24T17:21:41Z" Dec 03 14:10:19 crc kubenswrapper[4805]: I1203 14:10:19.305803 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-bdtdg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"53a0e9e3-b6d8-44b7-a32b-aa3450442865\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f557e17ad105d6972d3c1480656c6073f67e28c0f5773aa365549951f4d161c7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9ww54\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T14:09:51Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-bdtdg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T14:10:19Z is after 2025-08-24T17:21:41Z" Dec 03 14:10:19 crc kubenswrapper[4805]: I1203 14:10:19.329117 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b6d36671-7173-4e9c-843f-005b4dcf576c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a837d7d491eaacbe4279b4014bb5fcbe9b514ba5cc82c708227bcd9b25a346e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6e5f04faf3b23197bb8ea11d3e574dc86f31e21d798734962001ec8c95de7c84\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2ff1a979387da41b00e95eb9cc8b2c9de966380a12fae687591df3c5b18a4497\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1e66e1cb18573a14e3c3ae16af4dbd9f8997a6ad38680c3ab4505646ebd48e7c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7d5b5c85e3de43e7e95111aa967248550607c26200c783251b68696f202f92a1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://474f72dac37a808fc4b9832306e76979b151d7221807ff3086013466ed1419c3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://474f72dac37a808fc4b9832306e76979b151d7221807ff3086013466ed1419c3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T14:09:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T14:09:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://74c8cea1a61c4775a0ad2f4ab4bb8a36437cf23d2ee953311d6ea967e297510a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://74c8cea1a61c4775a0ad2f4ab4bb8a36437cf23d2ee953311d6ea967e297510a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T14:09:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T14:09:31Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://a4f2f56a809adc01fba82eba0652d78f58770c132f77b1424f8b17068b1927e3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a4f2f56a809adc01fba82eba0652d78f58770c132f77b1424f8b17068b1927e3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T14:09:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T14:09:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T14:09:30Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T14:10:19Z is after 2025-08-24T17:21:41Z" Dec 03 14:10:19 crc kubenswrapper[4805]: I1203 14:10:19.345286 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ced8291d-0ce5-4e74-9eaf-2bf033774039\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:10:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:10:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3e5a34636f0a8c4d9179074596d28bc2eb7a504573398c596a637defd829907a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://29cd95be3f964415943c873e7914ffcea4abe50dfdf17f8a9cd03fc38df35981\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://951173b3d3d581275a6f6e43209b5e0611a701ca7a10f81504d723e24a2d21a3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e64730957bbbb2feb4d506e0ed9731a0594bf66fa1e8fcdb8ec944ded6308eed\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5da55f769e23ad52bac9c77da62c78e70fa34db8ed1a435e4eaf4dad3fd2b68a\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-03T14:09:48Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1203 14:09:42.992179 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1203 14:09:42.993233 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1830926909/tls.crt::/tmp/serving-cert-1830926909/tls.key\\\\\\\"\\\\nI1203 14:09:48.622403 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1203 14:09:48.626476 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1203 14:09:48.626527 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1203 14:09:48.626574 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1203 14:09:48.626582 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1203 14:09:48.633990 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1203 14:09:48.634024 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 14:09:48.634029 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 14:09:48.634034 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1203 14:09:48.634038 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1203 14:09:48.634042 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1203 14:09:48.634045 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1203 14:09:48.634344 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1203 14:09:48.644497 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T14:09:32Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bd92a160eb530c5687ebd7cd5b83b41820a2b808a0fd9514e5ace9d237e1bd26\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:32Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://af9c940135012c629447528c61743ed1e58708ab57ce76604c6da0479aa459cf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://af9c940135012c629447528c61743ed1e58708ab57ce76604c6da0479aa459cf\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T14:09:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T14:09:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T14:09:30Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T14:10:19Z is after 2025-08-24T17:21:41Z" Dec 03 14:10:19 crc kubenswrapper[4805]: I1203 14:10:19.365445 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:49Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:49Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://38339331a3209d6787d1def81b72f4a8514bf91080a139cad6d8ac5cc6bef10e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T14:10:19Z is after 2025-08-24T17:21:41Z" Dec 03 14:10:19 crc kubenswrapper[4805]: I1203 14:10:19.381026 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:48Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:48Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T14:10:19Z is after 2025-08-24T17:21:41Z" Dec 03 14:10:19 crc kubenswrapper[4805]: I1203 14:10:19.397272 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"750d2b4b-738a-4992-8902-742528f6e875\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://abd2d0156113bb5182727a53aac594410ed051c19258beb8a07b808eff0b42a7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8673a29c0e16f4e707d79f0610722ccabb023ff2b4a18d7cf9d305183817d952\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://35766e7206b1be7a53d13d6890d759cf81dbd999643c12349268e4bf16fbbdb0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5b9844a11b2b903978d0d464b469e26c9211100da61d018dde9636dca95b684e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T14:09:30Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T14:10:19Z is after 2025-08-24T17:21:41Z" Dec 03 14:10:19 crc kubenswrapper[4805]: I1203 14:10:19.398659 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 14:10:19 crc kubenswrapper[4805]: I1203 14:10:19.398733 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 14:10:19 crc kubenswrapper[4805]: I1203 14:10:19.398746 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 14:10:19 crc kubenswrapper[4805]: I1203 14:10:19.398786 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 14:10:19 crc kubenswrapper[4805]: I1203 14:10:19.398800 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T14:10:19Z","lastTransitionTime":"2025-12-03T14:10:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 14:10:19 crc kubenswrapper[4805]: I1203 14:10:19.411065 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:48Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:48Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T14:10:19Z is after 2025-08-24T17:21:41Z" Dec 03 14:10:19 crc kubenswrapper[4805]: I1203 14:10:19.429365 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-k4mhs" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c09c1cdc-f461-45c9-8444-5b8764bccabc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3354c1a44f891182d04116fc76234a2b337d50567d032a3584f902a9516387cc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qdcd7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T14:09:50Z\\\"}}\" for pod \"openshift-multus\"/\"multus-k4mhs\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T14:10:19Z is after 2025-08-24T17:21:41Z" Dec 03 14:10:19 crc kubenswrapper[4805]: I1203 14:10:19.449986 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-bpk49" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"de00a81f-82e1-4223-9614-f98606b9a8ef\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://04f4ad27e02e2024ae9e7cd1b16116267e69cecda882dd723cdb274c41a0e7db\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-25d7t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7741e0e70416ebe094bfabc6009e98d7d4c0fbcba1fc82df13b9ba40774b456a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7741e0e70416ebe094bfabc6009e98d7d4c0fbcba1fc82df13b9ba40774b456a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T14:09:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T14:09:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-25d7t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ee3499c02eb52bbf08c447d2433d435a74dfb70670dc4cbaac44f5fbff9e8a29\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ee3499c02eb52bbf08c447d2433d435a74dfb70670dc4cbaac44f5fbff9e8a29\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T14:09:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T14:09:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-25d7t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ec8488b1d69799951be7ef40e5e028732265c4ddcdea2d453923e567864771f7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ec8488b1d69799951be7ef40e5e028732265c4ddcdea2d453923e567864771f7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T14:09:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T14:09:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-25d7t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://66707bb4e1fdacb265e89786df8a04605830f1c156394cacfcec9b54abe5d41a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://66707bb4e1fdacb265e89786df8a04605830f1c156394cacfcec9b54abe5d41a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T14:09:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T14:09:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-25d7t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6207ab6600970ff15c8713c547d4920b2c77eb9bea990bb5da497e24870dcc0f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6207ab6600970ff15c8713c547d4920b2c77eb9bea990bb5da497e24870dcc0f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T14:09:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T14:09:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-25d7t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8de65d4a205e2deec0b433e53c0e37feb94aa0ede6a4c58d7414b80829804505\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8de65d4a205e2deec0b433e53c0e37feb94aa0ede6a4c58d7414b80829804505\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T14:09:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T14:09:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-25d7t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T14:09:50Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-bpk49\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T14:10:19Z is after 2025-08-24T17:21:41Z" Dec 03 14:10:19 crc kubenswrapper[4805]: I1203 14:10:19.464151 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:48Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:48Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T14:10:19Z is after 2025-08-24T17:21:41Z" Dec 03 14:10:19 crc kubenswrapper[4805]: I1203 14:10:19.478735 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:49Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:49Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://541ace7ccbb66dbd191c3ca8c1e4479e3edbf80dfc67326bea8aaee81fdcfc3c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://821e3e9a902aacecf6c41e8b8d8aa1ac951ce2fbb6505b7148b16348a3754e55\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T14:10:19Z is after 2025-08-24T17:21:41Z" Dec 03 14:10:19 crc kubenswrapper[4805]: I1203 14:10:19.493191 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-gskh4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6098937f-e3f6-45e8-a647-4994a79cd711\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://28d5653f26ec61647808b3a733ff7eac70aa69ac184660f24147e8cfaf0eb802\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nwrw9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9ad6b68a42fd33ee167e67873c6eed3f11cdf6dc71d22e952c021436eb967dd9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nwrw9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T14:09:50Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-gskh4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T14:10:19Z is after 2025-08-24T17:21:41Z" Dec 03 14:10:19 crc kubenswrapper[4805]: I1203 14:10:19.501091 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 14:10:19 crc kubenswrapper[4805]: I1203 14:10:19.501114 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 14:10:19 crc kubenswrapper[4805]: I1203 14:10:19.501122 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 14:10:19 crc kubenswrapper[4805]: I1203 14:10:19.501140 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 14:10:19 crc kubenswrapper[4805]: I1203 14:10:19.501150 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T14:10:19Z","lastTransitionTime":"2025-12-03T14:10:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 14:10:19 crc kubenswrapper[4805]: I1203 14:10:19.512215 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-psjpm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"deeec30d-cad2-4dfd-adb6-64f6646581be\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:50Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:50Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4b8d00cdf0f37f2e5cd4e952cf85d3de8d72c43fdf786088fd9f26ed17b1e21f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wdlzx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bd67c6de01a8d196a1f51d4d90640c9e69de6f05d5e201b25e6eaa5c33440ea8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wdlzx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cb8bf05ef093eb0727a066d4bbbdf9af9422867cd4dadacafc6eeb016867a61e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wdlzx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0304467102ed053628c0d3e06548c74b6492336b6fd044246508d1e8bdb17bc3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wdlzx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6ce80ccafb3262a11fbe9923317a9d38b61e5dc5b3e7e7ae136f143d6b643fd7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wdlzx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://47d76962beb8c0f196d50af96fb3ec611097823ca80a1c8bd718f2d18840128c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wdlzx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a0272667e733f2c5eec0fce0eb9c31a16c048babdaed90635895d689ae9303f4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a0272667e733f2c5eec0fce0eb9c31a16c048babdaed90635895d689ae9303f4\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-03T14:10:16Z\\\",\\\"message\\\":\\\"o create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T14:10:16Z is after 2025-08-24T17:21:41Z]\\\\nI1203 14:10:16.717659 6478 services_controller.go:473] Services do not match for network=default, existing lbs: []services.LB{services.LB{Name:\\\\\\\"Service_openshift-network-console/networking-console-plugin_TCP_cluster\\\\\\\", UUID:\\\\\\\"ab0b1d51-5ec6-479b-8881-93dfa8d30337\\\\\\\", Protocol:\\\\\\\"tcp\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-network-console/networking-console-plugin\\\\\\\"}, Opts:services.LBOpts{Reject:false, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{}, Templates:services.TemplateMap{}, Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGr\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T14:10:15Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-psjpm_openshift-ovn-kubernetes(deeec30d-cad2-4dfd-adb6-64f6646581be)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wdlzx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ab98e365c1a4242a47b8cbe2dd85109d366d6002c4de1f18ddc950a46bea6c12\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wdlzx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2f875252fbe6a673a72fc9c56bb96899f62732305ad77b3108e3b244bfa91ecb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2f875252fbe6a673a72fc9c56bb96899f62732305ad77b3108e3b244bfa91ecb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T14:09:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T14:09:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wdlzx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T14:09:50Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-psjpm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T14:10:19Z is after 2025-08-24T17:21:41Z" Dec 03 14:10:19 crc kubenswrapper[4805]: I1203 14:10:19.522663 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-4gd9r" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6b36a3db-1047-487a-9d6c-4f51ffdd8cc1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:10:04Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:10:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:10:04Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:10:04Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4t6zl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4t6zl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T14:10:04Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-4gd9r\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T14:10:19Z is after 2025-08-24T17:21:41Z" Dec 03 14:10:19 crc kubenswrapper[4805]: I1203 14:10:19.603898 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 14:10:19 crc kubenswrapper[4805]: I1203 14:10:19.603939 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 14:10:19 crc kubenswrapper[4805]: I1203 14:10:19.603950 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 14:10:19 crc kubenswrapper[4805]: I1203 14:10:19.603969 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 14:10:19 crc kubenswrapper[4805]: I1203 14:10:19.603981 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T14:10:19Z","lastTransitionTime":"2025-12-03T14:10:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 14:10:19 crc kubenswrapper[4805]: I1203 14:10:19.706235 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 14:10:19 crc kubenswrapper[4805]: I1203 14:10:19.706480 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 14:10:19 crc kubenswrapper[4805]: I1203 14:10:19.706538 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 14:10:19 crc kubenswrapper[4805]: I1203 14:10:19.706620 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 14:10:19 crc kubenswrapper[4805]: I1203 14:10:19.706675 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T14:10:19Z","lastTransitionTime":"2025-12-03T14:10:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 14:10:19 crc kubenswrapper[4805]: I1203 14:10:19.776039 4805 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Dec 03 14:10:19 crc kubenswrapper[4805]: I1203 14:10:19.788954 4805 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-scheduler/openshift-kube-scheduler-crc"] Dec 03 14:10:19 crc kubenswrapper[4805]: I1203 14:10:19.801953 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-psjpm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"deeec30d-cad2-4dfd-adb6-64f6646581be\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:50Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:50Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4b8d00cdf0f37f2e5cd4e952cf85d3de8d72c43fdf786088fd9f26ed17b1e21f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wdlzx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bd67c6de01a8d196a1f51d4d90640c9e69de6f05d5e201b25e6eaa5c33440ea8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wdlzx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cb8bf05ef093eb0727a066d4bbbdf9af9422867cd4dadacafc6eeb016867a61e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wdlzx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0304467102ed053628c0d3e06548c74b6492336b6fd044246508d1e8bdb17bc3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wdlzx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6ce80ccafb3262a11fbe9923317a9d38b61e5dc5b3e7e7ae136f143d6b643fd7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wdlzx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://47d76962beb8c0f196d50af96fb3ec611097823ca80a1c8bd718f2d18840128c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wdlzx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a0272667e733f2c5eec0fce0eb9c31a16c048babdaed90635895d689ae9303f4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a0272667e733f2c5eec0fce0eb9c31a16c048babdaed90635895d689ae9303f4\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-03T14:10:16Z\\\",\\\"message\\\":\\\"o create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T14:10:16Z is after 2025-08-24T17:21:41Z]\\\\nI1203 14:10:16.717659 6478 services_controller.go:473] Services do not match for network=default, existing lbs: []services.LB{services.LB{Name:\\\\\\\"Service_openshift-network-console/networking-console-plugin_TCP_cluster\\\\\\\", UUID:\\\\\\\"ab0b1d51-5ec6-479b-8881-93dfa8d30337\\\\\\\", Protocol:\\\\\\\"tcp\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-network-console/networking-console-plugin\\\\\\\"}, Opts:services.LBOpts{Reject:false, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{}, Templates:services.TemplateMap{}, Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGr\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T14:10:15Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-psjpm_openshift-ovn-kubernetes(deeec30d-cad2-4dfd-adb6-64f6646581be)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wdlzx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ab98e365c1a4242a47b8cbe2dd85109d366d6002c4de1f18ddc950a46bea6c12\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wdlzx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2f875252fbe6a673a72fc9c56bb96899f62732305ad77b3108e3b244bfa91ecb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2f875252fbe6a673a72fc9c56bb96899f62732305ad77b3108e3b244bfa91ecb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T14:09:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T14:09:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wdlzx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T14:09:50Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-psjpm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T14:10:19Z is after 2025-08-24T17:21:41Z" Dec 03 14:10:19 crc kubenswrapper[4805]: I1203 14:10:19.808999 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 14:10:19 crc kubenswrapper[4805]: I1203 14:10:19.809106 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 14:10:19 crc kubenswrapper[4805]: I1203 14:10:19.809132 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 14:10:19 crc kubenswrapper[4805]: I1203 14:10:19.809157 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 14:10:19 crc kubenswrapper[4805]: I1203 14:10:19.809176 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T14:10:19Z","lastTransitionTime":"2025-12-03T14:10:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 14:10:19 crc kubenswrapper[4805]: I1203 14:10:19.814353 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-4gd9r" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6b36a3db-1047-487a-9d6c-4f51ffdd8cc1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:10:04Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:10:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:10:04Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:10:04Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4t6zl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4t6zl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T14:10:04Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-4gd9r\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T14:10:19Z is after 2025-08-24T17:21:41Z" Dec 03 14:10:19 crc kubenswrapper[4805]: I1203 14:10:19.827801 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:48Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:48Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T14:10:19Z is after 2025-08-24T17:21:41Z" Dec 03 14:10:19 crc kubenswrapper[4805]: I1203 14:10:19.848533 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:49Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:49Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://541ace7ccbb66dbd191c3ca8c1e4479e3edbf80dfc67326bea8aaee81fdcfc3c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://821e3e9a902aacecf6c41e8b8d8aa1ac951ce2fbb6505b7148b16348a3754e55\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T14:10:19Z is after 2025-08-24T17:21:41Z" Dec 03 14:10:19 crc kubenswrapper[4805]: I1203 14:10:19.861778 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-gskh4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6098937f-e3f6-45e8-a647-4994a79cd711\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://28d5653f26ec61647808b3a733ff7eac70aa69ac184660f24147e8cfaf0eb802\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nwrw9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9ad6b68a42fd33ee167e67873c6eed3f11cdf6dc71d22e952c021436eb967dd9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nwrw9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T14:09:50Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-gskh4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T14:10:19Z is after 2025-08-24T17:21:41Z" Dec 03 14:10:19 crc kubenswrapper[4805]: I1203 14:10:19.872944 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-pdzgk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"860769de-6a5b-4cb9-a752-264e7e2e785f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5e2f0a0cc21505d609f81918365d5e4f5d21959ac12ab147d6a51601c9de4c14\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ggn4n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T14:09:50Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-pdzgk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T14:10:19Z is after 2025-08-24T17:21:41Z" Dec 03 14:10:19 crc kubenswrapper[4805]: I1203 14:10:19.886111 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-69h75" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5d1c694f-9847-4086-8b37-3573ddeb3ace\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:10:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:10:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:10:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:10:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://abc1d9b8582ca9fe4a51fb608fb39129a179ddcdfde0797a6369a55b2e0124cf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:10:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s578b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c7c4d434b035cac2a91a4f0846a2f4fb5bd68f881c28487860adb6db85e8ce75\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:10:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s578b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T14:10:03Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-69h75\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T14:10:19Z is after 2025-08-24T17:21:41Z" Dec 03 14:10:19 crc kubenswrapper[4805]: I1203 14:10:19.900844 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:49Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:49Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://38339331a3209d6787d1def81b72f4a8514bf91080a139cad6d8ac5cc6bef10e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T14:10:19Z is after 2025-08-24T17:21:41Z" Dec 03 14:10:19 crc kubenswrapper[4805]: I1203 14:10:19.912047 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:48Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:48Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T14:10:19Z is after 2025-08-24T17:21:41Z" Dec 03 14:10:19 crc kubenswrapper[4805]: I1203 14:10:19.912486 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 14:10:19 crc kubenswrapper[4805]: I1203 14:10:19.912515 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 14:10:19 crc kubenswrapper[4805]: I1203 14:10:19.912527 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 14:10:19 crc kubenswrapper[4805]: I1203 14:10:19.912542 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 14:10:19 crc kubenswrapper[4805]: I1203 14:10:19.912553 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T14:10:19Z","lastTransitionTime":"2025-12-03T14:10:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 14:10:19 crc kubenswrapper[4805]: I1203 14:10:19.928961 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:51Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:51Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b3123d3fa879097db88ed2f3cfb05631f440c24c5405503b2de3546f12e2e899\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T14:10:19Z is after 2025-08-24T17:21:41Z" Dec 03 14:10:19 crc kubenswrapper[4805]: I1203 14:10:19.941718 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-bdtdg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"53a0e9e3-b6d8-44b7-a32b-aa3450442865\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f557e17ad105d6972d3c1480656c6073f67e28c0f5773aa365549951f4d161c7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9ww54\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T14:09:51Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-bdtdg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T14:10:19Z is after 2025-08-24T17:21:41Z" Dec 03 14:10:19 crc kubenswrapper[4805]: I1203 14:10:19.966015 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b6d36671-7173-4e9c-843f-005b4dcf576c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a837d7d491eaacbe4279b4014bb5fcbe9b514ba5cc82c708227bcd9b25a346e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6e5f04faf3b23197bb8ea11d3e574dc86f31e21d798734962001ec8c95de7c84\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2ff1a979387da41b00e95eb9cc8b2c9de966380a12fae687591df3c5b18a4497\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1e66e1cb18573a14e3c3ae16af4dbd9f8997a6ad38680c3ab4505646ebd48e7c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7d5b5c85e3de43e7e95111aa967248550607c26200c783251b68696f202f92a1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://474f72dac37a808fc4b9832306e76979b151d7221807ff3086013466ed1419c3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://474f72dac37a808fc4b9832306e76979b151d7221807ff3086013466ed1419c3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T14:09:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T14:09:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://74c8cea1a61c4775a0ad2f4ab4bb8a36437cf23d2ee953311d6ea967e297510a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://74c8cea1a61c4775a0ad2f4ab4bb8a36437cf23d2ee953311d6ea967e297510a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T14:09:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T14:09:31Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://a4f2f56a809adc01fba82eba0652d78f58770c132f77b1424f8b17068b1927e3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a4f2f56a809adc01fba82eba0652d78f58770c132f77b1424f8b17068b1927e3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T14:09:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T14:09:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T14:09:30Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T14:10:19Z is after 2025-08-24T17:21:41Z" Dec 03 14:10:19 crc kubenswrapper[4805]: I1203 14:10:19.983774 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ced8291d-0ce5-4e74-9eaf-2bf033774039\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:10:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:10:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3e5a34636f0a8c4d9179074596d28bc2eb7a504573398c596a637defd829907a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://29cd95be3f964415943c873e7914ffcea4abe50dfdf17f8a9cd03fc38df35981\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://951173b3d3d581275a6f6e43209b5e0611a701ca7a10f81504d723e24a2d21a3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e64730957bbbb2feb4d506e0ed9731a0594bf66fa1e8fcdb8ec944ded6308eed\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5da55f769e23ad52bac9c77da62c78e70fa34db8ed1a435e4eaf4dad3fd2b68a\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-03T14:09:48Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1203 14:09:42.992179 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1203 14:09:42.993233 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1830926909/tls.crt::/tmp/serving-cert-1830926909/tls.key\\\\\\\"\\\\nI1203 14:09:48.622403 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1203 14:09:48.626476 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1203 14:09:48.626527 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1203 14:09:48.626574 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1203 14:09:48.626582 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1203 14:09:48.633990 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1203 14:09:48.634024 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 14:09:48.634029 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 14:09:48.634034 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1203 14:09:48.634038 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1203 14:09:48.634042 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1203 14:09:48.634045 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1203 14:09:48.634344 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1203 14:09:48.644497 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T14:09:32Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bd92a160eb530c5687ebd7cd5b83b41820a2b808a0fd9514e5ace9d237e1bd26\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:32Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://af9c940135012c629447528c61743ed1e58708ab57ce76604c6da0479aa459cf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://af9c940135012c629447528c61743ed1e58708ab57ce76604c6da0479aa459cf\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T14:09:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T14:09:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T14:09:30Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T14:10:19Z is after 2025-08-24T17:21:41Z" Dec 03 14:10:20 crc kubenswrapper[4805]: I1203 14:10:20.005105 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-bpk49" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"de00a81f-82e1-4223-9614-f98606b9a8ef\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://04f4ad27e02e2024ae9e7cd1b16116267e69cecda882dd723cdb274c41a0e7db\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-25d7t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7741e0e70416ebe094bfabc6009e98d7d4c0fbcba1fc82df13b9ba40774b456a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7741e0e70416ebe094bfabc6009e98d7d4c0fbcba1fc82df13b9ba40774b456a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T14:09:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T14:09:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-25d7t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ee3499c02eb52bbf08c447d2433d435a74dfb70670dc4cbaac44f5fbff9e8a29\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ee3499c02eb52bbf08c447d2433d435a74dfb70670dc4cbaac44f5fbff9e8a29\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T14:09:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T14:09:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-25d7t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ec8488b1d69799951be7ef40e5e028732265c4ddcdea2d453923e567864771f7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ec8488b1d69799951be7ef40e5e028732265c4ddcdea2d453923e567864771f7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T14:09:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T14:09:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-25d7t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://66707bb4e1fdacb265e89786df8a04605830f1c156394cacfcec9b54abe5d41a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://66707bb4e1fdacb265e89786df8a04605830f1c156394cacfcec9b54abe5d41a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T14:09:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T14:09:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-25d7t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6207ab6600970ff15c8713c547d4920b2c77eb9bea990bb5da497e24870dcc0f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6207ab6600970ff15c8713c547d4920b2c77eb9bea990bb5da497e24870dcc0f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T14:09:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T14:09:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-25d7t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8de65d4a205e2deec0b433e53c0e37feb94aa0ede6a4c58d7414b80829804505\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8de65d4a205e2deec0b433e53c0e37feb94aa0ede6a4c58d7414b80829804505\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T14:09:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T14:09:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-25d7t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T14:09:50Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-bpk49\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T14:10:20Z is after 2025-08-24T17:21:41Z" Dec 03 14:10:20 crc kubenswrapper[4805]: I1203 14:10:20.014891 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 14:10:20 crc kubenswrapper[4805]: I1203 14:10:20.014955 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 14:10:20 crc kubenswrapper[4805]: I1203 14:10:20.014966 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 14:10:20 crc kubenswrapper[4805]: I1203 14:10:20.014990 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 14:10:20 crc kubenswrapper[4805]: I1203 14:10:20.015003 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T14:10:20Z","lastTransitionTime":"2025-12-03T14:10:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 14:10:20 crc kubenswrapper[4805]: I1203 14:10:20.023640 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"750d2b4b-738a-4992-8902-742528f6e875\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://abd2d0156113bb5182727a53aac594410ed051c19258beb8a07b808eff0b42a7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8673a29c0e16f4e707d79f0610722ccabb023ff2b4a18d7cf9d305183817d952\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://35766e7206b1be7a53d13d6890d759cf81dbd999643c12349268e4bf16fbbdb0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5b9844a11b2b903978d0d464b469e26c9211100da61d018dde9636dca95b684e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T14:09:30Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T14:10:20Z is after 2025-08-24T17:21:41Z" Dec 03 14:10:20 crc kubenswrapper[4805]: I1203 14:10:20.040702 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:48Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:48Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T14:10:20Z is after 2025-08-24T17:21:41Z" Dec 03 14:10:20 crc kubenswrapper[4805]: I1203 14:10:20.058455 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-k4mhs" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c09c1cdc-f461-45c9-8444-5b8764bccabc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3354c1a44f891182d04116fc76234a2b337d50567d032a3584f902a9516387cc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qdcd7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T14:09:50Z\\\"}}\" for pod \"openshift-multus\"/\"multus-k4mhs\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T14:10:20Z is after 2025-08-24T17:21:41Z" Dec 03 14:10:20 crc kubenswrapper[4805]: I1203 14:10:20.117066 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 14:10:20 crc kubenswrapper[4805]: I1203 14:10:20.117100 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 14:10:20 crc kubenswrapper[4805]: I1203 14:10:20.117108 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 14:10:20 crc kubenswrapper[4805]: I1203 14:10:20.117121 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 14:10:20 crc kubenswrapper[4805]: I1203 14:10:20.117130 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T14:10:20Z","lastTransitionTime":"2025-12-03T14:10:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 14:10:20 crc kubenswrapper[4805]: I1203 14:10:20.219954 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 14:10:20 crc kubenswrapper[4805]: I1203 14:10:20.220024 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 14:10:20 crc kubenswrapper[4805]: I1203 14:10:20.220048 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 14:10:20 crc kubenswrapper[4805]: I1203 14:10:20.220080 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 14:10:20 crc kubenswrapper[4805]: I1203 14:10:20.220152 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T14:10:20Z","lastTransitionTime":"2025-12-03T14:10:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 14:10:20 crc kubenswrapper[4805]: I1203 14:10:20.324313 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 14:10:20 crc kubenswrapper[4805]: I1203 14:10:20.324370 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 14:10:20 crc kubenswrapper[4805]: I1203 14:10:20.324386 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 14:10:20 crc kubenswrapper[4805]: I1203 14:10:20.324410 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 14:10:20 crc kubenswrapper[4805]: I1203 14:10:20.324428 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T14:10:20Z","lastTransitionTime":"2025-12-03T14:10:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 14:10:20 crc kubenswrapper[4805]: I1203 14:10:20.427077 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 14:10:20 crc kubenswrapper[4805]: I1203 14:10:20.427121 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 14:10:20 crc kubenswrapper[4805]: I1203 14:10:20.427132 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 14:10:20 crc kubenswrapper[4805]: I1203 14:10:20.427149 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 14:10:20 crc kubenswrapper[4805]: I1203 14:10:20.427162 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T14:10:20Z","lastTransitionTime":"2025-12-03T14:10:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 14:10:20 crc kubenswrapper[4805]: I1203 14:10:20.485299 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/6b36a3db-1047-487a-9d6c-4f51ffdd8cc1-metrics-certs\") pod \"network-metrics-daemon-4gd9r\" (UID: \"6b36a3db-1047-487a-9d6c-4f51ffdd8cc1\") " pod="openshift-multus/network-metrics-daemon-4gd9r" Dec 03 14:10:20 crc kubenswrapper[4805]: E1203 14:10:20.485492 4805 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Dec 03 14:10:20 crc kubenswrapper[4805]: E1203 14:10:20.485583 4805 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/6b36a3db-1047-487a-9d6c-4f51ffdd8cc1-metrics-certs podName:6b36a3db-1047-487a-9d6c-4f51ffdd8cc1 nodeName:}" failed. No retries permitted until 2025-12-03 14:10:36.485553881 +0000 UTC m=+66.148470844 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/6b36a3db-1047-487a-9d6c-4f51ffdd8cc1-metrics-certs") pod "network-metrics-daemon-4gd9r" (UID: "6b36a3db-1047-487a-9d6c-4f51ffdd8cc1") : object "openshift-multus"/"metrics-daemon-secret" not registered Dec 03 14:10:20 crc kubenswrapper[4805]: I1203 14:10:20.530126 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 14:10:20 crc kubenswrapper[4805]: I1203 14:10:20.530190 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 14:10:20 crc kubenswrapper[4805]: I1203 14:10:20.530213 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 14:10:20 crc kubenswrapper[4805]: I1203 14:10:20.530244 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 14:10:20 crc kubenswrapper[4805]: I1203 14:10:20.530266 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T14:10:20Z","lastTransitionTime":"2025-12-03T14:10:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 14:10:20 crc kubenswrapper[4805]: I1203 14:10:20.585936 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 14:10:20 crc kubenswrapper[4805]: E1203 14:10:20.586245 4805 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-03 14:10:52.586203652 +0000 UTC m=+82.249120625 (durationBeforeRetry 32s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 14:10:20 crc kubenswrapper[4805]: I1203 14:10:20.633660 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 14:10:20 crc kubenswrapper[4805]: I1203 14:10:20.633791 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 14:10:20 crc kubenswrapper[4805]: I1203 14:10:20.633890 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 14:10:20 crc kubenswrapper[4805]: I1203 14:10:20.633964 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 14:10:20 crc kubenswrapper[4805]: I1203 14:10:20.634030 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T14:10:20Z","lastTransitionTime":"2025-12-03T14:10:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 14:10:20 crc kubenswrapper[4805]: I1203 14:10:20.687811 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 14:10:20 crc kubenswrapper[4805]: I1203 14:10:20.688072 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 14:10:20 crc kubenswrapper[4805]: I1203 14:10:20.688210 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 14:10:20 crc kubenswrapper[4805]: E1203 14:10:20.688133 4805 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 03 14:10:20 crc kubenswrapper[4805]: E1203 14:10:20.688310 4805 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 03 14:10:20 crc kubenswrapper[4805]: E1203 14:10:20.688325 4805 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 03 14:10:20 crc kubenswrapper[4805]: E1203 14:10:20.688146 4805 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Dec 03 14:10:20 crc kubenswrapper[4805]: I1203 14:10:20.688425 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 14:10:20 crc kubenswrapper[4805]: E1203 14:10:20.688636 4805 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 03 14:10:20 crc kubenswrapper[4805]: E1203 14:10:20.688751 4805 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 03 14:10:20 crc kubenswrapper[4805]: E1203 14:10:20.688834 4805 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 03 14:10:20 crc kubenswrapper[4805]: E1203 14:10:20.688929 4805 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 03 14:10:20 crc kubenswrapper[4805]: E1203 14:10:20.689079 4805 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-12-03 14:10:52.688433467 +0000 UTC m=+82.351350380 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 03 14:10:20 crc kubenswrapper[4805]: E1203 14:10:20.689177 4805 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-03 14:10:52.689165388 +0000 UTC m=+82.352082321 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Dec 03 14:10:20 crc kubenswrapper[4805]: E1203 14:10:20.689259 4805 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-03 14:10:52.6892505 +0000 UTC m=+82.352167423 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 03 14:10:20 crc kubenswrapper[4805]: E1203 14:10:20.689328 4805 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-12-03 14:10:52.689321462 +0000 UTC m=+82.352238385 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 03 14:10:20 crc kubenswrapper[4805]: I1203 14:10:20.693745 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 14:10:20 crc kubenswrapper[4805]: E1203 14:10:20.693976 4805 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 03 14:10:20 crc kubenswrapper[4805]: I1203 14:10:20.694127 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-4gd9r" Dec 03 14:10:20 crc kubenswrapper[4805]: I1203 14:10:20.694172 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 14:10:20 crc kubenswrapper[4805]: E1203 14:10:20.694282 4805 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-4gd9r" podUID="6b36a3db-1047-487a-9d6c-4f51ffdd8cc1" Dec 03 14:10:20 crc kubenswrapper[4805]: I1203 14:10:20.694136 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 14:10:20 crc kubenswrapper[4805]: E1203 14:10:20.694417 4805 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 03 14:10:20 crc kubenswrapper[4805]: E1203 14:10:20.694559 4805 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 03 14:10:20 crc kubenswrapper[4805]: I1203 14:10:20.711191 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"750d2b4b-738a-4992-8902-742528f6e875\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://abd2d0156113bb5182727a53aac594410ed051c19258beb8a07b808eff0b42a7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8673a29c0e16f4e707d79f0610722ccabb023ff2b4a18d7cf9d305183817d952\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://35766e7206b1be7a53d13d6890d759cf81dbd999643c12349268e4bf16fbbdb0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5b9844a11b2b903978d0d464b469e26c9211100da61d018dde9636dca95b684e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T14:09:30Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T14:10:20Z is after 2025-08-24T17:21:41Z" Dec 03 14:10:20 crc kubenswrapper[4805]: I1203 14:10:20.724903 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:48Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:48Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T14:10:20Z is after 2025-08-24T17:21:41Z" Dec 03 14:10:20 crc kubenswrapper[4805]: I1203 14:10:20.736724 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 14:10:20 crc kubenswrapper[4805]: I1203 14:10:20.736988 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 14:10:20 crc kubenswrapper[4805]: I1203 14:10:20.737076 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 14:10:20 crc kubenswrapper[4805]: I1203 14:10:20.737143 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 14:10:20 crc kubenswrapper[4805]: I1203 14:10:20.737208 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T14:10:20Z","lastTransitionTime":"2025-12-03T14:10:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 14:10:20 crc kubenswrapper[4805]: I1203 14:10:20.740543 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-k4mhs" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c09c1cdc-f461-45c9-8444-5b8764bccabc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3354c1a44f891182d04116fc76234a2b337d50567d032a3584f902a9516387cc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qdcd7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T14:09:50Z\\\"}}\" for pod \"openshift-multus\"/\"multus-k4mhs\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T14:10:20Z is after 2025-08-24T17:21:41Z" Dec 03 14:10:20 crc kubenswrapper[4805]: I1203 14:10:20.761194 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-bpk49" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"de00a81f-82e1-4223-9614-f98606b9a8ef\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://04f4ad27e02e2024ae9e7cd1b16116267e69cecda882dd723cdb274c41a0e7db\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-25d7t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7741e0e70416ebe094bfabc6009e98d7d4c0fbcba1fc82df13b9ba40774b456a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7741e0e70416ebe094bfabc6009e98d7d4c0fbcba1fc82df13b9ba40774b456a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T14:09:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T14:09:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-25d7t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ee3499c02eb52bbf08c447d2433d435a74dfb70670dc4cbaac44f5fbff9e8a29\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ee3499c02eb52bbf08c447d2433d435a74dfb70670dc4cbaac44f5fbff9e8a29\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T14:09:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T14:09:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-25d7t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ec8488b1d69799951be7ef40e5e028732265c4ddcdea2d453923e567864771f7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ec8488b1d69799951be7ef40e5e028732265c4ddcdea2d453923e567864771f7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T14:09:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T14:09:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-25d7t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://66707bb4e1fdacb265e89786df8a04605830f1c156394cacfcec9b54abe5d41a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://66707bb4e1fdacb265e89786df8a04605830f1c156394cacfcec9b54abe5d41a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T14:09:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T14:09:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-25d7t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6207ab6600970ff15c8713c547d4920b2c77eb9bea990bb5da497e24870dcc0f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6207ab6600970ff15c8713c547d4920b2c77eb9bea990bb5da497e24870dcc0f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T14:09:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T14:09:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-25d7t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8de65d4a205e2deec0b433e53c0e37feb94aa0ede6a4c58d7414b80829804505\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8de65d4a205e2deec0b433e53c0e37feb94aa0ede6a4c58d7414b80829804505\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T14:09:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T14:09:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-25d7t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T14:09:50Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-bpk49\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T14:10:20Z is after 2025-08-24T17:21:41Z" Dec 03 14:10:20 crc kubenswrapper[4805]: I1203 14:10:20.778021 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:48Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:48Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T14:10:20Z is after 2025-08-24T17:21:41Z" Dec 03 14:10:20 crc kubenswrapper[4805]: I1203 14:10:20.803923 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:49Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:49Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://541ace7ccbb66dbd191c3ca8c1e4479e3edbf80dfc67326bea8aaee81fdcfc3c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://821e3e9a902aacecf6c41e8b8d8aa1ac951ce2fbb6505b7148b16348a3754e55\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T14:10:20Z is after 2025-08-24T17:21:41Z" Dec 03 14:10:20 crc kubenswrapper[4805]: I1203 14:10:20.815552 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-gskh4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6098937f-e3f6-45e8-a647-4994a79cd711\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://28d5653f26ec61647808b3a733ff7eac70aa69ac184660f24147e8cfaf0eb802\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nwrw9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9ad6b68a42fd33ee167e67873c6eed3f11cdf6dc71d22e952c021436eb967dd9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nwrw9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T14:09:50Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-gskh4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T14:10:20Z is after 2025-08-24T17:21:41Z" Dec 03 14:10:20 crc kubenswrapper[4805]: I1203 14:10:20.831880 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-psjpm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"deeec30d-cad2-4dfd-adb6-64f6646581be\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:50Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:50Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4b8d00cdf0f37f2e5cd4e952cf85d3de8d72c43fdf786088fd9f26ed17b1e21f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wdlzx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bd67c6de01a8d196a1f51d4d90640c9e69de6f05d5e201b25e6eaa5c33440ea8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wdlzx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cb8bf05ef093eb0727a066d4bbbdf9af9422867cd4dadacafc6eeb016867a61e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wdlzx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0304467102ed053628c0d3e06548c74b6492336b6fd044246508d1e8bdb17bc3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wdlzx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6ce80ccafb3262a11fbe9923317a9d38b61e5dc5b3e7e7ae136f143d6b643fd7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wdlzx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://47d76962beb8c0f196d50af96fb3ec611097823ca80a1c8bd718f2d18840128c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wdlzx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a0272667e733f2c5eec0fce0eb9c31a16c048babdaed90635895d689ae9303f4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a0272667e733f2c5eec0fce0eb9c31a16c048babdaed90635895d689ae9303f4\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-03T14:10:16Z\\\",\\\"message\\\":\\\"o create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T14:10:16Z is after 2025-08-24T17:21:41Z]\\\\nI1203 14:10:16.717659 6478 services_controller.go:473] Services do not match for network=default, existing lbs: []services.LB{services.LB{Name:\\\\\\\"Service_openshift-network-console/networking-console-plugin_TCP_cluster\\\\\\\", UUID:\\\\\\\"ab0b1d51-5ec6-479b-8881-93dfa8d30337\\\\\\\", Protocol:\\\\\\\"tcp\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-network-console/networking-console-plugin\\\\\\\"}, Opts:services.LBOpts{Reject:false, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{}, Templates:services.TemplateMap{}, Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGr\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T14:10:15Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-psjpm_openshift-ovn-kubernetes(deeec30d-cad2-4dfd-adb6-64f6646581be)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wdlzx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ab98e365c1a4242a47b8cbe2dd85109d366d6002c4de1f18ddc950a46bea6c12\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wdlzx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2f875252fbe6a673a72fc9c56bb96899f62732305ad77b3108e3b244bfa91ecb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2f875252fbe6a673a72fc9c56bb96899f62732305ad77b3108e3b244bfa91ecb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T14:09:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T14:09:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wdlzx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T14:09:50Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-psjpm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T14:10:20Z is after 2025-08-24T17:21:41Z" Dec 03 14:10:20 crc kubenswrapper[4805]: I1203 14:10:20.838784 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 14:10:20 crc kubenswrapper[4805]: I1203 14:10:20.838828 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 14:10:20 crc kubenswrapper[4805]: I1203 14:10:20.838867 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 14:10:20 crc kubenswrapper[4805]: I1203 14:10:20.838891 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 14:10:20 crc kubenswrapper[4805]: I1203 14:10:20.838907 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T14:10:20Z","lastTransitionTime":"2025-12-03T14:10:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 14:10:20 crc kubenswrapper[4805]: I1203 14:10:20.843207 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-4gd9r" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6b36a3db-1047-487a-9d6c-4f51ffdd8cc1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:10:04Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:10:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:10:04Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:10:04Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4t6zl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4t6zl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T14:10:04Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-4gd9r\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T14:10:20Z is after 2025-08-24T17:21:41Z" Dec 03 14:10:20 crc kubenswrapper[4805]: I1203 14:10:20.852787 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-pdzgk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"860769de-6a5b-4cb9-a752-264e7e2e785f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5e2f0a0cc21505d609f81918365d5e4f5d21959ac12ab147d6a51601c9de4c14\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ggn4n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T14:09:50Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-pdzgk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T14:10:20Z is after 2025-08-24T17:21:41Z" Dec 03 14:10:20 crc kubenswrapper[4805]: I1203 14:10:20.863096 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-69h75" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5d1c694f-9847-4086-8b37-3573ddeb3ace\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:10:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:10:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:10:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:10:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://abc1d9b8582ca9fe4a51fb608fb39129a179ddcdfde0797a6369a55b2e0124cf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:10:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s578b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c7c4d434b035cac2a91a4f0846a2f4fb5bd68f881c28487860adb6db85e8ce75\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:10:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s578b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T14:10:03Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-69h75\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T14:10:20Z is after 2025-08-24T17:21:41Z" Dec 03 14:10:20 crc kubenswrapper[4805]: I1203 14:10:20.875150 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:51Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:51Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b3123d3fa879097db88ed2f3cfb05631f440c24c5405503b2de3546f12e2e899\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T14:10:20Z is after 2025-08-24T17:21:41Z" Dec 03 14:10:20 crc kubenswrapper[4805]: I1203 14:10:20.884005 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-bdtdg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"53a0e9e3-b6d8-44b7-a32b-aa3450442865\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f557e17ad105d6972d3c1480656c6073f67e28c0f5773aa365549951f4d161c7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9ww54\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T14:09:51Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-bdtdg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T14:10:20Z is after 2025-08-24T17:21:41Z" Dec 03 14:10:20 crc kubenswrapper[4805]: I1203 14:10:20.906092 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b6d36671-7173-4e9c-843f-005b4dcf576c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a837d7d491eaacbe4279b4014bb5fcbe9b514ba5cc82c708227bcd9b25a346e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6e5f04faf3b23197bb8ea11d3e574dc86f31e21d798734962001ec8c95de7c84\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2ff1a979387da41b00e95eb9cc8b2c9de966380a12fae687591df3c5b18a4497\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1e66e1cb18573a14e3c3ae16af4dbd9f8997a6ad38680c3ab4505646ebd48e7c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7d5b5c85e3de43e7e95111aa967248550607c26200c783251b68696f202f92a1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://474f72dac37a808fc4b9832306e76979b151d7221807ff3086013466ed1419c3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://474f72dac37a808fc4b9832306e76979b151d7221807ff3086013466ed1419c3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T14:09:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T14:09:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://74c8cea1a61c4775a0ad2f4ab4bb8a36437cf23d2ee953311d6ea967e297510a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://74c8cea1a61c4775a0ad2f4ab4bb8a36437cf23d2ee953311d6ea967e297510a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T14:09:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T14:09:31Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://a4f2f56a809adc01fba82eba0652d78f58770c132f77b1424f8b17068b1927e3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a4f2f56a809adc01fba82eba0652d78f58770c132f77b1424f8b17068b1927e3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T14:09:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T14:09:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T14:09:30Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T14:10:20Z is after 2025-08-24T17:21:41Z" Dec 03 14:10:20 crc kubenswrapper[4805]: I1203 14:10:20.921384 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ced8291d-0ce5-4e74-9eaf-2bf033774039\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:10:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:10:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3e5a34636f0a8c4d9179074596d28bc2eb7a504573398c596a637defd829907a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://29cd95be3f964415943c873e7914ffcea4abe50dfdf17f8a9cd03fc38df35981\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://951173b3d3d581275a6f6e43209b5e0611a701ca7a10f81504d723e24a2d21a3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e64730957bbbb2feb4d506e0ed9731a0594bf66fa1e8fcdb8ec944ded6308eed\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5da55f769e23ad52bac9c77da62c78e70fa34db8ed1a435e4eaf4dad3fd2b68a\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-03T14:09:48Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1203 14:09:42.992179 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1203 14:09:42.993233 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1830926909/tls.crt::/tmp/serving-cert-1830926909/tls.key\\\\\\\"\\\\nI1203 14:09:48.622403 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1203 14:09:48.626476 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1203 14:09:48.626527 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1203 14:09:48.626574 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1203 14:09:48.626582 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1203 14:09:48.633990 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1203 14:09:48.634024 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 14:09:48.634029 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 14:09:48.634034 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1203 14:09:48.634038 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1203 14:09:48.634042 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1203 14:09:48.634045 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1203 14:09:48.634344 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1203 14:09:48.644497 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T14:09:32Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bd92a160eb530c5687ebd7cd5b83b41820a2b808a0fd9514e5ace9d237e1bd26\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:32Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://af9c940135012c629447528c61743ed1e58708ab57ce76604c6da0479aa459cf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://af9c940135012c629447528c61743ed1e58708ab57ce76604c6da0479aa459cf\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T14:09:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T14:09:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T14:09:30Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T14:10:20Z is after 2025-08-24T17:21:41Z" Dec 03 14:10:20 crc kubenswrapper[4805]: I1203 14:10:20.933218 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6d48b4f1-5dfc-4f6b-a4e8-4e465c651c7b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:10:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:10:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5ea698a1e0de3b5201e2f9032d834d81d9aab9c2fb17aca582184c7090f1bd2f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://66aa257562f8e113fc1cd2936b8aa18e927f90018ad4c11a184c1c56444be648\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2fd6d50f839b603a90e5d00ab6e36e832eab2e299b7af1dbb84498c35bf1c08e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://77c26dae6ccd328a8a1ce4fe6ffe94c78061d8d3ac3727adb80a24c16f4ab533\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://77c26dae6ccd328a8a1ce4fe6ffe94c78061d8d3ac3727adb80a24c16f4ab533\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T14:09:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T14:09:31Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T14:09:30Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T14:10:20Z is after 2025-08-24T17:21:41Z" Dec 03 14:10:20 crc kubenswrapper[4805]: I1203 14:10:20.940888 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 14:10:20 crc kubenswrapper[4805]: I1203 14:10:20.940929 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 14:10:20 crc kubenswrapper[4805]: I1203 14:10:20.940947 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 14:10:20 crc kubenswrapper[4805]: I1203 14:10:20.940964 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 14:10:20 crc kubenswrapper[4805]: I1203 14:10:20.940976 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T14:10:20Z","lastTransitionTime":"2025-12-03T14:10:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 14:10:20 crc kubenswrapper[4805]: I1203 14:10:20.946639 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:49Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:49Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://38339331a3209d6787d1def81b72f4a8514bf91080a139cad6d8ac5cc6bef10e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T14:10:20Z is after 2025-08-24T17:21:41Z" Dec 03 14:10:20 crc kubenswrapper[4805]: I1203 14:10:20.959355 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:48Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:48Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T14:10:20Z is after 2025-08-24T17:21:41Z" Dec 03 14:10:21 crc kubenswrapper[4805]: I1203 14:10:21.044222 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 14:10:21 crc kubenswrapper[4805]: I1203 14:10:21.044312 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 14:10:21 crc kubenswrapper[4805]: I1203 14:10:21.044326 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 14:10:21 crc kubenswrapper[4805]: I1203 14:10:21.044343 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 14:10:21 crc kubenswrapper[4805]: I1203 14:10:21.044384 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T14:10:21Z","lastTransitionTime":"2025-12-03T14:10:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 14:10:21 crc kubenswrapper[4805]: I1203 14:10:21.147284 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 14:10:21 crc kubenswrapper[4805]: I1203 14:10:21.147342 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 14:10:21 crc kubenswrapper[4805]: I1203 14:10:21.147360 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 14:10:21 crc kubenswrapper[4805]: I1203 14:10:21.147385 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 14:10:21 crc kubenswrapper[4805]: I1203 14:10:21.147404 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T14:10:21Z","lastTransitionTime":"2025-12-03T14:10:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 14:10:21 crc kubenswrapper[4805]: I1203 14:10:21.250467 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 14:10:21 crc kubenswrapper[4805]: I1203 14:10:21.250798 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 14:10:21 crc kubenswrapper[4805]: I1203 14:10:21.250812 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 14:10:21 crc kubenswrapper[4805]: I1203 14:10:21.250827 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 14:10:21 crc kubenswrapper[4805]: I1203 14:10:21.250841 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T14:10:21Z","lastTransitionTime":"2025-12-03T14:10:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 14:10:21 crc kubenswrapper[4805]: I1203 14:10:21.354100 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 14:10:21 crc kubenswrapper[4805]: I1203 14:10:21.354167 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 14:10:21 crc kubenswrapper[4805]: I1203 14:10:21.354191 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 14:10:21 crc kubenswrapper[4805]: I1203 14:10:21.354223 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 14:10:21 crc kubenswrapper[4805]: I1203 14:10:21.354245 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T14:10:21Z","lastTransitionTime":"2025-12-03T14:10:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 14:10:21 crc kubenswrapper[4805]: I1203 14:10:21.456571 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 14:10:21 crc kubenswrapper[4805]: I1203 14:10:21.456613 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 14:10:21 crc kubenswrapper[4805]: I1203 14:10:21.456623 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 14:10:21 crc kubenswrapper[4805]: I1203 14:10:21.456639 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 14:10:21 crc kubenswrapper[4805]: I1203 14:10:21.456651 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T14:10:21Z","lastTransitionTime":"2025-12-03T14:10:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 14:10:21 crc kubenswrapper[4805]: I1203 14:10:21.558875 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 14:10:21 crc kubenswrapper[4805]: I1203 14:10:21.558928 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 14:10:21 crc kubenswrapper[4805]: I1203 14:10:21.558936 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 14:10:21 crc kubenswrapper[4805]: I1203 14:10:21.558951 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 14:10:21 crc kubenswrapper[4805]: I1203 14:10:21.558962 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T14:10:21Z","lastTransitionTime":"2025-12-03T14:10:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 14:10:21 crc kubenswrapper[4805]: I1203 14:10:21.661671 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 14:10:21 crc kubenswrapper[4805]: I1203 14:10:21.661962 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 14:10:21 crc kubenswrapper[4805]: I1203 14:10:21.662067 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 14:10:21 crc kubenswrapper[4805]: I1203 14:10:21.662179 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 14:10:21 crc kubenswrapper[4805]: I1203 14:10:21.662283 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T14:10:21Z","lastTransitionTime":"2025-12-03T14:10:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 14:10:21 crc kubenswrapper[4805]: I1203 14:10:21.765144 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 14:10:21 crc kubenswrapper[4805]: I1203 14:10:21.765175 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 14:10:21 crc kubenswrapper[4805]: I1203 14:10:21.765186 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 14:10:21 crc kubenswrapper[4805]: I1203 14:10:21.765200 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 14:10:21 crc kubenswrapper[4805]: I1203 14:10:21.765209 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T14:10:21Z","lastTransitionTime":"2025-12-03T14:10:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 14:10:21 crc kubenswrapper[4805]: I1203 14:10:21.867617 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 14:10:21 crc kubenswrapper[4805]: I1203 14:10:21.867672 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 14:10:21 crc kubenswrapper[4805]: I1203 14:10:21.867689 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 14:10:21 crc kubenswrapper[4805]: I1203 14:10:21.867712 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 14:10:21 crc kubenswrapper[4805]: I1203 14:10:21.867733 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T14:10:21Z","lastTransitionTime":"2025-12-03T14:10:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 14:10:21 crc kubenswrapper[4805]: I1203 14:10:21.970832 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 14:10:21 crc kubenswrapper[4805]: I1203 14:10:21.970975 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 14:10:21 crc kubenswrapper[4805]: I1203 14:10:21.970997 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 14:10:21 crc kubenswrapper[4805]: I1203 14:10:21.971023 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 14:10:21 crc kubenswrapper[4805]: I1203 14:10:21.971039 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T14:10:21Z","lastTransitionTime":"2025-12-03T14:10:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 14:10:22 crc kubenswrapper[4805]: I1203 14:10:22.072641 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 14:10:22 crc kubenswrapper[4805]: I1203 14:10:22.072685 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 14:10:22 crc kubenswrapper[4805]: I1203 14:10:22.072696 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 14:10:22 crc kubenswrapper[4805]: I1203 14:10:22.072712 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 14:10:22 crc kubenswrapper[4805]: I1203 14:10:22.072724 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T14:10:22Z","lastTransitionTime":"2025-12-03T14:10:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 14:10:22 crc kubenswrapper[4805]: I1203 14:10:22.174919 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 14:10:22 crc kubenswrapper[4805]: I1203 14:10:22.174956 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 14:10:22 crc kubenswrapper[4805]: I1203 14:10:22.174965 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 14:10:22 crc kubenswrapper[4805]: I1203 14:10:22.174980 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 14:10:22 crc kubenswrapper[4805]: I1203 14:10:22.174989 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T14:10:22Z","lastTransitionTime":"2025-12-03T14:10:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 14:10:22 crc kubenswrapper[4805]: I1203 14:10:22.277795 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 14:10:22 crc kubenswrapper[4805]: I1203 14:10:22.277918 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 14:10:22 crc kubenswrapper[4805]: I1203 14:10:22.277941 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 14:10:22 crc kubenswrapper[4805]: I1203 14:10:22.277967 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 14:10:22 crc kubenswrapper[4805]: I1203 14:10:22.277989 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T14:10:22Z","lastTransitionTime":"2025-12-03T14:10:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 14:10:22 crc kubenswrapper[4805]: I1203 14:10:22.381417 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 14:10:22 crc kubenswrapper[4805]: I1203 14:10:22.381487 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 14:10:22 crc kubenswrapper[4805]: I1203 14:10:22.381510 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 14:10:22 crc kubenswrapper[4805]: I1203 14:10:22.381540 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 14:10:22 crc kubenswrapper[4805]: I1203 14:10:22.381562 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T14:10:22Z","lastTransitionTime":"2025-12-03T14:10:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 14:10:22 crc kubenswrapper[4805]: I1203 14:10:22.484617 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 14:10:22 crc kubenswrapper[4805]: I1203 14:10:22.484667 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 14:10:22 crc kubenswrapper[4805]: I1203 14:10:22.484679 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 14:10:22 crc kubenswrapper[4805]: I1203 14:10:22.484696 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 14:10:22 crc kubenswrapper[4805]: I1203 14:10:22.484711 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T14:10:22Z","lastTransitionTime":"2025-12-03T14:10:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 14:10:22 crc kubenswrapper[4805]: I1203 14:10:22.587495 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 14:10:22 crc kubenswrapper[4805]: I1203 14:10:22.587551 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 14:10:22 crc kubenswrapper[4805]: I1203 14:10:22.587572 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 14:10:22 crc kubenswrapper[4805]: I1203 14:10:22.587597 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 14:10:22 crc kubenswrapper[4805]: I1203 14:10:22.587620 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T14:10:22Z","lastTransitionTime":"2025-12-03T14:10:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 14:10:22 crc kubenswrapper[4805]: I1203 14:10:22.689544 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 14:10:22 crc kubenswrapper[4805]: I1203 14:10:22.689592 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 14:10:22 crc kubenswrapper[4805]: I1203 14:10:22.689601 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 14:10:22 crc kubenswrapper[4805]: I1203 14:10:22.689616 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 14:10:22 crc kubenswrapper[4805]: I1203 14:10:22.689626 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T14:10:22Z","lastTransitionTime":"2025-12-03T14:10:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 14:10:22 crc kubenswrapper[4805]: I1203 14:10:22.693764 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 14:10:22 crc kubenswrapper[4805]: I1203 14:10:22.693824 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 14:10:22 crc kubenswrapper[4805]: E1203 14:10:22.693863 4805 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 03 14:10:22 crc kubenswrapper[4805]: I1203 14:10:22.693899 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 14:10:22 crc kubenswrapper[4805]: E1203 14:10:22.693947 4805 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 03 14:10:22 crc kubenswrapper[4805]: I1203 14:10:22.693957 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-4gd9r" Dec 03 14:10:22 crc kubenswrapper[4805]: E1203 14:10:22.694023 4805 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 03 14:10:22 crc kubenswrapper[4805]: E1203 14:10:22.694124 4805 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-4gd9r" podUID="6b36a3db-1047-487a-9d6c-4f51ffdd8cc1" Dec 03 14:10:22 crc kubenswrapper[4805]: I1203 14:10:22.792138 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 14:10:22 crc kubenswrapper[4805]: I1203 14:10:22.792181 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 14:10:22 crc kubenswrapper[4805]: I1203 14:10:22.792191 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 14:10:22 crc kubenswrapper[4805]: I1203 14:10:22.792206 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 14:10:22 crc kubenswrapper[4805]: I1203 14:10:22.792215 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T14:10:22Z","lastTransitionTime":"2025-12-03T14:10:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 14:10:22 crc kubenswrapper[4805]: I1203 14:10:22.811886 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 14:10:22 crc kubenswrapper[4805]: I1203 14:10:22.811973 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 14:10:22 crc kubenswrapper[4805]: I1203 14:10:22.811991 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 14:10:22 crc kubenswrapper[4805]: I1203 14:10:22.812014 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 14:10:22 crc kubenswrapper[4805]: I1203 14:10:22.812030 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T14:10:22Z","lastTransitionTime":"2025-12-03T14:10:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 14:10:22 crc kubenswrapper[4805]: E1203 14:10:22.829321 4805 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T14:10:22Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T14:10:22Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T14:10:22Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T14:10:22Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T14:10:22Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T14:10:22Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T14:10:22Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T14:10:22Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"e4edfbcd-1de1-486e-9c0b-65f9fbba52ad\\\",\\\"systemUUID\\\":\\\"7a729b24-19ef-490e-ab4c-d31c52ed5658\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T14:10:22Z is after 2025-08-24T17:21:41Z" Dec 03 14:10:22 crc kubenswrapper[4805]: I1203 14:10:22.832549 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 14:10:22 crc kubenswrapper[4805]: I1203 14:10:22.832597 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 14:10:22 crc kubenswrapper[4805]: I1203 14:10:22.832608 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 14:10:22 crc kubenswrapper[4805]: I1203 14:10:22.832629 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 14:10:22 crc kubenswrapper[4805]: I1203 14:10:22.832642 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T14:10:22Z","lastTransitionTime":"2025-12-03T14:10:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 14:10:22 crc kubenswrapper[4805]: E1203 14:10:22.848870 4805 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T14:10:22Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T14:10:22Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T14:10:22Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T14:10:22Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T14:10:22Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T14:10:22Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T14:10:22Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T14:10:22Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"e4edfbcd-1de1-486e-9c0b-65f9fbba52ad\\\",\\\"systemUUID\\\":\\\"7a729b24-19ef-490e-ab4c-d31c52ed5658\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T14:10:22Z is after 2025-08-24T17:21:41Z" Dec 03 14:10:22 crc kubenswrapper[4805]: I1203 14:10:22.852221 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 14:10:22 crc kubenswrapper[4805]: I1203 14:10:22.852272 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 14:10:22 crc kubenswrapper[4805]: I1203 14:10:22.852315 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 14:10:22 crc kubenswrapper[4805]: I1203 14:10:22.852337 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 14:10:22 crc kubenswrapper[4805]: I1203 14:10:22.852352 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T14:10:22Z","lastTransitionTime":"2025-12-03T14:10:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 14:10:22 crc kubenswrapper[4805]: E1203 14:10:22.865992 4805 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T14:10:22Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T14:10:22Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T14:10:22Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T14:10:22Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T14:10:22Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T14:10:22Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T14:10:22Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T14:10:22Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"e4edfbcd-1de1-486e-9c0b-65f9fbba52ad\\\",\\\"systemUUID\\\":\\\"7a729b24-19ef-490e-ab4c-d31c52ed5658\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T14:10:22Z is after 2025-08-24T17:21:41Z" Dec 03 14:10:22 crc kubenswrapper[4805]: I1203 14:10:22.869264 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 14:10:22 crc kubenswrapper[4805]: I1203 14:10:22.869303 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 14:10:22 crc kubenswrapper[4805]: I1203 14:10:22.869315 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 14:10:22 crc kubenswrapper[4805]: I1203 14:10:22.869334 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 14:10:22 crc kubenswrapper[4805]: I1203 14:10:22.869346 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T14:10:22Z","lastTransitionTime":"2025-12-03T14:10:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 14:10:22 crc kubenswrapper[4805]: E1203 14:10:22.880490 4805 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T14:10:22Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T14:10:22Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T14:10:22Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T14:10:22Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T14:10:22Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T14:10:22Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T14:10:22Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T14:10:22Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"e4edfbcd-1de1-486e-9c0b-65f9fbba52ad\\\",\\\"systemUUID\\\":\\\"7a729b24-19ef-490e-ab4c-d31c52ed5658\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T14:10:22Z is after 2025-08-24T17:21:41Z" Dec 03 14:10:22 crc kubenswrapper[4805]: I1203 14:10:22.883795 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 14:10:22 crc kubenswrapper[4805]: I1203 14:10:22.883870 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 14:10:22 crc kubenswrapper[4805]: I1203 14:10:22.883885 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 14:10:22 crc kubenswrapper[4805]: I1203 14:10:22.883906 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 14:10:22 crc kubenswrapper[4805]: I1203 14:10:22.883922 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T14:10:22Z","lastTransitionTime":"2025-12-03T14:10:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 14:10:22 crc kubenswrapper[4805]: E1203 14:10:22.897439 4805 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T14:10:22Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T14:10:22Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T14:10:22Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T14:10:22Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T14:10:22Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T14:10:22Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T14:10:22Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T14:10:22Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"e4edfbcd-1de1-486e-9c0b-65f9fbba52ad\\\",\\\"systemUUID\\\":\\\"7a729b24-19ef-490e-ab4c-d31c52ed5658\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T14:10:22Z is after 2025-08-24T17:21:41Z" Dec 03 14:10:22 crc kubenswrapper[4805]: E1203 14:10:22.897596 4805 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Dec 03 14:10:22 crc kubenswrapper[4805]: I1203 14:10:22.899023 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 14:10:22 crc kubenswrapper[4805]: I1203 14:10:22.899057 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 14:10:22 crc kubenswrapper[4805]: I1203 14:10:22.899091 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 14:10:22 crc kubenswrapper[4805]: I1203 14:10:22.899105 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 14:10:22 crc kubenswrapper[4805]: I1203 14:10:22.899113 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T14:10:22Z","lastTransitionTime":"2025-12-03T14:10:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 14:10:23 crc kubenswrapper[4805]: I1203 14:10:23.002224 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 14:10:23 crc kubenswrapper[4805]: I1203 14:10:23.002315 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 14:10:23 crc kubenswrapper[4805]: I1203 14:10:23.002377 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 14:10:23 crc kubenswrapper[4805]: I1203 14:10:23.002405 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 14:10:23 crc kubenswrapper[4805]: I1203 14:10:23.002469 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T14:10:23Z","lastTransitionTime":"2025-12-03T14:10:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 14:10:23 crc kubenswrapper[4805]: I1203 14:10:23.105289 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 14:10:23 crc kubenswrapper[4805]: I1203 14:10:23.105335 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 14:10:23 crc kubenswrapper[4805]: I1203 14:10:23.105346 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 14:10:23 crc kubenswrapper[4805]: I1203 14:10:23.105362 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 14:10:23 crc kubenswrapper[4805]: I1203 14:10:23.105374 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T14:10:23Z","lastTransitionTime":"2025-12-03T14:10:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 14:10:23 crc kubenswrapper[4805]: I1203 14:10:23.207986 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 14:10:23 crc kubenswrapper[4805]: I1203 14:10:23.208041 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 14:10:23 crc kubenswrapper[4805]: I1203 14:10:23.208052 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 14:10:23 crc kubenswrapper[4805]: I1203 14:10:23.208424 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 14:10:23 crc kubenswrapper[4805]: I1203 14:10:23.208576 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T14:10:23Z","lastTransitionTime":"2025-12-03T14:10:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 14:10:23 crc kubenswrapper[4805]: I1203 14:10:23.312354 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 14:10:23 crc kubenswrapper[4805]: I1203 14:10:23.312427 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 14:10:23 crc kubenswrapper[4805]: I1203 14:10:23.312450 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 14:10:23 crc kubenswrapper[4805]: I1203 14:10:23.312480 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 14:10:23 crc kubenswrapper[4805]: I1203 14:10:23.312500 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T14:10:23Z","lastTransitionTime":"2025-12-03T14:10:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 14:10:23 crc kubenswrapper[4805]: I1203 14:10:23.415358 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 14:10:23 crc kubenswrapper[4805]: I1203 14:10:23.415394 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 14:10:23 crc kubenswrapper[4805]: I1203 14:10:23.415401 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 14:10:23 crc kubenswrapper[4805]: I1203 14:10:23.415414 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 14:10:23 crc kubenswrapper[4805]: I1203 14:10:23.415423 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T14:10:23Z","lastTransitionTime":"2025-12-03T14:10:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 14:10:23 crc kubenswrapper[4805]: I1203 14:10:23.517514 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 14:10:23 crc kubenswrapper[4805]: I1203 14:10:23.517550 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 14:10:23 crc kubenswrapper[4805]: I1203 14:10:23.517566 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 14:10:23 crc kubenswrapper[4805]: I1203 14:10:23.517583 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 14:10:23 crc kubenswrapper[4805]: I1203 14:10:23.517595 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T14:10:23Z","lastTransitionTime":"2025-12-03T14:10:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 14:10:23 crc kubenswrapper[4805]: I1203 14:10:23.620034 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 14:10:23 crc kubenswrapper[4805]: I1203 14:10:23.620066 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 14:10:23 crc kubenswrapper[4805]: I1203 14:10:23.620074 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 14:10:23 crc kubenswrapper[4805]: I1203 14:10:23.620088 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 14:10:23 crc kubenswrapper[4805]: I1203 14:10:23.620096 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T14:10:23Z","lastTransitionTime":"2025-12-03T14:10:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 14:10:23 crc kubenswrapper[4805]: I1203 14:10:23.723199 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 14:10:23 crc kubenswrapper[4805]: I1203 14:10:23.723748 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 14:10:23 crc kubenswrapper[4805]: I1203 14:10:23.723791 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 14:10:23 crc kubenswrapper[4805]: I1203 14:10:23.723832 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 14:10:23 crc kubenswrapper[4805]: I1203 14:10:23.723909 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T14:10:23Z","lastTransitionTime":"2025-12-03T14:10:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 14:10:23 crc kubenswrapper[4805]: I1203 14:10:23.826953 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 14:10:23 crc kubenswrapper[4805]: I1203 14:10:23.826998 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 14:10:23 crc kubenswrapper[4805]: I1203 14:10:23.827016 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 14:10:23 crc kubenswrapper[4805]: I1203 14:10:23.827038 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 14:10:23 crc kubenswrapper[4805]: I1203 14:10:23.827054 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T14:10:23Z","lastTransitionTime":"2025-12-03T14:10:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 14:10:23 crc kubenswrapper[4805]: I1203 14:10:23.933979 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 14:10:23 crc kubenswrapper[4805]: I1203 14:10:23.934050 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 14:10:23 crc kubenswrapper[4805]: I1203 14:10:23.934074 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 14:10:23 crc kubenswrapper[4805]: I1203 14:10:23.934102 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 14:10:23 crc kubenswrapper[4805]: I1203 14:10:23.934123 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T14:10:23Z","lastTransitionTime":"2025-12-03T14:10:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 14:10:24 crc kubenswrapper[4805]: I1203 14:10:24.037136 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 14:10:24 crc kubenswrapper[4805]: I1203 14:10:24.037209 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 14:10:24 crc kubenswrapper[4805]: I1203 14:10:24.037226 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 14:10:24 crc kubenswrapper[4805]: I1203 14:10:24.037252 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 14:10:24 crc kubenswrapper[4805]: I1203 14:10:24.037270 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T14:10:24Z","lastTransitionTime":"2025-12-03T14:10:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 14:10:24 crc kubenswrapper[4805]: I1203 14:10:24.140460 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 14:10:24 crc kubenswrapper[4805]: I1203 14:10:24.140527 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 14:10:24 crc kubenswrapper[4805]: I1203 14:10:24.140552 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 14:10:24 crc kubenswrapper[4805]: I1203 14:10:24.140581 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 14:10:24 crc kubenswrapper[4805]: I1203 14:10:24.140605 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T14:10:24Z","lastTransitionTime":"2025-12-03T14:10:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 14:10:24 crc kubenswrapper[4805]: I1203 14:10:24.243198 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 14:10:24 crc kubenswrapper[4805]: I1203 14:10:24.243247 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 14:10:24 crc kubenswrapper[4805]: I1203 14:10:24.243260 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 14:10:24 crc kubenswrapper[4805]: I1203 14:10:24.243278 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 14:10:24 crc kubenswrapper[4805]: I1203 14:10:24.243291 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T14:10:24Z","lastTransitionTime":"2025-12-03T14:10:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 14:10:24 crc kubenswrapper[4805]: I1203 14:10:24.345968 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 14:10:24 crc kubenswrapper[4805]: I1203 14:10:24.346036 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 14:10:24 crc kubenswrapper[4805]: I1203 14:10:24.346057 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 14:10:24 crc kubenswrapper[4805]: I1203 14:10:24.346085 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 14:10:24 crc kubenswrapper[4805]: I1203 14:10:24.346105 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T14:10:24Z","lastTransitionTime":"2025-12-03T14:10:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 14:10:24 crc kubenswrapper[4805]: I1203 14:10:24.448715 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 14:10:24 crc kubenswrapper[4805]: I1203 14:10:24.448740 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 14:10:24 crc kubenswrapper[4805]: I1203 14:10:24.448750 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 14:10:24 crc kubenswrapper[4805]: I1203 14:10:24.448763 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 14:10:24 crc kubenswrapper[4805]: I1203 14:10:24.448771 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T14:10:24Z","lastTransitionTime":"2025-12-03T14:10:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 14:10:24 crc kubenswrapper[4805]: I1203 14:10:24.551093 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 14:10:24 crc kubenswrapper[4805]: I1203 14:10:24.551144 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 14:10:24 crc kubenswrapper[4805]: I1203 14:10:24.551169 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 14:10:24 crc kubenswrapper[4805]: I1203 14:10:24.551185 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 14:10:24 crc kubenswrapper[4805]: I1203 14:10:24.551197 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T14:10:24Z","lastTransitionTime":"2025-12-03T14:10:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 14:10:24 crc kubenswrapper[4805]: I1203 14:10:24.653432 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 14:10:24 crc kubenswrapper[4805]: I1203 14:10:24.653542 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 14:10:24 crc kubenswrapper[4805]: I1203 14:10:24.653554 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 14:10:24 crc kubenswrapper[4805]: I1203 14:10:24.653582 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 14:10:24 crc kubenswrapper[4805]: I1203 14:10:24.653597 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T14:10:24Z","lastTransitionTime":"2025-12-03T14:10:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 14:10:24 crc kubenswrapper[4805]: I1203 14:10:24.694326 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 14:10:24 crc kubenswrapper[4805]: I1203 14:10:24.694415 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 14:10:24 crc kubenswrapper[4805]: E1203 14:10:24.694476 4805 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 03 14:10:24 crc kubenswrapper[4805]: E1203 14:10:24.694596 4805 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 03 14:10:24 crc kubenswrapper[4805]: I1203 14:10:24.694652 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 14:10:24 crc kubenswrapper[4805]: I1203 14:10:24.694702 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-4gd9r" Dec 03 14:10:24 crc kubenswrapper[4805]: E1203 14:10:24.694909 4805 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 03 14:10:24 crc kubenswrapper[4805]: E1203 14:10:24.695062 4805 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-4gd9r" podUID="6b36a3db-1047-487a-9d6c-4f51ffdd8cc1" Dec 03 14:10:24 crc kubenswrapper[4805]: I1203 14:10:24.757335 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 14:10:24 crc kubenswrapper[4805]: I1203 14:10:24.757416 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 14:10:24 crc kubenswrapper[4805]: I1203 14:10:24.757437 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 14:10:24 crc kubenswrapper[4805]: I1203 14:10:24.757463 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 14:10:24 crc kubenswrapper[4805]: I1203 14:10:24.757479 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T14:10:24Z","lastTransitionTime":"2025-12-03T14:10:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 14:10:24 crc kubenswrapper[4805]: I1203 14:10:24.860745 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 14:10:24 crc kubenswrapper[4805]: I1203 14:10:24.860833 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 14:10:24 crc kubenswrapper[4805]: I1203 14:10:24.860900 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 14:10:24 crc kubenswrapper[4805]: I1203 14:10:24.860931 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 14:10:24 crc kubenswrapper[4805]: I1203 14:10:24.860954 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T14:10:24Z","lastTransitionTime":"2025-12-03T14:10:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 14:10:24 crc kubenswrapper[4805]: I1203 14:10:24.964440 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 14:10:24 crc kubenswrapper[4805]: I1203 14:10:24.964506 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 14:10:24 crc kubenswrapper[4805]: I1203 14:10:24.964522 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 14:10:24 crc kubenswrapper[4805]: I1203 14:10:24.964551 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 14:10:24 crc kubenswrapper[4805]: I1203 14:10:24.964567 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T14:10:24Z","lastTransitionTime":"2025-12-03T14:10:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 14:10:25 crc kubenswrapper[4805]: I1203 14:10:25.066879 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 14:10:25 crc kubenswrapper[4805]: I1203 14:10:25.066921 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 14:10:25 crc kubenswrapper[4805]: I1203 14:10:25.066930 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 14:10:25 crc kubenswrapper[4805]: I1203 14:10:25.066947 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 14:10:25 crc kubenswrapper[4805]: I1203 14:10:25.066956 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T14:10:25Z","lastTransitionTime":"2025-12-03T14:10:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 14:10:25 crc kubenswrapper[4805]: I1203 14:10:25.169497 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 14:10:25 crc kubenswrapper[4805]: I1203 14:10:25.169534 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 14:10:25 crc kubenswrapper[4805]: I1203 14:10:25.169543 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 14:10:25 crc kubenswrapper[4805]: I1203 14:10:25.169557 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 14:10:25 crc kubenswrapper[4805]: I1203 14:10:25.169566 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T14:10:25Z","lastTransitionTime":"2025-12-03T14:10:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 14:10:25 crc kubenswrapper[4805]: I1203 14:10:25.272025 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 14:10:25 crc kubenswrapper[4805]: I1203 14:10:25.272086 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 14:10:25 crc kubenswrapper[4805]: I1203 14:10:25.272101 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 14:10:25 crc kubenswrapper[4805]: I1203 14:10:25.272122 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 14:10:25 crc kubenswrapper[4805]: I1203 14:10:25.272137 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T14:10:25Z","lastTransitionTime":"2025-12-03T14:10:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 14:10:25 crc kubenswrapper[4805]: I1203 14:10:25.375217 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 14:10:25 crc kubenswrapper[4805]: I1203 14:10:25.375263 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 14:10:25 crc kubenswrapper[4805]: I1203 14:10:25.375273 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 14:10:25 crc kubenswrapper[4805]: I1203 14:10:25.375289 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 14:10:25 crc kubenswrapper[4805]: I1203 14:10:25.375301 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T14:10:25Z","lastTransitionTime":"2025-12-03T14:10:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 14:10:25 crc kubenswrapper[4805]: I1203 14:10:25.478946 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 14:10:25 crc kubenswrapper[4805]: I1203 14:10:25.479006 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 14:10:25 crc kubenswrapper[4805]: I1203 14:10:25.479025 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 14:10:25 crc kubenswrapper[4805]: I1203 14:10:25.479054 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 14:10:25 crc kubenswrapper[4805]: I1203 14:10:25.479076 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T14:10:25Z","lastTransitionTime":"2025-12-03T14:10:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 14:10:25 crc kubenswrapper[4805]: I1203 14:10:25.581705 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 14:10:25 crc kubenswrapper[4805]: I1203 14:10:25.581774 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 14:10:25 crc kubenswrapper[4805]: I1203 14:10:25.581792 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 14:10:25 crc kubenswrapper[4805]: I1203 14:10:25.581869 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 14:10:25 crc kubenswrapper[4805]: I1203 14:10:25.581901 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T14:10:25Z","lastTransitionTime":"2025-12-03T14:10:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 14:10:25 crc kubenswrapper[4805]: I1203 14:10:25.684664 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 14:10:25 crc kubenswrapper[4805]: I1203 14:10:25.684965 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 14:10:25 crc kubenswrapper[4805]: I1203 14:10:25.685066 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 14:10:25 crc kubenswrapper[4805]: I1203 14:10:25.685158 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 14:10:25 crc kubenswrapper[4805]: I1203 14:10:25.685242 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T14:10:25Z","lastTransitionTime":"2025-12-03T14:10:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 14:10:25 crc kubenswrapper[4805]: I1203 14:10:25.787931 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 14:10:25 crc kubenswrapper[4805]: I1203 14:10:25.788181 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 14:10:25 crc kubenswrapper[4805]: I1203 14:10:25.788247 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 14:10:25 crc kubenswrapper[4805]: I1203 14:10:25.788313 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 14:10:25 crc kubenswrapper[4805]: I1203 14:10:25.788371 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T14:10:25Z","lastTransitionTime":"2025-12-03T14:10:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 14:10:25 crc kubenswrapper[4805]: I1203 14:10:25.890891 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 14:10:25 crc kubenswrapper[4805]: I1203 14:10:25.890946 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 14:10:25 crc kubenswrapper[4805]: I1203 14:10:25.890954 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 14:10:25 crc kubenswrapper[4805]: I1203 14:10:25.890970 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 14:10:25 crc kubenswrapper[4805]: I1203 14:10:25.890980 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T14:10:25Z","lastTransitionTime":"2025-12-03T14:10:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 14:10:25 crc kubenswrapper[4805]: I1203 14:10:25.993061 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 14:10:25 crc kubenswrapper[4805]: I1203 14:10:25.993099 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 14:10:25 crc kubenswrapper[4805]: I1203 14:10:25.993109 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 14:10:25 crc kubenswrapper[4805]: I1203 14:10:25.993124 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 14:10:25 crc kubenswrapper[4805]: I1203 14:10:25.993133 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T14:10:25Z","lastTransitionTime":"2025-12-03T14:10:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 14:10:26 crc kubenswrapper[4805]: I1203 14:10:26.095773 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 14:10:26 crc kubenswrapper[4805]: I1203 14:10:26.095875 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 14:10:26 crc kubenswrapper[4805]: I1203 14:10:26.095901 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 14:10:26 crc kubenswrapper[4805]: I1203 14:10:26.095930 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 14:10:26 crc kubenswrapper[4805]: I1203 14:10:26.095952 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T14:10:26Z","lastTransitionTime":"2025-12-03T14:10:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 14:10:26 crc kubenswrapper[4805]: I1203 14:10:26.198412 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 14:10:26 crc kubenswrapper[4805]: I1203 14:10:26.198473 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 14:10:26 crc kubenswrapper[4805]: I1203 14:10:26.198495 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 14:10:26 crc kubenswrapper[4805]: I1203 14:10:26.198524 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 14:10:26 crc kubenswrapper[4805]: I1203 14:10:26.198546 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T14:10:26Z","lastTransitionTime":"2025-12-03T14:10:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 14:10:26 crc kubenswrapper[4805]: I1203 14:10:26.301386 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 14:10:26 crc kubenswrapper[4805]: I1203 14:10:26.301468 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 14:10:26 crc kubenswrapper[4805]: I1203 14:10:26.301489 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 14:10:26 crc kubenswrapper[4805]: I1203 14:10:26.301520 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 14:10:26 crc kubenswrapper[4805]: I1203 14:10:26.301542 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T14:10:26Z","lastTransitionTime":"2025-12-03T14:10:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 14:10:26 crc kubenswrapper[4805]: I1203 14:10:26.404500 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 14:10:26 crc kubenswrapper[4805]: I1203 14:10:26.404590 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 14:10:26 crc kubenswrapper[4805]: I1203 14:10:26.404602 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 14:10:26 crc kubenswrapper[4805]: I1203 14:10:26.404622 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 14:10:26 crc kubenswrapper[4805]: I1203 14:10:26.404635 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T14:10:26Z","lastTransitionTime":"2025-12-03T14:10:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 14:10:26 crc kubenswrapper[4805]: I1203 14:10:26.507530 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 14:10:26 crc kubenswrapper[4805]: I1203 14:10:26.507560 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 14:10:26 crc kubenswrapper[4805]: I1203 14:10:26.507568 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 14:10:26 crc kubenswrapper[4805]: I1203 14:10:26.507596 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 14:10:26 crc kubenswrapper[4805]: I1203 14:10:26.507605 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T14:10:26Z","lastTransitionTime":"2025-12-03T14:10:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 14:10:26 crc kubenswrapper[4805]: I1203 14:10:26.610337 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 14:10:26 crc kubenswrapper[4805]: I1203 14:10:26.610428 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 14:10:26 crc kubenswrapper[4805]: I1203 14:10:26.610448 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 14:10:26 crc kubenswrapper[4805]: I1203 14:10:26.610468 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 14:10:26 crc kubenswrapper[4805]: I1203 14:10:26.610654 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T14:10:26Z","lastTransitionTime":"2025-12-03T14:10:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 14:10:26 crc kubenswrapper[4805]: I1203 14:10:26.693899 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 14:10:26 crc kubenswrapper[4805]: I1203 14:10:26.693921 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 14:10:26 crc kubenswrapper[4805]: I1203 14:10:26.694059 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 14:10:26 crc kubenswrapper[4805]: E1203 14:10:26.694201 4805 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 03 14:10:26 crc kubenswrapper[4805]: I1203 14:10:26.694265 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-4gd9r" Dec 03 14:10:26 crc kubenswrapper[4805]: E1203 14:10:26.694394 4805 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 03 14:10:26 crc kubenswrapper[4805]: E1203 14:10:26.694539 4805 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-4gd9r" podUID="6b36a3db-1047-487a-9d6c-4f51ffdd8cc1" Dec 03 14:10:26 crc kubenswrapper[4805]: E1203 14:10:26.694646 4805 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 03 14:10:26 crc kubenswrapper[4805]: I1203 14:10:26.713051 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 14:10:26 crc kubenswrapper[4805]: I1203 14:10:26.713112 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 14:10:26 crc kubenswrapper[4805]: I1203 14:10:26.713121 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 14:10:26 crc kubenswrapper[4805]: I1203 14:10:26.713136 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 14:10:26 crc kubenswrapper[4805]: I1203 14:10:26.713147 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T14:10:26Z","lastTransitionTime":"2025-12-03T14:10:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 14:10:26 crc kubenswrapper[4805]: I1203 14:10:26.815779 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 14:10:26 crc kubenswrapper[4805]: I1203 14:10:26.815817 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 14:10:26 crc kubenswrapper[4805]: I1203 14:10:26.815826 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 14:10:26 crc kubenswrapper[4805]: I1203 14:10:26.815857 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 14:10:26 crc kubenswrapper[4805]: I1203 14:10:26.815866 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T14:10:26Z","lastTransitionTime":"2025-12-03T14:10:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 14:10:26 crc kubenswrapper[4805]: I1203 14:10:26.918901 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 14:10:26 crc kubenswrapper[4805]: I1203 14:10:26.918970 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 14:10:26 crc kubenswrapper[4805]: I1203 14:10:26.918988 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 14:10:26 crc kubenswrapper[4805]: I1203 14:10:26.919012 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 14:10:26 crc kubenswrapper[4805]: I1203 14:10:26.919031 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T14:10:26Z","lastTransitionTime":"2025-12-03T14:10:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 14:10:27 crc kubenswrapper[4805]: I1203 14:10:27.022647 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 14:10:27 crc kubenswrapper[4805]: I1203 14:10:27.022729 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 14:10:27 crc kubenswrapper[4805]: I1203 14:10:27.022751 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 14:10:27 crc kubenswrapper[4805]: I1203 14:10:27.022781 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 14:10:27 crc kubenswrapper[4805]: I1203 14:10:27.022801 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T14:10:27Z","lastTransitionTime":"2025-12-03T14:10:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 14:10:27 crc kubenswrapper[4805]: I1203 14:10:27.125576 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 14:10:27 crc kubenswrapper[4805]: I1203 14:10:27.125632 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 14:10:27 crc kubenswrapper[4805]: I1203 14:10:27.125648 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 14:10:27 crc kubenswrapper[4805]: I1203 14:10:27.125668 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 14:10:27 crc kubenswrapper[4805]: I1203 14:10:27.125683 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T14:10:27Z","lastTransitionTime":"2025-12-03T14:10:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 14:10:27 crc kubenswrapper[4805]: I1203 14:10:27.228230 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 14:10:27 crc kubenswrapper[4805]: I1203 14:10:27.228276 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 14:10:27 crc kubenswrapper[4805]: I1203 14:10:27.228286 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 14:10:27 crc kubenswrapper[4805]: I1203 14:10:27.228303 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 14:10:27 crc kubenswrapper[4805]: I1203 14:10:27.228314 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T14:10:27Z","lastTransitionTime":"2025-12-03T14:10:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 14:10:27 crc kubenswrapper[4805]: I1203 14:10:27.330464 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 14:10:27 crc kubenswrapper[4805]: I1203 14:10:27.330523 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 14:10:27 crc kubenswrapper[4805]: I1203 14:10:27.330540 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 14:10:27 crc kubenswrapper[4805]: I1203 14:10:27.330558 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 14:10:27 crc kubenswrapper[4805]: I1203 14:10:27.330570 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T14:10:27Z","lastTransitionTime":"2025-12-03T14:10:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 14:10:27 crc kubenswrapper[4805]: I1203 14:10:27.433299 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 14:10:27 crc kubenswrapper[4805]: I1203 14:10:27.433351 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 14:10:27 crc kubenswrapper[4805]: I1203 14:10:27.433366 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 14:10:27 crc kubenswrapper[4805]: I1203 14:10:27.433382 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 14:10:27 crc kubenswrapper[4805]: I1203 14:10:27.433392 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T14:10:27Z","lastTransitionTime":"2025-12-03T14:10:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 14:10:27 crc kubenswrapper[4805]: I1203 14:10:27.535745 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 14:10:27 crc kubenswrapper[4805]: I1203 14:10:27.535793 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 14:10:27 crc kubenswrapper[4805]: I1203 14:10:27.535801 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 14:10:27 crc kubenswrapper[4805]: I1203 14:10:27.535814 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 14:10:27 crc kubenswrapper[4805]: I1203 14:10:27.535825 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T14:10:27Z","lastTransitionTime":"2025-12-03T14:10:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 14:10:27 crc kubenswrapper[4805]: I1203 14:10:27.638486 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 14:10:27 crc kubenswrapper[4805]: I1203 14:10:27.638566 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 14:10:27 crc kubenswrapper[4805]: I1203 14:10:27.638589 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 14:10:27 crc kubenswrapper[4805]: I1203 14:10:27.638614 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 14:10:27 crc kubenswrapper[4805]: I1203 14:10:27.638630 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T14:10:27Z","lastTransitionTime":"2025-12-03T14:10:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 14:10:27 crc kubenswrapper[4805]: I1203 14:10:27.741482 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 14:10:27 crc kubenswrapper[4805]: I1203 14:10:27.741597 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 14:10:27 crc kubenswrapper[4805]: I1203 14:10:27.741620 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 14:10:27 crc kubenswrapper[4805]: I1203 14:10:27.741643 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 14:10:27 crc kubenswrapper[4805]: I1203 14:10:27.741660 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T14:10:27Z","lastTransitionTime":"2025-12-03T14:10:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 14:10:27 crc kubenswrapper[4805]: I1203 14:10:27.844663 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 14:10:27 crc kubenswrapper[4805]: I1203 14:10:27.844796 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 14:10:27 crc kubenswrapper[4805]: I1203 14:10:27.844899 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 14:10:27 crc kubenswrapper[4805]: I1203 14:10:27.844943 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 14:10:27 crc kubenswrapper[4805]: I1203 14:10:27.844965 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T14:10:27Z","lastTransitionTime":"2025-12-03T14:10:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 14:10:27 crc kubenswrapper[4805]: I1203 14:10:27.949007 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 14:10:27 crc kubenswrapper[4805]: I1203 14:10:27.949073 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 14:10:27 crc kubenswrapper[4805]: I1203 14:10:27.949092 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 14:10:27 crc kubenswrapper[4805]: I1203 14:10:27.949119 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 14:10:27 crc kubenswrapper[4805]: I1203 14:10:27.949138 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T14:10:27Z","lastTransitionTime":"2025-12-03T14:10:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 14:10:28 crc kubenswrapper[4805]: I1203 14:10:28.053375 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 14:10:28 crc kubenswrapper[4805]: I1203 14:10:28.053434 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 14:10:28 crc kubenswrapper[4805]: I1203 14:10:28.053447 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 14:10:28 crc kubenswrapper[4805]: I1203 14:10:28.053468 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 14:10:28 crc kubenswrapper[4805]: I1203 14:10:28.053482 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T14:10:28Z","lastTransitionTime":"2025-12-03T14:10:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 14:10:28 crc kubenswrapper[4805]: I1203 14:10:28.156408 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 14:10:28 crc kubenswrapper[4805]: I1203 14:10:28.156453 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 14:10:28 crc kubenswrapper[4805]: I1203 14:10:28.156462 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 14:10:28 crc kubenswrapper[4805]: I1203 14:10:28.156475 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 14:10:28 crc kubenswrapper[4805]: I1203 14:10:28.156485 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T14:10:28Z","lastTransitionTime":"2025-12-03T14:10:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 14:10:28 crc kubenswrapper[4805]: I1203 14:10:28.259899 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 14:10:28 crc kubenswrapper[4805]: I1203 14:10:28.259939 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 14:10:28 crc kubenswrapper[4805]: I1203 14:10:28.259949 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 14:10:28 crc kubenswrapper[4805]: I1203 14:10:28.259986 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 14:10:28 crc kubenswrapper[4805]: I1203 14:10:28.259997 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T14:10:28Z","lastTransitionTime":"2025-12-03T14:10:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 14:10:28 crc kubenswrapper[4805]: I1203 14:10:28.362561 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 14:10:28 crc kubenswrapper[4805]: I1203 14:10:28.362626 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 14:10:28 crc kubenswrapper[4805]: I1203 14:10:28.362651 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 14:10:28 crc kubenswrapper[4805]: I1203 14:10:28.362679 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 14:10:28 crc kubenswrapper[4805]: I1203 14:10:28.362699 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T14:10:28Z","lastTransitionTime":"2025-12-03T14:10:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 14:10:28 crc kubenswrapper[4805]: I1203 14:10:28.464755 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 14:10:28 crc kubenswrapper[4805]: I1203 14:10:28.464818 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 14:10:28 crc kubenswrapper[4805]: I1203 14:10:28.464856 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 14:10:28 crc kubenswrapper[4805]: I1203 14:10:28.464878 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 14:10:28 crc kubenswrapper[4805]: I1203 14:10:28.464894 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T14:10:28Z","lastTransitionTime":"2025-12-03T14:10:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 14:10:28 crc kubenswrapper[4805]: I1203 14:10:28.567782 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 14:10:28 crc kubenswrapper[4805]: I1203 14:10:28.567816 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 14:10:28 crc kubenswrapper[4805]: I1203 14:10:28.567825 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 14:10:28 crc kubenswrapper[4805]: I1203 14:10:28.567851 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 14:10:28 crc kubenswrapper[4805]: I1203 14:10:28.567860 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T14:10:28Z","lastTransitionTime":"2025-12-03T14:10:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 14:10:28 crc kubenswrapper[4805]: I1203 14:10:28.670167 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 14:10:28 crc kubenswrapper[4805]: I1203 14:10:28.670220 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 14:10:28 crc kubenswrapper[4805]: I1203 14:10:28.670236 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 14:10:28 crc kubenswrapper[4805]: I1203 14:10:28.670259 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 14:10:28 crc kubenswrapper[4805]: I1203 14:10:28.670275 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T14:10:28Z","lastTransitionTime":"2025-12-03T14:10:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 14:10:28 crc kubenswrapper[4805]: I1203 14:10:28.694492 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 14:10:28 crc kubenswrapper[4805]: I1203 14:10:28.694545 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 14:10:28 crc kubenswrapper[4805]: I1203 14:10:28.694545 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-4gd9r" Dec 03 14:10:28 crc kubenswrapper[4805]: E1203 14:10:28.694617 4805 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 03 14:10:28 crc kubenswrapper[4805]: I1203 14:10:28.694658 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 14:10:28 crc kubenswrapper[4805]: E1203 14:10:28.694816 4805 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 03 14:10:28 crc kubenswrapper[4805]: E1203 14:10:28.694815 4805 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-4gd9r" podUID="6b36a3db-1047-487a-9d6c-4f51ffdd8cc1" Dec 03 14:10:28 crc kubenswrapper[4805]: E1203 14:10:28.695707 4805 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 03 14:10:28 crc kubenswrapper[4805]: I1203 14:10:28.772930 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 14:10:28 crc kubenswrapper[4805]: I1203 14:10:28.772995 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 14:10:28 crc kubenswrapper[4805]: I1203 14:10:28.773017 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 14:10:28 crc kubenswrapper[4805]: I1203 14:10:28.773046 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 14:10:28 crc kubenswrapper[4805]: I1203 14:10:28.773067 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T14:10:28Z","lastTransitionTime":"2025-12-03T14:10:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 14:10:28 crc kubenswrapper[4805]: I1203 14:10:28.876617 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 14:10:28 crc kubenswrapper[4805]: I1203 14:10:28.876659 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 14:10:28 crc kubenswrapper[4805]: I1203 14:10:28.876671 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 14:10:28 crc kubenswrapper[4805]: I1203 14:10:28.876690 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 14:10:28 crc kubenswrapper[4805]: I1203 14:10:28.876705 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T14:10:28Z","lastTransitionTime":"2025-12-03T14:10:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 14:10:28 crc kubenswrapper[4805]: I1203 14:10:28.979579 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 14:10:28 crc kubenswrapper[4805]: I1203 14:10:28.979626 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 14:10:28 crc kubenswrapper[4805]: I1203 14:10:28.979637 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 14:10:28 crc kubenswrapper[4805]: I1203 14:10:28.979654 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 14:10:28 crc kubenswrapper[4805]: I1203 14:10:28.979666 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T14:10:28Z","lastTransitionTime":"2025-12-03T14:10:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 14:10:29 crc kubenswrapper[4805]: I1203 14:10:29.081792 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 14:10:29 crc kubenswrapper[4805]: I1203 14:10:29.081887 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 14:10:29 crc kubenswrapper[4805]: I1203 14:10:29.081905 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 14:10:29 crc kubenswrapper[4805]: I1203 14:10:29.081930 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 14:10:29 crc kubenswrapper[4805]: I1203 14:10:29.081947 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T14:10:29Z","lastTransitionTime":"2025-12-03T14:10:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 14:10:29 crc kubenswrapper[4805]: I1203 14:10:29.185294 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 14:10:29 crc kubenswrapper[4805]: I1203 14:10:29.185342 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 14:10:29 crc kubenswrapper[4805]: I1203 14:10:29.185354 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 14:10:29 crc kubenswrapper[4805]: I1203 14:10:29.185371 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 14:10:29 crc kubenswrapper[4805]: I1203 14:10:29.185385 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T14:10:29Z","lastTransitionTime":"2025-12-03T14:10:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 14:10:29 crc kubenswrapper[4805]: I1203 14:10:29.288204 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 14:10:29 crc kubenswrapper[4805]: I1203 14:10:29.288255 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 14:10:29 crc kubenswrapper[4805]: I1203 14:10:29.288266 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 14:10:29 crc kubenswrapper[4805]: I1203 14:10:29.288284 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 14:10:29 crc kubenswrapper[4805]: I1203 14:10:29.288296 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T14:10:29Z","lastTransitionTime":"2025-12-03T14:10:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 14:10:29 crc kubenswrapper[4805]: I1203 14:10:29.390434 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 14:10:29 crc kubenswrapper[4805]: I1203 14:10:29.390611 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 14:10:29 crc kubenswrapper[4805]: I1203 14:10:29.390631 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 14:10:29 crc kubenswrapper[4805]: I1203 14:10:29.390675 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 14:10:29 crc kubenswrapper[4805]: I1203 14:10:29.390699 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T14:10:29Z","lastTransitionTime":"2025-12-03T14:10:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 14:10:29 crc kubenswrapper[4805]: I1203 14:10:29.494111 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 14:10:29 crc kubenswrapper[4805]: I1203 14:10:29.494434 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 14:10:29 crc kubenswrapper[4805]: I1203 14:10:29.494676 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 14:10:29 crc kubenswrapper[4805]: I1203 14:10:29.494986 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 14:10:29 crc kubenswrapper[4805]: I1203 14:10:29.495219 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T14:10:29Z","lastTransitionTime":"2025-12-03T14:10:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 14:10:29 crc kubenswrapper[4805]: I1203 14:10:29.598312 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 14:10:29 crc kubenswrapper[4805]: I1203 14:10:29.598598 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 14:10:29 crc kubenswrapper[4805]: I1203 14:10:29.598680 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 14:10:29 crc kubenswrapper[4805]: I1203 14:10:29.598778 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 14:10:29 crc kubenswrapper[4805]: I1203 14:10:29.598901 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T14:10:29Z","lastTransitionTime":"2025-12-03T14:10:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 14:10:29 crc kubenswrapper[4805]: I1203 14:10:29.702620 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 14:10:29 crc kubenswrapper[4805]: I1203 14:10:29.702693 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 14:10:29 crc kubenswrapper[4805]: I1203 14:10:29.702711 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 14:10:29 crc kubenswrapper[4805]: I1203 14:10:29.702741 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 14:10:29 crc kubenswrapper[4805]: I1203 14:10:29.702763 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T14:10:29Z","lastTransitionTime":"2025-12-03T14:10:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 14:10:29 crc kubenswrapper[4805]: I1203 14:10:29.805743 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 14:10:29 crc kubenswrapper[4805]: I1203 14:10:29.805781 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 14:10:29 crc kubenswrapper[4805]: I1203 14:10:29.805790 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 14:10:29 crc kubenswrapper[4805]: I1203 14:10:29.805804 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 14:10:29 crc kubenswrapper[4805]: I1203 14:10:29.805814 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T14:10:29Z","lastTransitionTime":"2025-12-03T14:10:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 14:10:29 crc kubenswrapper[4805]: I1203 14:10:29.908474 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 14:10:29 crc kubenswrapper[4805]: I1203 14:10:29.908530 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 14:10:29 crc kubenswrapper[4805]: I1203 14:10:29.908542 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 14:10:29 crc kubenswrapper[4805]: I1203 14:10:29.908561 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 14:10:29 crc kubenswrapper[4805]: I1203 14:10:29.908574 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T14:10:29Z","lastTransitionTime":"2025-12-03T14:10:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 14:10:30 crc kubenswrapper[4805]: I1203 14:10:30.011493 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 14:10:30 crc kubenswrapper[4805]: I1203 14:10:30.011534 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 14:10:30 crc kubenswrapper[4805]: I1203 14:10:30.011544 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 14:10:30 crc kubenswrapper[4805]: I1203 14:10:30.011558 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 14:10:30 crc kubenswrapper[4805]: I1203 14:10:30.011568 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T14:10:30Z","lastTransitionTime":"2025-12-03T14:10:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 14:10:30 crc kubenswrapper[4805]: I1203 14:10:30.114300 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 14:10:30 crc kubenswrapper[4805]: I1203 14:10:30.114645 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 14:10:30 crc kubenswrapper[4805]: I1203 14:10:30.114889 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 14:10:30 crc kubenswrapper[4805]: I1203 14:10:30.115123 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 14:10:30 crc kubenswrapper[4805]: I1203 14:10:30.115322 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T14:10:30Z","lastTransitionTime":"2025-12-03T14:10:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 14:10:30 crc kubenswrapper[4805]: I1203 14:10:30.218088 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 14:10:30 crc kubenswrapper[4805]: I1203 14:10:30.218172 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 14:10:30 crc kubenswrapper[4805]: I1203 14:10:30.218195 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 14:10:30 crc kubenswrapper[4805]: I1203 14:10:30.218224 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 14:10:30 crc kubenswrapper[4805]: I1203 14:10:30.218244 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T14:10:30Z","lastTransitionTime":"2025-12-03T14:10:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 14:10:30 crc kubenswrapper[4805]: I1203 14:10:30.321122 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 14:10:30 crc kubenswrapper[4805]: I1203 14:10:30.321231 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 14:10:30 crc kubenswrapper[4805]: I1203 14:10:30.321255 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 14:10:30 crc kubenswrapper[4805]: I1203 14:10:30.321284 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 14:10:30 crc kubenswrapper[4805]: I1203 14:10:30.321306 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T14:10:30Z","lastTransitionTime":"2025-12-03T14:10:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 14:10:30 crc kubenswrapper[4805]: I1203 14:10:30.423901 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 14:10:30 crc kubenswrapper[4805]: I1203 14:10:30.423934 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 14:10:30 crc kubenswrapper[4805]: I1203 14:10:30.423943 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 14:10:30 crc kubenswrapper[4805]: I1203 14:10:30.423957 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 14:10:30 crc kubenswrapper[4805]: I1203 14:10:30.423966 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T14:10:30Z","lastTransitionTime":"2025-12-03T14:10:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 14:10:30 crc kubenswrapper[4805]: I1203 14:10:30.526244 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 14:10:30 crc kubenswrapper[4805]: I1203 14:10:30.526283 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 14:10:30 crc kubenswrapper[4805]: I1203 14:10:30.526293 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 14:10:30 crc kubenswrapper[4805]: I1203 14:10:30.526308 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 14:10:30 crc kubenswrapper[4805]: I1203 14:10:30.526318 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T14:10:30Z","lastTransitionTime":"2025-12-03T14:10:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 14:10:30 crc kubenswrapper[4805]: I1203 14:10:30.628360 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 14:10:30 crc kubenswrapper[4805]: I1203 14:10:30.628427 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 14:10:30 crc kubenswrapper[4805]: I1203 14:10:30.628439 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 14:10:30 crc kubenswrapper[4805]: I1203 14:10:30.628458 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 14:10:30 crc kubenswrapper[4805]: I1203 14:10:30.628470 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T14:10:30Z","lastTransitionTime":"2025-12-03T14:10:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 14:10:30 crc kubenswrapper[4805]: I1203 14:10:30.693407 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 14:10:30 crc kubenswrapper[4805]: I1203 14:10:30.693468 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-4gd9r" Dec 03 14:10:30 crc kubenswrapper[4805]: I1203 14:10:30.693475 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 14:10:30 crc kubenswrapper[4805]: I1203 14:10:30.693424 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 14:10:30 crc kubenswrapper[4805]: E1203 14:10:30.693545 4805 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 03 14:10:30 crc kubenswrapper[4805]: E1203 14:10:30.693609 4805 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 03 14:10:30 crc kubenswrapper[4805]: E1203 14:10:30.693703 4805 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-4gd9r" podUID="6b36a3db-1047-487a-9d6c-4f51ffdd8cc1" Dec 03 14:10:30 crc kubenswrapper[4805]: E1203 14:10:30.693789 4805 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 03 14:10:30 crc kubenswrapper[4805]: I1203 14:10:30.718925 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b6d36671-7173-4e9c-843f-005b4dcf576c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a837d7d491eaacbe4279b4014bb5fcbe9b514ba5cc82c708227bcd9b25a346e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6e5f04faf3b23197bb8ea11d3e574dc86f31e21d798734962001ec8c95de7c84\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2ff1a979387da41b00e95eb9cc8b2c9de966380a12fae687591df3c5b18a4497\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1e66e1cb18573a14e3c3ae16af4dbd9f8997a6ad38680c3ab4505646ebd48e7c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7d5b5c85e3de43e7e95111aa967248550607c26200c783251b68696f202f92a1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://474f72dac37a808fc4b9832306e76979b151d7221807ff3086013466ed1419c3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://474f72dac37a808fc4b9832306e76979b151d7221807ff3086013466ed1419c3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T14:09:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T14:09:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://74c8cea1a61c4775a0ad2f4ab4bb8a36437cf23d2ee953311d6ea967e297510a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://74c8cea1a61c4775a0ad2f4ab4bb8a36437cf23d2ee953311d6ea967e297510a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T14:09:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T14:09:31Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://a4f2f56a809adc01fba82eba0652d78f58770c132f77b1424f8b17068b1927e3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a4f2f56a809adc01fba82eba0652d78f58770c132f77b1424f8b17068b1927e3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T14:09:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T14:09:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T14:09:30Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T14:10:30Z is after 2025-08-24T17:21:41Z" Dec 03 14:10:30 crc kubenswrapper[4805]: I1203 14:10:30.731627 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 14:10:30 crc kubenswrapper[4805]: I1203 14:10:30.731931 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 14:10:30 crc kubenswrapper[4805]: I1203 14:10:30.732112 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 14:10:30 crc kubenswrapper[4805]: I1203 14:10:30.732269 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 14:10:30 crc kubenswrapper[4805]: I1203 14:10:30.732389 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T14:10:30Z","lastTransitionTime":"2025-12-03T14:10:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 14:10:30 crc kubenswrapper[4805]: I1203 14:10:30.736122 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ced8291d-0ce5-4e74-9eaf-2bf033774039\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:10:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:10:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3e5a34636f0a8c4d9179074596d28bc2eb7a504573398c596a637defd829907a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://29cd95be3f964415943c873e7914ffcea4abe50dfdf17f8a9cd03fc38df35981\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://951173b3d3d581275a6f6e43209b5e0611a701ca7a10f81504d723e24a2d21a3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e64730957bbbb2feb4d506e0ed9731a0594bf66fa1e8fcdb8ec944ded6308eed\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5da55f769e23ad52bac9c77da62c78e70fa34db8ed1a435e4eaf4dad3fd2b68a\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-03T14:09:48Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1203 14:09:42.992179 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1203 14:09:42.993233 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1830926909/tls.crt::/tmp/serving-cert-1830926909/tls.key\\\\\\\"\\\\nI1203 14:09:48.622403 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1203 14:09:48.626476 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1203 14:09:48.626527 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1203 14:09:48.626574 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1203 14:09:48.626582 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1203 14:09:48.633990 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1203 14:09:48.634024 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 14:09:48.634029 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 14:09:48.634034 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1203 14:09:48.634038 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1203 14:09:48.634042 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1203 14:09:48.634045 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1203 14:09:48.634344 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1203 14:09:48.644497 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T14:09:32Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bd92a160eb530c5687ebd7cd5b83b41820a2b808a0fd9514e5ace9d237e1bd26\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:32Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://af9c940135012c629447528c61743ed1e58708ab57ce76604c6da0479aa459cf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://af9c940135012c629447528c61743ed1e58708ab57ce76604c6da0479aa459cf\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T14:09:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T14:09:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T14:09:30Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T14:10:30Z is after 2025-08-24T17:21:41Z" Dec 03 14:10:30 crc kubenswrapper[4805]: I1203 14:10:30.748988 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6d48b4f1-5dfc-4f6b-a4e8-4e465c651c7b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:10:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:10:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5ea698a1e0de3b5201e2f9032d834d81d9aab9c2fb17aca582184c7090f1bd2f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://66aa257562f8e113fc1cd2936b8aa18e927f90018ad4c11a184c1c56444be648\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2fd6d50f839b603a90e5d00ab6e36e832eab2e299b7af1dbb84498c35bf1c08e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://77c26dae6ccd328a8a1ce4fe6ffe94c78061d8d3ac3727adb80a24c16f4ab533\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://77c26dae6ccd328a8a1ce4fe6ffe94c78061d8d3ac3727adb80a24c16f4ab533\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T14:09:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T14:09:31Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T14:09:30Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T14:10:30Z is after 2025-08-24T17:21:41Z" Dec 03 14:10:30 crc kubenswrapper[4805]: I1203 14:10:30.765771 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:49Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:49Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://38339331a3209d6787d1def81b72f4a8514bf91080a139cad6d8ac5cc6bef10e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T14:10:30Z is after 2025-08-24T17:21:41Z" Dec 03 14:10:30 crc kubenswrapper[4805]: I1203 14:10:30.782149 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:48Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:48Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T14:10:30Z is after 2025-08-24T17:21:41Z" Dec 03 14:10:30 crc kubenswrapper[4805]: I1203 14:10:30.800109 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:51Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:51Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b3123d3fa879097db88ed2f3cfb05631f440c24c5405503b2de3546f12e2e899\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T14:10:30Z is after 2025-08-24T17:21:41Z" Dec 03 14:10:30 crc kubenswrapper[4805]: I1203 14:10:30.811091 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-bdtdg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"53a0e9e3-b6d8-44b7-a32b-aa3450442865\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f557e17ad105d6972d3c1480656c6073f67e28c0f5773aa365549951f4d161c7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9ww54\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T14:09:51Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-bdtdg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T14:10:30Z is after 2025-08-24T17:21:41Z" Dec 03 14:10:30 crc kubenswrapper[4805]: I1203 14:10:30.825685 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"750d2b4b-738a-4992-8902-742528f6e875\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://abd2d0156113bb5182727a53aac594410ed051c19258beb8a07b808eff0b42a7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8673a29c0e16f4e707d79f0610722ccabb023ff2b4a18d7cf9d305183817d952\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://35766e7206b1be7a53d13d6890d759cf81dbd999643c12349268e4bf16fbbdb0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5b9844a11b2b903978d0d464b469e26c9211100da61d018dde9636dca95b684e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T14:09:30Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T14:10:30Z is after 2025-08-24T17:21:41Z" Dec 03 14:10:30 crc kubenswrapper[4805]: I1203 14:10:30.834987 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 14:10:30 crc kubenswrapper[4805]: I1203 14:10:30.835033 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 14:10:30 crc kubenswrapper[4805]: I1203 14:10:30.835042 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 14:10:30 crc kubenswrapper[4805]: I1203 14:10:30.835058 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 14:10:30 crc kubenswrapper[4805]: I1203 14:10:30.835069 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T14:10:30Z","lastTransitionTime":"2025-12-03T14:10:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 14:10:30 crc kubenswrapper[4805]: I1203 14:10:30.840825 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:48Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:48Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T14:10:30Z is after 2025-08-24T17:21:41Z" Dec 03 14:10:30 crc kubenswrapper[4805]: I1203 14:10:30.857389 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-k4mhs" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c09c1cdc-f461-45c9-8444-5b8764bccabc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3354c1a44f891182d04116fc76234a2b337d50567d032a3584f902a9516387cc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qdcd7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T14:09:50Z\\\"}}\" for pod \"openshift-multus\"/\"multus-k4mhs\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T14:10:30Z is after 2025-08-24T17:21:41Z" Dec 03 14:10:30 crc kubenswrapper[4805]: I1203 14:10:30.871875 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-bpk49" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"de00a81f-82e1-4223-9614-f98606b9a8ef\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://04f4ad27e02e2024ae9e7cd1b16116267e69cecda882dd723cdb274c41a0e7db\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-25d7t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7741e0e70416ebe094bfabc6009e98d7d4c0fbcba1fc82df13b9ba40774b456a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7741e0e70416ebe094bfabc6009e98d7d4c0fbcba1fc82df13b9ba40774b456a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T14:09:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T14:09:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-25d7t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ee3499c02eb52bbf08c447d2433d435a74dfb70670dc4cbaac44f5fbff9e8a29\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ee3499c02eb52bbf08c447d2433d435a74dfb70670dc4cbaac44f5fbff9e8a29\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T14:09:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T14:09:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-25d7t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ec8488b1d69799951be7ef40e5e028732265c4ddcdea2d453923e567864771f7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ec8488b1d69799951be7ef40e5e028732265c4ddcdea2d453923e567864771f7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T14:09:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T14:09:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-25d7t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://66707bb4e1fdacb265e89786df8a04605830f1c156394cacfcec9b54abe5d41a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://66707bb4e1fdacb265e89786df8a04605830f1c156394cacfcec9b54abe5d41a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T14:09:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T14:09:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-25d7t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6207ab6600970ff15c8713c547d4920b2c77eb9bea990bb5da497e24870dcc0f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6207ab6600970ff15c8713c547d4920b2c77eb9bea990bb5da497e24870dcc0f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T14:09:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T14:09:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-25d7t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8de65d4a205e2deec0b433e53c0e37feb94aa0ede6a4c58d7414b80829804505\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8de65d4a205e2deec0b433e53c0e37feb94aa0ede6a4c58d7414b80829804505\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T14:09:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T14:09:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-25d7t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T14:09:50Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-bpk49\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T14:10:30Z is after 2025-08-24T17:21:41Z" Dec 03 14:10:30 crc kubenswrapper[4805]: I1203 14:10:30.885989 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:48Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:48Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T14:10:30Z is after 2025-08-24T17:21:41Z" Dec 03 14:10:30 crc kubenswrapper[4805]: I1203 14:10:30.900348 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:49Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:49Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://541ace7ccbb66dbd191c3ca8c1e4479e3edbf80dfc67326bea8aaee81fdcfc3c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://821e3e9a902aacecf6c41e8b8d8aa1ac951ce2fbb6505b7148b16348a3754e55\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T14:10:30Z is after 2025-08-24T17:21:41Z" Dec 03 14:10:30 crc kubenswrapper[4805]: I1203 14:10:30.910207 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-gskh4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6098937f-e3f6-45e8-a647-4994a79cd711\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://28d5653f26ec61647808b3a733ff7eac70aa69ac184660f24147e8cfaf0eb802\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nwrw9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9ad6b68a42fd33ee167e67873c6eed3f11cdf6dc71d22e952c021436eb967dd9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nwrw9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T14:09:50Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-gskh4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T14:10:30Z is after 2025-08-24T17:21:41Z" Dec 03 14:10:30 crc kubenswrapper[4805]: I1203 14:10:30.927585 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-psjpm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"deeec30d-cad2-4dfd-adb6-64f6646581be\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:50Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:50Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4b8d00cdf0f37f2e5cd4e952cf85d3de8d72c43fdf786088fd9f26ed17b1e21f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wdlzx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bd67c6de01a8d196a1f51d4d90640c9e69de6f05d5e201b25e6eaa5c33440ea8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wdlzx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cb8bf05ef093eb0727a066d4bbbdf9af9422867cd4dadacafc6eeb016867a61e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wdlzx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0304467102ed053628c0d3e06548c74b6492336b6fd044246508d1e8bdb17bc3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wdlzx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6ce80ccafb3262a11fbe9923317a9d38b61e5dc5b3e7e7ae136f143d6b643fd7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wdlzx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://47d76962beb8c0f196d50af96fb3ec611097823ca80a1c8bd718f2d18840128c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wdlzx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a0272667e733f2c5eec0fce0eb9c31a16c048babdaed90635895d689ae9303f4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a0272667e733f2c5eec0fce0eb9c31a16c048babdaed90635895d689ae9303f4\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-03T14:10:16Z\\\",\\\"message\\\":\\\"o create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T14:10:16Z is after 2025-08-24T17:21:41Z]\\\\nI1203 14:10:16.717659 6478 services_controller.go:473] Services do not match for network=default, existing lbs: []services.LB{services.LB{Name:\\\\\\\"Service_openshift-network-console/networking-console-plugin_TCP_cluster\\\\\\\", UUID:\\\\\\\"ab0b1d51-5ec6-479b-8881-93dfa8d30337\\\\\\\", Protocol:\\\\\\\"tcp\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-network-console/networking-console-plugin\\\\\\\"}, Opts:services.LBOpts{Reject:false, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{}, Templates:services.TemplateMap{}, Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGr\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T14:10:15Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-psjpm_openshift-ovn-kubernetes(deeec30d-cad2-4dfd-adb6-64f6646581be)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wdlzx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ab98e365c1a4242a47b8cbe2dd85109d366d6002c4de1f18ddc950a46bea6c12\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wdlzx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2f875252fbe6a673a72fc9c56bb96899f62732305ad77b3108e3b244bfa91ecb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2f875252fbe6a673a72fc9c56bb96899f62732305ad77b3108e3b244bfa91ecb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T14:09:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T14:09:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wdlzx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T14:09:50Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-psjpm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T14:10:30Z is after 2025-08-24T17:21:41Z" Dec 03 14:10:30 crc kubenswrapper[4805]: I1203 14:10:30.937498 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 14:10:30 crc kubenswrapper[4805]: I1203 14:10:30.937571 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 14:10:30 crc kubenswrapper[4805]: I1203 14:10:30.937586 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 14:10:30 crc kubenswrapper[4805]: I1203 14:10:30.937603 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 14:10:30 crc kubenswrapper[4805]: I1203 14:10:30.937636 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T14:10:30Z","lastTransitionTime":"2025-12-03T14:10:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 14:10:30 crc kubenswrapper[4805]: I1203 14:10:30.940920 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-4gd9r" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6b36a3db-1047-487a-9d6c-4f51ffdd8cc1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:10:04Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:10:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:10:04Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:10:04Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4t6zl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4t6zl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T14:10:04Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-4gd9r\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T14:10:30Z is after 2025-08-24T17:21:41Z" Dec 03 14:10:30 crc kubenswrapper[4805]: I1203 14:10:30.950222 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-pdzgk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"860769de-6a5b-4cb9-a752-264e7e2e785f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5e2f0a0cc21505d609f81918365d5e4f5d21959ac12ab147d6a51601c9de4c14\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ggn4n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T14:09:50Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-pdzgk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T14:10:30Z is after 2025-08-24T17:21:41Z" Dec 03 14:10:30 crc kubenswrapper[4805]: I1203 14:10:30.961889 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-69h75" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5d1c694f-9847-4086-8b37-3573ddeb3ace\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:10:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:10:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:10:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:10:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://abc1d9b8582ca9fe4a51fb608fb39129a179ddcdfde0797a6369a55b2e0124cf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:10:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s578b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c7c4d434b035cac2a91a4f0846a2f4fb5bd68f881c28487860adb6db85e8ce75\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:10:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s578b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T14:10:03Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-69h75\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T14:10:30Z is after 2025-08-24T17:21:41Z" Dec 03 14:10:31 crc kubenswrapper[4805]: I1203 14:10:31.040279 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 14:10:31 crc kubenswrapper[4805]: I1203 14:10:31.040363 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 14:10:31 crc kubenswrapper[4805]: I1203 14:10:31.040378 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 14:10:31 crc kubenswrapper[4805]: I1203 14:10:31.040401 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 14:10:31 crc kubenswrapper[4805]: I1203 14:10:31.040419 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T14:10:31Z","lastTransitionTime":"2025-12-03T14:10:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 14:10:31 crc kubenswrapper[4805]: I1203 14:10:31.143888 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 14:10:31 crc kubenswrapper[4805]: I1203 14:10:31.143937 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 14:10:31 crc kubenswrapper[4805]: I1203 14:10:31.143948 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 14:10:31 crc kubenswrapper[4805]: I1203 14:10:31.143965 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 14:10:31 crc kubenswrapper[4805]: I1203 14:10:31.143978 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T14:10:31Z","lastTransitionTime":"2025-12-03T14:10:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 14:10:31 crc kubenswrapper[4805]: I1203 14:10:31.246130 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 14:10:31 crc kubenswrapper[4805]: I1203 14:10:31.246162 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 14:10:31 crc kubenswrapper[4805]: I1203 14:10:31.246171 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 14:10:31 crc kubenswrapper[4805]: I1203 14:10:31.246184 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 14:10:31 crc kubenswrapper[4805]: I1203 14:10:31.246194 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T14:10:31Z","lastTransitionTime":"2025-12-03T14:10:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 14:10:31 crc kubenswrapper[4805]: I1203 14:10:31.349519 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 14:10:31 crc kubenswrapper[4805]: I1203 14:10:31.349558 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 14:10:31 crc kubenswrapper[4805]: I1203 14:10:31.349569 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 14:10:31 crc kubenswrapper[4805]: I1203 14:10:31.349584 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 14:10:31 crc kubenswrapper[4805]: I1203 14:10:31.349594 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T14:10:31Z","lastTransitionTime":"2025-12-03T14:10:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 14:10:31 crc kubenswrapper[4805]: I1203 14:10:31.452616 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 14:10:31 crc kubenswrapper[4805]: I1203 14:10:31.452684 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 14:10:31 crc kubenswrapper[4805]: I1203 14:10:31.452710 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 14:10:31 crc kubenswrapper[4805]: I1203 14:10:31.452743 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 14:10:31 crc kubenswrapper[4805]: I1203 14:10:31.452766 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T14:10:31Z","lastTransitionTime":"2025-12-03T14:10:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 14:10:31 crc kubenswrapper[4805]: I1203 14:10:31.554785 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 14:10:31 crc kubenswrapper[4805]: I1203 14:10:31.554905 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 14:10:31 crc kubenswrapper[4805]: I1203 14:10:31.554932 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 14:10:31 crc kubenswrapper[4805]: I1203 14:10:31.554963 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 14:10:31 crc kubenswrapper[4805]: I1203 14:10:31.554984 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T14:10:31Z","lastTransitionTime":"2025-12-03T14:10:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 14:10:31 crc kubenswrapper[4805]: I1203 14:10:31.657197 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 14:10:31 crc kubenswrapper[4805]: I1203 14:10:31.657262 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 14:10:31 crc kubenswrapper[4805]: I1203 14:10:31.657277 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 14:10:31 crc kubenswrapper[4805]: I1203 14:10:31.657294 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 14:10:31 crc kubenswrapper[4805]: I1203 14:10:31.657306 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T14:10:31Z","lastTransitionTime":"2025-12-03T14:10:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 14:10:31 crc kubenswrapper[4805]: I1203 14:10:31.760510 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 14:10:31 crc kubenswrapper[4805]: I1203 14:10:31.760555 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 14:10:31 crc kubenswrapper[4805]: I1203 14:10:31.760566 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 14:10:31 crc kubenswrapper[4805]: I1203 14:10:31.760583 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 14:10:31 crc kubenswrapper[4805]: I1203 14:10:31.760598 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T14:10:31Z","lastTransitionTime":"2025-12-03T14:10:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 14:10:31 crc kubenswrapper[4805]: I1203 14:10:31.863895 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 14:10:31 crc kubenswrapper[4805]: I1203 14:10:31.863974 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 14:10:31 crc kubenswrapper[4805]: I1203 14:10:31.863995 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 14:10:31 crc kubenswrapper[4805]: I1203 14:10:31.864017 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 14:10:31 crc kubenswrapper[4805]: I1203 14:10:31.864036 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T14:10:31Z","lastTransitionTime":"2025-12-03T14:10:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 14:10:31 crc kubenswrapper[4805]: I1203 14:10:31.966786 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 14:10:31 crc kubenswrapper[4805]: I1203 14:10:31.966859 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 14:10:31 crc kubenswrapper[4805]: I1203 14:10:31.966874 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 14:10:31 crc kubenswrapper[4805]: I1203 14:10:31.966892 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 14:10:31 crc kubenswrapper[4805]: I1203 14:10:31.966907 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T14:10:31Z","lastTransitionTime":"2025-12-03T14:10:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 14:10:32 crc kubenswrapper[4805]: I1203 14:10:32.068433 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 14:10:32 crc kubenswrapper[4805]: I1203 14:10:32.068477 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 14:10:32 crc kubenswrapper[4805]: I1203 14:10:32.068488 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 14:10:32 crc kubenswrapper[4805]: I1203 14:10:32.068504 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 14:10:32 crc kubenswrapper[4805]: I1203 14:10:32.068516 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T14:10:32Z","lastTransitionTime":"2025-12-03T14:10:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 14:10:32 crc kubenswrapper[4805]: I1203 14:10:32.171178 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 14:10:32 crc kubenswrapper[4805]: I1203 14:10:32.171242 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 14:10:32 crc kubenswrapper[4805]: I1203 14:10:32.171258 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 14:10:32 crc kubenswrapper[4805]: I1203 14:10:32.171276 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 14:10:32 crc kubenswrapper[4805]: I1203 14:10:32.171287 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T14:10:32Z","lastTransitionTime":"2025-12-03T14:10:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 14:10:32 crc kubenswrapper[4805]: I1203 14:10:32.274673 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 14:10:32 crc kubenswrapper[4805]: I1203 14:10:32.274773 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 14:10:32 crc kubenswrapper[4805]: I1203 14:10:32.274785 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 14:10:32 crc kubenswrapper[4805]: I1203 14:10:32.274803 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 14:10:32 crc kubenswrapper[4805]: I1203 14:10:32.274812 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T14:10:32Z","lastTransitionTime":"2025-12-03T14:10:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 14:10:32 crc kubenswrapper[4805]: I1203 14:10:32.377421 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 14:10:32 crc kubenswrapper[4805]: I1203 14:10:32.377466 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 14:10:32 crc kubenswrapper[4805]: I1203 14:10:32.377476 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 14:10:32 crc kubenswrapper[4805]: I1203 14:10:32.377493 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 14:10:32 crc kubenswrapper[4805]: I1203 14:10:32.377504 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T14:10:32Z","lastTransitionTime":"2025-12-03T14:10:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 14:10:32 crc kubenswrapper[4805]: I1203 14:10:32.480477 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 14:10:32 crc kubenswrapper[4805]: I1203 14:10:32.480546 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 14:10:32 crc kubenswrapper[4805]: I1203 14:10:32.480581 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 14:10:32 crc kubenswrapper[4805]: I1203 14:10:32.480612 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 14:10:32 crc kubenswrapper[4805]: I1203 14:10:32.480635 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T14:10:32Z","lastTransitionTime":"2025-12-03T14:10:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 14:10:32 crc kubenswrapper[4805]: I1203 14:10:32.583327 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 14:10:32 crc kubenswrapper[4805]: I1203 14:10:32.583369 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 14:10:32 crc kubenswrapper[4805]: I1203 14:10:32.583381 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 14:10:32 crc kubenswrapper[4805]: I1203 14:10:32.583396 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 14:10:32 crc kubenswrapper[4805]: I1203 14:10:32.583406 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T14:10:32Z","lastTransitionTime":"2025-12-03T14:10:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 14:10:32 crc kubenswrapper[4805]: I1203 14:10:32.686028 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 14:10:32 crc kubenswrapper[4805]: I1203 14:10:32.686085 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 14:10:32 crc kubenswrapper[4805]: I1203 14:10:32.686104 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 14:10:32 crc kubenswrapper[4805]: I1203 14:10:32.686130 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 14:10:32 crc kubenswrapper[4805]: I1203 14:10:32.686151 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T14:10:32Z","lastTransitionTime":"2025-12-03T14:10:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 14:10:32 crc kubenswrapper[4805]: I1203 14:10:32.693562 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-4gd9r" Dec 03 14:10:32 crc kubenswrapper[4805]: I1203 14:10:32.693617 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 14:10:32 crc kubenswrapper[4805]: I1203 14:10:32.693563 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 14:10:32 crc kubenswrapper[4805]: I1203 14:10:32.693691 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 14:10:32 crc kubenswrapper[4805]: E1203 14:10:32.693691 4805 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-4gd9r" podUID="6b36a3db-1047-487a-9d6c-4f51ffdd8cc1" Dec 03 14:10:32 crc kubenswrapper[4805]: E1203 14:10:32.693788 4805 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 03 14:10:32 crc kubenswrapper[4805]: E1203 14:10:32.693954 4805 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 03 14:10:32 crc kubenswrapper[4805]: E1203 14:10:32.694035 4805 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 03 14:10:32 crc kubenswrapper[4805]: I1203 14:10:32.694560 4805 scope.go:117] "RemoveContainer" containerID="a0272667e733f2c5eec0fce0eb9c31a16c048babdaed90635895d689ae9303f4" Dec 03 14:10:32 crc kubenswrapper[4805]: E1203 14:10:32.694754 4805 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-psjpm_openshift-ovn-kubernetes(deeec30d-cad2-4dfd-adb6-64f6646581be)\"" pod="openshift-ovn-kubernetes/ovnkube-node-psjpm" podUID="deeec30d-cad2-4dfd-adb6-64f6646581be" Dec 03 14:10:32 crc kubenswrapper[4805]: I1203 14:10:32.788177 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 14:10:32 crc kubenswrapper[4805]: I1203 14:10:32.788659 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 14:10:32 crc kubenswrapper[4805]: I1203 14:10:32.788751 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 14:10:32 crc kubenswrapper[4805]: I1203 14:10:32.788769 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 14:10:32 crc kubenswrapper[4805]: I1203 14:10:32.788778 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T14:10:32Z","lastTransitionTime":"2025-12-03T14:10:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 14:10:32 crc kubenswrapper[4805]: I1203 14:10:32.891608 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 14:10:32 crc kubenswrapper[4805]: I1203 14:10:32.891657 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 14:10:32 crc kubenswrapper[4805]: I1203 14:10:32.891669 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 14:10:32 crc kubenswrapper[4805]: I1203 14:10:32.891686 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 14:10:32 crc kubenswrapper[4805]: I1203 14:10:32.891698 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T14:10:32Z","lastTransitionTime":"2025-12-03T14:10:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 14:10:32 crc kubenswrapper[4805]: I1203 14:10:32.963861 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 14:10:32 crc kubenswrapper[4805]: I1203 14:10:32.963939 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 14:10:32 crc kubenswrapper[4805]: I1203 14:10:32.963949 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 14:10:32 crc kubenswrapper[4805]: I1203 14:10:32.963963 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 14:10:32 crc kubenswrapper[4805]: I1203 14:10:32.963975 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T14:10:32Z","lastTransitionTime":"2025-12-03T14:10:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 14:10:32 crc kubenswrapper[4805]: E1203 14:10:32.980003 4805 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T14:10:32Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T14:10:32Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T14:10:32Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T14:10:32Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T14:10:32Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T14:10:32Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T14:10:32Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T14:10:32Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"e4edfbcd-1de1-486e-9c0b-65f9fbba52ad\\\",\\\"systemUUID\\\":\\\"7a729b24-19ef-490e-ab4c-d31c52ed5658\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T14:10:32Z is after 2025-08-24T17:21:41Z" Dec 03 14:10:32 crc kubenswrapper[4805]: I1203 14:10:32.984109 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 14:10:32 crc kubenswrapper[4805]: I1203 14:10:32.984134 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 14:10:32 crc kubenswrapper[4805]: I1203 14:10:32.984141 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 14:10:32 crc kubenswrapper[4805]: I1203 14:10:32.984153 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 14:10:32 crc kubenswrapper[4805]: I1203 14:10:32.984161 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T14:10:32Z","lastTransitionTime":"2025-12-03T14:10:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 14:10:32 crc kubenswrapper[4805]: E1203 14:10:32.998573 4805 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T14:10:32Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T14:10:32Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T14:10:32Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T14:10:32Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T14:10:32Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T14:10:32Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T14:10:32Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T14:10:32Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"e4edfbcd-1de1-486e-9c0b-65f9fbba52ad\\\",\\\"systemUUID\\\":\\\"7a729b24-19ef-490e-ab4c-d31c52ed5658\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T14:10:32Z is after 2025-08-24T17:21:41Z" Dec 03 14:10:33 crc kubenswrapper[4805]: I1203 14:10:33.002320 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 14:10:33 crc kubenswrapper[4805]: I1203 14:10:33.002343 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 14:10:33 crc kubenswrapper[4805]: I1203 14:10:33.002350 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 14:10:33 crc kubenswrapper[4805]: I1203 14:10:33.002362 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 14:10:33 crc kubenswrapper[4805]: I1203 14:10:33.002371 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T14:10:33Z","lastTransitionTime":"2025-12-03T14:10:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 14:10:33 crc kubenswrapper[4805]: E1203 14:10:33.015814 4805 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T14:10:33Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T14:10:33Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T14:10:33Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T14:10:33Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T14:10:33Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T14:10:33Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T14:10:33Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T14:10:33Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"e4edfbcd-1de1-486e-9c0b-65f9fbba52ad\\\",\\\"systemUUID\\\":\\\"7a729b24-19ef-490e-ab4c-d31c52ed5658\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T14:10:33Z is after 2025-08-24T17:21:41Z" Dec 03 14:10:33 crc kubenswrapper[4805]: I1203 14:10:33.018913 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 14:10:33 crc kubenswrapper[4805]: I1203 14:10:33.018967 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 14:10:33 crc kubenswrapper[4805]: I1203 14:10:33.018984 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 14:10:33 crc kubenswrapper[4805]: I1203 14:10:33.019001 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 14:10:33 crc kubenswrapper[4805]: I1203 14:10:33.019012 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T14:10:33Z","lastTransitionTime":"2025-12-03T14:10:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 14:10:33 crc kubenswrapper[4805]: E1203 14:10:33.035061 4805 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T14:10:33Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T14:10:33Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T14:10:33Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T14:10:33Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T14:10:33Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T14:10:33Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T14:10:33Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T14:10:33Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"e4edfbcd-1de1-486e-9c0b-65f9fbba52ad\\\",\\\"systemUUID\\\":\\\"7a729b24-19ef-490e-ab4c-d31c52ed5658\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T14:10:33Z is after 2025-08-24T17:21:41Z" Dec 03 14:10:33 crc kubenswrapper[4805]: I1203 14:10:33.038412 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 14:10:33 crc kubenswrapper[4805]: I1203 14:10:33.038445 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 14:10:33 crc kubenswrapper[4805]: I1203 14:10:33.038454 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 14:10:33 crc kubenswrapper[4805]: I1203 14:10:33.038468 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 14:10:33 crc kubenswrapper[4805]: I1203 14:10:33.038478 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T14:10:33Z","lastTransitionTime":"2025-12-03T14:10:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 14:10:33 crc kubenswrapper[4805]: E1203 14:10:33.056130 4805 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T14:10:33Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T14:10:33Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T14:10:33Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T14:10:33Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T14:10:33Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T14:10:33Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T14:10:33Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T14:10:33Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"e4edfbcd-1de1-486e-9c0b-65f9fbba52ad\\\",\\\"systemUUID\\\":\\\"7a729b24-19ef-490e-ab4c-d31c52ed5658\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T14:10:33Z is after 2025-08-24T17:21:41Z" Dec 03 14:10:33 crc kubenswrapper[4805]: E1203 14:10:33.056302 4805 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Dec 03 14:10:33 crc kubenswrapper[4805]: I1203 14:10:33.057988 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 14:10:33 crc kubenswrapper[4805]: I1203 14:10:33.058019 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 14:10:33 crc kubenswrapper[4805]: I1203 14:10:33.058029 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 14:10:33 crc kubenswrapper[4805]: I1203 14:10:33.058047 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 14:10:33 crc kubenswrapper[4805]: I1203 14:10:33.058059 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T14:10:33Z","lastTransitionTime":"2025-12-03T14:10:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 14:10:33 crc kubenswrapper[4805]: I1203 14:10:33.160283 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 14:10:33 crc kubenswrapper[4805]: I1203 14:10:33.160311 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 14:10:33 crc kubenswrapper[4805]: I1203 14:10:33.160319 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 14:10:33 crc kubenswrapper[4805]: I1203 14:10:33.160330 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 14:10:33 crc kubenswrapper[4805]: I1203 14:10:33.160338 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T14:10:33Z","lastTransitionTime":"2025-12-03T14:10:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 14:10:33 crc kubenswrapper[4805]: I1203 14:10:33.263145 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 14:10:33 crc kubenswrapper[4805]: I1203 14:10:33.263192 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 14:10:33 crc kubenswrapper[4805]: I1203 14:10:33.263210 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 14:10:33 crc kubenswrapper[4805]: I1203 14:10:33.263233 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 14:10:33 crc kubenswrapper[4805]: I1203 14:10:33.263249 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T14:10:33Z","lastTransitionTime":"2025-12-03T14:10:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 14:10:33 crc kubenswrapper[4805]: I1203 14:10:33.372886 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 14:10:33 crc kubenswrapper[4805]: I1203 14:10:33.372949 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 14:10:33 crc kubenswrapper[4805]: I1203 14:10:33.372973 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 14:10:33 crc kubenswrapper[4805]: I1203 14:10:33.373002 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 14:10:33 crc kubenswrapper[4805]: I1203 14:10:33.373026 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T14:10:33Z","lastTransitionTime":"2025-12-03T14:10:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 14:10:33 crc kubenswrapper[4805]: I1203 14:10:33.475997 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 14:10:33 crc kubenswrapper[4805]: I1203 14:10:33.476319 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 14:10:33 crc kubenswrapper[4805]: I1203 14:10:33.476598 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 14:10:33 crc kubenswrapper[4805]: I1203 14:10:33.476699 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 14:10:33 crc kubenswrapper[4805]: I1203 14:10:33.477039 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T14:10:33Z","lastTransitionTime":"2025-12-03T14:10:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 14:10:33 crc kubenswrapper[4805]: I1203 14:10:33.580260 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 14:10:33 crc kubenswrapper[4805]: I1203 14:10:33.580306 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 14:10:33 crc kubenswrapper[4805]: I1203 14:10:33.580317 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 14:10:33 crc kubenswrapper[4805]: I1203 14:10:33.580335 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 14:10:33 crc kubenswrapper[4805]: I1203 14:10:33.580345 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T14:10:33Z","lastTransitionTime":"2025-12-03T14:10:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 14:10:33 crc kubenswrapper[4805]: I1203 14:10:33.682869 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 14:10:33 crc kubenswrapper[4805]: I1203 14:10:33.682938 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 14:10:33 crc kubenswrapper[4805]: I1203 14:10:33.682961 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 14:10:33 crc kubenswrapper[4805]: I1203 14:10:33.682991 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 14:10:33 crc kubenswrapper[4805]: I1203 14:10:33.683013 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T14:10:33Z","lastTransitionTime":"2025-12-03T14:10:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 14:10:33 crc kubenswrapper[4805]: I1203 14:10:33.785124 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 14:10:33 crc kubenswrapper[4805]: I1203 14:10:33.785166 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 14:10:33 crc kubenswrapper[4805]: I1203 14:10:33.785201 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 14:10:33 crc kubenswrapper[4805]: I1203 14:10:33.785217 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 14:10:33 crc kubenswrapper[4805]: I1203 14:10:33.785227 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T14:10:33Z","lastTransitionTime":"2025-12-03T14:10:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 14:10:33 crc kubenswrapper[4805]: I1203 14:10:33.887685 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 14:10:33 crc kubenswrapper[4805]: I1203 14:10:33.887736 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 14:10:33 crc kubenswrapper[4805]: I1203 14:10:33.887746 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 14:10:33 crc kubenswrapper[4805]: I1203 14:10:33.887763 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 14:10:33 crc kubenswrapper[4805]: I1203 14:10:33.887773 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T14:10:33Z","lastTransitionTime":"2025-12-03T14:10:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 14:10:33 crc kubenswrapper[4805]: I1203 14:10:33.990604 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 14:10:33 crc kubenswrapper[4805]: I1203 14:10:33.991009 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 14:10:33 crc kubenswrapper[4805]: I1203 14:10:33.991192 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 14:10:33 crc kubenswrapper[4805]: I1203 14:10:33.991413 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 14:10:33 crc kubenswrapper[4805]: I1203 14:10:33.991611 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T14:10:33Z","lastTransitionTime":"2025-12-03T14:10:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 14:10:34 crc kubenswrapper[4805]: I1203 14:10:34.093879 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 14:10:34 crc kubenswrapper[4805]: I1203 14:10:34.093930 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 14:10:34 crc kubenswrapper[4805]: I1203 14:10:34.093960 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 14:10:34 crc kubenswrapper[4805]: I1203 14:10:34.093977 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 14:10:34 crc kubenswrapper[4805]: I1203 14:10:34.093988 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T14:10:34Z","lastTransitionTime":"2025-12-03T14:10:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 14:10:34 crc kubenswrapper[4805]: I1203 14:10:34.196908 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 14:10:34 crc kubenswrapper[4805]: I1203 14:10:34.196949 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 14:10:34 crc kubenswrapper[4805]: I1203 14:10:34.196960 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 14:10:34 crc kubenswrapper[4805]: I1203 14:10:34.196976 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 14:10:34 crc kubenswrapper[4805]: I1203 14:10:34.196990 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T14:10:34Z","lastTransitionTime":"2025-12-03T14:10:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 14:10:34 crc kubenswrapper[4805]: I1203 14:10:34.299546 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 14:10:34 crc kubenswrapper[4805]: I1203 14:10:34.299591 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 14:10:34 crc kubenswrapper[4805]: I1203 14:10:34.299602 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 14:10:34 crc kubenswrapper[4805]: I1203 14:10:34.299618 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 14:10:34 crc kubenswrapper[4805]: I1203 14:10:34.299628 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T14:10:34Z","lastTransitionTime":"2025-12-03T14:10:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 14:10:34 crc kubenswrapper[4805]: I1203 14:10:34.402672 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 14:10:34 crc kubenswrapper[4805]: I1203 14:10:34.402721 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 14:10:34 crc kubenswrapper[4805]: I1203 14:10:34.402732 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 14:10:34 crc kubenswrapper[4805]: I1203 14:10:34.402749 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 14:10:34 crc kubenswrapper[4805]: I1203 14:10:34.402762 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T14:10:34Z","lastTransitionTime":"2025-12-03T14:10:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 14:10:34 crc kubenswrapper[4805]: I1203 14:10:34.505708 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 14:10:34 crc kubenswrapper[4805]: I1203 14:10:34.505754 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 14:10:34 crc kubenswrapper[4805]: I1203 14:10:34.505766 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 14:10:34 crc kubenswrapper[4805]: I1203 14:10:34.505781 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 14:10:34 crc kubenswrapper[4805]: I1203 14:10:34.505793 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T14:10:34Z","lastTransitionTime":"2025-12-03T14:10:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 14:10:34 crc kubenswrapper[4805]: I1203 14:10:34.607601 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 14:10:34 crc kubenswrapper[4805]: I1203 14:10:34.607634 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 14:10:34 crc kubenswrapper[4805]: I1203 14:10:34.607644 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 14:10:34 crc kubenswrapper[4805]: I1203 14:10:34.607658 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 14:10:34 crc kubenswrapper[4805]: I1203 14:10:34.607667 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T14:10:34Z","lastTransitionTime":"2025-12-03T14:10:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 14:10:34 crc kubenswrapper[4805]: I1203 14:10:34.694177 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 14:10:34 crc kubenswrapper[4805]: E1203 14:10:34.694322 4805 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 03 14:10:34 crc kubenswrapper[4805]: I1203 14:10:34.694532 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 14:10:34 crc kubenswrapper[4805]: E1203 14:10:34.694594 4805 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 03 14:10:34 crc kubenswrapper[4805]: I1203 14:10:34.694775 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 14:10:34 crc kubenswrapper[4805]: I1203 14:10:34.694905 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-4gd9r" Dec 03 14:10:34 crc kubenswrapper[4805]: E1203 14:10:34.695050 4805 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 03 14:10:34 crc kubenswrapper[4805]: E1203 14:10:34.695129 4805 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-4gd9r" podUID="6b36a3db-1047-487a-9d6c-4f51ffdd8cc1" Dec 03 14:10:34 crc kubenswrapper[4805]: I1203 14:10:34.709632 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 14:10:34 crc kubenswrapper[4805]: I1203 14:10:34.709653 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 14:10:34 crc kubenswrapper[4805]: I1203 14:10:34.709661 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 14:10:34 crc kubenswrapper[4805]: I1203 14:10:34.709672 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 14:10:34 crc kubenswrapper[4805]: I1203 14:10:34.709680 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T14:10:34Z","lastTransitionTime":"2025-12-03T14:10:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 14:10:34 crc kubenswrapper[4805]: I1203 14:10:34.812506 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 14:10:34 crc kubenswrapper[4805]: I1203 14:10:34.812541 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 14:10:34 crc kubenswrapper[4805]: I1203 14:10:34.812556 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 14:10:34 crc kubenswrapper[4805]: I1203 14:10:34.812573 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 14:10:34 crc kubenswrapper[4805]: I1203 14:10:34.812584 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T14:10:34Z","lastTransitionTime":"2025-12-03T14:10:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 14:10:34 crc kubenswrapper[4805]: I1203 14:10:34.915019 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 14:10:34 crc kubenswrapper[4805]: I1203 14:10:34.915059 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 14:10:34 crc kubenswrapper[4805]: I1203 14:10:34.915070 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 14:10:34 crc kubenswrapper[4805]: I1203 14:10:34.915088 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 14:10:34 crc kubenswrapper[4805]: I1203 14:10:34.915100 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T14:10:34Z","lastTransitionTime":"2025-12-03T14:10:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 14:10:35 crc kubenswrapper[4805]: I1203 14:10:35.017134 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 14:10:35 crc kubenswrapper[4805]: I1203 14:10:35.017170 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 14:10:35 crc kubenswrapper[4805]: I1203 14:10:35.017181 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 14:10:35 crc kubenswrapper[4805]: I1203 14:10:35.017197 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 14:10:35 crc kubenswrapper[4805]: I1203 14:10:35.017208 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T14:10:35Z","lastTransitionTime":"2025-12-03T14:10:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 14:10:35 crc kubenswrapper[4805]: I1203 14:10:35.119764 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 14:10:35 crc kubenswrapper[4805]: I1203 14:10:35.119815 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 14:10:35 crc kubenswrapper[4805]: I1203 14:10:35.119826 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 14:10:35 crc kubenswrapper[4805]: I1203 14:10:35.119875 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 14:10:35 crc kubenswrapper[4805]: I1203 14:10:35.119891 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T14:10:35Z","lastTransitionTime":"2025-12-03T14:10:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 14:10:35 crc kubenswrapper[4805]: I1203 14:10:35.223070 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 14:10:35 crc kubenswrapper[4805]: I1203 14:10:35.223234 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 14:10:35 crc kubenswrapper[4805]: I1203 14:10:35.223251 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 14:10:35 crc kubenswrapper[4805]: I1203 14:10:35.223273 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 14:10:35 crc kubenswrapper[4805]: I1203 14:10:35.223288 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T14:10:35Z","lastTransitionTime":"2025-12-03T14:10:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 14:10:35 crc kubenswrapper[4805]: I1203 14:10:35.326035 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 14:10:35 crc kubenswrapper[4805]: I1203 14:10:35.326105 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 14:10:35 crc kubenswrapper[4805]: I1203 14:10:35.326128 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 14:10:35 crc kubenswrapper[4805]: I1203 14:10:35.326154 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 14:10:35 crc kubenswrapper[4805]: I1203 14:10:35.326172 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T14:10:35Z","lastTransitionTime":"2025-12-03T14:10:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 14:10:35 crc kubenswrapper[4805]: I1203 14:10:35.428779 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 14:10:35 crc kubenswrapper[4805]: I1203 14:10:35.428882 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 14:10:35 crc kubenswrapper[4805]: I1203 14:10:35.428909 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 14:10:35 crc kubenswrapper[4805]: I1203 14:10:35.428939 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 14:10:35 crc kubenswrapper[4805]: I1203 14:10:35.428960 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T14:10:35Z","lastTransitionTime":"2025-12-03T14:10:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 14:10:35 crc kubenswrapper[4805]: I1203 14:10:35.531065 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 14:10:35 crc kubenswrapper[4805]: I1203 14:10:35.531103 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 14:10:35 crc kubenswrapper[4805]: I1203 14:10:35.531117 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 14:10:35 crc kubenswrapper[4805]: I1203 14:10:35.531134 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 14:10:35 crc kubenswrapper[4805]: I1203 14:10:35.531145 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T14:10:35Z","lastTransitionTime":"2025-12-03T14:10:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 14:10:35 crc kubenswrapper[4805]: I1203 14:10:35.634666 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 14:10:35 crc kubenswrapper[4805]: I1203 14:10:35.634702 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 14:10:35 crc kubenswrapper[4805]: I1203 14:10:35.634713 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 14:10:35 crc kubenswrapper[4805]: I1203 14:10:35.634730 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 14:10:35 crc kubenswrapper[4805]: I1203 14:10:35.634739 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T14:10:35Z","lastTransitionTime":"2025-12-03T14:10:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 14:10:35 crc kubenswrapper[4805]: I1203 14:10:35.737048 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 14:10:35 crc kubenswrapper[4805]: I1203 14:10:35.737094 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 14:10:35 crc kubenswrapper[4805]: I1203 14:10:35.737111 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 14:10:35 crc kubenswrapper[4805]: I1203 14:10:35.737128 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 14:10:35 crc kubenswrapper[4805]: I1203 14:10:35.737140 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T14:10:35Z","lastTransitionTime":"2025-12-03T14:10:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 14:10:35 crc kubenswrapper[4805]: I1203 14:10:35.839506 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 14:10:35 crc kubenswrapper[4805]: I1203 14:10:35.839540 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 14:10:35 crc kubenswrapper[4805]: I1203 14:10:35.839548 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 14:10:35 crc kubenswrapper[4805]: I1203 14:10:35.839561 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 14:10:35 crc kubenswrapper[4805]: I1203 14:10:35.839570 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T14:10:35Z","lastTransitionTime":"2025-12-03T14:10:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 14:10:35 crc kubenswrapper[4805]: I1203 14:10:35.941397 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 14:10:35 crc kubenswrapper[4805]: I1203 14:10:35.941428 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 14:10:35 crc kubenswrapper[4805]: I1203 14:10:35.941456 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 14:10:35 crc kubenswrapper[4805]: I1203 14:10:35.941476 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 14:10:35 crc kubenswrapper[4805]: I1203 14:10:35.941486 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T14:10:35Z","lastTransitionTime":"2025-12-03T14:10:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 14:10:36 crc kubenswrapper[4805]: I1203 14:10:36.043827 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 14:10:36 crc kubenswrapper[4805]: I1203 14:10:36.043889 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 14:10:36 crc kubenswrapper[4805]: I1203 14:10:36.043899 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 14:10:36 crc kubenswrapper[4805]: I1203 14:10:36.043914 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 14:10:36 crc kubenswrapper[4805]: I1203 14:10:36.043924 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T14:10:36Z","lastTransitionTime":"2025-12-03T14:10:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 14:10:36 crc kubenswrapper[4805]: I1203 14:10:36.146523 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 14:10:36 crc kubenswrapper[4805]: I1203 14:10:36.146571 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 14:10:36 crc kubenswrapper[4805]: I1203 14:10:36.146581 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 14:10:36 crc kubenswrapper[4805]: I1203 14:10:36.146599 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 14:10:36 crc kubenswrapper[4805]: I1203 14:10:36.146610 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T14:10:36Z","lastTransitionTime":"2025-12-03T14:10:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 14:10:36 crc kubenswrapper[4805]: I1203 14:10:36.249014 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 14:10:36 crc kubenswrapper[4805]: I1203 14:10:36.249063 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 14:10:36 crc kubenswrapper[4805]: I1203 14:10:36.249075 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 14:10:36 crc kubenswrapper[4805]: I1203 14:10:36.249092 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 14:10:36 crc kubenswrapper[4805]: I1203 14:10:36.249103 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T14:10:36Z","lastTransitionTime":"2025-12-03T14:10:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 14:10:36 crc kubenswrapper[4805]: I1203 14:10:36.350927 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 14:10:36 crc kubenswrapper[4805]: I1203 14:10:36.350961 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 14:10:36 crc kubenswrapper[4805]: I1203 14:10:36.350970 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 14:10:36 crc kubenswrapper[4805]: I1203 14:10:36.350982 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 14:10:36 crc kubenswrapper[4805]: I1203 14:10:36.350991 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T14:10:36Z","lastTransitionTime":"2025-12-03T14:10:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 14:10:36 crc kubenswrapper[4805]: I1203 14:10:36.453611 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 14:10:36 crc kubenswrapper[4805]: I1203 14:10:36.453668 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 14:10:36 crc kubenswrapper[4805]: I1203 14:10:36.453680 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 14:10:36 crc kubenswrapper[4805]: I1203 14:10:36.453699 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 14:10:36 crc kubenswrapper[4805]: I1203 14:10:36.453711 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T14:10:36Z","lastTransitionTime":"2025-12-03T14:10:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 14:10:36 crc kubenswrapper[4805]: I1203 14:10:36.549889 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/6b36a3db-1047-487a-9d6c-4f51ffdd8cc1-metrics-certs\") pod \"network-metrics-daemon-4gd9r\" (UID: \"6b36a3db-1047-487a-9d6c-4f51ffdd8cc1\") " pod="openshift-multus/network-metrics-daemon-4gd9r" Dec 03 14:10:36 crc kubenswrapper[4805]: E1203 14:10:36.550063 4805 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Dec 03 14:10:36 crc kubenswrapper[4805]: E1203 14:10:36.550151 4805 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/6b36a3db-1047-487a-9d6c-4f51ffdd8cc1-metrics-certs podName:6b36a3db-1047-487a-9d6c-4f51ffdd8cc1 nodeName:}" failed. No retries permitted until 2025-12-03 14:11:08.550128676 +0000 UTC m=+98.213045699 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/6b36a3db-1047-487a-9d6c-4f51ffdd8cc1-metrics-certs") pod "network-metrics-daemon-4gd9r" (UID: "6b36a3db-1047-487a-9d6c-4f51ffdd8cc1") : object "openshift-multus"/"metrics-daemon-secret" not registered Dec 03 14:10:36 crc kubenswrapper[4805]: I1203 14:10:36.556427 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 14:10:36 crc kubenswrapper[4805]: I1203 14:10:36.556475 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 14:10:36 crc kubenswrapper[4805]: I1203 14:10:36.556487 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 14:10:36 crc kubenswrapper[4805]: I1203 14:10:36.556507 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 14:10:36 crc kubenswrapper[4805]: I1203 14:10:36.556521 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T14:10:36Z","lastTransitionTime":"2025-12-03T14:10:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 14:10:36 crc kubenswrapper[4805]: I1203 14:10:36.659151 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 14:10:36 crc kubenswrapper[4805]: I1203 14:10:36.659205 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 14:10:36 crc kubenswrapper[4805]: I1203 14:10:36.659216 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 14:10:36 crc kubenswrapper[4805]: I1203 14:10:36.659235 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 14:10:36 crc kubenswrapper[4805]: I1203 14:10:36.659246 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T14:10:36Z","lastTransitionTime":"2025-12-03T14:10:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 14:10:36 crc kubenswrapper[4805]: I1203 14:10:36.694715 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 14:10:36 crc kubenswrapper[4805]: I1203 14:10:36.694727 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 14:10:36 crc kubenswrapper[4805]: I1203 14:10:36.694751 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 14:10:36 crc kubenswrapper[4805]: E1203 14:10:36.695171 4805 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 03 14:10:36 crc kubenswrapper[4805]: I1203 14:10:36.694786 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-4gd9r" Dec 03 14:10:36 crc kubenswrapper[4805]: E1203 14:10:36.695280 4805 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 03 14:10:36 crc kubenswrapper[4805]: E1203 14:10:36.695373 4805 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-4gd9r" podUID="6b36a3db-1047-487a-9d6c-4f51ffdd8cc1" Dec 03 14:10:36 crc kubenswrapper[4805]: E1203 14:10:36.695178 4805 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 03 14:10:36 crc kubenswrapper[4805]: I1203 14:10:36.761532 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 14:10:36 crc kubenswrapper[4805]: I1203 14:10:36.761575 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 14:10:36 crc kubenswrapper[4805]: I1203 14:10:36.761586 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 14:10:36 crc kubenswrapper[4805]: I1203 14:10:36.761601 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 14:10:36 crc kubenswrapper[4805]: I1203 14:10:36.761612 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T14:10:36Z","lastTransitionTime":"2025-12-03T14:10:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 14:10:36 crc kubenswrapper[4805]: I1203 14:10:36.863935 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 14:10:36 crc kubenswrapper[4805]: I1203 14:10:36.863977 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 14:10:36 crc kubenswrapper[4805]: I1203 14:10:36.863987 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 14:10:36 crc kubenswrapper[4805]: I1203 14:10:36.864005 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 14:10:36 crc kubenswrapper[4805]: I1203 14:10:36.864020 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T14:10:36Z","lastTransitionTime":"2025-12-03T14:10:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 14:10:36 crc kubenswrapper[4805]: I1203 14:10:36.966744 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 14:10:36 crc kubenswrapper[4805]: I1203 14:10:36.966855 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 14:10:36 crc kubenswrapper[4805]: I1203 14:10:36.966867 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 14:10:36 crc kubenswrapper[4805]: I1203 14:10:36.966884 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 14:10:36 crc kubenswrapper[4805]: I1203 14:10:36.966897 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T14:10:36Z","lastTransitionTime":"2025-12-03T14:10:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 14:10:37 crc kubenswrapper[4805]: I1203 14:10:37.069774 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 14:10:37 crc kubenswrapper[4805]: I1203 14:10:37.069826 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 14:10:37 crc kubenswrapper[4805]: I1203 14:10:37.069853 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 14:10:37 crc kubenswrapper[4805]: I1203 14:10:37.069871 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 14:10:37 crc kubenswrapper[4805]: I1203 14:10:37.069883 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T14:10:37Z","lastTransitionTime":"2025-12-03T14:10:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 14:10:37 crc kubenswrapper[4805]: I1203 14:10:37.082075 4805 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-k4mhs_c09c1cdc-f461-45c9-8444-5b8764bccabc/kube-multus/0.log" Dec 03 14:10:37 crc kubenswrapper[4805]: I1203 14:10:37.082142 4805 generic.go:334] "Generic (PLEG): container finished" podID="c09c1cdc-f461-45c9-8444-5b8764bccabc" containerID="3354c1a44f891182d04116fc76234a2b337d50567d032a3584f902a9516387cc" exitCode=1 Dec 03 14:10:37 crc kubenswrapper[4805]: I1203 14:10:37.082175 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-k4mhs" event={"ID":"c09c1cdc-f461-45c9-8444-5b8764bccabc","Type":"ContainerDied","Data":"3354c1a44f891182d04116fc76234a2b337d50567d032a3584f902a9516387cc"} Dec 03 14:10:37 crc kubenswrapper[4805]: I1203 14:10:37.082605 4805 scope.go:117] "RemoveContainer" containerID="3354c1a44f891182d04116fc76234a2b337d50567d032a3584f902a9516387cc" Dec 03 14:10:37 crc kubenswrapper[4805]: I1203 14:10:37.096486 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-bdtdg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"53a0e9e3-b6d8-44b7-a32b-aa3450442865\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f557e17ad105d6972d3c1480656c6073f67e28c0f5773aa365549951f4d161c7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9ww54\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T14:09:51Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-bdtdg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T14:10:37Z is after 2025-08-24T17:21:41Z" Dec 03 14:10:37 crc kubenswrapper[4805]: I1203 14:10:37.118801 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b6d36671-7173-4e9c-843f-005b4dcf576c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a837d7d491eaacbe4279b4014bb5fcbe9b514ba5cc82c708227bcd9b25a346e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6e5f04faf3b23197bb8ea11d3e574dc86f31e21d798734962001ec8c95de7c84\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2ff1a979387da41b00e95eb9cc8b2c9de966380a12fae687591df3c5b18a4497\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1e66e1cb18573a14e3c3ae16af4dbd9f8997a6ad38680c3ab4505646ebd48e7c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7d5b5c85e3de43e7e95111aa967248550607c26200c783251b68696f202f92a1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://474f72dac37a808fc4b9832306e76979b151d7221807ff3086013466ed1419c3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://474f72dac37a808fc4b9832306e76979b151d7221807ff3086013466ed1419c3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T14:09:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T14:09:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://74c8cea1a61c4775a0ad2f4ab4bb8a36437cf23d2ee953311d6ea967e297510a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://74c8cea1a61c4775a0ad2f4ab4bb8a36437cf23d2ee953311d6ea967e297510a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T14:09:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T14:09:31Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://a4f2f56a809adc01fba82eba0652d78f58770c132f77b1424f8b17068b1927e3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a4f2f56a809adc01fba82eba0652d78f58770c132f77b1424f8b17068b1927e3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T14:09:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T14:09:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T14:09:30Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T14:10:37Z is after 2025-08-24T17:21:41Z" Dec 03 14:10:37 crc kubenswrapper[4805]: I1203 14:10:37.136702 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ced8291d-0ce5-4e74-9eaf-2bf033774039\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:10:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:10:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3e5a34636f0a8c4d9179074596d28bc2eb7a504573398c596a637defd829907a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://29cd95be3f964415943c873e7914ffcea4abe50dfdf17f8a9cd03fc38df35981\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://951173b3d3d581275a6f6e43209b5e0611a701ca7a10f81504d723e24a2d21a3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e64730957bbbb2feb4d506e0ed9731a0594bf66fa1e8fcdb8ec944ded6308eed\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5da55f769e23ad52bac9c77da62c78e70fa34db8ed1a435e4eaf4dad3fd2b68a\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-03T14:09:48Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1203 14:09:42.992179 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1203 14:09:42.993233 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1830926909/tls.crt::/tmp/serving-cert-1830926909/tls.key\\\\\\\"\\\\nI1203 14:09:48.622403 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1203 14:09:48.626476 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1203 14:09:48.626527 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1203 14:09:48.626574 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1203 14:09:48.626582 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1203 14:09:48.633990 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1203 14:09:48.634024 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 14:09:48.634029 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 14:09:48.634034 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1203 14:09:48.634038 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1203 14:09:48.634042 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1203 14:09:48.634045 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1203 14:09:48.634344 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1203 14:09:48.644497 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T14:09:32Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bd92a160eb530c5687ebd7cd5b83b41820a2b808a0fd9514e5ace9d237e1bd26\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:32Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://af9c940135012c629447528c61743ed1e58708ab57ce76604c6da0479aa459cf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://af9c940135012c629447528c61743ed1e58708ab57ce76604c6da0479aa459cf\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T14:09:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T14:09:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T14:09:30Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T14:10:37Z is after 2025-08-24T17:21:41Z" Dec 03 14:10:37 crc kubenswrapper[4805]: I1203 14:10:37.149858 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6d48b4f1-5dfc-4f6b-a4e8-4e465c651c7b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:10:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:10:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5ea698a1e0de3b5201e2f9032d834d81d9aab9c2fb17aca582184c7090f1bd2f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://66aa257562f8e113fc1cd2936b8aa18e927f90018ad4c11a184c1c56444be648\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2fd6d50f839b603a90e5d00ab6e36e832eab2e299b7af1dbb84498c35bf1c08e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://77c26dae6ccd328a8a1ce4fe6ffe94c78061d8d3ac3727adb80a24c16f4ab533\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://77c26dae6ccd328a8a1ce4fe6ffe94c78061d8d3ac3727adb80a24c16f4ab533\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T14:09:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T14:09:31Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T14:09:30Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T14:10:37Z is after 2025-08-24T17:21:41Z" Dec 03 14:10:37 crc kubenswrapper[4805]: I1203 14:10:37.163577 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:49Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:49Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://38339331a3209d6787d1def81b72f4a8514bf91080a139cad6d8ac5cc6bef10e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T14:10:37Z is after 2025-08-24T17:21:41Z" Dec 03 14:10:37 crc kubenswrapper[4805]: I1203 14:10:37.172762 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 14:10:37 crc kubenswrapper[4805]: I1203 14:10:37.172823 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 14:10:37 crc kubenswrapper[4805]: I1203 14:10:37.172854 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 14:10:37 crc kubenswrapper[4805]: I1203 14:10:37.172879 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 14:10:37 crc kubenswrapper[4805]: I1203 14:10:37.172890 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T14:10:37Z","lastTransitionTime":"2025-12-03T14:10:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 14:10:37 crc kubenswrapper[4805]: I1203 14:10:37.174996 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:48Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:48Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T14:10:37Z is after 2025-08-24T17:21:41Z" Dec 03 14:10:37 crc kubenswrapper[4805]: I1203 14:10:37.186750 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:51Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:51Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b3123d3fa879097db88ed2f3cfb05631f440c24c5405503b2de3546f12e2e899\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T14:10:37Z is after 2025-08-24T17:21:41Z" Dec 03 14:10:37 crc kubenswrapper[4805]: I1203 14:10:37.201009 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"750d2b4b-738a-4992-8902-742528f6e875\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://abd2d0156113bb5182727a53aac594410ed051c19258beb8a07b808eff0b42a7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8673a29c0e16f4e707d79f0610722ccabb023ff2b4a18d7cf9d305183817d952\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://35766e7206b1be7a53d13d6890d759cf81dbd999643c12349268e4bf16fbbdb0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5b9844a11b2b903978d0d464b469e26c9211100da61d018dde9636dca95b684e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T14:09:30Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T14:10:37Z is after 2025-08-24T17:21:41Z" Dec 03 14:10:37 crc kubenswrapper[4805]: I1203 14:10:37.212829 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:48Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:48Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T14:10:37Z is after 2025-08-24T17:21:41Z" Dec 03 14:10:37 crc kubenswrapper[4805]: I1203 14:10:37.224660 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-k4mhs" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c09c1cdc-f461-45c9-8444-5b8764bccabc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:10:37Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:10:37Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3354c1a44f891182d04116fc76234a2b337d50567d032a3584f902a9516387cc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3354c1a44f891182d04116fc76234a2b337d50567d032a3584f902a9516387cc\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-03T14:10:36Z\\\",\\\"message\\\":\\\"2025-12-03T14:09:51+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_3e51ab9a-fbbe-4a78-9308-8a81e203fcad\\\\n2025-12-03T14:09:51+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_3e51ab9a-fbbe-4a78-9308-8a81e203fcad to /host/opt/cni/bin/\\\\n2025-12-03T14:09:51Z [verbose] multus-daemon started\\\\n2025-12-03T14:09:51Z [verbose] Readiness Indicator file check\\\\n2025-12-03T14:10:36Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T14:09:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qdcd7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T14:09:50Z\\\"}}\" for pod \"openshift-multus\"/\"multus-k4mhs\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T14:10:37Z is after 2025-08-24T17:21:41Z" Dec 03 14:10:37 crc kubenswrapper[4805]: I1203 14:10:37.259979 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-bpk49" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"de00a81f-82e1-4223-9614-f98606b9a8ef\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://04f4ad27e02e2024ae9e7cd1b16116267e69cecda882dd723cdb274c41a0e7db\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-25d7t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7741e0e70416ebe094bfabc6009e98d7d4c0fbcba1fc82df13b9ba40774b456a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7741e0e70416ebe094bfabc6009e98d7d4c0fbcba1fc82df13b9ba40774b456a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T14:09:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T14:09:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-25d7t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ee3499c02eb52bbf08c447d2433d435a74dfb70670dc4cbaac44f5fbff9e8a29\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ee3499c02eb52bbf08c447d2433d435a74dfb70670dc4cbaac44f5fbff9e8a29\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T14:09:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T14:09:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-25d7t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ec8488b1d69799951be7ef40e5e028732265c4ddcdea2d453923e567864771f7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ec8488b1d69799951be7ef40e5e028732265c4ddcdea2d453923e567864771f7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T14:09:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T14:09:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-25d7t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://66707bb4e1fdacb265e89786df8a04605830f1c156394cacfcec9b54abe5d41a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://66707bb4e1fdacb265e89786df8a04605830f1c156394cacfcec9b54abe5d41a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T14:09:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T14:09:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-25d7t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6207ab6600970ff15c8713c547d4920b2c77eb9bea990bb5da497e24870dcc0f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6207ab6600970ff15c8713c547d4920b2c77eb9bea990bb5da497e24870dcc0f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T14:09:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T14:09:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-25d7t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8de65d4a205e2deec0b433e53c0e37feb94aa0ede6a4c58d7414b80829804505\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8de65d4a205e2deec0b433e53c0e37feb94aa0ede6a4c58d7414b80829804505\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T14:09:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T14:09:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-25d7t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T14:09:50Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-bpk49\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T14:10:37Z is after 2025-08-24T17:21:41Z" Dec 03 14:10:37 crc kubenswrapper[4805]: I1203 14:10:37.274818 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 14:10:37 crc kubenswrapper[4805]: I1203 14:10:37.275097 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 14:10:37 crc kubenswrapper[4805]: I1203 14:10:37.275207 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 14:10:37 crc kubenswrapper[4805]: I1203 14:10:37.275341 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 14:10:37 crc kubenswrapper[4805]: I1203 14:10:37.275432 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T14:10:37Z","lastTransitionTime":"2025-12-03T14:10:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 14:10:37 crc kubenswrapper[4805]: I1203 14:10:37.292449 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:48Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:48Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T14:10:37Z is after 2025-08-24T17:21:41Z" Dec 03 14:10:37 crc kubenswrapper[4805]: I1203 14:10:37.320805 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:49Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:49Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://541ace7ccbb66dbd191c3ca8c1e4479e3edbf80dfc67326bea8aaee81fdcfc3c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://821e3e9a902aacecf6c41e8b8d8aa1ac951ce2fbb6505b7148b16348a3754e55\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T14:10:37Z is after 2025-08-24T17:21:41Z" Dec 03 14:10:37 crc kubenswrapper[4805]: I1203 14:10:37.332721 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-gskh4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6098937f-e3f6-45e8-a647-4994a79cd711\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://28d5653f26ec61647808b3a733ff7eac70aa69ac184660f24147e8cfaf0eb802\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nwrw9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9ad6b68a42fd33ee167e67873c6eed3f11cdf6dc71d22e952c021436eb967dd9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nwrw9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T14:09:50Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-gskh4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T14:10:37Z is after 2025-08-24T17:21:41Z" Dec 03 14:10:37 crc kubenswrapper[4805]: I1203 14:10:37.351493 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-psjpm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"deeec30d-cad2-4dfd-adb6-64f6646581be\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:50Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:50Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4b8d00cdf0f37f2e5cd4e952cf85d3de8d72c43fdf786088fd9f26ed17b1e21f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wdlzx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bd67c6de01a8d196a1f51d4d90640c9e69de6f05d5e201b25e6eaa5c33440ea8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wdlzx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cb8bf05ef093eb0727a066d4bbbdf9af9422867cd4dadacafc6eeb016867a61e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wdlzx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0304467102ed053628c0d3e06548c74b6492336b6fd044246508d1e8bdb17bc3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wdlzx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6ce80ccafb3262a11fbe9923317a9d38b61e5dc5b3e7e7ae136f143d6b643fd7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wdlzx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://47d76962beb8c0f196d50af96fb3ec611097823ca80a1c8bd718f2d18840128c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wdlzx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a0272667e733f2c5eec0fce0eb9c31a16c048babdaed90635895d689ae9303f4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a0272667e733f2c5eec0fce0eb9c31a16c048babdaed90635895d689ae9303f4\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-03T14:10:16Z\\\",\\\"message\\\":\\\"o create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T14:10:16Z is after 2025-08-24T17:21:41Z]\\\\nI1203 14:10:16.717659 6478 services_controller.go:473] Services do not match for network=default, existing lbs: []services.LB{services.LB{Name:\\\\\\\"Service_openshift-network-console/networking-console-plugin_TCP_cluster\\\\\\\", UUID:\\\\\\\"ab0b1d51-5ec6-479b-8881-93dfa8d30337\\\\\\\", Protocol:\\\\\\\"tcp\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-network-console/networking-console-plugin\\\\\\\"}, Opts:services.LBOpts{Reject:false, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{}, Templates:services.TemplateMap{}, Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGr\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T14:10:15Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-psjpm_openshift-ovn-kubernetes(deeec30d-cad2-4dfd-adb6-64f6646581be)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wdlzx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ab98e365c1a4242a47b8cbe2dd85109d366d6002c4de1f18ddc950a46bea6c12\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wdlzx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2f875252fbe6a673a72fc9c56bb96899f62732305ad77b3108e3b244bfa91ecb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2f875252fbe6a673a72fc9c56bb96899f62732305ad77b3108e3b244bfa91ecb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T14:09:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T14:09:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wdlzx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T14:09:50Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-psjpm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T14:10:37Z is after 2025-08-24T17:21:41Z" Dec 03 14:10:37 crc kubenswrapper[4805]: I1203 14:10:37.363232 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-4gd9r" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6b36a3db-1047-487a-9d6c-4f51ffdd8cc1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:10:04Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:10:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:10:04Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:10:04Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4t6zl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4t6zl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T14:10:04Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-4gd9r\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T14:10:37Z is after 2025-08-24T17:21:41Z" Dec 03 14:10:37 crc kubenswrapper[4805]: I1203 14:10:37.376869 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-pdzgk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"860769de-6a5b-4cb9-a752-264e7e2e785f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5e2f0a0cc21505d609f81918365d5e4f5d21959ac12ab147d6a51601c9de4c14\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ggn4n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T14:09:50Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-pdzgk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T14:10:37Z is after 2025-08-24T17:21:41Z" Dec 03 14:10:37 crc kubenswrapper[4805]: I1203 14:10:37.377986 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 14:10:37 crc kubenswrapper[4805]: I1203 14:10:37.378031 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 14:10:37 crc kubenswrapper[4805]: I1203 14:10:37.378057 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 14:10:37 crc kubenswrapper[4805]: I1203 14:10:37.378075 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 14:10:37 crc kubenswrapper[4805]: I1203 14:10:37.378086 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T14:10:37Z","lastTransitionTime":"2025-12-03T14:10:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 14:10:37 crc kubenswrapper[4805]: I1203 14:10:37.388969 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-69h75" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5d1c694f-9847-4086-8b37-3573ddeb3ace\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:10:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:10:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:10:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:10:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://abc1d9b8582ca9fe4a51fb608fb39129a179ddcdfde0797a6369a55b2e0124cf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:10:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s578b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c7c4d434b035cac2a91a4f0846a2f4fb5bd68f881c28487860adb6db85e8ce75\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:10:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s578b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T14:10:03Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-69h75\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T14:10:37Z is after 2025-08-24T17:21:41Z" Dec 03 14:10:37 crc kubenswrapper[4805]: I1203 14:10:37.480920 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 14:10:37 crc kubenswrapper[4805]: I1203 14:10:37.480955 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 14:10:37 crc kubenswrapper[4805]: I1203 14:10:37.480965 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 14:10:37 crc kubenswrapper[4805]: I1203 14:10:37.480978 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 14:10:37 crc kubenswrapper[4805]: I1203 14:10:37.480986 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T14:10:37Z","lastTransitionTime":"2025-12-03T14:10:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 14:10:37 crc kubenswrapper[4805]: I1203 14:10:37.583966 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 14:10:37 crc kubenswrapper[4805]: I1203 14:10:37.584227 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 14:10:37 crc kubenswrapper[4805]: I1203 14:10:37.584294 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 14:10:37 crc kubenswrapper[4805]: I1203 14:10:37.584357 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 14:10:37 crc kubenswrapper[4805]: I1203 14:10:37.584423 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T14:10:37Z","lastTransitionTime":"2025-12-03T14:10:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 14:10:37 crc kubenswrapper[4805]: I1203 14:10:37.686385 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 14:10:37 crc kubenswrapper[4805]: I1203 14:10:37.686415 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 14:10:37 crc kubenswrapper[4805]: I1203 14:10:37.686423 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 14:10:37 crc kubenswrapper[4805]: I1203 14:10:37.686435 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 14:10:37 crc kubenswrapper[4805]: I1203 14:10:37.686443 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T14:10:37Z","lastTransitionTime":"2025-12-03T14:10:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 14:10:37 crc kubenswrapper[4805]: I1203 14:10:37.790218 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 14:10:37 crc kubenswrapper[4805]: I1203 14:10:37.790268 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 14:10:37 crc kubenswrapper[4805]: I1203 14:10:37.790284 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 14:10:37 crc kubenswrapper[4805]: I1203 14:10:37.790307 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 14:10:37 crc kubenswrapper[4805]: I1203 14:10:37.790324 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T14:10:37Z","lastTransitionTime":"2025-12-03T14:10:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 14:10:37 crc kubenswrapper[4805]: I1203 14:10:37.892746 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 14:10:37 crc kubenswrapper[4805]: I1203 14:10:37.892781 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 14:10:37 crc kubenswrapper[4805]: I1203 14:10:37.892789 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 14:10:37 crc kubenswrapper[4805]: I1203 14:10:37.892804 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 14:10:37 crc kubenswrapper[4805]: I1203 14:10:37.892813 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T14:10:37Z","lastTransitionTime":"2025-12-03T14:10:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 14:10:37 crc kubenswrapper[4805]: I1203 14:10:37.995251 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 14:10:37 crc kubenswrapper[4805]: I1203 14:10:37.995313 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 14:10:37 crc kubenswrapper[4805]: I1203 14:10:37.995334 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 14:10:37 crc kubenswrapper[4805]: I1203 14:10:37.995362 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 14:10:37 crc kubenswrapper[4805]: I1203 14:10:37.995385 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T14:10:37Z","lastTransitionTime":"2025-12-03T14:10:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 14:10:38 crc kubenswrapper[4805]: I1203 14:10:38.087408 4805 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-k4mhs_c09c1cdc-f461-45c9-8444-5b8764bccabc/kube-multus/0.log" Dec 03 14:10:38 crc kubenswrapper[4805]: I1203 14:10:38.087697 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-k4mhs" event={"ID":"c09c1cdc-f461-45c9-8444-5b8764bccabc","Type":"ContainerStarted","Data":"7e077e19b6a5bf7d29b357982e12224d4b7058e0f9ac81b66b5a665c9bbf5014"} Dec 03 14:10:38 crc kubenswrapper[4805]: I1203 14:10:38.097670 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 14:10:38 crc kubenswrapper[4805]: I1203 14:10:38.097705 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 14:10:38 crc kubenswrapper[4805]: I1203 14:10:38.097716 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 14:10:38 crc kubenswrapper[4805]: I1203 14:10:38.097733 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 14:10:38 crc kubenswrapper[4805]: I1203 14:10:38.097744 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T14:10:38Z","lastTransitionTime":"2025-12-03T14:10:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 14:10:38 crc kubenswrapper[4805]: I1203 14:10:38.105294 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"750d2b4b-738a-4992-8902-742528f6e875\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://abd2d0156113bb5182727a53aac594410ed051c19258beb8a07b808eff0b42a7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8673a29c0e16f4e707d79f0610722ccabb023ff2b4a18d7cf9d305183817d952\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://35766e7206b1be7a53d13d6890d759cf81dbd999643c12349268e4bf16fbbdb0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5b9844a11b2b903978d0d464b469e26c9211100da61d018dde9636dca95b684e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T14:09:30Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T14:10:38Z is after 2025-08-24T17:21:41Z" Dec 03 14:10:38 crc kubenswrapper[4805]: I1203 14:10:38.118922 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:48Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:48Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T14:10:38Z is after 2025-08-24T17:21:41Z" Dec 03 14:10:38 crc kubenswrapper[4805]: I1203 14:10:38.134196 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-k4mhs" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c09c1cdc-f461-45c9-8444-5b8764bccabc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:10:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:10:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7e077e19b6a5bf7d29b357982e12224d4b7058e0f9ac81b66b5a665c9bbf5014\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3354c1a44f891182d04116fc76234a2b337d50567d032a3584f902a9516387cc\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-03T14:10:36Z\\\",\\\"message\\\":\\\"2025-12-03T14:09:51+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_3e51ab9a-fbbe-4a78-9308-8a81e203fcad\\\\n2025-12-03T14:09:51+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_3e51ab9a-fbbe-4a78-9308-8a81e203fcad to /host/opt/cni/bin/\\\\n2025-12-03T14:09:51Z [verbose] multus-daemon started\\\\n2025-12-03T14:09:51Z [verbose] Readiness Indicator file check\\\\n2025-12-03T14:10:36Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T14:09:50Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:10:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qdcd7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T14:09:50Z\\\"}}\" for pod \"openshift-multus\"/\"multus-k4mhs\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T14:10:38Z is after 2025-08-24T17:21:41Z" Dec 03 14:10:38 crc kubenswrapper[4805]: I1203 14:10:38.152212 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-bpk49" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"de00a81f-82e1-4223-9614-f98606b9a8ef\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://04f4ad27e02e2024ae9e7cd1b16116267e69cecda882dd723cdb274c41a0e7db\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-25d7t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7741e0e70416ebe094bfabc6009e98d7d4c0fbcba1fc82df13b9ba40774b456a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7741e0e70416ebe094bfabc6009e98d7d4c0fbcba1fc82df13b9ba40774b456a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T14:09:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T14:09:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-25d7t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ee3499c02eb52bbf08c447d2433d435a74dfb70670dc4cbaac44f5fbff9e8a29\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ee3499c02eb52bbf08c447d2433d435a74dfb70670dc4cbaac44f5fbff9e8a29\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T14:09:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T14:09:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-25d7t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ec8488b1d69799951be7ef40e5e028732265c4ddcdea2d453923e567864771f7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ec8488b1d69799951be7ef40e5e028732265c4ddcdea2d453923e567864771f7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T14:09:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T14:09:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-25d7t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://66707bb4e1fdacb265e89786df8a04605830f1c156394cacfcec9b54abe5d41a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://66707bb4e1fdacb265e89786df8a04605830f1c156394cacfcec9b54abe5d41a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T14:09:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T14:09:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-25d7t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6207ab6600970ff15c8713c547d4920b2c77eb9bea990bb5da497e24870dcc0f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6207ab6600970ff15c8713c547d4920b2c77eb9bea990bb5da497e24870dcc0f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T14:09:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T14:09:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-25d7t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8de65d4a205e2deec0b433e53c0e37feb94aa0ede6a4c58d7414b80829804505\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8de65d4a205e2deec0b433e53c0e37feb94aa0ede6a4c58d7414b80829804505\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T14:09:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T14:09:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-25d7t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T14:09:50Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-bpk49\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T14:10:38Z is after 2025-08-24T17:21:41Z" Dec 03 14:10:38 crc kubenswrapper[4805]: I1203 14:10:38.163600 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-4gd9r" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6b36a3db-1047-487a-9d6c-4f51ffdd8cc1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:10:04Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:10:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:10:04Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:10:04Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4t6zl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4t6zl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T14:10:04Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-4gd9r\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T14:10:38Z is after 2025-08-24T17:21:41Z" Dec 03 14:10:38 crc kubenswrapper[4805]: I1203 14:10:38.176542 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:48Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:48Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T14:10:38Z is after 2025-08-24T17:21:41Z" Dec 03 14:10:38 crc kubenswrapper[4805]: I1203 14:10:38.188524 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:49Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:49Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://541ace7ccbb66dbd191c3ca8c1e4479e3edbf80dfc67326bea8aaee81fdcfc3c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://821e3e9a902aacecf6c41e8b8d8aa1ac951ce2fbb6505b7148b16348a3754e55\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T14:10:38Z is after 2025-08-24T17:21:41Z" Dec 03 14:10:38 crc kubenswrapper[4805]: I1203 14:10:38.199783 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 14:10:38 crc kubenswrapper[4805]: I1203 14:10:38.199818 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 14:10:38 crc kubenswrapper[4805]: I1203 14:10:38.199827 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 14:10:38 crc kubenswrapper[4805]: I1203 14:10:38.199857 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 14:10:38 crc kubenswrapper[4805]: I1203 14:10:38.199867 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T14:10:38Z","lastTransitionTime":"2025-12-03T14:10:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 14:10:38 crc kubenswrapper[4805]: I1203 14:10:38.201159 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-gskh4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6098937f-e3f6-45e8-a647-4994a79cd711\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://28d5653f26ec61647808b3a733ff7eac70aa69ac184660f24147e8cfaf0eb802\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nwrw9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9ad6b68a42fd33ee167e67873c6eed3f11cdf6dc71d22e952c021436eb967dd9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nwrw9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T14:09:50Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-gskh4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T14:10:38Z is after 2025-08-24T17:21:41Z" Dec 03 14:10:38 crc kubenswrapper[4805]: I1203 14:10:38.219111 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-psjpm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"deeec30d-cad2-4dfd-adb6-64f6646581be\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:50Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:50Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4b8d00cdf0f37f2e5cd4e952cf85d3de8d72c43fdf786088fd9f26ed17b1e21f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wdlzx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bd67c6de01a8d196a1f51d4d90640c9e69de6f05d5e201b25e6eaa5c33440ea8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wdlzx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cb8bf05ef093eb0727a066d4bbbdf9af9422867cd4dadacafc6eeb016867a61e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wdlzx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0304467102ed053628c0d3e06548c74b6492336b6fd044246508d1e8bdb17bc3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wdlzx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6ce80ccafb3262a11fbe9923317a9d38b61e5dc5b3e7e7ae136f143d6b643fd7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wdlzx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://47d76962beb8c0f196d50af96fb3ec611097823ca80a1c8bd718f2d18840128c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wdlzx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a0272667e733f2c5eec0fce0eb9c31a16c048babdaed90635895d689ae9303f4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a0272667e733f2c5eec0fce0eb9c31a16c048babdaed90635895d689ae9303f4\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-03T14:10:16Z\\\",\\\"message\\\":\\\"o create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T14:10:16Z is after 2025-08-24T17:21:41Z]\\\\nI1203 14:10:16.717659 6478 services_controller.go:473] Services do not match for network=default, existing lbs: []services.LB{services.LB{Name:\\\\\\\"Service_openshift-network-console/networking-console-plugin_TCP_cluster\\\\\\\", UUID:\\\\\\\"ab0b1d51-5ec6-479b-8881-93dfa8d30337\\\\\\\", Protocol:\\\\\\\"tcp\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-network-console/networking-console-plugin\\\\\\\"}, Opts:services.LBOpts{Reject:false, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{}, Templates:services.TemplateMap{}, Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGr\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T14:10:15Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-psjpm_openshift-ovn-kubernetes(deeec30d-cad2-4dfd-adb6-64f6646581be)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wdlzx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ab98e365c1a4242a47b8cbe2dd85109d366d6002c4de1f18ddc950a46bea6c12\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wdlzx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2f875252fbe6a673a72fc9c56bb96899f62732305ad77b3108e3b244bfa91ecb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2f875252fbe6a673a72fc9c56bb96899f62732305ad77b3108e3b244bfa91ecb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T14:09:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T14:09:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wdlzx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T14:09:50Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-psjpm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T14:10:38Z is after 2025-08-24T17:21:41Z" Dec 03 14:10:38 crc kubenswrapper[4805]: I1203 14:10:38.229898 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-pdzgk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"860769de-6a5b-4cb9-a752-264e7e2e785f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5e2f0a0cc21505d609f81918365d5e4f5d21959ac12ab147d6a51601c9de4c14\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ggn4n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T14:09:50Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-pdzgk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T14:10:38Z is after 2025-08-24T17:21:41Z" Dec 03 14:10:38 crc kubenswrapper[4805]: I1203 14:10:38.242769 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-69h75" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5d1c694f-9847-4086-8b37-3573ddeb3ace\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:10:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:10:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:10:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:10:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://abc1d9b8582ca9fe4a51fb608fb39129a179ddcdfde0797a6369a55b2e0124cf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:10:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s578b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c7c4d434b035cac2a91a4f0846a2f4fb5bd68f881c28487860adb6db85e8ce75\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:10:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s578b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T14:10:03Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-69h75\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T14:10:38Z is after 2025-08-24T17:21:41Z" Dec 03 14:10:38 crc kubenswrapper[4805]: I1203 14:10:38.256308 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:48Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:48Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T14:10:38Z is after 2025-08-24T17:21:41Z" Dec 03 14:10:38 crc kubenswrapper[4805]: I1203 14:10:38.267387 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:51Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:51Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b3123d3fa879097db88ed2f3cfb05631f440c24c5405503b2de3546f12e2e899\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T14:10:38Z is after 2025-08-24T17:21:41Z" Dec 03 14:10:38 crc kubenswrapper[4805]: I1203 14:10:38.278257 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-bdtdg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"53a0e9e3-b6d8-44b7-a32b-aa3450442865\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f557e17ad105d6972d3c1480656c6073f67e28c0f5773aa365549951f4d161c7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9ww54\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T14:09:51Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-bdtdg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T14:10:38Z is after 2025-08-24T17:21:41Z" Dec 03 14:10:38 crc kubenswrapper[4805]: I1203 14:10:38.302260 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 14:10:38 crc kubenswrapper[4805]: I1203 14:10:38.302307 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 14:10:38 crc kubenswrapper[4805]: I1203 14:10:38.302323 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 14:10:38 crc kubenswrapper[4805]: I1203 14:10:38.302376 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 14:10:38 crc kubenswrapper[4805]: I1203 14:10:38.302391 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T14:10:38Z","lastTransitionTime":"2025-12-03T14:10:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 14:10:38 crc kubenswrapper[4805]: I1203 14:10:38.302901 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b6d36671-7173-4e9c-843f-005b4dcf576c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a837d7d491eaacbe4279b4014bb5fcbe9b514ba5cc82c708227bcd9b25a346e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6e5f04faf3b23197bb8ea11d3e574dc86f31e21d798734962001ec8c95de7c84\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2ff1a979387da41b00e95eb9cc8b2c9de966380a12fae687591df3c5b18a4497\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1e66e1cb18573a14e3c3ae16af4dbd9f8997a6ad38680c3ab4505646ebd48e7c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7d5b5c85e3de43e7e95111aa967248550607c26200c783251b68696f202f92a1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://474f72dac37a808fc4b9832306e76979b151d7221807ff3086013466ed1419c3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://474f72dac37a808fc4b9832306e76979b151d7221807ff3086013466ed1419c3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T14:09:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T14:09:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://74c8cea1a61c4775a0ad2f4ab4bb8a36437cf23d2ee953311d6ea967e297510a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://74c8cea1a61c4775a0ad2f4ab4bb8a36437cf23d2ee953311d6ea967e297510a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T14:09:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T14:09:31Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://a4f2f56a809adc01fba82eba0652d78f58770c132f77b1424f8b17068b1927e3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a4f2f56a809adc01fba82eba0652d78f58770c132f77b1424f8b17068b1927e3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T14:09:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T14:09:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T14:09:30Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T14:10:38Z is after 2025-08-24T17:21:41Z" Dec 03 14:10:38 crc kubenswrapper[4805]: I1203 14:10:38.321294 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ced8291d-0ce5-4e74-9eaf-2bf033774039\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:10:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:10:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3e5a34636f0a8c4d9179074596d28bc2eb7a504573398c596a637defd829907a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://29cd95be3f964415943c873e7914ffcea4abe50dfdf17f8a9cd03fc38df35981\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://951173b3d3d581275a6f6e43209b5e0611a701ca7a10f81504d723e24a2d21a3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e64730957bbbb2feb4d506e0ed9731a0594bf66fa1e8fcdb8ec944ded6308eed\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5da55f769e23ad52bac9c77da62c78e70fa34db8ed1a435e4eaf4dad3fd2b68a\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-03T14:09:48Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1203 14:09:42.992179 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1203 14:09:42.993233 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1830926909/tls.crt::/tmp/serving-cert-1830926909/tls.key\\\\\\\"\\\\nI1203 14:09:48.622403 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1203 14:09:48.626476 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1203 14:09:48.626527 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1203 14:09:48.626574 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1203 14:09:48.626582 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1203 14:09:48.633990 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1203 14:09:48.634024 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 14:09:48.634029 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 14:09:48.634034 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1203 14:09:48.634038 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1203 14:09:48.634042 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1203 14:09:48.634045 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1203 14:09:48.634344 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1203 14:09:48.644497 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T14:09:32Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bd92a160eb530c5687ebd7cd5b83b41820a2b808a0fd9514e5ace9d237e1bd26\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:32Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://af9c940135012c629447528c61743ed1e58708ab57ce76604c6da0479aa459cf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://af9c940135012c629447528c61743ed1e58708ab57ce76604c6da0479aa459cf\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T14:09:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T14:09:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T14:09:30Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T14:10:38Z is after 2025-08-24T17:21:41Z" Dec 03 14:10:38 crc kubenswrapper[4805]: I1203 14:10:38.332410 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6d48b4f1-5dfc-4f6b-a4e8-4e465c651c7b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:10:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:10:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5ea698a1e0de3b5201e2f9032d834d81d9aab9c2fb17aca582184c7090f1bd2f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://66aa257562f8e113fc1cd2936b8aa18e927f90018ad4c11a184c1c56444be648\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2fd6d50f839b603a90e5d00ab6e36e832eab2e299b7af1dbb84498c35bf1c08e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://77c26dae6ccd328a8a1ce4fe6ffe94c78061d8d3ac3727adb80a24c16f4ab533\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://77c26dae6ccd328a8a1ce4fe6ffe94c78061d8d3ac3727adb80a24c16f4ab533\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T14:09:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T14:09:31Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T14:09:30Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T14:10:38Z is after 2025-08-24T17:21:41Z" Dec 03 14:10:38 crc kubenswrapper[4805]: I1203 14:10:38.344721 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:49Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:49Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://38339331a3209d6787d1def81b72f4a8514bf91080a139cad6d8ac5cc6bef10e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T14:10:38Z is after 2025-08-24T17:21:41Z" Dec 03 14:10:38 crc kubenswrapper[4805]: I1203 14:10:38.404377 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 14:10:38 crc kubenswrapper[4805]: I1203 14:10:38.404433 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 14:10:38 crc kubenswrapper[4805]: I1203 14:10:38.404446 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 14:10:38 crc kubenswrapper[4805]: I1203 14:10:38.404462 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 14:10:38 crc kubenswrapper[4805]: I1203 14:10:38.404471 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T14:10:38Z","lastTransitionTime":"2025-12-03T14:10:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 14:10:38 crc kubenswrapper[4805]: I1203 14:10:38.506674 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 14:10:38 crc kubenswrapper[4805]: I1203 14:10:38.506733 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 14:10:38 crc kubenswrapper[4805]: I1203 14:10:38.506748 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 14:10:38 crc kubenswrapper[4805]: I1203 14:10:38.506768 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 14:10:38 crc kubenswrapper[4805]: I1203 14:10:38.506780 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T14:10:38Z","lastTransitionTime":"2025-12-03T14:10:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 14:10:38 crc kubenswrapper[4805]: I1203 14:10:38.608747 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 14:10:38 crc kubenswrapper[4805]: I1203 14:10:38.608808 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 14:10:38 crc kubenswrapper[4805]: I1203 14:10:38.608819 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 14:10:38 crc kubenswrapper[4805]: I1203 14:10:38.608873 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 14:10:38 crc kubenswrapper[4805]: I1203 14:10:38.608889 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T14:10:38Z","lastTransitionTime":"2025-12-03T14:10:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 14:10:38 crc kubenswrapper[4805]: I1203 14:10:38.693880 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 14:10:38 crc kubenswrapper[4805]: I1203 14:10:38.693921 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-4gd9r" Dec 03 14:10:38 crc kubenswrapper[4805]: I1203 14:10:38.693933 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 14:10:38 crc kubenswrapper[4805]: E1203 14:10:38.694068 4805 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 03 14:10:38 crc kubenswrapper[4805]: I1203 14:10:38.694100 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 14:10:38 crc kubenswrapper[4805]: E1203 14:10:38.694196 4805 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-4gd9r" podUID="6b36a3db-1047-487a-9d6c-4f51ffdd8cc1" Dec 03 14:10:38 crc kubenswrapper[4805]: E1203 14:10:38.694285 4805 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 03 14:10:38 crc kubenswrapper[4805]: E1203 14:10:38.694332 4805 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 03 14:10:38 crc kubenswrapper[4805]: I1203 14:10:38.711998 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 14:10:38 crc kubenswrapper[4805]: I1203 14:10:38.712083 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 14:10:38 crc kubenswrapper[4805]: I1203 14:10:38.712096 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 14:10:38 crc kubenswrapper[4805]: I1203 14:10:38.712114 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 14:10:38 crc kubenswrapper[4805]: I1203 14:10:38.712126 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T14:10:38Z","lastTransitionTime":"2025-12-03T14:10:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 14:10:38 crc kubenswrapper[4805]: I1203 14:10:38.815101 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 14:10:38 crc kubenswrapper[4805]: I1203 14:10:38.815159 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 14:10:38 crc kubenswrapper[4805]: I1203 14:10:38.815175 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 14:10:38 crc kubenswrapper[4805]: I1203 14:10:38.815193 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 14:10:38 crc kubenswrapper[4805]: I1203 14:10:38.815204 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T14:10:38Z","lastTransitionTime":"2025-12-03T14:10:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 14:10:38 crc kubenswrapper[4805]: I1203 14:10:38.917128 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 14:10:38 crc kubenswrapper[4805]: I1203 14:10:38.917169 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 14:10:38 crc kubenswrapper[4805]: I1203 14:10:38.917181 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 14:10:38 crc kubenswrapper[4805]: I1203 14:10:38.917196 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 14:10:38 crc kubenswrapper[4805]: I1203 14:10:38.917207 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T14:10:38Z","lastTransitionTime":"2025-12-03T14:10:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 14:10:39 crc kubenswrapper[4805]: I1203 14:10:39.020385 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 14:10:39 crc kubenswrapper[4805]: I1203 14:10:39.020424 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 14:10:39 crc kubenswrapper[4805]: I1203 14:10:39.020438 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 14:10:39 crc kubenswrapper[4805]: I1203 14:10:39.020459 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 14:10:39 crc kubenswrapper[4805]: I1203 14:10:39.020473 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T14:10:39Z","lastTransitionTime":"2025-12-03T14:10:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 14:10:39 crc kubenswrapper[4805]: I1203 14:10:39.122725 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 14:10:39 crc kubenswrapper[4805]: I1203 14:10:39.123064 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 14:10:39 crc kubenswrapper[4805]: I1203 14:10:39.123161 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 14:10:39 crc kubenswrapper[4805]: I1203 14:10:39.123264 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 14:10:39 crc kubenswrapper[4805]: I1203 14:10:39.123345 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T14:10:39Z","lastTransitionTime":"2025-12-03T14:10:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 14:10:39 crc kubenswrapper[4805]: I1203 14:10:39.225318 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 14:10:39 crc kubenswrapper[4805]: I1203 14:10:39.225350 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 14:10:39 crc kubenswrapper[4805]: I1203 14:10:39.225358 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 14:10:39 crc kubenswrapper[4805]: I1203 14:10:39.225370 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 14:10:39 crc kubenswrapper[4805]: I1203 14:10:39.225381 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T14:10:39Z","lastTransitionTime":"2025-12-03T14:10:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 14:10:39 crc kubenswrapper[4805]: I1203 14:10:39.330222 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 14:10:39 crc kubenswrapper[4805]: I1203 14:10:39.330308 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 14:10:39 crc kubenswrapper[4805]: I1203 14:10:39.330322 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 14:10:39 crc kubenswrapper[4805]: I1203 14:10:39.330339 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 14:10:39 crc kubenswrapper[4805]: I1203 14:10:39.330355 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T14:10:39Z","lastTransitionTime":"2025-12-03T14:10:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 14:10:39 crc kubenswrapper[4805]: I1203 14:10:39.433586 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 14:10:39 crc kubenswrapper[4805]: I1203 14:10:39.433638 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 14:10:39 crc kubenswrapper[4805]: I1203 14:10:39.433648 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 14:10:39 crc kubenswrapper[4805]: I1203 14:10:39.433664 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 14:10:39 crc kubenswrapper[4805]: I1203 14:10:39.433675 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T14:10:39Z","lastTransitionTime":"2025-12-03T14:10:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 14:10:39 crc kubenswrapper[4805]: I1203 14:10:39.536368 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 14:10:39 crc kubenswrapper[4805]: I1203 14:10:39.536407 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 14:10:39 crc kubenswrapper[4805]: I1203 14:10:39.536418 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 14:10:39 crc kubenswrapper[4805]: I1203 14:10:39.536432 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 14:10:39 crc kubenswrapper[4805]: I1203 14:10:39.536443 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T14:10:39Z","lastTransitionTime":"2025-12-03T14:10:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 14:10:39 crc kubenswrapper[4805]: I1203 14:10:39.638417 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 14:10:39 crc kubenswrapper[4805]: I1203 14:10:39.638801 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 14:10:39 crc kubenswrapper[4805]: I1203 14:10:39.638963 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 14:10:39 crc kubenswrapper[4805]: I1203 14:10:39.638983 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 14:10:39 crc kubenswrapper[4805]: I1203 14:10:39.638999 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T14:10:39Z","lastTransitionTime":"2025-12-03T14:10:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 14:10:39 crc kubenswrapper[4805]: I1203 14:10:39.740900 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 14:10:39 crc kubenswrapper[4805]: I1203 14:10:39.740940 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 14:10:39 crc kubenswrapper[4805]: I1203 14:10:39.740948 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 14:10:39 crc kubenswrapper[4805]: I1203 14:10:39.740963 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 14:10:39 crc kubenswrapper[4805]: I1203 14:10:39.740971 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T14:10:39Z","lastTransitionTime":"2025-12-03T14:10:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 14:10:39 crc kubenswrapper[4805]: I1203 14:10:39.843495 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 14:10:39 crc kubenswrapper[4805]: I1203 14:10:39.843539 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 14:10:39 crc kubenswrapper[4805]: I1203 14:10:39.843548 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 14:10:39 crc kubenswrapper[4805]: I1203 14:10:39.843563 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 14:10:39 crc kubenswrapper[4805]: I1203 14:10:39.843572 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T14:10:39Z","lastTransitionTime":"2025-12-03T14:10:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 14:10:39 crc kubenswrapper[4805]: I1203 14:10:39.946116 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 14:10:39 crc kubenswrapper[4805]: I1203 14:10:39.946154 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 14:10:39 crc kubenswrapper[4805]: I1203 14:10:39.946163 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 14:10:39 crc kubenswrapper[4805]: I1203 14:10:39.946176 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 14:10:39 crc kubenswrapper[4805]: I1203 14:10:39.946184 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T14:10:39Z","lastTransitionTime":"2025-12-03T14:10:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 14:10:40 crc kubenswrapper[4805]: I1203 14:10:40.048093 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 14:10:40 crc kubenswrapper[4805]: I1203 14:10:40.048132 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 14:10:40 crc kubenswrapper[4805]: I1203 14:10:40.048145 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 14:10:40 crc kubenswrapper[4805]: I1203 14:10:40.048159 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 14:10:40 crc kubenswrapper[4805]: I1203 14:10:40.048171 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T14:10:40Z","lastTransitionTime":"2025-12-03T14:10:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 14:10:40 crc kubenswrapper[4805]: I1203 14:10:40.150568 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 14:10:40 crc kubenswrapper[4805]: I1203 14:10:40.150603 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 14:10:40 crc kubenswrapper[4805]: I1203 14:10:40.150613 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 14:10:40 crc kubenswrapper[4805]: I1203 14:10:40.150626 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 14:10:40 crc kubenswrapper[4805]: I1203 14:10:40.150636 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T14:10:40Z","lastTransitionTime":"2025-12-03T14:10:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 14:10:40 crc kubenswrapper[4805]: I1203 14:10:40.252936 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 14:10:40 crc kubenswrapper[4805]: I1203 14:10:40.253175 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 14:10:40 crc kubenswrapper[4805]: I1203 14:10:40.253256 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 14:10:40 crc kubenswrapper[4805]: I1203 14:10:40.253372 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 14:10:40 crc kubenswrapper[4805]: I1203 14:10:40.253459 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T14:10:40Z","lastTransitionTime":"2025-12-03T14:10:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 14:10:40 crc kubenswrapper[4805]: I1203 14:10:40.355706 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 14:10:40 crc kubenswrapper[4805]: I1203 14:10:40.355781 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 14:10:40 crc kubenswrapper[4805]: I1203 14:10:40.355792 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 14:10:40 crc kubenswrapper[4805]: I1203 14:10:40.355808 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 14:10:40 crc kubenswrapper[4805]: I1203 14:10:40.355817 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T14:10:40Z","lastTransitionTime":"2025-12-03T14:10:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 14:10:40 crc kubenswrapper[4805]: I1203 14:10:40.457595 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 14:10:40 crc kubenswrapper[4805]: I1203 14:10:40.457663 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 14:10:40 crc kubenswrapper[4805]: I1203 14:10:40.457675 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 14:10:40 crc kubenswrapper[4805]: I1203 14:10:40.457692 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 14:10:40 crc kubenswrapper[4805]: I1203 14:10:40.457726 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T14:10:40Z","lastTransitionTime":"2025-12-03T14:10:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 14:10:40 crc kubenswrapper[4805]: I1203 14:10:40.559994 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 14:10:40 crc kubenswrapper[4805]: I1203 14:10:40.560045 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 14:10:40 crc kubenswrapper[4805]: I1203 14:10:40.560061 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 14:10:40 crc kubenswrapper[4805]: I1203 14:10:40.560086 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 14:10:40 crc kubenswrapper[4805]: I1203 14:10:40.560105 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T14:10:40Z","lastTransitionTime":"2025-12-03T14:10:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 14:10:40 crc kubenswrapper[4805]: I1203 14:10:40.662092 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 14:10:40 crc kubenswrapper[4805]: I1203 14:10:40.662158 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 14:10:40 crc kubenswrapper[4805]: I1203 14:10:40.662169 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 14:10:40 crc kubenswrapper[4805]: I1203 14:10:40.662184 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 14:10:40 crc kubenswrapper[4805]: I1203 14:10:40.662195 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T14:10:40Z","lastTransitionTime":"2025-12-03T14:10:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 14:10:40 crc kubenswrapper[4805]: I1203 14:10:40.694095 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 14:10:40 crc kubenswrapper[4805]: I1203 14:10:40.694139 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 14:10:40 crc kubenswrapper[4805]: I1203 14:10:40.694174 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 14:10:40 crc kubenswrapper[4805]: I1203 14:10:40.694189 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-4gd9r" Dec 03 14:10:40 crc kubenswrapper[4805]: E1203 14:10:40.694324 4805 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 03 14:10:40 crc kubenswrapper[4805]: E1203 14:10:40.694405 4805 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 03 14:10:40 crc kubenswrapper[4805]: E1203 14:10:40.694498 4805 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 03 14:10:40 crc kubenswrapper[4805]: E1203 14:10:40.694595 4805 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-4gd9r" podUID="6b36a3db-1047-487a-9d6c-4f51ffdd8cc1" Dec 03 14:10:40 crc kubenswrapper[4805]: I1203 14:10:40.705267 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6d48b4f1-5dfc-4f6b-a4e8-4e465c651c7b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:10:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:10:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5ea698a1e0de3b5201e2f9032d834d81d9aab9c2fb17aca582184c7090f1bd2f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://66aa257562f8e113fc1cd2936b8aa18e927f90018ad4c11a184c1c56444be648\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2fd6d50f839b603a90e5d00ab6e36e832eab2e299b7af1dbb84498c35bf1c08e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://77c26dae6ccd328a8a1ce4fe6ffe94c78061d8d3ac3727adb80a24c16f4ab533\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://77c26dae6ccd328a8a1ce4fe6ffe94c78061d8d3ac3727adb80a24c16f4ab533\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T14:09:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T14:09:31Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T14:09:30Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T14:10:40Z is after 2025-08-24T17:21:41Z" Dec 03 14:10:40 crc kubenswrapper[4805]: I1203 14:10:40.716592 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:49Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:49Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://38339331a3209d6787d1def81b72f4a8514bf91080a139cad6d8ac5cc6bef10e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T14:10:40Z is after 2025-08-24T17:21:41Z" Dec 03 14:10:40 crc kubenswrapper[4805]: I1203 14:10:40.727855 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:48Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:48Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T14:10:40Z is after 2025-08-24T17:21:41Z" Dec 03 14:10:40 crc kubenswrapper[4805]: I1203 14:10:40.739663 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:51Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:51Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b3123d3fa879097db88ed2f3cfb05631f440c24c5405503b2de3546f12e2e899\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T14:10:40Z is after 2025-08-24T17:21:41Z" Dec 03 14:10:40 crc kubenswrapper[4805]: I1203 14:10:40.750993 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-bdtdg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"53a0e9e3-b6d8-44b7-a32b-aa3450442865\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f557e17ad105d6972d3c1480656c6073f67e28c0f5773aa365549951f4d161c7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9ww54\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T14:09:51Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-bdtdg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T14:10:40Z is after 2025-08-24T17:21:41Z" Dec 03 14:10:40 crc kubenswrapper[4805]: I1203 14:10:40.764823 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 14:10:40 crc kubenswrapper[4805]: I1203 14:10:40.764884 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 14:10:40 crc kubenswrapper[4805]: I1203 14:10:40.764895 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 14:10:40 crc kubenswrapper[4805]: I1203 14:10:40.764910 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 14:10:40 crc kubenswrapper[4805]: I1203 14:10:40.764939 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T14:10:40Z","lastTransitionTime":"2025-12-03T14:10:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 14:10:40 crc kubenswrapper[4805]: I1203 14:10:40.772132 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b6d36671-7173-4e9c-843f-005b4dcf576c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a837d7d491eaacbe4279b4014bb5fcbe9b514ba5cc82c708227bcd9b25a346e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6e5f04faf3b23197bb8ea11d3e574dc86f31e21d798734962001ec8c95de7c84\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2ff1a979387da41b00e95eb9cc8b2c9de966380a12fae687591df3c5b18a4497\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1e66e1cb18573a14e3c3ae16af4dbd9f8997a6ad38680c3ab4505646ebd48e7c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7d5b5c85e3de43e7e95111aa967248550607c26200c783251b68696f202f92a1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://474f72dac37a808fc4b9832306e76979b151d7221807ff3086013466ed1419c3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://474f72dac37a808fc4b9832306e76979b151d7221807ff3086013466ed1419c3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T14:09:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T14:09:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://74c8cea1a61c4775a0ad2f4ab4bb8a36437cf23d2ee953311d6ea967e297510a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://74c8cea1a61c4775a0ad2f4ab4bb8a36437cf23d2ee953311d6ea967e297510a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T14:09:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T14:09:31Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://a4f2f56a809adc01fba82eba0652d78f58770c132f77b1424f8b17068b1927e3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a4f2f56a809adc01fba82eba0652d78f58770c132f77b1424f8b17068b1927e3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T14:09:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T14:09:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T14:09:30Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T14:10:40Z is after 2025-08-24T17:21:41Z" Dec 03 14:10:40 crc kubenswrapper[4805]: I1203 14:10:40.791687 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ced8291d-0ce5-4e74-9eaf-2bf033774039\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:10:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:10:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3e5a34636f0a8c4d9179074596d28bc2eb7a504573398c596a637defd829907a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://29cd95be3f964415943c873e7914ffcea4abe50dfdf17f8a9cd03fc38df35981\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://951173b3d3d581275a6f6e43209b5e0611a701ca7a10f81504d723e24a2d21a3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e64730957bbbb2feb4d506e0ed9731a0594bf66fa1e8fcdb8ec944ded6308eed\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5da55f769e23ad52bac9c77da62c78e70fa34db8ed1a435e4eaf4dad3fd2b68a\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-03T14:09:48Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1203 14:09:42.992179 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1203 14:09:42.993233 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1830926909/tls.crt::/tmp/serving-cert-1830926909/tls.key\\\\\\\"\\\\nI1203 14:09:48.622403 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1203 14:09:48.626476 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1203 14:09:48.626527 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1203 14:09:48.626574 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1203 14:09:48.626582 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1203 14:09:48.633990 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1203 14:09:48.634024 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 14:09:48.634029 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 14:09:48.634034 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1203 14:09:48.634038 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1203 14:09:48.634042 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1203 14:09:48.634045 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1203 14:09:48.634344 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1203 14:09:48.644497 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T14:09:32Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bd92a160eb530c5687ebd7cd5b83b41820a2b808a0fd9514e5ace9d237e1bd26\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:32Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://af9c940135012c629447528c61743ed1e58708ab57ce76604c6da0479aa459cf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://af9c940135012c629447528c61743ed1e58708ab57ce76604c6da0479aa459cf\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T14:09:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T14:09:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T14:09:30Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T14:10:40Z is after 2025-08-24T17:21:41Z" Dec 03 14:10:40 crc kubenswrapper[4805]: I1203 14:10:40.804893 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-k4mhs" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c09c1cdc-f461-45c9-8444-5b8764bccabc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:10:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:10:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7e077e19b6a5bf7d29b357982e12224d4b7058e0f9ac81b66b5a665c9bbf5014\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3354c1a44f891182d04116fc76234a2b337d50567d032a3584f902a9516387cc\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-03T14:10:36Z\\\",\\\"message\\\":\\\"2025-12-03T14:09:51+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_3e51ab9a-fbbe-4a78-9308-8a81e203fcad\\\\n2025-12-03T14:09:51+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_3e51ab9a-fbbe-4a78-9308-8a81e203fcad to /host/opt/cni/bin/\\\\n2025-12-03T14:09:51Z [verbose] multus-daemon started\\\\n2025-12-03T14:09:51Z [verbose] Readiness Indicator file check\\\\n2025-12-03T14:10:36Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T14:09:50Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:10:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qdcd7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T14:09:50Z\\\"}}\" for pod \"openshift-multus\"/\"multus-k4mhs\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T14:10:40Z is after 2025-08-24T17:21:41Z" Dec 03 14:10:40 crc kubenswrapper[4805]: I1203 14:10:40.818626 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-bpk49" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"de00a81f-82e1-4223-9614-f98606b9a8ef\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://04f4ad27e02e2024ae9e7cd1b16116267e69cecda882dd723cdb274c41a0e7db\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-25d7t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7741e0e70416ebe094bfabc6009e98d7d4c0fbcba1fc82df13b9ba40774b456a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7741e0e70416ebe094bfabc6009e98d7d4c0fbcba1fc82df13b9ba40774b456a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T14:09:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T14:09:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-25d7t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ee3499c02eb52bbf08c447d2433d435a74dfb70670dc4cbaac44f5fbff9e8a29\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ee3499c02eb52bbf08c447d2433d435a74dfb70670dc4cbaac44f5fbff9e8a29\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T14:09:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T14:09:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-25d7t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ec8488b1d69799951be7ef40e5e028732265c4ddcdea2d453923e567864771f7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ec8488b1d69799951be7ef40e5e028732265c4ddcdea2d453923e567864771f7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T14:09:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T14:09:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-25d7t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://66707bb4e1fdacb265e89786df8a04605830f1c156394cacfcec9b54abe5d41a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://66707bb4e1fdacb265e89786df8a04605830f1c156394cacfcec9b54abe5d41a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T14:09:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T14:09:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-25d7t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6207ab6600970ff15c8713c547d4920b2c77eb9bea990bb5da497e24870dcc0f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6207ab6600970ff15c8713c547d4920b2c77eb9bea990bb5da497e24870dcc0f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T14:09:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T14:09:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-25d7t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8de65d4a205e2deec0b433e53c0e37feb94aa0ede6a4c58d7414b80829804505\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8de65d4a205e2deec0b433e53c0e37feb94aa0ede6a4c58d7414b80829804505\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T14:09:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T14:09:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-25d7t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T14:09:50Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-bpk49\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T14:10:40Z is after 2025-08-24T17:21:41Z" Dec 03 14:10:40 crc kubenswrapper[4805]: I1203 14:10:40.832618 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"750d2b4b-738a-4992-8902-742528f6e875\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://abd2d0156113bb5182727a53aac594410ed051c19258beb8a07b808eff0b42a7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8673a29c0e16f4e707d79f0610722ccabb023ff2b4a18d7cf9d305183817d952\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://35766e7206b1be7a53d13d6890d759cf81dbd999643c12349268e4bf16fbbdb0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5b9844a11b2b903978d0d464b469e26c9211100da61d018dde9636dca95b684e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T14:09:30Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T14:10:40Z is after 2025-08-24T17:21:41Z" Dec 03 14:10:40 crc kubenswrapper[4805]: I1203 14:10:40.845461 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:48Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:48Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T14:10:40Z is after 2025-08-24T17:21:41Z" Dec 03 14:10:40 crc kubenswrapper[4805]: I1203 14:10:40.856958 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-gskh4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6098937f-e3f6-45e8-a647-4994a79cd711\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://28d5653f26ec61647808b3a733ff7eac70aa69ac184660f24147e8cfaf0eb802\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nwrw9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9ad6b68a42fd33ee167e67873c6eed3f11cdf6dc71d22e952c021436eb967dd9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nwrw9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T14:09:50Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-gskh4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T14:10:40Z is after 2025-08-24T17:21:41Z" Dec 03 14:10:40 crc kubenswrapper[4805]: I1203 14:10:40.870699 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 14:10:40 crc kubenswrapper[4805]: I1203 14:10:40.870732 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 14:10:40 crc kubenswrapper[4805]: I1203 14:10:40.870747 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 14:10:40 crc kubenswrapper[4805]: I1203 14:10:40.870920 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 14:10:40 crc kubenswrapper[4805]: I1203 14:10:40.871084 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T14:10:40Z","lastTransitionTime":"2025-12-03T14:10:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 14:10:40 crc kubenswrapper[4805]: I1203 14:10:40.878876 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-psjpm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"deeec30d-cad2-4dfd-adb6-64f6646581be\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:50Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:50Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4b8d00cdf0f37f2e5cd4e952cf85d3de8d72c43fdf786088fd9f26ed17b1e21f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wdlzx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bd67c6de01a8d196a1f51d4d90640c9e69de6f05d5e201b25e6eaa5c33440ea8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wdlzx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cb8bf05ef093eb0727a066d4bbbdf9af9422867cd4dadacafc6eeb016867a61e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wdlzx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0304467102ed053628c0d3e06548c74b6492336b6fd044246508d1e8bdb17bc3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wdlzx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6ce80ccafb3262a11fbe9923317a9d38b61e5dc5b3e7e7ae136f143d6b643fd7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wdlzx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://47d76962beb8c0f196d50af96fb3ec611097823ca80a1c8bd718f2d18840128c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wdlzx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a0272667e733f2c5eec0fce0eb9c31a16c048babdaed90635895d689ae9303f4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a0272667e733f2c5eec0fce0eb9c31a16c048babdaed90635895d689ae9303f4\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-03T14:10:16Z\\\",\\\"message\\\":\\\"o create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T14:10:16Z is after 2025-08-24T17:21:41Z]\\\\nI1203 14:10:16.717659 6478 services_controller.go:473] Services do not match for network=default, existing lbs: []services.LB{services.LB{Name:\\\\\\\"Service_openshift-network-console/networking-console-plugin_TCP_cluster\\\\\\\", UUID:\\\\\\\"ab0b1d51-5ec6-479b-8881-93dfa8d30337\\\\\\\", Protocol:\\\\\\\"tcp\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-network-console/networking-console-plugin\\\\\\\"}, Opts:services.LBOpts{Reject:false, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{}, Templates:services.TemplateMap{}, Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGr\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T14:10:15Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-psjpm_openshift-ovn-kubernetes(deeec30d-cad2-4dfd-adb6-64f6646581be)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wdlzx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ab98e365c1a4242a47b8cbe2dd85109d366d6002c4de1f18ddc950a46bea6c12\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wdlzx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2f875252fbe6a673a72fc9c56bb96899f62732305ad77b3108e3b244bfa91ecb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2f875252fbe6a673a72fc9c56bb96899f62732305ad77b3108e3b244bfa91ecb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T14:09:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T14:09:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wdlzx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T14:09:50Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-psjpm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T14:10:40Z is after 2025-08-24T17:21:41Z" Dec 03 14:10:40 crc kubenswrapper[4805]: I1203 14:10:40.888021 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-4gd9r" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6b36a3db-1047-487a-9d6c-4f51ffdd8cc1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:10:04Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:10:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:10:04Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:10:04Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4t6zl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4t6zl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T14:10:04Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-4gd9r\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T14:10:40Z is after 2025-08-24T17:21:41Z" Dec 03 14:10:40 crc kubenswrapper[4805]: I1203 14:10:40.898159 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:48Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:48Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T14:10:40Z is after 2025-08-24T17:21:41Z" Dec 03 14:10:40 crc kubenswrapper[4805]: I1203 14:10:40.908201 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:49Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:49Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://541ace7ccbb66dbd191c3ca8c1e4479e3edbf80dfc67326bea8aaee81fdcfc3c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://821e3e9a902aacecf6c41e8b8d8aa1ac951ce2fbb6505b7148b16348a3754e55\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T14:10:40Z is after 2025-08-24T17:21:41Z" Dec 03 14:10:40 crc kubenswrapper[4805]: I1203 14:10:40.921652 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-69h75" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5d1c694f-9847-4086-8b37-3573ddeb3ace\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:10:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:10:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:10:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:10:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://abc1d9b8582ca9fe4a51fb608fb39129a179ddcdfde0797a6369a55b2e0124cf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:10:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s578b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c7c4d434b035cac2a91a4f0846a2f4fb5bd68f881c28487860adb6db85e8ce75\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:10:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s578b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T14:10:03Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-69h75\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T14:10:40Z is after 2025-08-24T17:21:41Z" Dec 03 14:10:40 crc kubenswrapper[4805]: I1203 14:10:40.933296 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-pdzgk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"860769de-6a5b-4cb9-a752-264e7e2e785f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5e2f0a0cc21505d609f81918365d5e4f5d21959ac12ab147d6a51601c9de4c14\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ggn4n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T14:09:50Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-pdzgk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T14:10:40Z is after 2025-08-24T17:21:41Z" Dec 03 14:10:40 crc kubenswrapper[4805]: I1203 14:10:40.974769 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 14:10:40 crc kubenswrapper[4805]: I1203 14:10:40.974820 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 14:10:40 crc kubenswrapper[4805]: I1203 14:10:40.974831 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 14:10:40 crc kubenswrapper[4805]: I1203 14:10:40.974866 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 14:10:40 crc kubenswrapper[4805]: I1203 14:10:40.974879 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T14:10:40Z","lastTransitionTime":"2025-12-03T14:10:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 14:10:41 crc kubenswrapper[4805]: I1203 14:10:41.078102 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 14:10:41 crc kubenswrapper[4805]: I1203 14:10:41.078155 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 14:10:41 crc kubenswrapper[4805]: I1203 14:10:41.078165 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 14:10:41 crc kubenswrapper[4805]: I1203 14:10:41.078184 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 14:10:41 crc kubenswrapper[4805]: I1203 14:10:41.078193 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T14:10:41Z","lastTransitionTime":"2025-12-03T14:10:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 14:10:41 crc kubenswrapper[4805]: I1203 14:10:41.180718 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 14:10:41 crc kubenswrapper[4805]: I1203 14:10:41.180996 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 14:10:41 crc kubenswrapper[4805]: I1203 14:10:41.181100 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 14:10:41 crc kubenswrapper[4805]: I1203 14:10:41.181203 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 14:10:41 crc kubenswrapper[4805]: I1203 14:10:41.181304 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T14:10:41Z","lastTransitionTime":"2025-12-03T14:10:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 14:10:41 crc kubenswrapper[4805]: I1203 14:10:41.283241 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 14:10:41 crc kubenswrapper[4805]: I1203 14:10:41.283280 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 14:10:41 crc kubenswrapper[4805]: I1203 14:10:41.283288 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 14:10:41 crc kubenswrapper[4805]: I1203 14:10:41.283301 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 14:10:41 crc kubenswrapper[4805]: I1203 14:10:41.283310 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T14:10:41Z","lastTransitionTime":"2025-12-03T14:10:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 14:10:41 crc kubenswrapper[4805]: I1203 14:10:41.385218 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 14:10:41 crc kubenswrapper[4805]: I1203 14:10:41.385258 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 14:10:41 crc kubenswrapper[4805]: I1203 14:10:41.385269 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 14:10:41 crc kubenswrapper[4805]: I1203 14:10:41.385284 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 14:10:41 crc kubenswrapper[4805]: I1203 14:10:41.385294 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T14:10:41Z","lastTransitionTime":"2025-12-03T14:10:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 14:10:41 crc kubenswrapper[4805]: I1203 14:10:41.487684 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 14:10:41 crc kubenswrapper[4805]: I1203 14:10:41.487765 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 14:10:41 crc kubenswrapper[4805]: I1203 14:10:41.487792 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 14:10:41 crc kubenswrapper[4805]: I1203 14:10:41.487823 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 14:10:41 crc kubenswrapper[4805]: I1203 14:10:41.487883 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T14:10:41Z","lastTransitionTime":"2025-12-03T14:10:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 14:10:41 crc kubenswrapper[4805]: I1203 14:10:41.589752 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 14:10:41 crc kubenswrapper[4805]: I1203 14:10:41.589795 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 14:10:41 crc kubenswrapper[4805]: I1203 14:10:41.589806 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 14:10:41 crc kubenswrapper[4805]: I1203 14:10:41.589823 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 14:10:41 crc kubenswrapper[4805]: I1203 14:10:41.589860 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T14:10:41Z","lastTransitionTime":"2025-12-03T14:10:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 14:10:41 crc kubenswrapper[4805]: I1203 14:10:41.693825 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 14:10:41 crc kubenswrapper[4805]: I1203 14:10:41.693891 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 14:10:41 crc kubenswrapper[4805]: I1203 14:10:41.693904 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 14:10:41 crc kubenswrapper[4805]: I1203 14:10:41.693922 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 14:10:41 crc kubenswrapper[4805]: I1203 14:10:41.693934 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T14:10:41Z","lastTransitionTime":"2025-12-03T14:10:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 14:10:41 crc kubenswrapper[4805]: I1203 14:10:41.796892 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 14:10:41 crc kubenswrapper[4805]: I1203 14:10:41.797555 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 14:10:41 crc kubenswrapper[4805]: I1203 14:10:41.797635 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 14:10:41 crc kubenswrapper[4805]: I1203 14:10:41.797745 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 14:10:41 crc kubenswrapper[4805]: I1203 14:10:41.797833 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T14:10:41Z","lastTransitionTime":"2025-12-03T14:10:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 14:10:41 crc kubenswrapper[4805]: I1203 14:10:41.900959 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 14:10:41 crc kubenswrapper[4805]: I1203 14:10:41.901032 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 14:10:41 crc kubenswrapper[4805]: I1203 14:10:41.901049 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 14:10:41 crc kubenswrapper[4805]: I1203 14:10:41.901069 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 14:10:41 crc kubenswrapper[4805]: I1203 14:10:41.901107 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T14:10:41Z","lastTransitionTime":"2025-12-03T14:10:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 14:10:42 crc kubenswrapper[4805]: I1203 14:10:42.005179 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 14:10:42 crc kubenswrapper[4805]: I1203 14:10:42.005629 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 14:10:42 crc kubenswrapper[4805]: I1203 14:10:42.005724 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 14:10:42 crc kubenswrapper[4805]: I1203 14:10:42.005856 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 14:10:42 crc kubenswrapper[4805]: I1203 14:10:42.005956 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T14:10:42Z","lastTransitionTime":"2025-12-03T14:10:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 14:10:42 crc kubenswrapper[4805]: I1203 14:10:42.108040 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 14:10:42 crc kubenswrapper[4805]: I1203 14:10:42.108091 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 14:10:42 crc kubenswrapper[4805]: I1203 14:10:42.108109 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 14:10:42 crc kubenswrapper[4805]: I1203 14:10:42.108127 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 14:10:42 crc kubenswrapper[4805]: I1203 14:10:42.108138 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T14:10:42Z","lastTransitionTime":"2025-12-03T14:10:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 14:10:42 crc kubenswrapper[4805]: I1203 14:10:42.211389 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 14:10:42 crc kubenswrapper[4805]: I1203 14:10:42.211437 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 14:10:42 crc kubenswrapper[4805]: I1203 14:10:42.211448 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 14:10:42 crc kubenswrapper[4805]: I1203 14:10:42.211475 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 14:10:42 crc kubenswrapper[4805]: I1203 14:10:42.211507 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T14:10:42Z","lastTransitionTime":"2025-12-03T14:10:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 14:10:42 crc kubenswrapper[4805]: I1203 14:10:42.313184 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 14:10:42 crc kubenswrapper[4805]: I1203 14:10:42.313229 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 14:10:42 crc kubenswrapper[4805]: I1203 14:10:42.313241 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 14:10:42 crc kubenswrapper[4805]: I1203 14:10:42.313257 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 14:10:42 crc kubenswrapper[4805]: I1203 14:10:42.313268 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T14:10:42Z","lastTransitionTime":"2025-12-03T14:10:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 14:10:42 crc kubenswrapper[4805]: I1203 14:10:42.416134 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 14:10:42 crc kubenswrapper[4805]: I1203 14:10:42.416199 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 14:10:42 crc kubenswrapper[4805]: I1203 14:10:42.416217 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 14:10:42 crc kubenswrapper[4805]: I1203 14:10:42.416244 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 14:10:42 crc kubenswrapper[4805]: I1203 14:10:42.416266 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T14:10:42Z","lastTransitionTime":"2025-12-03T14:10:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 14:10:42 crc kubenswrapper[4805]: I1203 14:10:42.519626 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 14:10:42 crc kubenswrapper[4805]: I1203 14:10:42.519683 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 14:10:42 crc kubenswrapper[4805]: I1203 14:10:42.519695 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 14:10:42 crc kubenswrapper[4805]: I1203 14:10:42.519718 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 14:10:42 crc kubenswrapper[4805]: I1203 14:10:42.519731 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T14:10:42Z","lastTransitionTime":"2025-12-03T14:10:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 14:10:42 crc kubenswrapper[4805]: I1203 14:10:42.622032 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 14:10:42 crc kubenswrapper[4805]: I1203 14:10:42.622098 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 14:10:42 crc kubenswrapper[4805]: I1203 14:10:42.622114 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 14:10:42 crc kubenswrapper[4805]: I1203 14:10:42.622135 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 14:10:42 crc kubenswrapper[4805]: I1203 14:10:42.622151 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T14:10:42Z","lastTransitionTime":"2025-12-03T14:10:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 14:10:42 crc kubenswrapper[4805]: I1203 14:10:42.694553 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-4gd9r" Dec 03 14:10:42 crc kubenswrapper[4805]: I1203 14:10:42.694682 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 14:10:42 crc kubenswrapper[4805]: I1203 14:10:42.694682 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 14:10:42 crc kubenswrapper[4805]: E1203 14:10:42.694739 4805 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-4gd9r" podUID="6b36a3db-1047-487a-9d6c-4f51ffdd8cc1" Dec 03 14:10:42 crc kubenswrapper[4805]: I1203 14:10:42.694827 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 14:10:42 crc kubenswrapper[4805]: E1203 14:10:42.695014 4805 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 03 14:10:42 crc kubenswrapper[4805]: E1203 14:10:42.695145 4805 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 03 14:10:42 crc kubenswrapper[4805]: E1203 14:10:42.695748 4805 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 03 14:10:42 crc kubenswrapper[4805]: I1203 14:10:42.725482 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 14:10:42 crc kubenswrapper[4805]: I1203 14:10:42.725766 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 14:10:42 crc kubenswrapper[4805]: I1203 14:10:42.725854 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 14:10:42 crc kubenswrapper[4805]: I1203 14:10:42.725937 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 14:10:42 crc kubenswrapper[4805]: I1203 14:10:42.726027 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T14:10:42Z","lastTransitionTime":"2025-12-03T14:10:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 14:10:42 crc kubenswrapper[4805]: I1203 14:10:42.829620 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 14:10:42 crc kubenswrapper[4805]: I1203 14:10:42.829699 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 14:10:42 crc kubenswrapper[4805]: I1203 14:10:42.829724 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 14:10:42 crc kubenswrapper[4805]: I1203 14:10:42.829753 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 14:10:42 crc kubenswrapper[4805]: I1203 14:10:42.829810 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T14:10:42Z","lastTransitionTime":"2025-12-03T14:10:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 14:10:42 crc kubenswrapper[4805]: I1203 14:10:42.934943 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 14:10:42 crc kubenswrapper[4805]: I1203 14:10:42.935730 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 14:10:42 crc kubenswrapper[4805]: I1203 14:10:42.935816 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 14:10:42 crc kubenswrapper[4805]: I1203 14:10:42.935937 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 14:10:42 crc kubenswrapper[4805]: I1203 14:10:42.936026 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T14:10:42Z","lastTransitionTime":"2025-12-03T14:10:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 14:10:43 crc kubenswrapper[4805]: I1203 14:10:43.039239 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 14:10:43 crc kubenswrapper[4805]: I1203 14:10:43.039513 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 14:10:43 crc kubenswrapper[4805]: I1203 14:10:43.039585 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 14:10:43 crc kubenswrapper[4805]: I1203 14:10:43.039662 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 14:10:43 crc kubenswrapper[4805]: I1203 14:10:43.039720 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T14:10:43Z","lastTransitionTime":"2025-12-03T14:10:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 14:10:43 crc kubenswrapper[4805]: I1203 14:10:43.062756 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 14:10:43 crc kubenswrapper[4805]: I1203 14:10:43.062823 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 14:10:43 crc kubenswrapper[4805]: I1203 14:10:43.062854 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 14:10:43 crc kubenswrapper[4805]: I1203 14:10:43.062875 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 14:10:43 crc kubenswrapper[4805]: I1203 14:10:43.062889 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T14:10:43Z","lastTransitionTime":"2025-12-03T14:10:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 14:10:43 crc kubenswrapper[4805]: E1203 14:10:43.075177 4805 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T14:10:43Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T14:10:43Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T14:10:43Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T14:10:43Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T14:10:43Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T14:10:43Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T14:10:43Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T14:10:43Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"e4edfbcd-1de1-486e-9c0b-65f9fbba52ad\\\",\\\"systemUUID\\\":\\\"7a729b24-19ef-490e-ab4c-d31c52ed5658\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T14:10:43Z is after 2025-08-24T17:21:41Z" Dec 03 14:10:43 crc kubenswrapper[4805]: I1203 14:10:43.078608 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 14:10:43 crc kubenswrapper[4805]: I1203 14:10:43.078663 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 14:10:43 crc kubenswrapper[4805]: I1203 14:10:43.078674 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 14:10:43 crc kubenswrapper[4805]: I1203 14:10:43.078697 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 14:10:43 crc kubenswrapper[4805]: I1203 14:10:43.078711 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T14:10:43Z","lastTransitionTime":"2025-12-03T14:10:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 14:10:43 crc kubenswrapper[4805]: E1203 14:10:43.094625 4805 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T14:10:43Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T14:10:43Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T14:10:43Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T14:10:43Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T14:10:43Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T14:10:43Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T14:10:43Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T14:10:43Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"e4edfbcd-1de1-486e-9c0b-65f9fbba52ad\\\",\\\"systemUUID\\\":\\\"7a729b24-19ef-490e-ab4c-d31c52ed5658\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T14:10:43Z is after 2025-08-24T17:21:41Z" Dec 03 14:10:43 crc kubenswrapper[4805]: I1203 14:10:43.099093 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 14:10:43 crc kubenswrapper[4805]: I1203 14:10:43.099160 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 14:10:43 crc kubenswrapper[4805]: I1203 14:10:43.099174 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 14:10:43 crc kubenswrapper[4805]: I1203 14:10:43.099198 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 14:10:43 crc kubenswrapper[4805]: I1203 14:10:43.099213 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T14:10:43Z","lastTransitionTime":"2025-12-03T14:10:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 14:10:43 crc kubenswrapper[4805]: E1203 14:10:43.113088 4805 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T14:10:43Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T14:10:43Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T14:10:43Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T14:10:43Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T14:10:43Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T14:10:43Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T14:10:43Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T14:10:43Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"e4edfbcd-1de1-486e-9c0b-65f9fbba52ad\\\",\\\"systemUUID\\\":\\\"7a729b24-19ef-490e-ab4c-d31c52ed5658\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T14:10:43Z is after 2025-08-24T17:21:41Z" Dec 03 14:10:43 crc kubenswrapper[4805]: I1203 14:10:43.117976 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 14:10:43 crc kubenswrapper[4805]: I1203 14:10:43.118050 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 14:10:43 crc kubenswrapper[4805]: I1203 14:10:43.118067 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 14:10:43 crc kubenswrapper[4805]: I1203 14:10:43.118095 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 14:10:43 crc kubenswrapper[4805]: I1203 14:10:43.118113 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T14:10:43Z","lastTransitionTime":"2025-12-03T14:10:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 14:10:43 crc kubenswrapper[4805]: E1203 14:10:43.130520 4805 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T14:10:43Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T14:10:43Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T14:10:43Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T14:10:43Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T14:10:43Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T14:10:43Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T14:10:43Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T14:10:43Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"e4edfbcd-1de1-486e-9c0b-65f9fbba52ad\\\",\\\"systemUUID\\\":\\\"7a729b24-19ef-490e-ab4c-d31c52ed5658\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T14:10:43Z is after 2025-08-24T17:21:41Z" Dec 03 14:10:43 crc kubenswrapper[4805]: I1203 14:10:43.135778 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 14:10:43 crc kubenswrapper[4805]: I1203 14:10:43.135855 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 14:10:43 crc kubenswrapper[4805]: I1203 14:10:43.135868 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 14:10:43 crc kubenswrapper[4805]: I1203 14:10:43.135891 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 14:10:43 crc kubenswrapper[4805]: I1203 14:10:43.135906 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T14:10:43Z","lastTransitionTime":"2025-12-03T14:10:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 14:10:43 crc kubenswrapper[4805]: E1203 14:10:43.148513 4805 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T14:10:43Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T14:10:43Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T14:10:43Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T14:10:43Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T14:10:43Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T14:10:43Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T14:10:43Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T14:10:43Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"e4edfbcd-1de1-486e-9c0b-65f9fbba52ad\\\",\\\"systemUUID\\\":\\\"7a729b24-19ef-490e-ab4c-d31c52ed5658\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T14:10:43Z is after 2025-08-24T17:21:41Z" Dec 03 14:10:43 crc kubenswrapper[4805]: E1203 14:10:43.148674 4805 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Dec 03 14:10:43 crc kubenswrapper[4805]: I1203 14:10:43.150276 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 14:10:43 crc kubenswrapper[4805]: I1203 14:10:43.150321 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 14:10:43 crc kubenswrapper[4805]: I1203 14:10:43.150338 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 14:10:43 crc kubenswrapper[4805]: I1203 14:10:43.150360 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 14:10:43 crc kubenswrapper[4805]: I1203 14:10:43.150374 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T14:10:43Z","lastTransitionTime":"2025-12-03T14:10:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 14:10:43 crc kubenswrapper[4805]: I1203 14:10:43.252521 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 14:10:43 crc kubenswrapper[4805]: I1203 14:10:43.252586 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 14:10:43 crc kubenswrapper[4805]: I1203 14:10:43.252612 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 14:10:43 crc kubenswrapper[4805]: I1203 14:10:43.252641 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 14:10:43 crc kubenswrapper[4805]: I1203 14:10:43.252658 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T14:10:43Z","lastTransitionTime":"2025-12-03T14:10:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 14:10:43 crc kubenswrapper[4805]: I1203 14:10:43.355580 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 14:10:43 crc kubenswrapper[4805]: I1203 14:10:43.355767 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 14:10:43 crc kubenswrapper[4805]: I1203 14:10:43.355794 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 14:10:43 crc kubenswrapper[4805]: I1203 14:10:43.355818 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 14:10:43 crc kubenswrapper[4805]: I1203 14:10:43.355836 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T14:10:43Z","lastTransitionTime":"2025-12-03T14:10:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 14:10:43 crc kubenswrapper[4805]: I1203 14:10:43.458338 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 14:10:43 crc kubenswrapper[4805]: I1203 14:10:43.458395 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 14:10:43 crc kubenswrapper[4805]: I1203 14:10:43.458407 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 14:10:43 crc kubenswrapper[4805]: I1203 14:10:43.458430 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 14:10:43 crc kubenswrapper[4805]: I1203 14:10:43.458442 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T14:10:43Z","lastTransitionTime":"2025-12-03T14:10:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 14:10:43 crc kubenswrapper[4805]: I1203 14:10:43.560868 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 14:10:43 crc kubenswrapper[4805]: I1203 14:10:43.560933 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 14:10:43 crc kubenswrapper[4805]: I1203 14:10:43.560944 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 14:10:43 crc kubenswrapper[4805]: I1203 14:10:43.560960 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 14:10:43 crc kubenswrapper[4805]: I1203 14:10:43.560971 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T14:10:43Z","lastTransitionTime":"2025-12-03T14:10:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 14:10:43 crc kubenswrapper[4805]: I1203 14:10:43.663293 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 14:10:43 crc kubenswrapper[4805]: I1203 14:10:43.663339 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 14:10:43 crc kubenswrapper[4805]: I1203 14:10:43.663351 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 14:10:43 crc kubenswrapper[4805]: I1203 14:10:43.663369 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 14:10:43 crc kubenswrapper[4805]: I1203 14:10:43.663381 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T14:10:43Z","lastTransitionTime":"2025-12-03T14:10:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 14:10:43 crc kubenswrapper[4805]: I1203 14:10:43.766017 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 14:10:43 crc kubenswrapper[4805]: I1203 14:10:43.766096 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 14:10:43 crc kubenswrapper[4805]: I1203 14:10:43.766110 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 14:10:43 crc kubenswrapper[4805]: I1203 14:10:43.766128 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 14:10:43 crc kubenswrapper[4805]: I1203 14:10:43.766138 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T14:10:43Z","lastTransitionTime":"2025-12-03T14:10:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 14:10:43 crc kubenswrapper[4805]: I1203 14:10:43.869184 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 14:10:43 crc kubenswrapper[4805]: I1203 14:10:43.869217 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 14:10:43 crc kubenswrapper[4805]: I1203 14:10:43.869225 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 14:10:43 crc kubenswrapper[4805]: I1203 14:10:43.869237 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 14:10:43 crc kubenswrapper[4805]: I1203 14:10:43.869245 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T14:10:43Z","lastTransitionTime":"2025-12-03T14:10:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 14:10:43 crc kubenswrapper[4805]: I1203 14:10:43.972794 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 14:10:43 crc kubenswrapper[4805]: I1203 14:10:43.972879 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 14:10:43 crc kubenswrapper[4805]: I1203 14:10:43.972891 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 14:10:43 crc kubenswrapper[4805]: I1203 14:10:43.972914 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 14:10:43 crc kubenswrapper[4805]: I1203 14:10:43.972926 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T14:10:43Z","lastTransitionTime":"2025-12-03T14:10:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 14:10:44 crc kubenswrapper[4805]: I1203 14:10:44.075951 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 14:10:44 crc kubenswrapper[4805]: I1203 14:10:44.076004 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 14:10:44 crc kubenswrapper[4805]: I1203 14:10:44.076015 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 14:10:44 crc kubenswrapper[4805]: I1203 14:10:44.076036 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 14:10:44 crc kubenswrapper[4805]: I1203 14:10:44.076046 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T14:10:44Z","lastTransitionTime":"2025-12-03T14:10:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 14:10:44 crc kubenswrapper[4805]: I1203 14:10:44.178559 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 14:10:44 crc kubenswrapper[4805]: I1203 14:10:44.178627 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 14:10:44 crc kubenswrapper[4805]: I1203 14:10:44.178639 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 14:10:44 crc kubenswrapper[4805]: I1203 14:10:44.178662 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 14:10:44 crc kubenswrapper[4805]: I1203 14:10:44.178675 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T14:10:44Z","lastTransitionTime":"2025-12-03T14:10:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 14:10:44 crc kubenswrapper[4805]: I1203 14:10:44.281793 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 14:10:44 crc kubenswrapper[4805]: I1203 14:10:44.281864 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 14:10:44 crc kubenswrapper[4805]: I1203 14:10:44.281875 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 14:10:44 crc kubenswrapper[4805]: I1203 14:10:44.281892 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 14:10:44 crc kubenswrapper[4805]: I1203 14:10:44.281904 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T14:10:44Z","lastTransitionTime":"2025-12-03T14:10:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 14:10:44 crc kubenswrapper[4805]: I1203 14:10:44.384671 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 14:10:44 crc kubenswrapper[4805]: I1203 14:10:44.384747 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 14:10:44 crc kubenswrapper[4805]: I1203 14:10:44.384769 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 14:10:44 crc kubenswrapper[4805]: I1203 14:10:44.384809 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 14:10:44 crc kubenswrapper[4805]: I1203 14:10:44.384831 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T14:10:44Z","lastTransitionTime":"2025-12-03T14:10:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 14:10:44 crc kubenswrapper[4805]: I1203 14:10:44.487666 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 14:10:44 crc kubenswrapper[4805]: I1203 14:10:44.487698 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 14:10:44 crc kubenswrapper[4805]: I1203 14:10:44.487707 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 14:10:44 crc kubenswrapper[4805]: I1203 14:10:44.487722 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 14:10:44 crc kubenswrapper[4805]: I1203 14:10:44.487731 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T14:10:44Z","lastTransitionTime":"2025-12-03T14:10:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 14:10:44 crc kubenswrapper[4805]: I1203 14:10:44.590814 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 14:10:44 crc kubenswrapper[4805]: I1203 14:10:44.590904 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 14:10:44 crc kubenswrapper[4805]: I1203 14:10:44.590918 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 14:10:44 crc kubenswrapper[4805]: I1203 14:10:44.590940 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 14:10:44 crc kubenswrapper[4805]: I1203 14:10:44.590953 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T14:10:44Z","lastTransitionTime":"2025-12-03T14:10:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 14:10:44 crc kubenswrapper[4805]: I1203 14:10:44.693384 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-4gd9r" Dec 03 14:10:44 crc kubenswrapper[4805]: I1203 14:10:44.693403 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 14:10:44 crc kubenswrapper[4805]: I1203 14:10:44.693426 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 14:10:44 crc kubenswrapper[4805]: I1203 14:10:44.693437 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 14:10:44 crc kubenswrapper[4805]: I1203 14:10:44.693450 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 14:10:44 crc kubenswrapper[4805]: I1203 14:10:44.693462 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T14:10:44Z","lastTransitionTime":"2025-12-03T14:10:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 14:10:44 crc kubenswrapper[4805]: I1203 14:10:44.693525 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 14:10:44 crc kubenswrapper[4805]: I1203 14:10:44.694131 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 14:10:44 crc kubenswrapper[4805]: I1203 14:10:44.694147 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 14:10:44 crc kubenswrapper[4805]: E1203 14:10:44.694238 4805 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 03 14:10:44 crc kubenswrapper[4805]: E1203 14:10:44.694361 4805 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 03 14:10:44 crc kubenswrapper[4805]: E1203 14:10:44.693660 4805 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-4gd9r" podUID="6b36a3db-1047-487a-9d6c-4f51ffdd8cc1" Dec 03 14:10:44 crc kubenswrapper[4805]: E1203 14:10:44.694441 4805 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 03 14:10:44 crc kubenswrapper[4805]: I1203 14:10:44.796139 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 14:10:44 crc kubenswrapper[4805]: I1203 14:10:44.796255 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 14:10:44 crc kubenswrapper[4805]: I1203 14:10:44.796280 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 14:10:44 crc kubenswrapper[4805]: I1203 14:10:44.796311 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 14:10:44 crc kubenswrapper[4805]: I1203 14:10:44.796331 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T14:10:44Z","lastTransitionTime":"2025-12-03T14:10:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 14:10:44 crc kubenswrapper[4805]: I1203 14:10:44.899207 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 14:10:44 crc kubenswrapper[4805]: I1203 14:10:44.899251 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 14:10:44 crc kubenswrapper[4805]: I1203 14:10:44.899260 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 14:10:44 crc kubenswrapper[4805]: I1203 14:10:44.899276 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 14:10:44 crc kubenswrapper[4805]: I1203 14:10:44.899285 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T14:10:44Z","lastTransitionTime":"2025-12-03T14:10:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 14:10:45 crc kubenswrapper[4805]: I1203 14:10:45.002135 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 14:10:45 crc kubenswrapper[4805]: I1203 14:10:45.002180 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 14:10:45 crc kubenswrapper[4805]: I1203 14:10:45.002198 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 14:10:45 crc kubenswrapper[4805]: I1203 14:10:45.002220 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 14:10:45 crc kubenswrapper[4805]: I1203 14:10:45.002236 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T14:10:45Z","lastTransitionTime":"2025-12-03T14:10:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 14:10:45 crc kubenswrapper[4805]: I1203 14:10:45.104355 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 14:10:45 crc kubenswrapper[4805]: I1203 14:10:45.104410 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 14:10:45 crc kubenswrapper[4805]: I1203 14:10:45.104427 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 14:10:45 crc kubenswrapper[4805]: I1203 14:10:45.104449 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 14:10:45 crc kubenswrapper[4805]: I1203 14:10:45.104464 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T14:10:45Z","lastTransitionTime":"2025-12-03T14:10:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 14:10:45 crc kubenswrapper[4805]: I1203 14:10:45.207106 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 14:10:45 crc kubenswrapper[4805]: I1203 14:10:45.207133 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 14:10:45 crc kubenswrapper[4805]: I1203 14:10:45.207145 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 14:10:45 crc kubenswrapper[4805]: I1203 14:10:45.207159 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 14:10:45 crc kubenswrapper[4805]: I1203 14:10:45.207171 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T14:10:45Z","lastTransitionTime":"2025-12-03T14:10:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 14:10:45 crc kubenswrapper[4805]: I1203 14:10:45.310006 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 14:10:45 crc kubenswrapper[4805]: I1203 14:10:45.310055 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 14:10:45 crc kubenswrapper[4805]: I1203 14:10:45.310067 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 14:10:45 crc kubenswrapper[4805]: I1203 14:10:45.310084 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 14:10:45 crc kubenswrapper[4805]: I1203 14:10:45.310096 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T14:10:45Z","lastTransitionTime":"2025-12-03T14:10:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 14:10:45 crc kubenswrapper[4805]: I1203 14:10:45.412363 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 14:10:45 crc kubenswrapper[4805]: I1203 14:10:45.412440 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 14:10:45 crc kubenswrapper[4805]: I1203 14:10:45.412461 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 14:10:45 crc kubenswrapper[4805]: I1203 14:10:45.412489 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 14:10:45 crc kubenswrapper[4805]: I1203 14:10:45.412507 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T14:10:45Z","lastTransitionTime":"2025-12-03T14:10:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 14:10:45 crc kubenswrapper[4805]: I1203 14:10:45.516025 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 14:10:45 crc kubenswrapper[4805]: I1203 14:10:45.516088 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 14:10:45 crc kubenswrapper[4805]: I1203 14:10:45.516106 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 14:10:45 crc kubenswrapper[4805]: I1203 14:10:45.516129 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 14:10:45 crc kubenswrapper[4805]: I1203 14:10:45.516146 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T14:10:45Z","lastTransitionTime":"2025-12-03T14:10:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 14:10:45 crc kubenswrapper[4805]: I1203 14:10:45.619001 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 14:10:45 crc kubenswrapper[4805]: I1203 14:10:45.619047 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 14:10:45 crc kubenswrapper[4805]: I1203 14:10:45.619063 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 14:10:45 crc kubenswrapper[4805]: I1203 14:10:45.619085 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 14:10:45 crc kubenswrapper[4805]: I1203 14:10:45.619101 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T14:10:45Z","lastTransitionTime":"2025-12-03T14:10:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 14:10:45 crc kubenswrapper[4805]: I1203 14:10:45.723061 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 14:10:45 crc kubenswrapper[4805]: I1203 14:10:45.723133 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 14:10:45 crc kubenswrapper[4805]: I1203 14:10:45.723157 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 14:10:45 crc kubenswrapper[4805]: I1203 14:10:45.723187 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 14:10:45 crc kubenswrapper[4805]: I1203 14:10:45.723209 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T14:10:45Z","lastTransitionTime":"2025-12-03T14:10:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 14:10:45 crc kubenswrapper[4805]: I1203 14:10:45.826309 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 14:10:45 crc kubenswrapper[4805]: I1203 14:10:45.826353 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 14:10:45 crc kubenswrapper[4805]: I1203 14:10:45.826365 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 14:10:45 crc kubenswrapper[4805]: I1203 14:10:45.826379 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 14:10:45 crc kubenswrapper[4805]: I1203 14:10:45.826389 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T14:10:45Z","lastTransitionTime":"2025-12-03T14:10:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 14:10:45 crc kubenswrapper[4805]: I1203 14:10:45.929464 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 14:10:45 crc kubenswrapper[4805]: I1203 14:10:45.929517 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 14:10:45 crc kubenswrapper[4805]: I1203 14:10:45.929533 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 14:10:45 crc kubenswrapper[4805]: I1203 14:10:45.929556 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 14:10:45 crc kubenswrapper[4805]: I1203 14:10:45.929574 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T14:10:45Z","lastTransitionTime":"2025-12-03T14:10:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 14:10:46 crc kubenswrapper[4805]: I1203 14:10:46.031147 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 14:10:46 crc kubenswrapper[4805]: I1203 14:10:46.031186 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 14:10:46 crc kubenswrapper[4805]: I1203 14:10:46.031196 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 14:10:46 crc kubenswrapper[4805]: I1203 14:10:46.031213 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 14:10:46 crc kubenswrapper[4805]: I1203 14:10:46.031226 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T14:10:46Z","lastTransitionTime":"2025-12-03T14:10:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 14:10:46 crc kubenswrapper[4805]: I1203 14:10:46.133386 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 14:10:46 crc kubenswrapper[4805]: I1203 14:10:46.133425 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 14:10:46 crc kubenswrapper[4805]: I1203 14:10:46.133439 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 14:10:46 crc kubenswrapper[4805]: I1203 14:10:46.133486 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 14:10:46 crc kubenswrapper[4805]: I1203 14:10:46.133500 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T14:10:46Z","lastTransitionTime":"2025-12-03T14:10:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 14:10:46 crc kubenswrapper[4805]: I1203 14:10:46.236467 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 14:10:46 crc kubenswrapper[4805]: I1203 14:10:46.236538 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 14:10:46 crc kubenswrapper[4805]: I1203 14:10:46.236561 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 14:10:46 crc kubenswrapper[4805]: I1203 14:10:46.236590 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 14:10:46 crc kubenswrapper[4805]: I1203 14:10:46.236610 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T14:10:46Z","lastTransitionTime":"2025-12-03T14:10:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 14:10:46 crc kubenswrapper[4805]: I1203 14:10:46.340139 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 14:10:46 crc kubenswrapper[4805]: I1203 14:10:46.340198 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 14:10:46 crc kubenswrapper[4805]: I1203 14:10:46.340215 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 14:10:46 crc kubenswrapper[4805]: I1203 14:10:46.340237 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 14:10:46 crc kubenswrapper[4805]: I1203 14:10:46.340253 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T14:10:46Z","lastTransitionTime":"2025-12-03T14:10:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 14:10:46 crc kubenswrapper[4805]: I1203 14:10:46.443286 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 14:10:46 crc kubenswrapper[4805]: I1203 14:10:46.443328 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 14:10:46 crc kubenswrapper[4805]: I1203 14:10:46.443336 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 14:10:46 crc kubenswrapper[4805]: I1203 14:10:46.443354 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 14:10:46 crc kubenswrapper[4805]: I1203 14:10:46.443364 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T14:10:46Z","lastTransitionTime":"2025-12-03T14:10:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 14:10:46 crc kubenswrapper[4805]: I1203 14:10:46.546478 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 14:10:46 crc kubenswrapper[4805]: I1203 14:10:46.546524 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 14:10:46 crc kubenswrapper[4805]: I1203 14:10:46.546542 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 14:10:46 crc kubenswrapper[4805]: I1203 14:10:46.546561 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 14:10:46 crc kubenswrapper[4805]: I1203 14:10:46.546575 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T14:10:46Z","lastTransitionTime":"2025-12-03T14:10:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 14:10:46 crc kubenswrapper[4805]: I1203 14:10:46.648564 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 14:10:46 crc kubenswrapper[4805]: I1203 14:10:46.648634 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 14:10:46 crc kubenswrapper[4805]: I1203 14:10:46.648648 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 14:10:46 crc kubenswrapper[4805]: I1203 14:10:46.648667 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 14:10:46 crc kubenswrapper[4805]: I1203 14:10:46.648678 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T14:10:46Z","lastTransitionTime":"2025-12-03T14:10:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 14:10:46 crc kubenswrapper[4805]: I1203 14:10:46.694052 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 14:10:46 crc kubenswrapper[4805]: I1203 14:10:46.694209 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 14:10:46 crc kubenswrapper[4805]: I1203 14:10:46.694415 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 14:10:46 crc kubenswrapper[4805]: I1203 14:10:46.694445 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-4gd9r" Dec 03 14:10:46 crc kubenswrapper[4805]: E1203 14:10:46.694435 4805 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 03 14:10:46 crc kubenswrapper[4805]: E1203 14:10:46.694565 4805 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 03 14:10:46 crc kubenswrapper[4805]: E1203 14:10:46.694637 4805 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 03 14:10:46 crc kubenswrapper[4805]: E1203 14:10:46.694691 4805 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-4gd9r" podUID="6b36a3db-1047-487a-9d6c-4f51ffdd8cc1" Dec 03 14:10:46 crc kubenswrapper[4805]: I1203 14:10:46.751463 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 14:10:46 crc kubenswrapper[4805]: I1203 14:10:46.751520 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 14:10:46 crc kubenswrapper[4805]: I1203 14:10:46.751538 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 14:10:46 crc kubenswrapper[4805]: I1203 14:10:46.751563 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 14:10:46 crc kubenswrapper[4805]: I1203 14:10:46.751580 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T14:10:46Z","lastTransitionTime":"2025-12-03T14:10:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 14:10:46 crc kubenswrapper[4805]: I1203 14:10:46.854558 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 14:10:46 crc kubenswrapper[4805]: I1203 14:10:46.854613 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 14:10:46 crc kubenswrapper[4805]: I1203 14:10:46.854624 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 14:10:46 crc kubenswrapper[4805]: I1203 14:10:46.854643 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 14:10:46 crc kubenswrapper[4805]: I1203 14:10:46.854654 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T14:10:46Z","lastTransitionTime":"2025-12-03T14:10:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 14:10:46 crc kubenswrapper[4805]: I1203 14:10:46.957413 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 14:10:46 crc kubenswrapper[4805]: I1203 14:10:46.957465 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 14:10:46 crc kubenswrapper[4805]: I1203 14:10:46.957474 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 14:10:46 crc kubenswrapper[4805]: I1203 14:10:46.957490 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 14:10:46 crc kubenswrapper[4805]: I1203 14:10:46.957502 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T14:10:46Z","lastTransitionTime":"2025-12-03T14:10:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 14:10:47 crc kubenswrapper[4805]: I1203 14:10:47.059915 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 14:10:47 crc kubenswrapper[4805]: I1203 14:10:47.059980 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 14:10:47 crc kubenswrapper[4805]: I1203 14:10:47.059994 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 14:10:47 crc kubenswrapper[4805]: I1203 14:10:47.060017 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 14:10:47 crc kubenswrapper[4805]: I1203 14:10:47.060030 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T14:10:47Z","lastTransitionTime":"2025-12-03T14:10:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 14:10:47 crc kubenswrapper[4805]: I1203 14:10:47.163038 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 14:10:47 crc kubenswrapper[4805]: I1203 14:10:47.163079 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 14:10:47 crc kubenswrapper[4805]: I1203 14:10:47.163089 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 14:10:47 crc kubenswrapper[4805]: I1203 14:10:47.163103 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 14:10:47 crc kubenswrapper[4805]: I1203 14:10:47.163113 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T14:10:47Z","lastTransitionTime":"2025-12-03T14:10:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 14:10:47 crc kubenswrapper[4805]: I1203 14:10:47.266211 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 14:10:47 crc kubenswrapper[4805]: I1203 14:10:47.266247 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 14:10:47 crc kubenswrapper[4805]: I1203 14:10:47.266256 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 14:10:47 crc kubenswrapper[4805]: I1203 14:10:47.266268 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 14:10:47 crc kubenswrapper[4805]: I1203 14:10:47.266279 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T14:10:47Z","lastTransitionTime":"2025-12-03T14:10:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 14:10:47 crc kubenswrapper[4805]: I1203 14:10:47.368810 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 14:10:47 crc kubenswrapper[4805]: I1203 14:10:47.368904 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 14:10:47 crc kubenswrapper[4805]: I1203 14:10:47.368929 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 14:10:47 crc kubenswrapper[4805]: I1203 14:10:47.368947 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 14:10:47 crc kubenswrapper[4805]: I1203 14:10:47.368959 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T14:10:47Z","lastTransitionTime":"2025-12-03T14:10:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 14:10:47 crc kubenswrapper[4805]: I1203 14:10:47.471515 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 14:10:47 crc kubenswrapper[4805]: I1203 14:10:47.471574 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 14:10:47 crc kubenswrapper[4805]: I1203 14:10:47.471592 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 14:10:47 crc kubenswrapper[4805]: I1203 14:10:47.471618 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 14:10:47 crc kubenswrapper[4805]: I1203 14:10:47.471659 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T14:10:47Z","lastTransitionTime":"2025-12-03T14:10:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 14:10:47 crc kubenswrapper[4805]: I1203 14:10:47.574420 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 14:10:47 crc kubenswrapper[4805]: I1203 14:10:47.574486 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 14:10:47 crc kubenswrapper[4805]: I1203 14:10:47.574506 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 14:10:47 crc kubenswrapper[4805]: I1203 14:10:47.574533 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 14:10:47 crc kubenswrapper[4805]: I1203 14:10:47.574553 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T14:10:47Z","lastTransitionTime":"2025-12-03T14:10:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 14:10:47 crc kubenswrapper[4805]: I1203 14:10:47.677381 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 14:10:47 crc kubenswrapper[4805]: I1203 14:10:47.677422 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 14:10:47 crc kubenswrapper[4805]: I1203 14:10:47.677436 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 14:10:47 crc kubenswrapper[4805]: I1203 14:10:47.677457 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 14:10:47 crc kubenswrapper[4805]: I1203 14:10:47.677472 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T14:10:47Z","lastTransitionTime":"2025-12-03T14:10:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 14:10:47 crc kubenswrapper[4805]: I1203 14:10:47.695007 4805 scope.go:117] "RemoveContainer" containerID="a0272667e733f2c5eec0fce0eb9c31a16c048babdaed90635895d689ae9303f4" Dec 03 14:10:47 crc kubenswrapper[4805]: I1203 14:10:47.779881 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 14:10:47 crc kubenswrapper[4805]: I1203 14:10:47.779928 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 14:10:47 crc kubenswrapper[4805]: I1203 14:10:47.779940 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 14:10:47 crc kubenswrapper[4805]: I1203 14:10:47.779960 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 14:10:47 crc kubenswrapper[4805]: I1203 14:10:47.779971 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T14:10:47Z","lastTransitionTime":"2025-12-03T14:10:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 14:10:47 crc kubenswrapper[4805]: I1203 14:10:47.883022 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 14:10:47 crc kubenswrapper[4805]: I1203 14:10:47.883076 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 14:10:47 crc kubenswrapper[4805]: I1203 14:10:47.883093 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 14:10:47 crc kubenswrapper[4805]: I1203 14:10:47.883117 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 14:10:47 crc kubenswrapper[4805]: I1203 14:10:47.883134 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T14:10:47Z","lastTransitionTime":"2025-12-03T14:10:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 14:10:47 crc kubenswrapper[4805]: I1203 14:10:47.986545 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 14:10:47 crc kubenswrapper[4805]: I1203 14:10:47.986621 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 14:10:47 crc kubenswrapper[4805]: I1203 14:10:47.986646 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 14:10:47 crc kubenswrapper[4805]: I1203 14:10:47.986676 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 14:10:47 crc kubenswrapper[4805]: I1203 14:10:47.986698 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T14:10:47Z","lastTransitionTime":"2025-12-03T14:10:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 14:10:48 crc kubenswrapper[4805]: I1203 14:10:48.090074 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 14:10:48 crc kubenswrapper[4805]: I1203 14:10:48.090118 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 14:10:48 crc kubenswrapper[4805]: I1203 14:10:48.090131 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 14:10:48 crc kubenswrapper[4805]: I1203 14:10:48.090149 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 14:10:48 crc kubenswrapper[4805]: I1203 14:10:48.090160 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T14:10:48Z","lastTransitionTime":"2025-12-03T14:10:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 14:10:48 crc kubenswrapper[4805]: I1203 14:10:48.135054 4805 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-psjpm_deeec30d-cad2-4dfd-adb6-64f6646581be/ovnkube-controller/2.log" Dec 03 14:10:48 crc kubenswrapper[4805]: I1203 14:10:48.139329 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-psjpm" event={"ID":"deeec30d-cad2-4dfd-adb6-64f6646581be","Type":"ContainerStarted","Data":"8f75988e25825e00abdb9e4b2292c5cc8ccab3dd028c882c2a38c796c0fd19c5"} Dec 03 14:10:48 crc kubenswrapper[4805]: I1203 14:10:48.139904 4805 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-psjpm" Dec 03 14:10:48 crc kubenswrapper[4805]: I1203 14:10:48.158078 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-psjpm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"deeec30d-cad2-4dfd-adb6-64f6646581be\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:50Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:50Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4b8d00cdf0f37f2e5cd4e952cf85d3de8d72c43fdf786088fd9f26ed17b1e21f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wdlzx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bd67c6de01a8d196a1f51d4d90640c9e69de6f05d5e201b25e6eaa5c33440ea8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wdlzx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cb8bf05ef093eb0727a066d4bbbdf9af9422867cd4dadacafc6eeb016867a61e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wdlzx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0304467102ed053628c0d3e06548c74b6492336b6fd044246508d1e8bdb17bc3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wdlzx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6ce80ccafb3262a11fbe9923317a9d38b61e5dc5b3e7e7ae136f143d6b643fd7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wdlzx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://47d76962beb8c0f196d50af96fb3ec611097823ca80a1c8bd718f2d18840128c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wdlzx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8f75988e25825e00abdb9e4b2292c5cc8ccab3dd028c882c2a38c796c0fd19c5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a0272667e733f2c5eec0fce0eb9c31a16c048babdaed90635895d689ae9303f4\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-03T14:10:16Z\\\",\\\"message\\\":\\\"o create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T14:10:16Z is after 2025-08-24T17:21:41Z]\\\\nI1203 14:10:16.717659 6478 services_controller.go:473] Services do not match for network=default, existing lbs: []services.LB{services.LB{Name:\\\\\\\"Service_openshift-network-console/networking-console-plugin_TCP_cluster\\\\\\\", UUID:\\\\\\\"ab0b1d51-5ec6-479b-8881-93dfa8d30337\\\\\\\", Protocol:\\\\\\\"tcp\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-network-console/networking-console-plugin\\\\\\\"}, Opts:services.LBOpts{Reject:false, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{}, Templates:services.TemplateMap{}, Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGr\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T14:10:15Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:10:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wdlzx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ab98e365c1a4242a47b8cbe2dd85109d366d6002c4de1f18ddc950a46bea6c12\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wdlzx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2f875252fbe6a673a72fc9c56bb96899f62732305ad77b3108e3b244bfa91ecb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2f875252fbe6a673a72fc9c56bb96899f62732305ad77b3108e3b244bfa91ecb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T14:09:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T14:09:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wdlzx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T14:09:50Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-psjpm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T14:10:48Z is after 2025-08-24T17:21:41Z" Dec 03 14:10:48 crc kubenswrapper[4805]: I1203 14:10:48.168580 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-4gd9r" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6b36a3db-1047-487a-9d6c-4f51ffdd8cc1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:10:04Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:10:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:10:04Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:10:04Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4t6zl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4t6zl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T14:10:04Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-4gd9r\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T14:10:48Z is after 2025-08-24T17:21:41Z" Dec 03 14:10:48 crc kubenswrapper[4805]: I1203 14:10:48.186794 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:48Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:48Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T14:10:48Z is after 2025-08-24T17:21:41Z" Dec 03 14:10:48 crc kubenswrapper[4805]: I1203 14:10:48.193954 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 14:10:48 crc kubenswrapper[4805]: I1203 14:10:48.194008 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 14:10:48 crc kubenswrapper[4805]: I1203 14:10:48.194020 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 14:10:48 crc kubenswrapper[4805]: I1203 14:10:48.194041 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 14:10:48 crc kubenswrapper[4805]: I1203 14:10:48.194054 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T14:10:48Z","lastTransitionTime":"2025-12-03T14:10:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 14:10:48 crc kubenswrapper[4805]: I1203 14:10:48.204481 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:49Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:49Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://541ace7ccbb66dbd191c3ca8c1e4479e3edbf80dfc67326bea8aaee81fdcfc3c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://821e3e9a902aacecf6c41e8b8d8aa1ac951ce2fbb6505b7148b16348a3754e55\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T14:10:48Z is after 2025-08-24T17:21:41Z" Dec 03 14:10:48 crc kubenswrapper[4805]: I1203 14:10:48.220390 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-gskh4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6098937f-e3f6-45e8-a647-4994a79cd711\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://28d5653f26ec61647808b3a733ff7eac70aa69ac184660f24147e8cfaf0eb802\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nwrw9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9ad6b68a42fd33ee167e67873c6eed3f11cdf6dc71d22e952c021436eb967dd9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nwrw9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T14:09:50Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-gskh4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T14:10:48Z is after 2025-08-24T17:21:41Z" Dec 03 14:10:48 crc kubenswrapper[4805]: I1203 14:10:48.234903 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-pdzgk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"860769de-6a5b-4cb9-a752-264e7e2e785f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5e2f0a0cc21505d609f81918365d5e4f5d21959ac12ab147d6a51601c9de4c14\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ggn4n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T14:09:50Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-pdzgk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T14:10:48Z is after 2025-08-24T17:21:41Z" Dec 03 14:10:48 crc kubenswrapper[4805]: I1203 14:10:48.251968 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-69h75" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5d1c694f-9847-4086-8b37-3573ddeb3ace\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:10:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:10:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:10:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:10:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://abc1d9b8582ca9fe4a51fb608fb39129a179ddcdfde0797a6369a55b2e0124cf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:10:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s578b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c7c4d434b035cac2a91a4f0846a2f4fb5bd68f881c28487860adb6db85e8ce75\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:10:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s578b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T14:10:03Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-69h75\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T14:10:48Z is after 2025-08-24T17:21:41Z" Dec 03 14:10:48 crc kubenswrapper[4805]: I1203 14:10:48.267413 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:49Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:49Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://38339331a3209d6787d1def81b72f4a8514bf91080a139cad6d8ac5cc6bef10e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T14:10:48Z is after 2025-08-24T17:21:41Z" Dec 03 14:10:48 crc kubenswrapper[4805]: I1203 14:10:48.283505 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:48Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:48Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T14:10:48Z is after 2025-08-24T17:21:41Z" Dec 03 14:10:48 crc kubenswrapper[4805]: I1203 14:10:48.295984 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 14:10:48 crc kubenswrapper[4805]: I1203 14:10:48.296039 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 14:10:48 crc kubenswrapper[4805]: I1203 14:10:48.296049 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 14:10:48 crc kubenswrapper[4805]: I1203 14:10:48.296066 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 14:10:48 crc kubenswrapper[4805]: I1203 14:10:48.296078 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T14:10:48Z","lastTransitionTime":"2025-12-03T14:10:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 14:10:48 crc kubenswrapper[4805]: I1203 14:10:48.298482 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:51Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:51Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b3123d3fa879097db88ed2f3cfb05631f440c24c5405503b2de3546f12e2e899\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T14:10:48Z is after 2025-08-24T17:21:41Z" Dec 03 14:10:48 crc kubenswrapper[4805]: I1203 14:10:48.310900 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-bdtdg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"53a0e9e3-b6d8-44b7-a32b-aa3450442865\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f557e17ad105d6972d3c1480656c6073f67e28c0f5773aa365549951f4d161c7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9ww54\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T14:09:51Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-bdtdg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T14:10:48Z is after 2025-08-24T17:21:41Z" Dec 03 14:10:48 crc kubenswrapper[4805]: I1203 14:10:48.335096 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b6d36671-7173-4e9c-843f-005b4dcf576c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a837d7d491eaacbe4279b4014bb5fcbe9b514ba5cc82c708227bcd9b25a346e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6e5f04faf3b23197bb8ea11d3e574dc86f31e21d798734962001ec8c95de7c84\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2ff1a979387da41b00e95eb9cc8b2c9de966380a12fae687591df3c5b18a4497\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1e66e1cb18573a14e3c3ae16af4dbd9f8997a6ad38680c3ab4505646ebd48e7c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7d5b5c85e3de43e7e95111aa967248550607c26200c783251b68696f202f92a1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://474f72dac37a808fc4b9832306e76979b151d7221807ff3086013466ed1419c3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://474f72dac37a808fc4b9832306e76979b151d7221807ff3086013466ed1419c3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T14:09:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T14:09:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://74c8cea1a61c4775a0ad2f4ab4bb8a36437cf23d2ee953311d6ea967e297510a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://74c8cea1a61c4775a0ad2f4ab4bb8a36437cf23d2ee953311d6ea967e297510a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T14:09:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T14:09:31Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://a4f2f56a809adc01fba82eba0652d78f58770c132f77b1424f8b17068b1927e3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a4f2f56a809adc01fba82eba0652d78f58770c132f77b1424f8b17068b1927e3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T14:09:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T14:09:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T14:09:30Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T14:10:48Z is after 2025-08-24T17:21:41Z" Dec 03 14:10:48 crc kubenswrapper[4805]: I1203 14:10:48.352157 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ced8291d-0ce5-4e74-9eaf-2bf033774039\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:10:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:10:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3e5a34636f0a8c4d9179074596d28bc2eb7a504573398c596a637defd829907a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://29cd95be3f964415943c873e7914ffcea4abe50dfdf17f8a9cd03fc38df35981\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://951173b3d3d581275a6f6e43209b5e0611a701ca7a10f81504d723e24a2d21a3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e64730957bbbb2feb4d506e0ed9731a0594bf66fa1e8fcdb8ec944ded6308eed\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5da55f769e23ad52bac9c77da62c78e70fa34db8ed1a435e4eaf4dad3fd2b68a\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-03T14:09:48Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1203 14:09:42.992179 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1203 14:09:42.993233 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1830926909/tls.crt::/tmp/serving-cert-1830926909/tls.key\\\\\\\"\\\\nI1203 14:09:48.622403 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1203 14:09:48.626476 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1203 14:09:48.626527 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1203 14:09:48.626574 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1203 14:09:48.626582 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1203 14:09:48.633990 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1203 14:09:48.634024 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 14:09:48.634029 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 14:09:48.634034 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1203 14:09:48.634038 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1203 14:09:48.634042 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1203 14:09:48.634045 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1203 14:09:48.634344 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1203 14:09:48.644497 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T14:09:32Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bd92a160eb530c5687ebd7cd5b83b41820a2b808a0fd9514e5ace9d237e1bd26\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:32Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://af9c940135012c629447528c61743ed1e58708ab57ce76604c6da0479aa459cf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://af9c940135012c629447528c61743ed1e58708ab57ce76604c6da0479aa459cf\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T14:09:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T14:09:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T14:09:30Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T14:10:48Z is after 2025-08-24T17:21:41Z" Dec 03 14:10:48 crc kubenswrapper[4805]: I1203 14:10:48.365963 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6d48b4f1-5dfc-4f6b-a4e8-4e465c651c7b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:10:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:10:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5ea698a1e0de3b5201e2f9032d834d81d9aab9c2fb17aca582184c7090f1bd2f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://66aa257562f8e113fc1cd2936b8aa18e927f90018ad4c11a184c1c56444be648\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2fd6d50f839b603a90e5d00ab6e36e832eab2e299b7af1dbb84498c35bf1c08e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://77c26dae6ccd328a8a1ce4fe6ffe94c78061d8d3ac3727adb80a24c16f4ab533\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://77c26dae6ccd328a8a1ce4fe6ffe94c78061d8d3ac3727adb80a24c16f4ab533\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T14:09:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T14:09:31Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T14:09:30Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T14:10:48Z is after 2025-08-24T17:21:41Z" Dec 03 14:10:48 crc kubenswrapper[4805]: I1203 14:10:48.387577 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-bpk49" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"de00a81f-82e1-4223-9614-f98606b9a8ef\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://04f4ad27e02e2024ae9e7cd1b16116267e69cecda882dd723cdb274c41a0e7db\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-25d7t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7741e0e70416ebe094bfabc6009e98d7d4c0fbcba1fc82df13b9ba40774b456a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7741e0e70416ebe094bfabc6009e98d7d4c0fbcba1fc82df13b9ba40774b456a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T14:09:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T14:09:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-25d7t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ee3499c02eb52bbf08c447d2433d435a74dfb70670dc4cbaac44f5fbff9e8a29\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ee3499c02eb52bbf08c447d2433d435a74dfb70670dc4cbaac44f5fbff9e8a29\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T14:09:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T14:09:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-25d7t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ec8488b1d69799951be7ef40e5e028732265c4ddcdea2d453923e567864771f7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ec8488b1d69799951be7ef40e5e028732265c4ddcdea2d453923e567864771f7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T14:09:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T14:09:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-25d7t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://66707bb4e1fdacb265e89786df8a04605830f1c156394cacfcec9b54abe5d41a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://66707bb4e1fdacb265e89786df8a04605830f1c156394cacfcec9b54abe5d41a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T14:09:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T14:09:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-25d7t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6207ab6600970ff15c8713c547d4920b2c77eb9bea990bb5da497e24870dcc0f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6207ab6600970ff15c8713c547d4920b2c77eb9bea990bb5da497e24870dcc0f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T14:09:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T14:09:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-25d7t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8de65d4a205e2deec0b433e53c0e37feb94aa0ede6a4c58d7414b80829804505\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8de65d4a205e2deec0b433e53c0e37feb94aa0ede6a4c58d7414b80829804505\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T14:09:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T14:09:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-25d7t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T14:09:50Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-bpk49\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T14:10:48Z is after 2025-08-24T17:21:41Z" Dec 03 14:10:48 crc kubenswrapper[4805]: I1203 14:10:48.398391 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 14:10:48 crc kubenswrapper[4805]: I1203 14:10:48.398468 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 14:10:48 crc kubenswrapper[4805]: I1203 14:10:48.398487 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 14:10:48 crc kubenswrapper[4805]: I1203 14:10:48.398515 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 14:10:48 crc kubenswrapper[4805]: I1203 14:10:48.398537 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T14:10:48Z","lastTransitionTime":"2025-12-03T14:10:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 14:10:48 crc kubenswrapper[4805]: I1203 14:10:48.406937 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"750d2b4b-738a-4992-8902-742528f6e875\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://abd2d0156113bb5182727a53aac594410ed051c19258beb8a07b808eff0b42a7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8673a29c0e16f4e707d79f0610722ccabb023ff2b4a18d7cf9d305183817d952\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://35766e7206b1be7a53d13d6890d759cf81dbd999643c12349268e4bf16fbbdb0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5b9844a11b2b903978d0d464b469e26c9211100da61d018dde9636dca95b684e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T14:09:30Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T14:10:48Z is after 2025-08-24T17:21:41Z" Dec 03 14:10:48 crc kubenswrapper[4805]: I1203 14:10:48.423793 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:48Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:48Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T14:10:48Z is after 2025-08-24T17:21:41Z" Dec 03 14:10:48 crc kubenswrapper[4805]: I1203 14:10:48.437369 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-k4mhs" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c09c1cdc-f461-45c9-8444-5b8764bccabc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:10:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:10:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7e077e19b6a5bf7d29b357982e12224d4b7058e0f9ac81b66b5a665c9bbf5014\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3354c1a44f891182d04116fc76234a2b337d50567d032a3584f902a9516387cc\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-03T14:10:36Z\\\",\\\"message\\\":\\\"2025-12-03T14:09:51+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_3e51ab9a-fbbe-4a78-9308-8a81e203fcad\\\\n2025-12-03T14:09:51+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_3e51ab9a-fbbe-4a78-9308-8a81e203fcad to /host/opt/cni/bin/\\\\n2025-12-03T14:09:51Z [verbose] multus-daemon started\\\\n2025-12-03T14:09:51Z [verbose] Readiness Indicator file check\\\\n2025-12-03T14:10:36Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T14:09:50Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:10:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qdcd7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T14:09:50Z\\\"}}\" for pod \"openshift-multus\"/\"multus-k4mhs\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T14:10:48Z is after 2025-08-24T17:21:41Z" Dec 03 14:10:48 crc kubenswrapper[4805]: I1203 14:10:48.501534 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 14:10:48 crc kubenswrapper[4805]: I1203 14:10:48.501575 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 14:10:48 crc kubenswrapper[4805]: I1203 14:10:48.501586 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 14:10:48 crc kubenswrapper[4805]: I1203 14:10:48.501604 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 14:10:48 crc kubenswrapper[4805]: I1203 14:10:48.501616 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T14:10:48Z","lastTransitionTime":"2025-12-03T14:10:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 14:10:48 crc kubenswrapper[4805]: I1203 14:10:48.603876 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 14:10:48 crc kubenswrapper[4805]: I1203 14:10:48.603965 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 14:10:48 crc kubenswrapper[4805]: I1203 14:10:48.604023 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 14:10:48 crc kubenswrapper[4805]: I1203 14:10:48.604058 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 14:10:48 crc kubenswrapper[4805]: I1203 14:10:48.604080 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T14:10:48Z","lastTransitionTime":"2025-12-03T14:10:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 14:10:48 crc kubenswrapper[4805]: I1203 14:10:48.693964 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-4gd9r" Dec 03 14:10:48 crc kubenswrapper[4805]: I1203 14:10:48.693963 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 14:10:48 crc kubenswrapper[4805]: I1203 14:10:48.694261 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 14:10:48 crc kubenswrapper[4805]: E1203 14:10:48.694412 4805 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-4gd9r" podUID="6b36a3db-1047-487a-9d6c-4f51ffdd8cc1" Dec 03 14:10:48 crc kubenswrapper[4805]: I1203 14:10:48.694531 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 14:10:48 crc kubenswrapper[4805]: E1203 14:10:48.694652 4805 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 03 14:10:48 crc kubenswrapper[4805]: E1203 14:10:48.694722 4805 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 03 14:10:48 crc kubenswrapper[4805]: E1203 14:10:48.694756 4805 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 03 14:10:48 crc kubenswrapper[4805]: I1203 14:10:48.707299 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 14:10:48 crc kubenswrapper[4805]: I1203 14:10:48.707429 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 14:10:48 crc kubenswrapper[4805]: I1203 14:10:48.707506 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 14:10:48 crc kubenswrapper[4805]: I1203 14:10:48.707539 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 14:10:48 crc kubenswrapper[4805]: I1203 14:10:48.707616 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T14:10:48Z","lastTransitionTime":"2025-12-03T14:10:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 14:10:48 crc kubenswrapper[4805]: I1203 14:10:48.810284 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 14:10:48 crc kubenswrapper[4805]: I1203 14:10:48.810341 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 14:10:48 crc kubenswrapper[4805]: I1203 14:10:48.810351 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 14:10:48 crc kubenswrapper[4805]: I1203 14:10:48.810374 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 14:10:48 crc kubenswrapper[4805]: I1203 14:10:48.810389 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T14:10:48Z","lastTransitionTime":"2025-12-03T14:10:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 14:10:48 crc kubenswrapper[4805]: I1203 14:10:48.912105 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 14:10:48 crc kubenswrapper[4805]: I1203 14:10:48.912148 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 14:10:48 crc kubenswrapper[4805]: I1203 14:10:48.912159 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 14:10:48 crc kubenswrapper[4805]: I1203 14:10:48.912174 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 14:10:48 crc kubenswrapper[4805]: I1203 14:10:48.912186 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T14:10:48Z","lastTransitionTime":"2025-12-03T14:10:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 14:10:49 crc kubenswrapper[4805]: I1203 14:10:49.015292 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 14:10:49 crc kubenswrapper[4805]: I1203 14:10:49.015377 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 14:10:49 crc kubenswrapper[4805]: I1203 14:10:49.015396 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 14:10:49 crc kubenswrapper[4805]: I1203 14:10:49.015418 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 14:10:49 crc kubenswrapper[4805]: I1203 14:10:49.015430 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T14:10:49Z","lastTransitionTime":"2025-12-03T14:10:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 14:10:49 crc kubenswrapper[4805]: I1203 14:10:49.118546 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 14:10:49 crc kubenswrapper[4805]: I1203 14:10:49.118592 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 14:10:49 crc kubenswrapper[4805]: I1203 14:10:49.118603 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 14:10:49 crc kubenswrapper[4805]: I1203 14:10:49.118622 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 14:10:49 crc kubenswrapper[4805]: I1203 14:10:49.118635 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T14:10:49Z","lastTransitionTime":"2025-12-03T14:10:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 14:10:49 crc kubenswrapper[4805]: I1203 14:10:49.220466 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 14:10:49 crc kubenswrapper[4805]: I1203 14:10:49.220512 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 14:10:49 crc kubenswrapper[4805]: I1203 14:10:49.220527 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 14:10:49 crc kubenswrapper[4805]: I1203 14:10:49.220544 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 14:10:49 crc kubenswrapper[4805]: I1203 14:10:49.220555 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T14:10:49Z","lastTransitionTime":"2025-12-03T14:10:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 14:10:49 crc kubenswrapper[4805]: I1203 14:10:49.323243 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 14:10:49 crc kubenswrapper[4805]: I1203 14:10:49.323278 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 14:10:49 crc kubenswrapper[4805]: I1203 14:10:49.323286 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 14:10:49 crc kubenswrapper[4805]: I1203 14:10:49.323301 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 14:10:49 crc kubenswrapper[4805]: I1203 14:10:49.323309 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T14:10:49Z","lastTransitionTime":"2025-12-03T14:10:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 14:10:49 crc kubenswrapper[4805]: I1203 14:10:49.425387 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 14:10:49 crc kubenswrapper[4805]: I1203 14:10:49.425452 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 14:10:49 crc kubenswrapper[4805]: I1203 14:10:49.425470 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 14:10:49 crc kubenswrapper[4805]: I1203 14:10:49.425485 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 14:10:49 crc kubenswrapper[4805]: I1203 14:10:49.425495 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T14:10:49Z","lastTransitionTime":"2025-12-03T14:10:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 14:10:49 crc kubenswrapper[4805]: I1203 14:10:49.528009 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 14:10:49 crc kubenswrapper[4805]: I1203 14:10:49.528074 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 14:10:49 crc kubenswrapper[4805]: I1203 14:10:49.528084 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 14:10:49 crc kubenswrapper[4805]: I1203 14:10:49.528099 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 14:10:49 crc kubenswrapper[4805]: I1203 14:10:49.528111 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T14:10:49Z","lastTransitionTime":"2025-12-03T14:10:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 14:10:49 crc kubenswrapper[4805]: I1203 14:10:49.630988 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 14:10:49 crc kubenswrapper[4805]: I1203 14:10:49.631058 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 14:10:49 crc kubenswrapper[4805]: I1203 14:10:49.631075 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 14:10:49 crc kubenswrapper[4805]: I1203 14:10:49.631101 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 14:10:49 crc kubenswrapper[4805]: I1203 14:10:49.631125 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T14:10:49Z","lastTransitionTime":"2025-12-03T14:10:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 14:10:49 crc kubenswrapper[4805]: I1203 14:10:49.733880 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 14:10:49 crc kubenswrapper[4805]: I1203 14:10:49.734260 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 14:10:49 crc kubenswrapper[4805]: I1203 14:10:49.734270 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 14:10:49 crc kubenswrapper[4805]: I1203 14:10:49.734285 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 14:10:49 crc kubenswrapper[4805]: I1203 14:10:49.734297 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T14:10:49Z","lastTransitionTime":"2025-12-03T14:10:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 14:10:49 crc kubenswrapper[4805]: I1203 14:10:49.837328 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 14:10:49 crc kubenswrapper[4805]: I1203 14:10:49.837414 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 14:10:49 crc kubenswrapper[4805]: I1203 14:10:49.837440 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 14:10:49 crc kubenswrapper[4805]: I1203 14:10:49.837468 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 14:10:49 crc kubenswrapper[4805]: I1203 14:10:49.837490 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T14:10:49Z","lastTransitionTime":"2025-12-03T14:10:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 14:10:49 crc kubenswrapper[4805]: I1203 14:10:49.941338 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 14:10:49 crc kubenswrapper[4805]: I1203 14:10:49.941411 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 14:10:49 crc kubenswrapper[4805]: I1203 14:10:49.941429 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 14:10:49 crc kubenswrapper[4805]: I1203 14:10:49.941454 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 14:10:49 crc kubenswrapper[4805]: I1203 14:10:49.941475 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T14:10:49Z","lastTransitionTime":"2025-12-03T14:10:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 14:10:50 crc kubenswrapper[4805]: I1203 14:10:50.043874 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 14:10:50 crc kubenswrapper[4805]: I1203 14:10:50.043934 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 14:10:50 crc kubenswrapper[4805]: I1203 14:10:50.043949 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 14:10:50 crc kubenswrapper[4805]: I1203 14:10:50.043973 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 14:10:50 crc kubenswrapper[4805]: I1203 14:10:50.043989 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T14:10:50Z","lastTransitionTime":"2025-12-03T14:10:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 14:10:50 crc kubenswrapper[4805]: I1203 14:10:50.145894 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 14:10:50 crc kubenswrapper[4805]: I1203 14:10:50.145957 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 14:10:50 crc kubenswrapper[4805]: I1203 14:10:50.146001 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 14:10:50 crc kubenswrapper[4805]: I1203 14:10:50.146026 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 14:10:50 crc kubenswrapper[4805]: I1203 14:10:50.146042 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T14:10:50Z","lastTransitionTime":"2025-12-03T14:10:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 14:10:50 crc kubenswrapper[4805]: I1203 14:10:50.147750 4805 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-psjpm_deeec30d-cad2-4dfd-adb6-64f6646581be/ovnkube-controller/3.log" Dec 03 14:10:50 crc kubenswrapper[4805]: I1203 14:10:50.148602 4805 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-psjpm_deeec30d-cad2-4dfd-adb6-64f6646581be/ovnkube-controller/2.log" Dec 03 14:10:50 crc kubenswrapper[4805]: I1203 14:10:50.152647 4805 generic.go:334] "Generic (PLEG): container finished" podID="deeec30d-cad2-4dfd-adb6-64f6646581be" containerID="8f75988e25825e00abdb9e4b2292c5cc8ccab3dd028c882c2a38c796c0fd19c5" exitCode=1 Dec 03 14:10:50 crc kubenswrapper[4805]: I1203 14:10:50.152724 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-psjpm" event={"ID":"deeec30d-cad2-4dfd-adb6-64f6646581be","Type":"ContainerDied","Data":"8f75988e25825e00abdb9e4b2292c5cc8ccab3dd028c882c2a38c796c0fd19c5"} Dec 03 14:10:50 crc kubenswrapper[4805]: I1203 14:10:50.152785 4805 scope.go:117] "RemoveContainer" containerID="a0272667e733f2c5eec0fce0eb9c31a16c048babdaed90635895d689ae9303f4" Dec 03 14:10:50 crc kubenswrapper[4805]: I1203 14:10:50.153786 4805 scope.go:117] "RemoveContainer" containerID="8f75988e25825e00abdb9e4b2292c5cc8ccab3dd028c882c2a38c796c0fd19c5" Dec 03 14:10:50 crc kubenswrapper[4805]: E1203 14:10:50.154042 4805 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-psjpm_openshift-ovn-kubernetes(deeec30d-cad2-4dfd-adb6-64f6646581be)\"" pod="openshift-ovn-kubernetes/ovnkube-node-psjpm" podUID="deeec30d-cad2-4dfd-adb6-64f6646581be" Dec 03 14:10:50 crc kubenswrapper[4805]: I1203 14:10:50.168693 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"750d2b4b-738a-4992-8902-742528f6e875\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://abd2d0156113bb5182727a53aac594410ed051c19258beb8a07b808eff0b42a7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8673a29c0e16f4e707d79f0610722ccabb023ff2b4a18d7cf9d305183817d952\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://35766e7206b1be7a53d13d6890d759cf81dbd999643c12349268e4bf16fbbdb0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5b9844a11b2b903978d0d464b469e26c9211100da61d018dde9636dca95b684e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T14:09:30Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T14:10:50Z is after 2025-08-24T17:21:41Z" Dec 03 14:10:50 crc kubenswrapper[4805]: I1203 14:10:50.181521 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:48Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:48Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T14:10:50Z is after 2025-08-24T17:21:41Z" Dec 03 14:10:50 crc kubenswrapper[4805]: I1203 14:10:50.199042 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-k4mhs" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c09c1cdc-f461-45c9-8444-5b8764bccabc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:10:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:10:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7e077e19b6a5bf7d29b357982e12224d4b7058e0f9ac81b66b5a665c9bbf5014\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3354c1a44f891182d04116fc76234a2b337d50567d032a3584f902a9516387cc\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-03T14:10:36Z\\\",\\\"message\\\":\\\"2025-12-03T14:09:51+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_3e51ab9a-fbbe-4a78-9308-8a81e203fcad\\\\n2025-12-03T14:09:51+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_3e51ab9a-fbbe-4a78-9308-8a81e203fcad to /host/opt/cni/bin/\\\\n2025-12-03T14:09:51Z [verbose] multus-daemon started\\\\n2025-12-03T14:09:51Z [verbose] Readiness Indicator file check\\\\n2025-12-03T14:10:36Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T14:09:50Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:10:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qdcd7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T14:09:50Z\\\"}}\" for pod \"openshift-multus\"/\"multus-k4mhs\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T14:10:50Z is after 2025-08-24T17:21:41Z" Dec 03 14:10:50 crc kubenswrapper[4805]: I1203 14:10:50.219659 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-bpk49" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"de00a81f-82e1-4223-9614-f98606b9a8ef\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://04f4ad27e02e2024ae9e7cd1b16116267e69cecda882dd723cdb274c41a0e7db\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-25d7t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7741e0e70416ebe094bfabc6009e98d7d4c0fbcba1fc82df13b9ba40774b456a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7741e0e70416ebe094bfabc6009e98d7d4c0fbcba1fc82df13b9ba40774b456a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T14:09:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T14:09:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-25d7t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ee3499c02eb52bbf08c447d2433d435a74dfb70670dc4cbaac44f5fbff9e8a29\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ee3499c02eb52bbf08c447d2433d435a74dfb70670dc4cbaac44f5fbff9e8a29\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T14:09:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T14:09:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-25d7t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ec8488b1d69799951be7ef40e5e028732265c4ddcdea2d453923e567864771f7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ec8488b1d69799951be7ef40e5e028732265c4ddcdea2d453923e567864771f7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T14:09:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T14:09:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-25d7t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://66707bb4e1fdacb265e89786df8a04605830f1c156394cacfcec9b54abe5d41a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://66707bb4e1fdacb265e89786df8a04605830f1c156394cacfcec9b54abe5d41a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T14:09:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T14:09:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-25d7t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6207ab6600970ff15c8713c547d4920b2c77eb9bea990bb5da497e24870dcc0f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6207ab6600970ff15c8713c547d4920b2c77eb9bea990bb5da497e24870dcc0f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T14:09:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T14:09:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-25d7t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8de65d4a205e2deec0b433e53c0e37feb94aa0ede6a4c58d7414b80829804505\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8de65d4a205e2deec0b433e53c0e37feb94aa0ede6a4c58d7414b80829804505\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T14:09:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T14:09:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-25d7t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T14:09:50Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-bpk49\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T14:10:50Z is after 2025-08-24T17:21:41Z" Dec 03 14:10:50 crc kubenswrapper[4805]: I1203 14:10:50.236363 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:48Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:48Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T14:10:50Z is after 2025-08-24T17:21:41Z" Dec 03 14:10:50 crc kubenswrapper[4805]: I1203 14:10:50.248366 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 14:10:50 crc kubenswrapper[4805]: I1203 14:10:50.248442 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 14:10:50 crc kubenswrapper[4805]: I1203 14:10:50.248466 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 14:10:50 crc kubenswrapper[4805]: I1203 14:10:50.248498 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 14:10:50 crc kubenswrapper[4805]: I1203 14:10:50.248521 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T14:10:50Z","lastTransitionTime":"2025-12-03T14:10:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 14:10:50 crc kubenswrapper[4805]: I1203 14:10:50.253743 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:49Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:49Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://541ace7ccbb66dbd191c3ca8c1e4479e3edbf80dfc67326bea8aaee81fdcfc3c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://821e3e9a902aacecf6c41e8b8d8aa1ac951ce2fbb6505b7148b16348a3754e55\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T14:10:50Z is after 2025-08-24T17:21:41Z" Dec 03 14:10:50 crc kubenswrapper[4805]: I1203 14:10:50.273093 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-gskh4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6098937f-e3f6-45e8-a647-4994a79cd711\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://28d5653f26ec61647808b3a733ff7eac70aa69ac184660f24147e8cfaf0eb802\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nwrw9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9ad6b68a42fd33ee167e67873c6eed3f11cdf6dc71d22e952c021436eb967dd9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nwrw9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T14:09:50Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-gskh4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T14:10:50Z is after 2025-08-24T17:21:41Z" Dec 03 14:10:50 crc kubenswrapper[4805]: I1203 14:10:50.309051 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-psjpm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"deeec30d-cad2-4dfd-adb6-64f6646581be\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:50Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:50Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4b8d00cdf0f37f2e5cd4e952cf85d3de8d72c43fdf786088fd9f26ed17b1e21f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wdlzx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bd67c6de01a8d196a1f51d4d90640c9e69de6f05d5e201b25e6eaa5c33440ea8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wdlzx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cb8bf05ef093eb0727a066d4bbbdf9af9422867cd4dadacafc6eeb016867a61e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wdlzx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0304467102ed053628c0d3e06548c74b6492336b6fd044246508d1e8bdb17bc3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wdlzx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6ce80ccafb3262a11fbe9923317a9d38b61e5dc5b3e7e7ae136f143d6b643fd7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wdlzx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://47d76962beb8c0f196d50af96fb3ec611097823ca80a1c8bd718f2d18840128c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wdlzx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8f75988e25825e00abdb9e4b2292c5cc8ccab3dd028c882c2a38c796c0fd19c5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a0272667e733f2c5eec0fce0eb9c31a16c048babdaed90635895d689ae9303f4\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-03T14:10:16Z\\\",\\\"message\\\":\\\"o create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T14:10:16Z is after 2025-08-24T17:21:41Z]\\\\nI1203 14:10:16.717659 6478 services_controller.go:473] Services do not match for network=default, existing lbs: []services.LB{services.LB{Name:\\\\\\\"Service_openshift-network-console/networking-console-plugin_TCP_cluster\\\\\\\", UUID:\\\\\\\"ab0b1d51-5ec6-479b-8881-93dfa8d30337\\\\\\\", Protocol:\\\\\\\"tcp\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-network-console/networking-console-plugin\\\\\\\"}, Opts:services.LBOpts{Reject:false, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{}, Templates:services.TemplateMap{}, Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGr\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T14:10:15Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8f75988e25825e00abdb9e4b2292c5cc8ccab3dd028c882c2a38c796c0fd19c5\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-03T14:10:49Z\\\",\\\"message\\\":\\\"shift-kube-storage-version-migrator-operator/metrics]} name:Service_openshift-kube-storage-version-migrator-operator/metrics_TCP_cluster options:{GoMap:map[event:false hairpin_snat_ip:169.254.0.5 fd69::5 neighbor_responder:none reject:true skip_snat:false]} protocol:{GoSet:[tcp]} selection_fields:{GoSet:[]} vips:{GoMap:map[10.217.5.36:443:]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {7f9b8f25-db1a-4d02-a423-9afc5c2fb83c}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nF1203 14:10:49.548775 6841 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T14:10:49Z is after 2025-08-24T17:\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T14:10:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wdlzx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ab98e365c1a4242a47b8cbe2dd85109d366d6002c4de1f18ddc950a46bea6c12\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wdlzx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2f875252fbe6a673a72fc9c56bb96899f62732305ad77b3108e3b244bfa91ecb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2f875252fbe6a673a72fc9c56bb96899f62732305ad77b3108e3b244bfa91ecb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T14:09:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T14:09:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wdlzx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T14:09:50Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-psjpm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T14:10:50Z is after 2025-08-24T17:21:41Z" Dec 03 14:10:50 crc kubenswrapper[4805]: I1203 14:10:50.326360 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-4gd9r" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6b36a3db-1047-487a-9d6c-4f51ffdd8cc1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:10:04Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:10:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:10:04Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:10:04Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4t6zl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4t6zl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T14:10:04Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-4gd9r\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T14:10:50Z is after 2025-08-24T17:21:41Z" Dec 03 14:10:50 crc kubenswrapper[4805]: I1203 14:10:50.342492 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-pdzgk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"860769de-6a5b-4cb9-a752-264e7e2e785f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5e2f0a0cc21505d609f81918365d5e4f5d21959ac12ab147d6a51601c9de4c14\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ggn4n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T14:09:50Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-pdzgk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T14:10:50Z is after 2025-08-24T17:21:41Z" Dec 03 14:10:50 crc kubenswrapper[4805]: I1203 14:10:50.351933 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 14:10:50 crc kubenswrapper[4805]: I1203 14:10:50.351996 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 14:10:50 crc kubenswrapper[4805]: I1203 14:10:50.352017 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 14:10:50 crc kubenswrapper[4805]: I1203 14:10:50.352047 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 14:10:50 crc kubenswrapper[4805]: I1203 14:10:50.352068 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T14:10:50Z","lastTransitionTime":"2025-12-03T14:10:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 14:10:50 crc kubenswrapper[4805]: I1203 14:10:50.360798 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-69h75" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5d1c694f-9847-4086-8b37-3573ddeb3ace\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:10:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:10:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:10:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:10:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://abc1d9b8582ca9fe4a51fb608fb39129a179ddcdfde0797a6369a55b2e0124cf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:10:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s578b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c7c4d434b035cac2a91a4f0846a2f4fb5bd68f881c28487860adb6db85e8ce75\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:10:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s578b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T14:10:03Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-69h75\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T14:10:50Z is after 2025-08-24T17:21:41Z" Dec 03 14:10:50 crc kubenswrapper[4805]: I1203 14:10:50.377054 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:51Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:51Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b3123d3fa879097db88ed2f3cfb05631f440c24c5405503b2de3546f12e2e899\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T14:10:50Z is after 2025-08-24T17:21:41Z" Dec 03 14:10:50 crc kubenswrapper[4805]: I1203 14:10:50.392441 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-bdtdg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"53a0e9e3-b6d8-44b7-a32b-aa3450442865\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f557e17ad105d6972d3c1480656c6073f67e28c0f5773aa365549951f4d161c7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9ww54\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T14:09:51Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-bdtdg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T14:10:50Z is after 2025-08-24T17:21:41Z" Dec 03 14:10:50 crc kubenswrapper[4805]: I1203 14:10:50.424765 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b6d36671-7173-4e9c-843f-005b4dcf576c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a837d7d491eaacbe4279b4014bb5fcbe9b514ba5cc82c708227bcd9b25a346e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6e5f04faf3b23197bb8ea11d3e574dc86f31e21d798734962001ec8c95de7c84\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2ff1a979387da41b00e95eb9cc8b2c9de966380a12fae687591df3c5b18a4497\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1e66e1cb18573a14e3c3ae16af4dbd9f8997a6ad38680c3ab4505646ebd48e7c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7d5b5c85e3de43e7e95111aa967248550607c26200c783251b68696f202f92a1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://474f72dac37a808fc4b9832306e76979b151d7221807ff3086013466ed1419c3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://474f72dac37a808fc4b9832306e76979b151d7221807ff3086013466ed1419c3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T14:09:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T14:09:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://74c8cea1a61c4775a0ad2f4ab4bb8a36437cf23d2ee953311d6ea967e297510a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://74c8cea1a61c4775a0ad2f4ab4bb8a36437cf23d2ee953311d6ea967e297510a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T14:09:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T14:09:31Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://a4f2f56a809adc01fba82eba0652d78f58770c132f77b1424f8b17068b1927e3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a4f2f56a809adc01fba82eba0652d78f58770c132f77b1424f8b17068b1927e3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T14:09:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T14:09:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T14:09:30Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T14:10:50Z is after 2025-08-24T17:21:41Z" Dec 03 14:10:50 crc kubenswrapper[4805]: I1203 14:10:50.444803 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ced8291d-0ce5-4e74-9eaf-2bf033774039\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:10:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:10:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3e5a34636f0a8c4d9179074596d28bc2eb7a504573398c596a637defd829907a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://29cd95be3f964415943c873e7914ffcea4abe50dfdf17f8a9cd03fc38df35981\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://951173b3d3d581275a6f6e43209b5e0611a701ca7a10f81504d723e24a2d21a3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e64730957bbbb2feb4d506e0ed9731a0594bf66fa1e8fcdb8ec944ded6308eed\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5da55f769e23ad52bac9c77da62c78e70fa34db8ed1a435e4eaf4dad3fd2b68a\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-03T14:09:48Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1203 14:09:42.992179 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1203 14:09:42.993233 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1830926909/tls.crt::/tmp/serving-cert-1830926909/tls.key\\\\\\\"\\\\nI1203 14:09:48.622403 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1203 14:09:48.626476 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1203 14:09:48.626527 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1203 14:09:48.626574 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1203 14:09:48.626582 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1203 14:09:48.633990 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1203 14:09:48.634024 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 14:09:48.634029 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 14:09:48.634034 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1203 14:09:48.634038 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1203 14:09:48.634042 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1203 14:09:48.634045 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1203 14:09:48.634344 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1203 14:09:48.644497 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T14:09:32Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bd92a160eb530c5687ebd7cd5b83b41820a2b808a0fd9514e5ace9d237e1bd26\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:32Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://af9c940135012c629447528c61743ed1e58708ab57ce76604c6da0479aa459cf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://af9c940135012c629447528c61743ed1e58708ab57ce76604c6da0479aa459cf\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T14:09:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T14:09:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T14:09:30Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T14:10:50Z is after 2025-08-24T17:21:41Z" Dec 03 14:10:50 crc kubenswrapper[4805]: I1203 14:10:50.455498 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 14:10:50 crc kubenswrapper[4805]: I1203 14:10:50.455556 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 14:10:50 crc kubenswrapper[4805]: I1203 14:10:50.455579 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 14:10:50 crc kubenswrapper[4805]: I1203 14:10:50.455608 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 14:10:50 crc kubenswrapper[4805]: I1203 14:10:50.455630 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T14:10:50Z","lastTransitionTime":"2025-12-03T14:10:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 14:10:50 crc kubenswrapper[4805]: I1203 14:10:50.463182 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6d48b4f1-5dfc-4f6b-a4e8-4e465c651c7b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:10:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:10:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5ea698a1e0de3b5201e2f9032d834d81d9aab9c2fb17aca582184c7090f1bd2f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://66aa257562f8e113fc1cd2936b8aa18e927f90018ad4c11a184c1c56444be648\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2fd6d50f839b603a90e5d00ab6e36e832eab2e299b7af1dbb84498c35bf1c08e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://77c26dae6ccd328a8a1ce4fe6ffe94c78061d8d3ac3727adb80a24c16f4ab533\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://77c26dae6ccd328a8a1ce4fe6ffe94c78061d8d3ac3727adb80a24c16f4ab533\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T14:09:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T14:09:31Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T14:09:30Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T14:10:50Z is after 2025-08-24T17:21:41Z" Dec 03 14:10:50 crc kubenswrapper[4805]: I1203 14:10:50.481802 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:49Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:49Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://38339331a3209d6787d1def81b72f4a8514bf91080a139cad6d8ac5cc6bef10e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T14:10:50Z is after 2025-08-24T17:21:41Z" Dec 03 14:10:50 crc kubenswrapper[4805]: I1203 14:10:50.499034 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:48Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:48Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T14:10:50Z is after 2025-08-24T17:21:41Z" Dec 03 14:10:50 crc kubenswrapper[4805]: I1203 14:10:50.558299 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 14:10:50 crc kubenswrapper[4805]: I1203 14:10:50.558369 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 14:10:50 crc kubenswrapper[4805]: I1203 14:10:50.558395 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 14:10:50 crc kubenswrapper[4805]: I1203 14:10:50.558425 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 14:10:50 crc kubenswrapper[4805]: I1203 14:10:50.558442 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T14:10:50Z","lastTransitionTime":"2025-12-03T14:10:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 14:10:50 crc kubenswrapper[4805]: I1203 14:10:50.661962 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 14:10:50 crc kubenswrapper[4805]: I1203 14:10:50.662032 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 14:10:50 crc kubenswrapper[4805]: I1203 14:10:50.662049 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 14:10:50 crc kubenswrapper[4805]: I1203 14:10:50.662073 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 14:10:50 crc kubenswrapper[4805]: I1203 14:10:50.662091 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T14:10:50Z","lastTransitionTime":"2025-12-03T14:10:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 14:10:50 crc kubenswrapper[4805]: I1203 14:10:50.693993 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 14:10:50 crc kubenswrapper[4805]: I1203 14:10:50.694113 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-4gd9r" Dec 03 14:10:50 crc kubenswrapper[4805]: I1203 14:10:50.694019 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 14:10:50 crc kubenswrapper[4805]: E1203 14:10:50.694205 4805 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 03 14:10:50 crc kubenswrapper[4805]: E1203 14:10:50.694288 4805 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-4gd9r" podUID="6b36a3db-1047-487a-9d6c-4f51ffdd8cc1" Dec 03 14:10:50 crc kubenswrapper[4805]: E1203 14:10:50.694457 4805 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 03 14:10:50 crc kubenswrapper[4805]: I1203 14:10:50.694524 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 14:10:50 crc kubenswrapper[4805]: E1203 14:10:50.694685 4805 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 03 14:10:50 crc kubenswrapper[4805]: I1203 14:10:50.725680 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-psjpm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"deeec30d-cad2-4dfd-adb6-64f6646581be\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:50Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:50Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4b8d00cdf0f37f2e5cd4e952cf85d3de8d72c43fdf786088fd9f26ed17b1e21f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wdlzx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bd67c6de01a8d196a1f51d4d90640c9e69de6f05d5e201b25e6eaa5c33440ea8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wdlzx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cb8bf05ef093eb0727a066d4bbbdf9af9422867cd4dadacafc6eeb016867a61e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wdlzx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0304467102ed053628c0d3e06548c74b6492336b6fd044246508d1e8bdb17bc3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wdlzx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6ce80ccafb3262a11fbe9923317a9d38b61e5dc5b3e7e7ae136f143d6b643fd7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wdlzx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://47d76962beb8c0f196d50af96fb3ec611097823ca80a1c8bd718f2d18840128c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wdlzx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8f75988e25825e00abdb9e4b2292c5cc8ccab3dd028c882c2a38c796c0fd19c5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a0272667e733f2c5eec0fce0eb9c31a16c048babdaed90635895d689ae9303f4\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-03T14:10:16Z\\\",\\\"message\\\":\\\"o create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T14:10:16Z is after 2025-08-24T17:21:41Z]\\\\nI1203 14:10:16.717659 6478 services_controller.go:473] Services do not match for network=default, existing lbs: []services.LB{services.LB{Name:\\\\\\\"Service_openshift-network-console/networking-console-plugin_TCP_cluster\\\\\\\", UUID:\\\\\\\"ab0b1d51-5ec6-479b-8881-93dfa8d30337\\\\\\\", Protocol:\\\\\\\"tcp\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-network-console/networking-console-plugin\\\\\\\"}, Opts:services.LBOpts{Reject:false, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{}, Templates:services.TemplateMap{}, Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGr\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T14:10:15Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8f75988e25825e00abdb9e4b2292c5cc8ccab3dd028c882c2a38c796c0fd19c5\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-03T14:10:49Z\\\",\\\"message\\\":\\\"shift-kube-storage-version-migrator-operator/metrics]} name:Service_openshift-kube-storage-version-migrator-operator/metrics_TCP_cluster options:{GoMap:map[event:false hairpin_snat_ip:169.254.0.5 fd69::5 neighbor_responder:none reject:true skip_snat:false]} protocol:{GoSet:[tcp]} selection_fields:{GoSet:[]} vips:{GoMap:map[10.217.5.36:443:]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {7f9b8f25-db1a-4d02-a423-9afc5c2fb83c}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nF1203 14:10:49.548775 6841 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T14:10:49Z is after 2025-08-24T17:\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T14:10:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wdlzx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ab98e365c1a4242a47b8cbe2dd85109d366d6002c4de1f18ddc950a46bea6c12\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wdlzx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2f875252fbe6a673a72fc9c56bb96899f62732305ad77b3108e3b244bfa91ecb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2f875252fbe6a673a72fc9c56bb96899f62732305ad77b3108e3b244bfa91ecb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T14:09:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T14:09:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wdlzx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T14:09:50Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-psjpm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T14:10:50Z is after 2025-08-24T17:21:41Z" Dec 03 14:10:50 crc kubenswrapper[4805]: I1203 14:10:50.741102 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-4gd9r" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6b36a3db-1047-487a-9d6c-4f51ffdd8cc1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:10:04Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:10:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:10:04Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:10:04Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4t6zl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4t6zl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T14:10:04Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-4gd9r\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T14:10:50Z is after 2025-08-24T17:21:41Z" Dec 03 14:10:50 crc kubenswrapper[4805]: I1203 14:10:50.762143 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:48Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:48Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T14:10:50Z is after 2025-08-24T17:21:41Z" Dec 03 14:10:50 crc kubenswrapper[4805]: I1203 14:10:50.765285 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 14:10:50 crc kubenswrapper[4805]: I1203 14:10:50.765359 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 14:10:50 crc kubenswrapper[4805]: I1203 14:10:50.765385 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 14:10:50 crc kubenswrapper[4805]: I1203 14:10:50.765417 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 14:10:50 crc kubenswrapper[4805]: I1203 14:10:50.765440 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T14:10:50Z","lastTransitionTime":"2025-12-03T14:10:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 14:10:50 crc kubenswrapper[4805]: I1203 14:10:50.780419 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:49Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:49Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://541ace7ccbb66dbd191c3ca8c1e4479e3edbf80dfc67326bea8aaee81fdcfc3c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://821e3e9a902aacecf6c41e8b8d8aa1ac951ce2fbb6505b7148b16348a3754e55\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T14:10:50Z is after 2025-08-24T17:21:41Z" Dec 03 14:10:50 crc kubenswrapper[4805]: I1203 14:10:50.797976 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-gskh4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6098937f-e3f6-45e8-a647-4994a79cd711\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://28d5653f26ec61647808b3a733ff7eac70aa69ac184660f24147e8cfaf0eb802\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nwrw9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9ad6b68a42fd33ee167e67873c6eed3f11cdf6dc71d22e952c021436eb967dd9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nwrw9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T14:09:50Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-gskh4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T14:10:50Z is after 2025-08-24T17:21:41Z" Dec 03 14:10:50 crc kubenswrapper[4805]: I1203 14:10:50.812185 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-pdzgk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"860769de-6a5b-4cb9-a752-264e7e2e785f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5e2f0a0cc21505d609f81918365d5e4f5d21959ac12ab147d6a51601c9de4c14\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ggn4n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T14:09:50Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-pdzgk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T14:10:50Z is after 2025-08-24T17:21:41Z" Dec 03 14:10:50 crc kubenswrapper[4805]: I1203 14:10:50.829511 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-69h75" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5d1c694f-9847-4086-8b37-3573ddeb3ace\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:10:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:10:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:10:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:10:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://abc1d9b8582ca9fe4a51fb608fb39129a179ddcdfde0797a6369a55b2e0124cf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:10:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s578b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c7c4d434b035cac2a91a4f0846a2f4fb5bd68f881c28487860adb6db85e8ce75\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:10:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s578b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T14:10:03Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-69h75\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T14:10:50Z is after 2025-08-24T17:21:41Z" Dec 03 14:10:50 crc kubenswrapper[4805]: I1203 14:10:50.848232 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:49Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:49Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://38339331a3209d6787d1def81b72f4a8514bf91080a139cad6d8ac5cc6bef10e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T14:10:50Z is after 2025-08-24T17:21:41Z" Dec 03 14:10:50 crc kubenswrapper[4805]: I1203 14:10:50.867051 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:48Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:48Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T14:10:50Z is after 2025-08-24T17:21:41Z" Dec 03 14:10:50 crc kubenswrapper[4805]: I1203 14:10:50.867961 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 14:10:50 crc kubenswrapper[4805]: I1203 14:10:50.867990 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 14:10:50 crc kubenswrapper[4805]: I1203 14:10:50.867998 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 14:10:50 crc kubenswrapper[4805]: I1203 14:10:50.868010 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 14:10:50 crc kubenswrapper[4805]: I1203 14:10:50.868020 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T14:10:50Z","lastTransitionTime":"2025-12-03T14:10:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 14:10:50 crc kubenswrapper[4805]: I1203 14:10:50.885362 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:51Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:51Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b3123d3fa879097db88ed2f3cfb05631f440c24c5405503b2de3546f12e2e899\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T14:10:50Z is after 2025-08-24T17:21:41Z" Dec 03 14:10:50 crc kubenswrapper[4805]: I1203 14:10:50.900293 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-bdtdg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"53a0e9e3-b6d8-44b7-a32b-aa3450442865\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f557e17ad105d6972d3c1480656c6073f67e28c0f5773aa365549951f4d161c7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9ww54\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T14:09:51Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-bdtdg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T14:10:50Z is after 2025-08-24T17:21:41Z" Dec 03 14:10:50 crc kubenswrapper[4805]: I1203 14:10:50.923378 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b6d36671-7173-4e9c-843f-005b4dcf576c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a837d7d491eaacbe4279b4014bb5fcbe9b514ba5cc82c708227bcd9b25a346e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6e5f04faf3b23197bb8ea11d3e574dc86f31e21d798734962001ec8c95de7c84\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2ff1a979387da41b00e95eb9cc8b2c9de966380a12fae687591df3c5b18a4497\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1e66e1cb18573a14e3c3ae16af4dbd9f8997a6ad38680c3ab4505646ebd48e7c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7d5b5c85e3de43e7e95111aa967248550607c26200c783251b68696f202f92a1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://474f72dac37a808fc4b9832306e76979b151d7221807ff3086013466ed1419c3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://474f72dac37a808fc4b9832306e76979b151d7221807ff3086013466ed1419c3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T14:09:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T14:09:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://74c8cea1a61c4775a0ad2f4ab4bb8a36437cf23d2ee953311d6ea967e297510a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://74c8cea1a61c4775a0ad2f4ab4bb8a36437cf23d2ee953311d6ea967e297510a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T14:09:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T14:09:31Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://a4f2f56a809adc01fba82eba0652d78f58770c132f77b1424f8b17068b1927e3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a4f2f56a809adc01fba82eba0652d78f58770c132f77b1424f8b17068b1927e3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T14:09:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T14:09:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T14:09:30Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T14:10:50Z is after 2025-08-24T17:21:41Z" Dec 03 14:10:50 crc kubenswrapper[4805]: I1203 14:10:50.939897 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ced8291d-0ce5-4e74-9eaf-2bf033774039\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:10:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:10:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3e5a34636f0a8c4d9179074596d28bc2eb7a504573398c596a637defd829907a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://29cd95be3f964415943c873e7914ffcea4abe50dfdf17f8a9cd03fc38df35981\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://951173b3d3d581275a6f6e43209b5e0611a701ca7a10f81504d723e24a2d21a3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e64730957bbbb2feb4d506e0ed9731a0594bf66fa1e8fcdb8ec944ded6308eed\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5da55f769e23ad52bac9c77da62c78e70fa34db8ed1a435e4eaf4dad3fd2b68a\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-03T14:09:48Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1203 14:09:42.992179 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1203 14:09:42.993233 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1830926909/tls.crt::/tmp/serving-cert-1830926909/tls.key\\\\\\\"\\\\nI1203 14:09:48.622403 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1203 14:09:48.626476 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1203 14:09:48.626527 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1203 14:09:48.626574 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1203 14:09:48.626582 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1203 14:09:48.633990 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1203 14:09:48.634024 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 14:09:48.634029 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 14:09:48.634034 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1203 14:09:48.634038 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1203 14:09:48.634042 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1203 14:09:48.634045 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1203 14:09:48.634344 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1203 14:09:48.644497 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T14:09:32Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bd92a160eb530c5687ebd7cd5b83b41820a2b808a0fd9514e5ace9d237e1bd26\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:32Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://af9c940135012c629447528c61743ed1e58708ab57ce76604c6da0479aa459cf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://af9c940135012c629447528c61743ed1e58708ab57ce76604c6da0479aa459cf\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T14:09:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T14:09:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T14:09:30Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T14:10:50Z is after 2025-08-24T17:21:41Z" Dec 03 14:10:50 crc kubenswrapper[4805]: I1203 14:10:50.955591 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6d48b4f1-5dfc-4f6b-a4e8-4e465c651c7b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:10:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:10:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5ea698a1e0de3b5201e2f9032d834d81d9aab9c2fb17aca582184c7090f1bd2f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://66aa257562f8e113fc1cd2936b8aa18e927f90018ad4c11a184c1c56444be648\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2fd6d50f839b603a90e5d00ab6e36e832eab2e299b7af1dbb84498c35bf1c08e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://77c26dae6ccd328a8a1ce4fe6ffe94c78061d8d3ac3727adb80a24c16f4ab533\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://77c26dae6ccd328a8a1ce4fe6ffe94c78061d8d3ac3727adb80a24c16f4ab533\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T14:09:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T14:09:31Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T14:09:30Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T14:10:50Z is after 2025-08-24T17:21:41Z" Dec 03 14:10:50 crc kubenswrapper[4805]: I1203 14:10:50.971271 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 14:10:50 crc kubenswrapper[4805]: I1203 14:10:50.971329 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 14:10:50 crc kubenswrapper[4805]: I1203 14:10:50.971340 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 14:10:50 crc kubenswrapper[4805]: I1203 14:10:50.971360 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 14:10:50 crc kubenswrapper[4805]: I1203 14:10:50.971385 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T14:10:50Z","lastTransitionTime":"2025-12-03T14:10:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 14:10:50 crc kubenswrapper[4805]: I1203 14:10:50.979131 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-bpk49" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"de00a81f-82e1-4223-9614-f98606b9a8ef\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://04f4ad27e02e2024ae9e7cd1b16116267e69cecda882dd723cdb274c41a0e7db\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-25d7t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7741e0e70416ebe094bfabc6009e98d7d4c0fbcba1fc82df13b9ba40774b456a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7741e0e70416ebe094bfabc6009e98d7d4c0fbcba1fc82df13b9ba40774b456a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T14:09:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T14:09:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-25d7t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ee3499c02eb52bbf08c447d2433d435a74dfb70670dc4cbaac44f5fbff9e8a29\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ee3499c02eb52bbf08c447d2433d435a74dfb70670dc4cbaac44f5fbff9e8a29\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T14:09:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T14:09:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-25d7t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ec8488b1d69799951be7ef40e5e028732265c4ddcdea2d453923e567864771f7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ec8488b1d69799951be7ef40e5e028732265c4ddcdea2d453923e567864771f7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T14:09:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T14:09:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-25d7t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://66707bb4e1fdacb265e89786df8a04605830f1c156394cacfcec9b54abe5d41a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://66707bb4e1fdacb265e89786df8a04605830f1c156394cacfcec9b54abe5d41a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T14:09:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T14:09:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-25d7t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6207ab6600970ff15c8713c547d4920b2c77eb9bea990bb5da497e24870dcc0f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6207ab6600970ff15c8713c547d4920b2c77eb9bea990bb5da497e24870dcc0f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T14:09:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T14:09:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-25d7t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8de65d4a205e2deec0b433e53c0e37feb94aa0ede6a4c58d7414b80829804505\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8de65d4a205e2deec0b433e53c0e37feb94aa0ede6a4c58d7414b80829804505\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T14:09:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T14:09:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-25d7t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T14:09:50Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-bpk49\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T14:10:50Z is after 2025-08-24T17:21:41Z" Dec 03 14:10:50 crc kubenswrapper[4805]: I1203 14:10:50.998579 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"750d2b4b-738a-4992-8902-742528f6e875\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://abd2d0156113bb5182727a53aac594410ed051c19258beb8a07b808eff0b42a7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8673a29c0e16f4e707d79f0610722ccabb023ff2b4a18d7cf9d305183817d952\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://35766e7206b1be7a53d13d6890d759cf81dbd999643c12349268e4bf16fbbdb0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5b9844a11b2b903978d0d464b469e26c9211100da61d018dde9636dca95b684e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T14:09:30Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T14:10:50Z is after 2025-08-24T17:21:41Z" Dec 03 14:10:51 crc kubenswrapper[4805]: I1203 14:10:51.014893 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:48Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:48Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T14:10:51Z is after 2025-08-24T17:21:41Z" Dec 03 14:10:51 crc kubenswrapper[4805]: I1203 14:10:51.031217 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-k4mhs" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c09c1cdc-f461-45c9-8444-5b8764bccabc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:10:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:10:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7e077e19b6a5bf7d29b357982e12224d4b7058e0f9ac81b66b5a665c9bbf5014\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3354c1a44f891182d04116fc76234a2b337d50567d032a3584f902a9516387cc\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-03T14:10:36Z\\\",\\\"message\\\":\\\"2025-12-03T14:09:51+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_3e51ab9a-fbbe-4a78-9308-8a81e203fcad\\\\n2025-12-03T14:09:51+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_3e51ab9a-fbbe-4a78-9308-8a81e203fcad to /host/opt/cni/bin/\\\\n2025-12-03T14:09:51Z [verbose] multus-daemon started\\\\n2025-12-03T14:09:51Z [verbose] Readiness Indicator file check\\\\n2025-12-03T14:10:36Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T14:09:50Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:10:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qdcd7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T14:09:50Z\\\"}}\" for pod \"openshift-multus\"/\"multus-k4mhs\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T14:10:51Z is after 2025-08-24T17:21:41Z" Dec 03 14:10:51 crc kubenswrapper[4805]: I1203 14:10:51.074781 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 14:10:51 crc kubenswrapper[4805]: I1203 14:10:51.074860 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 14:10:51 crc kubenswrapper[4805]: I1203 14:10:51.074871 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 14:10:51 crc kubenswrapper[4805]: I1203 14:10:51.074889 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 14:10:51 crc kubenswrapper[4805]: I1203 14:10:51.074903 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T14:10:51Z","lastTransitionTime":"2025-12-03T14:10:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 14:10:51 crc kubenswrapper[4805]: I1203 14:10:51.159730 4805 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-psjpm_deeec30d-cad2-4dfd-adb6-64f6646581be/ovnkube-controller/3.log" Dec 03 14:10:51 crc kubenswrapper[4805]: I1203 14:10:51.180459 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 14:10:51 crc kubenswrapper[4805]: I1203 14:10:51.180599 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 14:10:51 crc kubenswrapper[4805]: I1203 14:10:51.180620 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 14:10:51 crc kubenswrapper[4805]: I1203 14:10:51.180647 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 14:10:51 crc kubenswrapper[4805]: I1203 14:10:51.180664 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T14:10:51Z","lastTransitionTime":"2025-12-03T14:10:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 14:10:51 crc kubenswrapper[4805]: I1203 14:10:51.283079 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 14:10:51 crc kubenswrapper[4805]: I1203 14:10:51.283120 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 14:10:51 crc kubenswrapper[4805]: I1203 14:10:51.283135 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 14:10:51 crc kubenswrapper[4805]: I1203 14:10:51.283154 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 14:10:51 crc kubenswrapper[4805]: I1203 14:10:51.283169 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T14:10:51Z","lastTransitionTime":"2025-12-03T14:10:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 14:10:51 crc kubenswrapper[4805]: I1203 14:10:51.386067 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 14:10:51 crc kubenswrapper[4805]: I1203 14:10:51.386134 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 14:10:51 crc kubenswrapper[4805]: I1203 14:10:51.386151 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 14:10:51 crc kubenswrapper[4805]: I1203 14:10:51.386181 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 14:10:51 crc kubenswrapper[4805]: I1203 14:10:51.386203 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T14:10:51Z","lastTransitionTime":"2025-12-03T14:10:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 14:10:51 crc kubenswrapper[4805]: I1203 14:10:51.489049 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 14:10:51 crc kubenswrapper[4805]: I1203 14:10:51.489100 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 14:10:51 crc kubenswrapper[4805]: I1203 14:10:51.489116 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 14:10:51 crc kubenswrapper[4805]: I1203 14:10:51.489135 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 14:10:51 crc kubenswrapper[4805]: I1203 14:10:51.489147 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T14:10:51Z","lastTransitionTime":"2025-12-03T14:10:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 14:10:51 crc kubenswrapper[4805]: I1203 14:10:51.592463 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 14:10:51 crc kubenswrapper[4805]: I1203 14:10:51.592524 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 14:10:51 crc kubenswrapper[4805]: I1203 14:10:51.592545 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 14:10:51 crc kubenswrapper[4805]: I1203 14:10:51.592570 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 14:10:51 crc kubenswrapper[4805]: I1203 14:10:51.592587 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T14:10:51Z","lastTransitionTime":"2025-12-03T14:10:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 14:10:51 crc kubenswrapper[4805]: I1203 14:10:51.696232 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 14:10:51 crc kubenswrapper[4805]: I1203 14:10:51.696272 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 14:10:51 crc kubenswrapper[4805]: I1203 14:10:51.696281 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 14:10:51 crc kubenswrapper[4805]: I1203 14:10:51.696315 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 14:10:51 crc kubenswrapper[4805]: I1203 14:10:51.696326 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T14:10:51Z","lastTransitionTime":"2025-12-03T14:10:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 14:10:51 crc kubenswrapper[4805]: I1203 14:10:51.799448 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 14:10:51 crc kubenswrapper[4805]: I1203 14:10:51.799483 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 14:10:51 crc kubenswrapper[4805]: I1203 14:10:51.799491 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 14:10:51 crc kubenswrapper[4805]: I1203 14:10:51.799504 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 14:10:51 crc kubenswrapper[4805]: I1203 14:10:51.799513 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T14:10:51Z","lastTransitionTime":"2025-12-03T14:10:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 14:10:51 crc kubenswrapper[4805]: I1203 14:10:51.902261 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 14:10:51 crc kubenswrapper[4805]: I1203 14:10:51.902335 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 14:10:51 crc kubenswrapper[4805]: I1203 14:10:51.902393 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 14:10:51 crc kubenswrapper[4805]: I1203 14:10:51.902422 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 14:10:51 crc kubenswrapper[4805]: I1203 14:10:51.902439 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T14:10:51Z","lastTransitionTime":"2025-12-03T14:10:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 14:10:52 crc kubenswrapper[4805]: I1203 14:10:52.005648 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 14:10:52 crc kubenswrapper[4805]: I1203 14:10:52.005705 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 14:10:52 crc kubenswrapper[4805]: I1203 14:10:52.005725 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 14:10:52 crc kubenswrapper[4805]: I1203 14:10:52.005748 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 14:10:52 crc kubenswrapper[4805]: I1203 14:10:52.005765 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T14:10:52Z","lastTransitionTime":"2025-12-03T14:10:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 14:10:52 crc kubenswrapper[4805]: I1203 14:10:52.109147 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 14:10:52 crc kubenswrapper[4805]: I1203 14:10:52.109212 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 14:10:52 crc kubenswrapper[4805]: I1203 14:10:52.109234 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 14:10:52 crc kubenswrapper[4805]: I1203 14:10:52.109264 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 14:10:52 crc kubenswrapper[4805]: I1203 14:10:52.109286 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T14:10:52Z","lastTransitionTime":"2025-12-03T14:10:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 14:10:52 crc kubenswrapper[4805]: I1203 14:10:52.211779 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 14:10:52 crc kubenswrapper[4805]: I1203 14:10:52.211819 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 14:10:52 crc kubenswrapper[4805]: I1203 14:10:52.211834 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 14:10:52 crc kubenswrapper[4805]: I1203 14:10:52.211876 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 14:10:52 crc kubenswrapper[4805]: I1203 14:10:52.211887 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T14:10:52Z","lastTransitionTime":"2025-12-03T14:10:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 14:10:52 crc kubenswrapper[4805]: I1203 14:10:52.314252 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 14:10:52 crc kubenswrapper[4805]: I1203 14:10:52.314318 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 14:10:52 crc kubenswrapper[4805]: I1203 14:10:52.314334 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 14:10:52 crc kubenswrapper[4805]: I1203 14:10:52.314354 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 14:10:52 crc kubenswrapper[4805]: I1203 14:10:52.314371 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T14:10:52Z","lastTransitionTime":"2025-12-03T14:10:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 14:10:52 crc kubenswrapper[4805]: I1203 14:10:52.417070 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 14:10:52 crc kubenswrapper[4805]: I1203 14:10:52.417100 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 14:10:52 crc kubenswrapper[4805]: I1203 14:10:52.417109 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 14:10:52 crc kubenswrapper[4805]: I1203 14:10:52.417123 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 14:10:52 crc kubenswrapper[4805]: I1203 14:10:52.417132 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T14:10:52Z","lastTransitionTime":"2025-12-03T14:10:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 14:10:52 crc kubenswrapper[4805]: I1203 14:10:52.520279 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 14:10:52 crc kubenswrapper[4805]: I1203 14:10:52.520329 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 14:10:52 crc kubenswrapper[4805]: I1203 14:10:52.520341 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 14:10:52 crc kubenswrapper[4805]: I1203 14:10:52.520360 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 14:10:52 crc kubenswrapper[4805]: I1203 14:10:52.520372 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T14:10:52Z","lastTransitionTime":"2025-12-03T14:10:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 14:10:52 crc kubenswrapper[4805]: I1203 14:10:52.623273 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 14:10:52 crc kubenswrapper[4805]: I1203 14:10:52.623333 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 14:10:52 crc kubenswrapper[4805]: I1203 14:10:52.623349 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 14:10:52 crc kubenswrapper[4805]: I1203 14:10:52.623378 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 14:10:52 crc kubenswrapper[4805]: I1203 14:10:52.623397 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T14:10:52Z","lastTransitionTime":"2025-12-03T14:10:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 14:10:52 crc kubenswrapper[4805]: I1203 14:10:52.623484 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 14:10:52 crc kubenswrapper[4805]: E1203 14:10:52.623699 4805 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-03 14:11:56.623675316 +0000 UTC m=+146.286592279 (durationBeforeRetry 1m4s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 14:10:52 crc kubenswrapper[4805]: I1203 14:10:52.693492 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 14:10:52 crc kubenswrapper[4805]: I1203 14:10:52.693568 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 14:10:52 crc kubenswrapper[4805]: I1203 14:10:52.693586 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 14:10:52 crc kubenswrapper[4805]: E1203 14:10:52.693649 4805 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 03 14:10:52 crc kubenswrapper[4805]: I1203 14:10:52.693698 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-4gd9r" Dec 03 14:10:52 crc kubenswrapper[4805]: E1203 14:10:52.693874 4805 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-4gd9r" podUID="6b36a3db-1047-487a-9d6c-4f51ffdd8cc1" Dec 03 14:10:52 crc kubenswrapper[4805]: E1203 14:10:52.693875 4805 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 03 14:10:52 crc kubenswrapper[4805]: E1203 14:10:52.693962 4805 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 03 14:10:52 crc kubenswrapper[4805]: I1203 14:10:52.724614 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 14:10:52 crc kubenswrapper[4805]: I1203 14:10:52.724676 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 14:10:52 crc kubenswrapper[4805]: I1203 14:10:52.724710 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 14:10:52 crc kubenswrapper[4805]: I1203 14:10:52.724737 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 14:10:52 crc kubenswrapper[4805]: E1203 14:10:52.724816 4805 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Dec 03 14:10:52 crc kubenswrapper[4805]: E1203 14:10:52.724871 4805 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 03 14:10:52 crc kubenswrapper[4805]: E1203 14:10:52.724894 4805 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-03 14:11:56.724877701 +0000 UTC m=+146.387794624 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Dec 03 14:10:52 crc kubenswrapper[4805]: E1203 14:10:52.724891 4805 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 03 14:10:52 crc kubenswrapper[4805]: E1203 14:10:52.724919 4805 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 03 14:10:52 crc kubenswrapper[4805]: E1203 14:10:52.724931 4805 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 03 14:10:52 crc kubenswrapper[4805]: E1203 14:10:52.724938 4805 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-03 14:11:56.724919202 +0000 UTC m=+146.387836155 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 03 14:10:52 crc kubenswrapper[4805]: E1203 14:10:52.724980 4805 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-12-03 14:11:56.724962543 +0000 UTC m=+146.387879526 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 03 14:10:52 crc kubenswrapper[4805]: E1203 14:10:52.725062 4805 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 03 14:10:52 crc kubenswrapper[4805]: E1203 14:10:52.725101 4805 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 03 14:10:52 crc kubenswrapper[4805]: E1203 14:10:52.725123 4805 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 03 14:10:52 crc kubenswrapper[4805]: E1203 14:10:52.725211 4805 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-12-03 14:11:56.725187279 +0000 UTC m=+146.388104242 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 03 14:10:52 crc kubenswrapper[4805]: I1203 14:10:52.726335 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 14:10:52 crc kubenswrapper[4805]: I1203 14:10:52.726390 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 14:10:52 crc kubenswrapper[4805]: I1203 14:10:52.726408 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 14:10:52 crc kubenswrapper[4805]: I1203 14:10:52.726431 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 14:10:52 crc kubenswrapper[4805]: I1203 14:10:52.726448 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T14:10:52Z","lastTransitionTime":"2025-12-03T14:10:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 14:10:52 crc kubenswrapper[4805]: I1203 14:10:52.829512 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 14:10:52 crc kubenswrapper[4805]: I1203 14:10:52.829549 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 14:10:52 crc kubenswrapper[4805]: I1203 14:10:52.829567 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 14:10:52 crc kubenswrapper[4805]: I1203 14:10:52.829584 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 14:10:52 crc kubenswrapper[4805]: I1203 14:10:52.829594 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T14:10:52Z","lastTransitionTime":"2025-12-03T14:10:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 14:10:52 crc kubenswrapper[4805]: I1203 14:10:52.935036 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 14:10:52 crc kubenswrapper[4805]: I1203 14:10:52.935356 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 14:10:52 crc kubenswrapper[4805]: I1203 14:10:52.935664 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 14:10:52 crc kubenswrapper[4805]: I1203 14:10:52.935691 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 14:10:52 crc kubenswrapper[4805]: I1203 14:10:52.935708 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T14:10:52Z","lastTransitionTime":"2025-12-03T14:10:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 14:10:53 crc kubenswrapper[4805]: I1203 14:10:53.037542 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 14:10:53 crc kubenswrapper[4805]: I1203 14:10:53.037591 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 14:10:53 crc kubenswrapper[4805]: I1203 14:10:53.037603 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 14:10:53 crc kubenswrapper[4805]: I1203 14:10:53.037619 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 14:10:53 crc kubenswrapper[4805]: I1203 14:10:53.037629 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T14:10:53Z","lastTransitionTime":"2025-12-03T14:10:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 14:10:53 crc kubenswrapper[4805]: I1203 14:10:53.139955 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 14:10:53 crc kubenswrapper[4805]: I1203 14:10:53.140000 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 14:10:53 crc kubenswrapper[4805]: I1203 14:10:53.140013 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 14:10:53 crc kubenswrapper[4805]: I1203 14:10:53.140030 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 14:10:53 crc kubenswrapper[4805]: I1203 14:10:53.140041 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T14:10:53Z","lastTransitionTime":"2025-12-03T14:10:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 14:10:53 crc kubenswrapper[4805]: I1203 14:10:53.242704 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 14:10:53 crc kubenswrapper[4805]: I1203 14:10:53.242778 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 14:10:53 crc kubenswrapper[4805]: I1203 14:10:53.242791 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 14:10:53 crc kubenswrapper[4805]: I1203 14:10:53.242810 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 14:10:53 crc kubenswrapper[4805]: I1203 14:10:53.242822 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T14:10:53Z","lastTransitionTime":"2025-12-03T14:10:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 14:10:53 crc kubenswrapper[4805]: I1203 14:10:53.344688 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 14:10:53 crc kubenswrapper[4805]: I1203 14:10:53.344954 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 14:10:53 crc kubenswrapper[4805]: I1203 14:10:53.344962 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 14:10:53 crc kubenswrapper[4805]: I1203 14:10:53.344975 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 14:10:53 crc kubenswrapper[4805]: I1203 14:10:53.344984 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T14:10:53Z","lastTransitionTime":"2025-12-03T14:10:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 14:10:53 crc kubenswrapper[4805]: I1203 14:10:53.386480 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 14:10:53 crc kubenswrapper[4805]: I1203 14:10:53.386536 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 14:10:53 crc kubenswrapper[4805]: I1203 14:10:53.386548 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 14:10:53 crc kubenswrapper[4805]: I1203 14:10:53.386563 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 14:10:53 crc kubenswrapper[4805]: I1203 14:10:53.386580 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T14:10:53Z","lastTransitionTime":"2025-12-03T14:10:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 14:10:53 crc kubenswrapper[4805]: E1203 14:10:53.404099 4805 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T14:10:53Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T14:10:53Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T14:10:53Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T14:10:53Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T14:10:53Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T14:10:53Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T14:10:53Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T14:10:53Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"e4edfbcd-1de1-486e-9c0b-65f9fbba52ad\\\",\\\"systemUUID\\\":\\\"7a729b24-19ef-490e-ab4c-d31c52ed5658\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T14:10:53Z is after 2025-08-24T17:21:41Z" Dec 03 14:10:53 crc kubenswrapper[4805]: I1203 14:10:53.409178 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 14:10:53 crc kubenswrapper[4805]: I1203 14:10:53.409234 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 14:10:53 crc kubenswrapper[4805]: I1203 14:10:53.409245 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 14:10:53 crc kubenswrapper[4805]: I1203 14:10:53.409260 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 14:10:53 crc kubenswrapper[4805]: I1203 14:10:53.409270 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T14:10:53Z","lastTransitionTime":"2025-12-03T14:10:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 14:10:53 crc kubenswrapper[4805]: E1203 14:10:53.424830 4805 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T14:10:53Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T14:10:53Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T14:10:53Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T14:10:53Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T14:10:53Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T14:10:53Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T14:10:53Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T14:10:53Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"e4edfbcd-1de1-486e-9c0b-65f9fbba52ad\\\",\\\"systemUUID\\\":\\\"7a729b24-19ef-490e-ab4c-d31c52ed5658\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T14:10:53Z is after 2025-08-24T17:21:41Z" Dec 03 14:10:53 crc kubenswrapper[4805]: I1203 14:10:53.429530 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 14:10:53 crc kubenswrapper[4805]: I1203 14:10:53.429596 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 14:10:53 crc kubenswrapper[4805]: I1203 14:10:53.429608 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 14:10:53 crc kubenswrapper[4805]: I1203 14:10:53.429643 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 14:10:53 crc kubenswrapper[4805]: I1203 14:10:53.429654 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T14:10:53Z","lastTransitionTime":"2025-12-03T14:10:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 14:10:53 crc kubenswrapper[4805]: E1203 14:10:53.450533 4805 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T14:10:53Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T14:10:53Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T14:10:53Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T14:10:53Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T14:10:53Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T14:10:53Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T14:10:53Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T14:10:53Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"e4edfbcd-1de1-486e-9c0b-65f9fbba52ad\\\",\\\"systemUUID\\\":\\\"7a729b24-19ef-490e-ab4c-d31c52ed5658\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T14:10:53Z is after 2025-08-24T17:21:41Z" Dec 03 14:10:53 crc kubenswrapper[4805]: I1203 14:10:53.456346 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 14:10:53 crc kubenswrapper[4805]: I1203 14:10:53.456403 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 14:10:53 crc kubenswrapper[4805]: I1203 14:10:53.456420 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 14:10:53 crc kubenswrapper[4805]: I1203 14:10:53.456444 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 14:10:53 crc kubenswrapper[4805]: I1203 14:10:53.456462 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T14:10:53Z","lastTransitionTime":"2025-12-03T14:10:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 14:10:53 crc kubenswrapper[4805]: E1203 14:10:53.473179 4805 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T14:10:53Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T14:10:53Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T14:10:53Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T14:10:53Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T14:10:53Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T14:10:53Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T14:10:53Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T14:10:53Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"e4edfbcd-1de1-486e-9c0b-65f9fbba52ad\\\",\\\"systemUUID\\\":\\\"7a729b24-19ef-490e-ab4c-d31c52ed5658\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T14:10:53Z is after 2025-08-24T17:21:41Z" Dec 03 14:10:53 crc kubenswrapper[4805]: I1203 14:10:53.478694 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 14:10:53 crc kubenswrapper[4805]: I1203 14:10:53.478748 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 14:10:53 crc kubenswrapper[4805]: I1203 14:10:53.478757 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 14:10:53 crc kubenswrapper[4805]: I1203 14:10:53.478773 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 14:10:53 crc kubenswrapper[4805]: I1203 14:10:53.478784 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T14:10:53Z","lastTransitionTime":"2025-12-03T14:10:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 14:10:53 crc kubenswrapper[4805]: E1203 14:10:53.493589 4805 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T14:10:53Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T14:10:53Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T14:10:53Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T14:10:53Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T14:10:53Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T14:10:53Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T14:10:53Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T14:10:53Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"e4edfbcd-1de1-486e-9c0b-65f9fbba52ad\\\",\\\"systemUUID\\\":\\\"7a729b24-19ef-490e-ab4c-d31c52ed5658\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T14:10:53Z is after 2025-08-24T17:21:41Z" Dec 03 14:10:53 crc kubenswrapper[4805]: E1203 14:10:53.493742 4805 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Dec 03 14:10:53 crc kubenswrapper[4805]: I1203 14:10:53.495755 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 14:10:53 crc kubenswrapper[4805]: I1203 14:10:53.495788 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 14:10:53 crc kubenswrapper[4805]: I1203 14:10:53.495798 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 14:10:53 crc kubenswrapper[4805]: I1203 14:10:53.495814 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 14:10:53 crc kubenswrapper[4805]: I1203 14:10:53.495828 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T14:10:53Z","lastTransitionTime":"2025-12-03T14:10:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 14:10:53 crc kubenswrapper[4805]: I1203 14:10:53.597977 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 14:10:53 crc kubenswrapper[4805]: I1203 14:10:53.598015 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 14:10:53 crc kubenswrapper[4805]: I1203 14:10:53.598026 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 14:10:53 crc kubenswrapper[4805]: I1203 14:10:53.598041 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 14:10:53 crc kubenswrapper[4805]: I1203 14:10:53.598052 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T14:10:53Z","lastTransitionTime":"2025-12-03T14:10:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 14:10:53 crc kubenswrapper[4805]: I1203 14:10:53.701538 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 14:10:53 crc kubenswrapper[4805]: I1203 14:10:53.701578 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 14:10:53 crc kubenswrapper[4805]: I1203 14:10:53.701588 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 14:10:53 crc kubenswrapper[4805]: I1203 14:10:53.701603 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 14:10:53 crc kubenswrapper[4805]: I1203 14:10:53.701614 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T14:10:53Z","lastTransitionTime":"2025-12-03T14:10:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 14:10:53 crc kubenswrapper[4805]: I1203 14:10:53.709264 4805 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/kube-rbac-proxy-crio-crc"] Dec 03 14:10:53 crc kubenswrapper[4805]: I1203 14:10:53.804402 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 14:10:53 crc kubenswrapper[4805]: I1203 14:10:53.804451 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 14:10:53 crc kubenswrapper[4805]: I1203 14:10:53.804463 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 14:10:53 crc kubenswrapper[4805]: I1203 14:10:53.804479 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 14:10:53 crc kubenswrapper[4805]: I1203 14:10:53.804493 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T14:10:53Z","lastTransitionTime":"2025-12-03T14:10:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 14:10:53 crc kubenswrapper[4805]: I1203 14:10:53.906505 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 14:10:53 crc kubenswrapper[4805]: I1203 14:10:53.906540 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 14:10:53 crc kubenswrapper[4805]: I1203 14:10:53.906550 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 14:10:53 crc kubenswrapper[4805]: I1203 14:10:53.906567 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 14:10:53 crc kubenswrapper[4805]: I1203 14:10:53.906577 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T14:10:53Z","lastTransitionTime":"2025-12-03T14:10:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 14:10:54 crc kubenswrapper[4805]: I1203 14:10:54.009787 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 14:10:54 crc kubenswrapper[4805]: I1203 14:10:54.009903 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 14:10:54 crc kubenswrapper[4805]: I1203 14:10:54.009931 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 14:10:54 crc kubenswrapper[4805]: I1203 14:10:54.009961 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 14:10:54 crc kubenswrapper[4805]: I1203 14:10:54.009979 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T14:10:54Z","lastTransitionTime":"2025-12-03T14:10:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 14:10:54 crc kubenswrapper[4805]: I1203 14:10:54.113100 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 14:10:54 crc kubenswrapper[4805]: I1203 14:10:54.113180 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 14:10:54 crc kubenswrapper[4805]: I1203 14:10:54.113208 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 14:10:54 crc kubenswrapper[4805]: I1203 14:10:54.113231 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 14:10:54 crc kubenswrapper[4805]: I1203 14:10:54.113246 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T14:10:54Z","lastTransitionTime":"2025-12-03T14:10:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 14:10:54 crc kubenswrapper[4805]: I1203 14:10:54.215670 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 14:10:54 crc kubenswrapper[4805]: I1203 14:10:54.215728 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 14:10:54 crc kubenswrapper[4805]: I1203 14:10:54.215741 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 14:10:54 crc kubenswrapper[4805]: I1203 14:10:54.215755 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 14:10:54 crc kubenswrapper[4805]: I1203 14:10:54.215763 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T14:10:54Z","lastTransitionTime":"2025-12-03T14:10:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 14:10:54 crc kubenswrapper[4805]: I1203 14:10:54.317731 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 14:10:54 crc kubenswrapper[4805]: I1203 14:10:54.317778 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 14:10:54 crc kubenswrapper[4805]: I1203 14:10:54.317790 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 14:10:54 crc kubenswrapper[4805]: I1203 14:10:54.317806 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 14:10:54 crc kubenswrapper[4805]: I1203 14:10:54.317818 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T14:10:54Z","lastTransitionTime":"2025-12-03T14:10:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 14:10:54 crc kubenswrapper[4805]: I1203 14:10:54.420310 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 14:10:54 crc kubenswrapper[4805]: I1203 14:10:54.420363 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 14:10:54 crc kubenswrapper[4805]: I1203 14:10:54.420374 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 14:10:54 crc kubenswrapper[4805]: I1203 14:10:54.420388 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 14:10:54 crc kubenswrapper[4805]: I1203 14:10:54.420399 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T14:10:54Z","lastTransitionTime":"2025-12-03T14:10:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 14:10:54 crc kubenswrapper[4805]: I1203 14:10:54.523860 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 14:10:54 crc kubenswrapper[4805]: I1203 14:10:54.523903 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 14:10:54 crc kubenswrapper[4805]: I1203 14:10:54.523914 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 14:10:54 crc kubenswrapper[4805]: I1203 14:10:54.523931 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 14:10:54 crc kubenswrapper[4805]: I1203 14:10:54.523943 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T14:10:54Z","lastTransitionTime":"2025-12-03T14:10:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 14:10:54 crc kubenswrapper[4805]: I1203 14:10:54.626469 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 14:10:54 crc kubenswrapper[4805]: I1203 14:10:54.626505 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 14:10:54 crc kubenswrapper[4805]: I1203 14:10:54.626515 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 14:10:54 crc kubenswrapper[4805]: I1203 14:10:54.626528 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 14:10:54 crc kubenswrapper[4805]: I1203 14:10:54.626536 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T14:10:54Z","lastTransitionTime":"2025-12-03T14:10:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 14:10:54 crc kubenswrapper[4805]: I1203 14:10:54.693997 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 14:10:54 crc kubenswrapper[4805]: I1203 14:10:54.694077 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 14:10:54 crc kubenswrapper[4805]: I1203 14:10:54.694025 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 14:10:54 crc kubenswrapper[4805]: I1203 14:10:54.693998 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-4gd9r" Dec 03 14:10:54 crc kubenswrapper[4805]: E1203 14:10:54.694212 4805 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 03 14:10:54 crc kubenswrapper[4805]: E1203 14:10:54.694629 4805 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 03 14:10:54 crc kubenswrapper[4805]: E1203 14:10:54.694427 4805 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-4gd9r" podUID="6b36a3db-1047-487a-9d6c-4f51ffdd8cc1" Dec 03 14:10:54 crc kubenswrapper[4805]: E1203 14:10:54.694758 4805 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 03 14:10:54 crc kubenswrapper[4805]: I1203 14:10:54.729571 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 14:10:54 crc kubenswrapper[4805]: I1203 14:10:54.729669 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 14:10:54 crc kubenswrapper[4805]: I1203 14:10:54.729719 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 14:10:54 crc kubenswrapper[4805]: I1203 14:10:54.729744 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 14:10:54 crc kubenswrapper[4805]: I1203 14:10:54.729762 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T14:10:54Z","lastTransitionTime":"2025-12-03T14:10:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 14:10:54 crc kubenswrapper[4805]: I1203 14:10:54.834381 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 14:10:54 crc kubenswrapper[4805]: I1203 14:10:54.834440 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 14:10:54 crc kubenswrapper[4805]: I1203 14:10:54.834453 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 14:10:54 crc kubenswrapper[4805]: I1203 14:10:54.834472 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 14:10:54 crc kubenswrapper[4805]: I1203 14:10:54.834506 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T14:10:54Z","lastTransitionTime":"2025-12-03T14:10:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 14:10:54 crc kubenswrapper[4805]: I1203 14:10:54.938003 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 14:10:54 crc kubenswrapper[4805]: I1203 14:10:54.938087 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 14:10:54 crc kubenswrapper[4805]: I1203 14:10:54.938109 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 14:10:54 crc kubenswrapper[4805]: I1203 14:10:54.938138 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 14:10:54 crc kubenswrapper[4805]: I1203 14:10:54.938160 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T14:10:54Z","lastTransitionTime":"2025-12-03T14:10:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 14:10:55 crc kubenswrapper[4805]: I1203 14:10:55.040803 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 14:10:55 crc kubenswrapper[4805]: I1203 14:10:55.040890 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 14:10:55 crc kubenswrapper[4805]: I1203 14:10:55.040904 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 14:10:55 crc kubenswrapper[4805]: I1203 14:10:55.040920 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 14:10:55 crc kubenswrapper[4805]: I1203 14:10:55.041248 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T14:10:55Z","lastTransitionTime":"2025-12-03T14:10:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 14:10:55 crc kubenswrapper[4805]: I1203 14:10:55.144276 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 14:10:55 crc kubenswrapper[4805]: I1203 14:10:55.144320 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 14:10:55 crc kubenswrapper[4805]: I1203 14:10:55.144328 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 14:10:55 crc kubenswrapper[4805]: I1203 14:10:55.144342 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 14:10:55 crc kubenswrapper[4805]: I1203 14:10:55.144351 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T14:10:55Z","lastTransitionTime":"2025-12-03T14:10:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 14:10:55 crc kubenswrapper[4805]: I1203 14:10:55.246727 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 14:10:55 crc kubenswrapper[4805]: I1203 14:10:55.246800 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 14:10:55 crc kubenswrapper[4805]: I1203 14:10:55.246813 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 14:10:55 crc kubenswrapper[4805]: I1203 14:10:55.246828 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 14:10:55 crc kubenswrapper[4805]: I1203 14:10:55.246858 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T14:10:55Z","lastTransitionTime":"2025-12-03T14:10:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 14:10:55 crc kubenswrapper[4805]: I1203 14:10:55.349398 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 14:10:55 crc kubenswrapper[4805]: I1203 14:10:55.349479 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 14:10:55 crc kubenswrapper[4805]: I1203 14:10:55.349510 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 14:10:55 crc kubenswrapper[4805]: I1203 14:10:55.349542 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 14:10:55 crc kubenswrapper[4805]: I1203 14:10:55.349562 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T14:10:55Z","lastTransitionTime":"2025-12-03T14:10:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 14:10:55 crc kubenswrapper[4805]: I1203 14:10:55.452017 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 14:10:55 crc kubenswrapper[4805]: I1203 14:10:55.452085 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 14:10:55 crc kubenswrapper[4805]: I1203 14:10:55.452102 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 14:10:55 crc kubenswrapper[4805]: I1203 14:10:55.452123 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 14:10:55 crc kubenswrapper[4805]: I1203 14:10:55.452141 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T14:10:55Z","lastTransitionTime":"2025-12-03T14:10:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 14:10:55 crc kubenswrapper[4805]: I1203 14:10:55.554517 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 14:10:55 crc kubenswrapper[4805]: I1203 14:10:55.554548 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 14:10:55 crc kubenswrapper[4805]: I1203 14:10:55.554558 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 14:10:55 crc kubenswrapper[4805]: I1203 14:10:55.554570 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 14:10:55 crc kubenswrapper[4805]: I1203 14:10:55.554578 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T14:10:55Z","lastTransitionTime":"2025-12-03T14:10:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 14:10:55 crc kubenswrapper[4805]: I1203 14:10:55.656501 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 14:10:55 crc kubenswrapper[4805]: I1203 14:10:55.656532 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 14:10:55 crc kubenswrapper[4805]: I1203 14:10:55.656541 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 14:10:55 crc kubenswrapper[4805]: I1203 14:10:55.656553 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 14:10:55 crc kubenswrapper[4805]: I1203 14:10:55.656562 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T14:10:55Z","lastTransitionTime":"2025-12-03T14:10:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 14:10:55 crc kubenswrapper[4805]: I1203 14:10:55.759575 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 14:10:55 crc kubenswrapper[4805]: I1203 14:10:55.759624 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 14:10:55 crc kubenswrapper[4805]: I1203 14:10:55.759640 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 14:10:55 crc kubenswrapper[4805]: I1203 14:10:55.759663 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 14:10:55 crc kubenswrapper[4805]: I1203 14:10:55.759680 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T14:10:55Z","lastTransitionTime":"2025-12-03T14:10:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 14:10:55 crc kubenswrapper[4805]: I1203 14:10:55.862972 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 14:10:55 crc kubenswrapper[4805]: I1203 14:10:55.863054 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 14:10:55 crc kubenswrapper[4805]: I1203 14:10:55.863081 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 14:10:55 crc kubenswrapper[4805]: I1203 14:10:55.863113 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 14:10:55 crc kubenswrapper[4805]: I1203 14:10:55.863136 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T14:10:55Z","lastTransitionTime":"2025-12-03T14:10:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 14:10:55 crc kubenswrapper[4805]: I1203 14:10:55.966759 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 14:10:55 crc kubenswrapper[4805]: I1203 14:10:55.966936 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 14:10:55 crc kubenswrapper[4805]: I1203 14:10:55.966963 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 14:10:55 crc kubenswrapper[4805]: I1203 14:10:55.966989 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 14:10:55 crc kubenswrapper[4805]: I1203 14:10:55.967010 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T14:10:55Z","lastTransitionTime":"2025-12-03T14:10:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 14:10:56 crc kubenswrapper[4805]: I1203 14:10:56.069754 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 14:10:56 crc kubenswrapper[4805]: I1203 14:10:56.069829 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 14:10:56 crc kubenswrapper[4805]: I1203 14:10:56.069888 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 14:10:56 crc kubenswrapper[4805]: I1203 14:10:56.069918 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 14:10:56 crc kubenswrapper[4805]: I1203 14:10:56.069939 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T14:10:56Z","lastTransitionTime":"2025-12-03T14:10:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 14:10:56 crc kubenswrapper[4805]: I1203 14:10:56.173331 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 14:10:56 crc kubenswrapper[4805]: I1203 14:10:56.173399 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 14:10:56 crc kubenswrapper[4805]: I1203 14:10:56.173417 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 14:10:56 crc kubenswrapper[4805]: I1203 14:10:56.173442 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 14:10:56 crc kubenswrapper[4805]: I1203 14:10:56.173458 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T14:10:56Z","lastTransitionTime":"2025-12-03T14:10:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 14:10:56 crc kubenswrapper[4805]: I1203 14:10:56.276413 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 14:10:56 crc kubenswrapper[4805]: I1203 14:10:56.276468 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 14:10:56 crc kubenswrapper[4805]: I1203 14:10:56.276486 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 14:10:56 crc kubenswrapper[4805]: I1203 14:10:56.276512 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 14:10:56 crc kubenswrapper[4805]: I1203 14:10:56.276531 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T14:10:56Z","lastTransitionTime":"2025-12-03T14:10:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 14:10:56 crc kubenswrapper[4805]: I1203 14:10:56.380002 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 14:10:56 crc kubenswrapper[4805]: I1203 14:10:56.380075 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 14:10:56 crc kubenswrapper[4805]: I1203 14:10:56.380096 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 14:10:56 crc kubenswrapper[4805]: I1203 14:10:56.380126 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 14:10:56 crc kubenswrapper[4805]: I1203 14:10:56.380147 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T14:10:56Z","lastTransitionTime":"2025-12-03T14:10:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 14:10:56 crc kubenswrapper[4805]: I1203 14:10:56.483466 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 14:10:56 crc kubenswrapper[4805]: I1203 14:10:56.483535 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 14:10:56 crc kubenswrapper[4805]: I1203 14:10:56.483557 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 14:10:56 crc kubenswrapper[4805]: I1203 14:10:56.483587 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 14:10:56 crc kubenswrapper[4805]: I1203 14:10:56.483611 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T14:10:56Z","lastTransitionTime":"2025-12-03T14:10:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 14:10:56 crc kubenswrapper[4805]: I1203 14:10:56.586673 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 14:10:56 crc kubenswrapper[4805]: I1203 14:10:56.586732 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 14:10:56 crc kubenswrapper[4805]: I1203 14:10:56.586748 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 14:10:56 crc kubenswrapper[4805]: I1203 14:10:56.586770 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 14:10:56 crc kubenswrapper[4805]: I1203 14:10:56.586787 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T14:10:56Z","lastTransitionTime":"2025-12-03T14:10:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 14:10:56 crc kubenswrapper[4805]: I1203 14:10:56.689788 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 14:10:56 crc kubenswrapper[4805]: I1203 14:10:56.689826 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 14:10:56 crc kubenswrapper[4805]: I1203 14:10:56.689857 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 14:10:56 crc kubenswrapper[4805]: I1203 14:10:56.689876 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 14:10:56 crc kubenswrapper[4805]: I1203 14:10:56.689887 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T14:10:56Z","lastTransitionTime":"2025-12-03T14:10:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 14:10:56 crc kubenswrapper[4805]: I1203 14:10:56.694512 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 14:10:56 crc kubenswrapper[4805]: I1203 14:10:56.694517 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 14:10:56 crc kubenswrapper[4805]: E1203 14:10:56.694595 4805 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 03 14:10:56 crc kubenswrapper[4805]: I1203 14:10:56.694529 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 14:10:56 crc kubenswrapper[4805]: E1203 14:10:56.694686 4805 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 03 14:10:56 crc kubenswrapper[4805]: I1203 14:10:56.694878 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-4gd9r" Dec 03 14:10:56 crc kubenswrapper[4805]: E1203 14:10:56.694911 4805 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 03 14:10:56 crc kubenswrapper[4805]: E1203 14:10:56.694953 4805 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-4gd9r" podUID="6b36a3db-1047-487a-9d6c-4f51ffdd8cc1" Dec 03 14:10:56 crc kubenswrapper[4805]: I1203 14:10:56.792408 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 14:10:56 crc kubenswrapper[4805]: I1203 14:10:56.792441 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 14:10:56 crc kubenswrapper[4805]: I1203 14:10:56.792451 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 14:10:56 crc kubenswrapper[4805]: I1203 14:10:56.792465 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 14:10:56 crc kubenswrapper[4805]: I1203 14:10:56.792476 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T14:10:56Z","lastTransitionTime":"2025-12-03T14:10:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 14:10:56 crc kubenswrapper[4805]: I1203 14:10:56.895341 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 14:10:56 crc kubenswrapper[4805]: I1203 14:10:56.895372 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 14:10:56 crc kubenswrapper[4805]: I1203 14:10:56.895383 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 14:10:56 crc kubenswrapper[4805]: I1203 14:10:56.895397 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 14:10:56 crc kubenswrapper[4805]: I1203 14:10:56.895407 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T14:10:56Z","lastTransitionTime":"2025-12-03T14:10:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 14:10:56 crc kubenswrapper[4805]: I1203 14:10:56.998280 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 14:10:56 crc kubenswrapper[4805]: I1203 14:10:56.998351 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 14:10:56 crc kubenswrapper[4805]: I1203 14:10:56.998371 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 14:10:56 crc kubenswrapper[4805]: I1203 14:10:56.998405 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 14:10:56 crc kubenswrapper[4805]: I1203 14:10:56.998422 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T14:10:56Z","lastTransitionTime":"2025-12-03T14:10:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 14:10:57 crc kubenswrapper[4805]: I1203 14:10:57.101726 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 14:10:57 crc kubenswrapper[4805]: I1203 14:10:57.101804 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 14:10:57 crc kubenswrapper[4805]: I1203 14:10:57.101823 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 14:10:57 crc kubenswrapper[4805]: I1203 14:10:57.102224 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 14:10:57 crc kubenswrapper[4805]: I1203 14:10:57.102270 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T14:10:57Z","lastTransitionTime":"2025-12-03T14:10:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 14:10:57 crc kubenswrapper[4805]: I1203 14:10:57.205033 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 14:10:57 crc kubenswrapper[4805]: I1203 14:10:57.205117 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 14:10:57 crc kubenswrapper[4805]: I1203 14:10:57.205134 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 14:10:57 crc kubenswrapper[4805]: I1203 14:10:57.205161 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 14:10:57 crc kubenswrapper[4805]: I1203 14:10:57.205181 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T14:10:57Z","lastTransitionTime":"2025-12-03T14:10:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 14:10:57 crc kubenswrapper[4805]: I1203 14:10:57.307826 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 14:10:57 crc kubenswrapper[4805]: I1203 14:10:57.307920 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 14:10:57 crc kubenswrapper[4805]: I1203 14:10:57.307941 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 14:10:57 crc kubenswrapper[4805]: I1203 14:10:57.307968 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 14:10:57 crc kubenswrapper[4805]: I1203 14:10:57.307988 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T14:10:57Z","lastTransitionTime":"2025-12-03T14:10:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 14:10:57 crc kubenswrapper[4805]: I1203 14:10:57.410688 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 14:10:57 crc kubenswrapper[4805]: I1203 14:10:57.410754 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 14:10:57 crc kubenswrapper[4805]: I1203 14:10:57.410775 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 14:10:57 crc kubenswrapper[4805]: I1203 14:10:57.410802 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 14:10:57 crc kubenswrapper[4805]: I1203 14:10:57.410820 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T14:10:57Z","lastTransitionTime":"2025-12-03T14:10:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 14:10:57 crc kubenswrapper[4805]: I1203 14:10:57.513506 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 14:10:57 crc kubenswrapper[4805]: I1203 14:10:57.513575 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 14:10:57 crc kubenswrapper[4805]: I1203 14:10:57.513596 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 14:10:57 crc kubenswrapper[4805]: I1203 14:10:57.513621 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 14:10:57 crc kubenswrapper[4805]: I1203 14:10:57.513641 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T14:10:57Z","lastTransitionTime":"2025-12-03T14:10:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 14:10:57 crc kubenswrapper[4805]: I1203 14:10:57.617461 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 14:10:57 crc kubenswrapper[4805]: I1203 14:10:57.617549 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 14:10:57 crc kubenswrapper[4805]: I1203 14:10:57.617576 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 14:10:57 crc kubenswrapper[4805]: I1203 14:10:57.617606 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 14:10:57 crc kubenswrapper[4805]: I1203 14:10:57.617670 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T14:10:57Z","lastTransitionTime":"2025-12-03T14:10:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 14:10:57 crc kubenswrapper[4805]: I1203 14:10:57.721950 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 14:10:57 crc kubenswrapper[4805]: I1203 14:10:57.722011 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 14:10:57 crc kubenswrapper[4805]: I1203 14:10:57.722019 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 14:10:57 crc kubenswrapper[4805]: I1203 14:10:57.722033 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 14:10:57 crc kubenswrapper[4805]: I1203 14:10:57.722042 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T14:10:57Z","lastTransitionTime":"2025-12-03T14:10:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 14:10:57 crc kubenswrapper[4805]: I1203 14:10:57.824679 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 14:10:57 crc kubenswrapper[4805]: I1203 14:10:57.824740 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 14:10:57 crc kubenswrapper[4805]: I1203 14:10:57.824758 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 14:10:57 crc kubenswrapper[4805]: I1203 14:10:57.824781 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 14:10:57 crc kubenswrapper[4805]: I1203 14:10:57.824798 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T14:10:57Z","lastTransitionTime":"2025-12-03T14:10:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 14:10:57 crc kubenswrapper[4805]: I1203 14:10:57.927707 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 14:10:57 crc kubenswrapper[4805]: I1203 14:10:57.927748 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 14:10:57 crc kubenswrapper[4805]: I1203 14:10:57.927762 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 14:10:57 crc kubenswrapper[4805]: I1203 14:10:57.927779 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 14:10:57 crc kubenswrapper[4805]: I1203 14:10:57.927791 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T14:10:57Z","lastTransitionTime":"2025-12-03T14:10:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 14:10:58 crc kubenswrapper[4805]: I1203 14:10:58.030105 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 14:10:58 crc kubenswrapper[4805]: I1203 14:10:58.030208 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 14:10:58 crc kubenswrapper[4805]: I1203 14:10:58.030223 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 14:10:58 crc kubenswrapper[4805]: I1203 14:10:58.030246 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 14:10:58 crc kubenswrapper[4805]: I1203 14:10:58.030263 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T14:10:58Z","lastTransitionTime":"2025-12-03T14:10:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 14:10:58 crc kubenswrapper[4805]: I1203 14:10:58.132723 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 14:10:58 crc kubenswrapper[4805]: I1203 14:10:58.132766 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 14:10:58 crc kubenswrapper[4805]: I1203 14:10:58.132776 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 14:10:58 crc kubenswrapper[4805]: I1203 14:10:58.132791 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 14:10:58 crc kubenswrapper[4805]: I1203 14:10:58.132802 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T14:10:58Z","lastTransitionTime":"2025-12-03T14:10:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 14:10:58 crc kubenswrapper[4805]: I1203 14:10:58.235551 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 14:10:58 crc kubenswrapper[4805]: I1203 14:10:58.235598 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 14:10:58 crc kubenswrapper[4805]: I1203 14:10:58.235608 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 14:10:58 crc kubenswrapper[4805]: I1203 14:10:58.235622 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 14:10:58 crc kubenswrapper[4805]: I1203 14:10:58.235632 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T14:10:58Z","lastTransitionTime":"2025-12-03T14:10:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 14:10:58 crc kubenswrapper[4805]: I1203 14:10:58.338113 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 14:10:58 crc kubenswrapper[4805]: I1203 14:10:58.338182 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 14:10:58 crc kubenswrapper[4805]: I1203 14:10:58.338206 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 14:10:58 crc kubenswrapper[4805]: I1203 14:10:58.338239 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 14:10:58 crc kubenswrapper[4805]: I1203 14:10:58.338266 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T14:10:58Z","lastTransitionTime":"2025-12-03T14:10:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 14:10:58 crc kubenswrapper[4805]: I1203 14:10:58.440057 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 14:10:58 crc kubenswrapper[4805]: I1203 14:10:58.440099 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 14:10:58 crc kubenswrapper[4805]: I1203 14:10:58.440110 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 14:10:58 crc kubenswrapper[4805]: I1203 14:10:58.440128 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 14:10:58 crc kubenswrapper[4805]: I1203 14:10:58.440140 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T14:10:58Z","lastTransitionTime":"2025-12-03T14:10:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 14:10:58 crc kubenswrapper[4805]: I1203 14:10:58.542460 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 14:10:58 crc kubenswrapper[4805]: I1203 14:10:58.542507 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 14:10:58 crc kubenswrapper[4805]: I1203 14:10:58.542519 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 14:10:58 crc kubenswrapper[4805]: I1203 14:10:58.542538 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 14:10:58 crc kubenswrapper[4805]: I1203 14:10:58.542552 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T14:10:58Z","lastTransitionTime":"2025-12-03T14:10:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 14:10:58 crc kubenswrapper[4805]: I1203 14:10:58.645643 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 14:10:58 crc kubenswrapper[4805]: I1203 14:10:58.645690 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 14:10:58 crc kubenswrapper[4805]: I1203 14:10:58.645702 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 14:10:58 crc kubenswrapper[4805]: I1203 14:10:58.645718 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 14:10:58 crc kubenswrapper[4805]: I1203 14:10:58.645733 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T14:10:58Z","lastTransitionTime":"2025-12-03T14:10:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 14:10:58 crc kubenswrapper[4805]: I1203 14:10:58.695007 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-4gd9r" Dec 03 14:10:58 crc kubenswrapper[4805]: I1203 14:10:58.695032 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 14:10:58 crc kubenswrapper[4805]: I1203 14:10:58.695140 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 14:10:58 crc kubenswrapper[4805]: E1203 14:10:58.695303 4805 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-4gd9r" podUID="6b36a3db-1047-487a-9d6c-4f51ffdd8cc1" Dec 03 14:10:58 crc kubenswrapper[4805]: I1203 14:10:58.695319 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 14:10:58 crc kubenswrapper[4805]: E1203 14:10:58.695431 4805 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 03 14:10:58 crc kubenswrapper[4805]: E1203 14:10:58.695523 4805 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 03 14:10:58 crc kubenswrapper[4805]: E1203 14:10:58.695588 4805 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 03 14:10:58 crc kubenswrapper[4805]: I1203 14:10:58.747781 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 14:10:58 crc kubenswrapper[4805]: I1203 14:10:58.747897 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 14:10:58 crc kubenswrapper[4805]: I1203 14:10:58.747920 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 14:10:58 crc kubenswrapper[4805]: I1203 14:10:58.747948 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 14:10:58 crc kubenswrapper[4805]: I1203 14:10:58.747969 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T14:10:58Z","lastTransitionTime":"2025-12-03T14:10:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 14:10:58 crc kubenswrapper[4805]: I1203 14:10:58.850585 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 14:10:58 crc kubenswrapper[4805]: I1203 14:10:58.850640 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 14:10:58 crc kubenswrapper[4805]: I1203 14:10:58.850650 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 14:10:58 crc kubenswrapper[4805]: I1203 14:10:58.850671 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 14:10:58 crc kubenswrapper[4805]: I1203 14:10:58.850685 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T14:10:58Z","lastTransitionTime":"2025-12-03T14:10:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 14:10:58 crc kubenswrapper[4805]: I1203 14:10:58.952977 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 14:10:58 crc kubenswrapper[4805]: I1203 14:10:58.953017 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 14:10:58 crc kubenswrapper[4805]: I1203 14:10:58.953030 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 14:10:58 crc kubenswrapper[4805]: I1203 14:10:58.953049 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 14:10:58 crc kubenswrapper[4805]: I1203 14:10:58.953061 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T14:10:58Z","lastTransitionTime":"2025-12-03T14:10:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 14:10:59 crc kubenswrapper[4805]: I1203 14:10:59.055720 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 14:10:59 crc kubenswrapper[4805]: I1203 14:10:59.056010 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 14:10:59 crc kubenswrapper[4805]: I1203 14:10:59.056087 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 14:10:59 crc kubenswrapper[4805]: I1203 14:10:59.056158 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 14:10:59 crc kubenswrapper[4805]: I1203 14:10:59.056214 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T14:10:59Z","lastTransitionTime":"2025-12-03T14:10:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 14:10:59 crc kubenswrapper[4805]: I1203 14:10:59.159098 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 14:10:59 crc kubenswrapper[4805]: I1203 14:10:59.159148 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 14:10:59 crc kubenswrapper[4805]: I1203 14:10:59.159157 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 14:10:59 crc kubenswrapper[4805]: I1203 14:10:59.159173 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 14:10:59 crc kubenswrapper[4805]: I1203 14:10:59.159184 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T14:10:59Z","lastTransitionTime":"2025-12-03T14:10:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 14:10:59 crc kubenswrapper[4805]: I1203 14:10:59.261470 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 14:10:59 crc kubenswrapper[4805]: I1203 14:10:59.261523 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 14:10:59 crc kubenswrapper[4805]: I1203 14:10:59.261535 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 14:10:59 crc kubenswrapper[4805]: I1203 14:10:59.261551 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 14:10:59 crc kubenswrapper[4805]: I1203 14:10:59.261561 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T14:10:59Z","lastTransitionTime":"2025-12-03T14:10:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 14:10:59 crc kubenswrapper[4805]: I1203 14:10:59.364325 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 14:10:59 crc kubenswrapper[4805]: I1203 14:10:59.364384 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 14:10:59 crc kubenswrapper[4805]: I1203 14:10:59.364401 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 14:10:59 crc kubenswrapper[4805]: I1203 14:10:59.364424 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 14:10:59 crc kubenswrapper[4805]: I1203 14:10:59.364443 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T14:10:59Z","lastTransitionTime":"2025-12-03T14:10:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 14:10:59 crc kubenswrapper[4805]: I1203 14:10:59.468161 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 14:10:59 crc kubenswrapper[4805]: I1203 14:10:59.468208 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 14:10:59 crc kubenswrapper[4805]: I1203 14:10:59.468221 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 14:10:59 crc kubenswrapper[4805]: I1203 14:10:59.468238 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 14:10:59 crc kubenswrapper[4805]: I1203 14:10:59.468250 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T14:10:59Z","lastTransitionTime":"2025-12-03T14:10:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 14:10:59 crc kubenswrapper[4805]: I1203 14:10:59.571551 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 14:10:59 crc kubenswrapper[4805]: I1203 14:10:59.571608 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 14:10:59 crc kubenswrapper[4805]: I1203 14:10:59.571620 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 14:10:59 crc kubenswrapper[4805]: I1203 14:10:59.571637 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 14:10:59 crc kubenswrapper[4805]: I1203 14:10:59.571654 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T14:10:59Z","lastTransitionTime":"2025-12-03T14:10:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 14:10:59 crc kubenswrapper[4805]: I1203 14:10:59.674719 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 14:10:59 crc kubenswrapper[4805]: I1203 14:10:59.674779 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 14:10:59 crc kubenswrapper[4805]: I1203 14:10:59.674789 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 14:10:59 crc kubenswrapper[4805]: I1203 14:10:59.674806 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 14:10:59 crc kubenswrapper[4805]: I1203 14:10:59.674814 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T14:10:59Z","lastTransitionTime":"2025-12-03T14:10:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 14:10:59 crc kubenswrapper[4805]: I1203 14:10:59.778430 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 14:10:59 crc kubenswrapper[4805]: I1203 14:10:59.778486 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 14:10:59 crc kubenswrapper[4805]: I1203 14:10:59.778503 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 14:10:59 crc kubenswrapper[4805]: I1203 14:10:59.778525 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 14:10:59 crc kubenswrapper[4805]: I1203 14:10:59.778543 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T14:10:59Z","lastTransitionTime":"2025-12-03T14:10:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 14:10:59 crc kubenswrapper[4805]: I1203 14:10:59.881184 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 14:10:59 crc kubenswrapper[4805]: I1203 14:10:59.881288 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 14:10:59 crc kubenswrapper[4805]: I1203 14:10:59.881309 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 14:10:59 crc kubenswrapper[4805]: I1203 14:10:59.881369 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 14:10:59 crc kubenswrapper[4805]: I1203 14:10:59.881392 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T14:10:59Z","lastTransitionTime":"2025-12-03T14:10:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 14:10:59 crc kubenswrapper[4805]: I1203 14:10:59.984278 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 14:10:59 crc kubenswrapper[4805]: I1203 14:10:59.984336 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 14:10:59 crc kubenswrapper[4805]: I1203 14:10:59.984349 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 14:10:59 crc kubenswrapper[4805]: I1203 14:10:59.984364 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 14:10:59 crc kubenswrapper[4805]: I1203 14:10:59.984373 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T14:10:59Z","lastTransitionTime":"2025-12-03T14:10:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 14:11:00 crc kubenswrapper[4805]: I1203 14:11:00.086943 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 14:11:00 crc kubenswrapper[4805]: I1203 14:11:00.086995 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 14:11:00 crc kubenswrapper[4805]: I1203 14:11:00.087004 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 14:11:00 crc kubenswrapper[4805]: I1203 14:11:00.087020 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 14:11:00 crc kubenswrapper[4805]: I1203 14:11:00.087031 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T14:11:00Z","lastTransitionTime":"2025-12-03T14:11:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 14:11:00 crc kubenswrapper[4805]: I1203 14:11:00.189158 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 14:11:00 crc kubenswrapper[4805]: I1203 14:11:00.189214 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 14:11:00 crc kubenswrapper[4805]: I1203 14:11:00.189225 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 14:11:00 crc kubenswrapper[4805]: I1203 14:11:00.189241 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 14:11:00 crc kubenswrapper[4805]: I1203 14:11:00.189253 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T14:11:00Z","lastTransitionTime":"2025-12-03T14:11:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 14:11:00 crc kubenswrapper[4805]: I1203 14:11:00.291808 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 14:11:00 crc kubenswrapper[4805]: I1203 14:11:00.291879 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 14:11:00 crc kubenswrapper[4805]: I1203 14:11:00.291888 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 14:11:00 crc kubenswrapper[4805]: I1203 14:11:00.291903 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 14:11:00 crc kubenswrapper[4805]: I1203 14:11:00.291912 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T14:11:00Z","lastTransitionTime":"2025-12-03T14:11:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 14:11:00 crc kubenswrapper[4805]: I1203 14:11:00.394744 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 14:11:00 crc kubenswrapper[4805]: I1203 14:11:00.394786 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 14:11:00 crc kubenswrapper[4805]: I1203 14:11:00.394798 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 14:11:00 crc kubenswrapper[4805]: I1203 14:11:00.394816 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 14:11:00 crc kubenswrapper[4805]: I1203 14:11:00.394827 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T14:11:00Z","lastTransitionTime":"2025-12-03T14:11:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 14:11:00 crc kubenswrapper[4805]: I1203 14:11:00.498147 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 14:11:00 crc kubenswrapper[4805]: I1203 14:11:00.498211 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 14:11:00 crc kubenswrapper[4805]: I1203 14:11:00.498227 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 14:11:00 crc kubenswrapper[4805]: I1203 14:11:00.498250 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 14:11:00 crc kubenswrapper[4805]: I1203 14:11:00.498267 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T14:11:00Z","lastTransitionTime":"2025-12-03T14:11:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 14:11:00 crc kubenswrapper[4805]: I1203 14:11:00.604905 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 14:11:00 crc kubenswrapper[4805]: I1203 14:11:00.604980 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 14:11:00 crc kubenswrapper[4805]: I1203 14:11:00.605003 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 14:11:00 crc kubenswrapper[4805]: I1203 14:11:00.605031 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 14:11:00 crc kubenswrapper[4805]: I1203 14:11:00.605053 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T14:11:00Z","lastTransitionTime":"2025-12-03T14:11:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 14:11:00 crc kubenswrapper[4805]: I1203 14:11:00.694144 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 14:11:00 crc kubenswrapper[4805]: I1203 14:11:00.694223 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-4gd9r" Dec 03 14:11:00 crc kubenswrapper[4805]: E1203 14:11:00.694635 4805 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 03 14:11:00 crc kubenswrapper[4805]: I1203 14:11:00.694715 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 14:11:00 crc kubenswrapper[4805]: I1203 14:11:00.694862 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 14:11:00 crc kubenswrapper[4805]: E1203 14:11:00.694924 4805 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-4gd9r" podUID="6b36a3db-1047-487a-9d6c-4f51ffdd8cc1" Dec 03 14:11:00 crc kubenswrapper[4805]: E1203 14:11:00.695047 4805 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 03 14:11:00 crc kubenswrapper[4805]: E1203 14:11:00.695154 4805 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 03 14:11:00 crc kubenswrapper[4805]: I1203 14:11:00.696164 4805 scope.go:117] "RemoveContainer" containerID="8f75988e25825e00abdb9e4b2292c5cc8ccab3dd028c882c2a38c796c0fd19c5" Dec 03 14:11:00 crc kubenswrapper[4805]: E1203 14:11:00.696435 4805 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-psjpm_openshift-ovn-kubernetes(deeec30d-cad2-4dfd-adb6-64f6646581be)\"" pod="openshift-ovn-kubernetes/ovnkube-node-psjpm" podUID="deeec30d-cad2-4dfd-adb6-64f6646581be" Dec 03 14:11:00 crc kubenswrapper[4805]: I1203 14:11:00.707377 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 14:11:00 crc kubenswrapper[4805]: I1203 14:11:00.707443 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 14:11:00 crc kubenswrapper[4805]: I1203 14:11:00.707465 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 14:11:00 crc kubenswrapper[4805]: I1203 14:11:00.707492 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 14:11:00 crc kubenswrapper[4805]: I1203 14:11:00.707514 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T14:11:00Z","lastTransitionTime":"2025-12-03T14:11:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 14:11:00 crc kubenswrapper[4805]: I1203 14:11:00.725997 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b6d36671-7173-4e9c-843f-005b4dcf576c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a837d7d491eaacbe4279b4014bb5fcbe9b514ba5cc82c708227bcd9b25a346e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6e5f04faf3b23197bb8ea11d3e574dc86f31e21d798734962001ec8c95de7c84\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2ff1a979387da41b00e95eb9cc8b2c9de966380a12fae687591df3c5b18a4497\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1e66e1cb18573a14e3c3ae16af4dbd9f8997a6ad38680c3ab4505646ebd48e7c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7d5b5c85e3de43e7e95111aa967248550607c26200c783251b68696f202f92a1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://474f72dac37a808fc4b9832306e76979b151d7221807ff3086013466ed1419c3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://474f72dac37a808fc4b9832306e76979b151d7221807ff3086013466ed1419c3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T14:09:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T14:09:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://74c8cea1a61c4775a0ad2f4ab4bb8a36437cf23d2ee953311d6ea967e297510a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://74c8cea1a61c4775a0ad2f4ab4bb8a36437cf23d2ee953311d6ea967e297510a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T14:09:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T14:09:31Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://a4f2f56a809adc01fba82eba0652d78f58770c132f77b1424f8b17068b1927e3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a4f2f56a809adc01fba82eba0652d78f58770c132f77b1424f8b17068b1927e3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T14:09:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T14:09:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T14:09:30Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T14:11:00Z is after 2025-08-24T17:21:41Z" Dec 03 14:11:00 crc kubenswrapper[4805]: I1203 14:11:00.744352 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ced8291d-0ce5-4e74-9eaf-2bf033774039\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:10:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:10:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3e5a34636f0a8c4d9179074596d28bc2eb7a504573398c596a637defd829907a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://29cd95be3f964415943c873e7914ffcea4abe50dfdf17f8a9cd03fc38df35981\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://951173b3d3d581275a6f6e43209b5e0611a701ca7a10f81504d723e24a2d21a3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e64730957bbbb2feb4d506e0ed9731a0594bf66fa1e8fcdb8ec944ded6308eed\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5da55f769e23ad52bac9c77da62c78e70fa34db8ed1a435e4eaf4dad3fd2b68a\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-03T14:09:48Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1203 14:09:42.992179 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1203 14:09:42.993233 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1830926909/tls.crt::/tmp/serving-cert-1830926909/tls.key\\\\\\\"\\\\nI1203 14:09:48.622403 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1203 14:09:48.626476 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1203 14:09:48.626527 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1203 14:09:48.626574 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1203 14:09:48.626582 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1203 14:09:48.633990 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1203 14:09:48.634024 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 14:09:48.634029 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 14:09:48.634034 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1203 14:09:48.634038 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1203 14:09:48.634042 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1203 14:09:48.634045 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1203 14:09:48.634344 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1203 14:09:48.644497 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T14:09:32Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bd92a160eb530c5687ebd7cd5b83b41820a2b808a0fd9514e5ace9d237e1bd26\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:32Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://af9c940135012c629447528c61743ed1e58708ab57ce76604c6da0479aa459cf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://af9c940135012c629447528c61743ed1e58708ab57ce76604c6da0479aa459cf\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T14:09:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T14:09:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T14:09:30Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T14:11:00Z is after 2025-08-24T17:21:41Z" Dec 03 14:11:00 crc kubenswrapper[4805]: I1203 14:11:00.761795 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6d48b4f1-5dfc-4f6b-a4e8-4e465c651c7b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:10:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:10:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5ea698a1e0de3b5201e2f9032d834d81d9aab9c2fb17aca582184c7090f1bd2f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://66aa257562f8e113fc1cd2936b8aa18e927f90018ad4c11a184c1c56444be648\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2fd6d50f839b603a90e5d00ab6e36e832eab2e299b7af1dbb84498c35bf1c08e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://77c26dae6ccd328a8a1ce4fe6ffe94c78061d8d3ac3727adb80a24c16f4ab533\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://77c26dae6ccd328a8a1ce4fe6ffe94c78061d8d3ac3727adb80a24c16f4ab533\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T14:09:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T14:09:31Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T14:09:30Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T14:11:00Z is after 2025-08-24T17:21:41Z" Dec 03 14:11:00 crc kubenswrapper[4805]: I1203 14:11:00.783073 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:49Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:49Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://38339331a3209d6787d1def81b72f4a8514bf91080a139cad6d8ac5cc6bef10e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T14:11:00Z is after 2025-08-24T17:21:41Z" Dec 03 14:11:00 crc kubenswrapper[4805]: I1203 14:11:00.808391 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:48Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:48Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T14:11:00Z is after 2025-08-24T17:21:41Z" Dec 03 14:11:00 crc kubenswrapper[4805]: I1203 14:11:00.810154 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 14:11:00 crc kubenswrapper[4805]: I1203 14:11:00.810215 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 14:11:00 crc kubenswrapper[4805]: I1203 14:11:00.810237 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 14:11:00 crc kubenswrapper[4805]: I1203 14:11:00.810265 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 14:11:00 crc kubenswrapper[4805]: I1203 14:11:00.810285 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T14:11:00Z","lastTransitionTime":"2025-12-03T14:11:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 14:11:00 crc kubenswrapper[4805]: I1203 14:11:00.824735 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:51Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:51Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b3123d3fa879097db88ed2f3cfb05631f440c24c5405503b2de3546f12e2e899\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T14:11:00Z is after 2025-08-24T17:21:41Z" Dec 03 14:11:00 crc kubenswrapper[4805]: I1203 14:11:00.836709 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-bdtdg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"53a0e9e3-b6d8-44b7-a32b-aa3450442865\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f557e17ad105d6972d3c1480656c6073f67e28c0f5773aa365549951f4d161c7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9ww54\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T14:09:51Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-bdtdg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T14:11:00Z is after 2025-08-24T17:21:41Z" Dec 03 14:11:00 crc kubenswrapper[4805]: I1203 14:11:00.852430 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"750d2b4b-738a-4992-8902-742528f6e875\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://abd2d0156113bb5182727a53aac594410ed051c19258beb8a07b808eff0b42a7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8673a29c0e16f4e707d79f0610722ccabb023ff2b4a18d7cf9d305183817d952\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://35766e7206b1be7a53d13d6890d759cf81dbd999643c12349268e4bf16fbbdb0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5b9844a11b2b903978d0d464b469e26c9211100da61d018dde9636dca95b684e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T14:09:30Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T14:11:00Z is after 2025-08-24T17:21:41Z" Dec 03 14:11:00 crc kubenswrapper[4805]: I1203 14:11:00.869895 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:48Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:48Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T14:11:00Z is after 2025-08-24T17:21:41Z" Dec 03 14:11:00 crc kubenswrapper[4805]: I1203 14:11:00.888184 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-k4mhs" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c09c1cdc-f461-45c9-8444-5b8764bccabc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:10:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:10:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7e077e19b6a5bf7d29b357982e12224d4b7058e0f9ac81b66b5a665c9bbf5014\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3354c1a44f891182d04116fc76234a2b337d50567d032a3584f902a9516387cc\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-03T14:10:36Z\\\",\\\"message\\\":\\\"2025-12-03T14:09:51+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_3e51ab9a-fbbe-4a78-9308-8a81e203fcad\\\\n2025-12-03T14:09:51+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_3e51ab9a-fbbe-4a78-9308-8a81e203fcad to /host/opt/cni/bin/\\\\n2025-12-03T14:09:51Z [verbose] multus-daemon started\\\\n2025-12-03T14:09:51Z [verbose] Readiness Indicator file check\\\\n2025-12-03T14:10:36Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T14:09:50Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:10:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qdcd7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T14:09:50Z\\\"}}\" for pod \"openshift-multus\"/\"multus-k4mhs\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T14:11:00Z is after 2025-08-24T17:21:41Z" Dec 03 14:11:00 crc kubenswrapper[4805]: I1203 14:11:00.909559 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-bpk49" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"de00a81f-82e1-4223-9614-f98606b9a8ef\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://04f4ad27e02e2024ae9e7cd1b16116267e69cecda882dd723cdb274c41a0e7db\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-25d7t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7741e0e70416ebe094bfabc6009e98d7d4c0fbcba1fc82df13b9ba40774b456a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7741e0e70416ebe094bfabc6009e98d7d4c0fbcba1fc82df13b9ba40774b456a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T14:09:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T14:09:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-25d7t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ee3499c02eb52bbf08c447d2433d435a74dfb70670dc4cbaac44f5fbff9e8a29\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ee3499c02eb52bbf08c447d2433d435a74dfb70670dc4cbaac44f5fbff9e8a29\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T14:09:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T14:09:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-25d7t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ec8488b1d69799951be7ef40e5e028732265c4ddcdea2d453923e567864771f7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ec8488b1d69799951be7ef40e5e028732265c4ddcdea2d453923e567864771f7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T14:09:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T14:09:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-25d7t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://66707bb4e1fdacb265e89786df8a04605830f1c156394cacfcec9b54abe5d41a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://66707bb4e1fdacb265e89786df8a04605830f1c156394cacfcec9b54abe5d41a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T14:09:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T14:09:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-25d7t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6207ab6600970ff15c8713c547d4920b2c77eb9bea990bb5da497e24870dcc0f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6207ab6600970ff15c8713c547d4920b2c77eb9bea990bb5da497e24870dcc0f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T14:09:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T14:09:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-25d7t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8de65d4a205e2deec0b433e53c0e37feb94aa0ede6a4c58d7414b80829804505\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8de65d4a205e2deec0b433e53c0e37feb94aa0ede6a4c58d7414b80829804505\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T14:09:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T14:09:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-25d7t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T14:09:50Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-bpk49\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T14:11:00Z is after 2025-08-24T17:21:41Z" Dec 03 14:11:00 crc kubenswrapper[4805]: I1203 14:11:00.913484 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 14:11:00 crc kubenswrapper[4805]: I1203 14:11:00.913532 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 14:11:00 crc kubenswrapper[4805]: I1203 14:11:00.913542 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 14:11:00 crc kubenswrapper[4805]: I1203 14:11:00.913561 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 14:11:00 crc kubenswrapper[4805]: I1203 14:11:00.913577 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T14:11:00Z","lastTransitionTime":"2025-12-03T14:11:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 14:11:00 crc kubenswrapper[4805]: I1203 14:11:00.927311 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:48Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:48Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T14:11:00Z is after 2025-08-24T17:21:41Z" Dec 03 14:11:00 crc kubenswrapper[4805]: I1203 14:11:00.938498 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:49Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:49Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://541ace7ccbb66dbd191c3ca8c1e4479e3edbf80dfc67326bea8aaee81fdcfc3c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://821e3e9a902aacecf6c41e8b8d8aa1ac951ce2fbb6505b7148b16348a3754e55\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T14:11:00Z is after 2025-08-24T17:21:41Z" Dec 03 14:11:00 crc kubenswrapper[4805]: I1203 14:11:00.948115 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-gskh4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6098937f-e3f6-45e8-a647-4994a79cd711\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://28d5653f26ec61647808b3a733ff7eac70aa69ac184660f24147e8cfaf0eb802\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nwrw9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9ad6b68a42fd33ee167e67873c6eed3f11cdf6dc71d22e952c021436eb967dd9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nwrw9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T14:09:50Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-gskh4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T14:11:00Z is after 2025-08-24T17:21:41Z" Dec 03 14:11:00 crc kubenswrapper[4805]: I1203 14:11:00.967966 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-psjpm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"deeec30d-cad2-4dfd-adb6-64f6646581be\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:50Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:50Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4b8d00cdf0f37f2e5cd4e952cf85d3de8d72c43fdf786088fd9f26ed17b1e21f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wdlzx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bd67c6de01a8d196a1f51d4d90640c9e69de6f05d5e201b25e6eaa5c33440ea8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wdlzx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cb8bf05ef093eb0727a066d4bbbdf9af9422867cd4dadacafc6eeb016867a61e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wdlzx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0304467102ed053628c0d3e06548c74b6492336b6fd044246508d1e8bdb17bc3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wdlzx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6ce80ccafb3262a11fbe9923317a9d38b61e5dc5b3e7e7ae136f143d6b643fd7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wdlzx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://47d76962beb8c0f196d50af96fb3ec611097823ca80a1c8bd718f2d18840128c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wdlzx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8f75988e25825e00abdb9e4b2292c5cc8ccab3dd028c882c2a38c796c0fd19c5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a0272667e733f2c5eec0fce0eb9c31a16c048babdaed90635895d689ae9303f4\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-03T14:10:16Z\\\",\\\"message\\\":\\\"o create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T14:10:16Z is after 2025-08-24T17:21:41Z]\\\\nI1203 14:10:16.717659 6478 services_controller.go:473] Services do not match for network=default, existing lbs: []services.LB{services.LB{Name:\\\\\\\"Service_openshift-network-console/networking-console-plugin_TCP_cluster\\\\\\\", UUID:\\\\\\\"ab0b1d51-5ec6-479b-8881-93dfa8d30337\\\\\\\", Protocol:\\\\\\\"tcp\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-network-console/networking-console-plugin\\\\\\\"}, Opts:services.LBOpts{Reject:false, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{}, Templates:services.TemplateMap{}, Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGr\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T14:10:15Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8f75988e25825e00abdb9e4b2292c5cc8ccab3dd028c882c2a38c796c0fd19c5\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-03T14:10:49Z\\\",\\\"message\\\":\\\"shift-kube-storage-version-migrator-operator/metrics]} name:Service_openshift-kube-storage-version-migrator-operator/metrics_TCP_cluster options:{GoMap:map[event:false hairpin_snat_ip:169.254.0.5 fd69::5 neighbor_responder:none reject:true skip_snat:false]} protocol:{GoSet:[tcp]} selection_fields:{GoSet:[]} vips:{GoMap:map[10.217.5.36:443:]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {7f9b8f25-db1a-4d02-a423-9afc5c2fb83c}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nF1203 14:10:49.548775 6841 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T14:10:49Z is after 2025-08-24T17:\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T14:10:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wdlzx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ab98e365c1a4242a47b8cbe2dd85109d366d6002c4de1f18ddc950a46bea6c12\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wdlzx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2f875252fbe6a673a72fc9c56bb96899f62732305ad77b3108e3b244bfa91ecb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2f875252fbe6a673a72fc9c56bb96899f62732305ad77b3108e3b244bfa91ecb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T14:09:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T14:09:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wdlzx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T14:09:50Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-psjpm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T14:11:00Z is after 2025-08-24T17:21:41Z" Dec 03 14:11:00 crc kubenswrapper[4805]: I1203 14:11:00.982285 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-4gd9r" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6b36a3db-1047-487a-9d6c-4f51ffdd8cc1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:10:04Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:10:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:10:04Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:10:04Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4t6zl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4t6zl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T14:10:04Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-4gd9r\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T14:11:00Z is after 2025-08-24T17:21:41Z" Dec 03 14:11:00 crc kubenswrapper[4805]: I1203 14:11:00.995069 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7c462309-e492-44c8-923b-4e37fd6a98f1\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f3942cab663eda9769efb606b01dba99630de2e30b77a7888c8a57a2a844ee0d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0d86b4e22a0aec3b88cf3d60d37796252dc122c2c89dfa9f23123110c7b6e9be\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0d86b4e22a0aec3b88cf3d60d37796252dc122c2c89dfa9f23123110c7b6e9be\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T14:09:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T14:09:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T14:09:30Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T14:11:00Z is after 2025-08-24T17:21:41Z" Dec 03 14:11:01 crc kubenswrapper[4805]: I1203 14:11:01.007159 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-pdzgk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"860769de-6a5b-4cb9-a752-264e7e2e785f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5e2f0a0cc21505d609f81918365d5e4f5d21959ac12ab147d6a51601c9de4c14\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ggn4n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T14:09:50Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-pdzgk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T14:11:01Z is after 2025-08-24T17:21:41Z" Dec 03 14:11:01 crc kubenswrapper[4805]: I1203 14:11:01.015122 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 14:11:01 crc kubenswrapper[4805]: I1203 14:11:01.015304 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 14:11:01 crc kubenswrapper[4805]: I1203 14:11:01.015385 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 14:11:01 crc kubenswrapper[4805]: I1203 14:11:01.015470 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 14:11:01 crc kubenswrapper[4805]: I1203 14:11:01.015553 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T14:11:01Z","lastTransitionTime":"2025-12-03T14:11:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 14:11:01 crc kubenswrapper[4805]: I1203 14:11:01.019454 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-69h75" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5d1c694f-9847-4086-8b37-3573ddeb3ace\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:10:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:10:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:10:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:10:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://abc1d9b8582ca9fe4a51fb608fb39129a179ddcdfde0797a6369a55b2e0124cf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:10:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s578b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c7c4d434b035cac2a91a4f0846a2f4fb5bd68f881c28487860adb6db85e8ce75\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:10:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s578b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T14:10:03Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-69h75\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T14:11:01Z is after 2025-08-24T17:21:41Z" Dec 03 14:11:01 crc kubenswrapper[4805]: I1203 14:11:01.031796 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"750d2b4b-738a-4992-8902-742528f6e875\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://abd2d0156113bb5182727a53aac594410ed051c19258beb8a07b808eff0b42a7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8673a29c0e16f4e707d79f0610722ccabb023ff2b4a18d7cf9d305183817d952\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://35766e7206b1be7a53d13d6890d759cf81dbd999643c12349268e4bf16fbbdb0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5b9844a11b2b903978d0d464b469e26c9211100da61d018dde9636dca95b684e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T14:09:30Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T14:11:01Z is after 2025-08-24T17:21:41Z" Dec 03 14:11:01 crc kubenswrapper[4805]: I1203 14:11:01.046692 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:48Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:48Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T14:11:01Z is after 2025-08-24T17:21:41Z" Dec 03 14:11:01 crc kubenswrapper[4805]: I1203 14:11:01.061880 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-k4mhs" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c09c1cdc-f461-45c9-8444-5b8764bccabc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:10:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:10:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7e077e19b6a5bf7d29b357982e12224d4b7058e0f9ac81b66b5a665c9bbf5014\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3354c1a44f891182d04116fc76234a2b337d50567d032a3584f902a9516387cc\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-03T14:10:36Z\\\",\\\"message\\\":\\\"2025-12-03T14:09:51+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_3e51ab9a-fbbe-4a78-9308-8a81e203fcad\\\\n2025-12-03T14:09:51+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_3e51ab9a-fbbe-4a78-9308-8a81e203fcad to /host/opt/cni/bin/\\\\n2025-12-03T14:09:51Z [verbose] multus-daemon started\\\\n2025-12-03T14:09:51Z [verbose] Readiness Indicator file check\\\\n2025-12-03T14:10:36Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T14:09:50Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:10:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qdcd7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T14:09:50Z\\\"}}\" for pod \"openshift-multus\"/\"multus-k4mhs\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T14:11:01Z is after 2025-08-24T17:21:41Z" Dec 03 14:11:01 crc kubenswrapper[4805]: I1203 14:11:01.083336 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-bpk49" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"de00a81f-82e1-4223-9614-f98606b9a8ef\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://04f4ad27e02e2024ae9e7cd1b16116267e69cecda882dd723cdb274c41a0e7db\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-25d7t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7741e0e70416ebe094bfabc6009e98d7d4c0fbcba1fc82df13b9ba40774b456a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7741e0e70416ebe094bfabc6009e98d7d4c0fbcba1fc82df13b9ba40774b456a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T14:09:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T14:09:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-25d7t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ee3499c02eb52bbf08c447d2433d435a74dfb70670dc4cbaac44f5fbff9e8a29\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ee3499c02eb52bbf08c447d2433d435a74dfb70670dc4cbaac44f5fbff9e8a29\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T14:09:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T14:09:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-25d7t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ec8488b1d69799951be7ef40e5e028732265c4ddcdea2d453923e567864771f7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ec8488b1d69799951be7ef40e5e028732265c4ddcdea2d453923e567864771f7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T14:09:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T14:09:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-25d7t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://66707bb4e1fdacb265e89786df8a04605830f1c156394cacfcec9b54abe5d41a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://66707bb4e1fdacb265e89786df8a04605830f1c156394cacfcec9b54abe5d41a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T14:09:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T14:09:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-25d7t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6207ab6600970ff15c8713c547d4920b2c77eb9bea990bb5da497e24870dcc0f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6207ab6600970ff15c8713c547d4920b2c77eb9bea990bb5da497e24870dcc0f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T14:09:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T14:09:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-25d7t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8de65d4a205e2deec0b433e53c0e37feb94aa0ede6a4c58d7414b80829804505\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8de65d4a205e2deec0b433e53c0e37feb94aa0ede6a4c58d7414b80829804505\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T14:09:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T14:09:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-25d7t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T14:09:50Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-bpk49\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T14:11:01Z is after 2025-08-24T17:21:41Z" Dec 03 14:11:01 crc kubenswrapper[4805]: I1203 14:11:01.100473 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:48Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:48Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T14:11:01Z is after 2025-08-24T17:21:41Z" Dec 03 14:11:01 crc kubenswrapper[4805]: I1203 14:11:01.118147 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 14:11:01 crc kubenswrapper[4805]: I1203 14:11:01.118193 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 14:11:01 crc kubenswrapper[4805]: I1203 14:11:01.118201 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 14:11:01 crc kubenswrapper[4805]: I1203 14:11:01.118216 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 14:11:01 crc kubenswrapper[4805]: I1203 14:11:01.118225 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T14:11:01Z","lastTransitionTime":"2025-12-03T14:11:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 14:11:01 crc kubenswrapper[4805]: I1203 14:11:01.120821 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:49Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:49Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://541ace7ccbb66dbd191c3ca8c1e4479e3edbf80dfc67326bea8aaee81fdcfc3c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://821e3e9a902aacecf6c41e8b8d8aa1ac951ce2fbb6505b7148b16348a3754e55\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T14:11:01Z is after 2025-08-24T17:21:41Z" Dec 03 14:11:01 crc kubenswrapper[4805]: I1203 14:11:01.133788 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-gskh4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6098937f-e3f6-45e8-a647-4994a79cd711\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://28d5653f26ec61647808b3a733ff7eac70aa69ac184660f24147e8cfaf0eb802\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nwrw9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9ad6b68a42fd33ee167e67873c6eed3f11cdf6dc71d22e952c021436eb967dd9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nwrw9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T14:09:50Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-gskh4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T14:11:01Z is after 2025-08-24T17:21:41Z" Dec 03 14:11:01 crc kubenswrapper[4805]: I1203 14:11:01.161769 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-psjpm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"deeec30d-cad2-4dfd-adb6-64f6646581be\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:50Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:50Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4b8d00cdf0f37f2e5cd4e952cf85d3de8d72c43fdf786088fd9f26ed17b1e21f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wdlzx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bd67c6de01a8d196a1f51d4d90640c9e69de6f05d5e201b25e6eaa5c33440ea8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wdlzx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cb8bf05ef093eb0727a066d4bbbdf9af9422867cd4dadacafc6eeb016867a61e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wdlzx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0304467102ed053628c0d3e06548c74b6492336b6fd044246508d1e8bdb17bc3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wdlzx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6ce80ccafb3262a11fbe9923317a9d38b61e5dc5b3e7e7ae136f143d6b643fd7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wdlzx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://47d76962beb8c0f196d50af96fb3ec611097823ca80a1c8bd718f2d18840128c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wdlzx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8f75988e25825e00abdb9e4b2292c5cc8ccab3dd028c882c2a38c796c0fd19c5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8f75988e25825e00abdb9e4b2292c5cc8ccab3dd028c882c2a38c796c0fd19c5\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-03T14:10:49Z\\\",\\\"message\\\":\\\"shift-kube-storage-version-migrator-operator/metrics]} name:Service_openshift-kube-storage-version-migrator-operator/metrics_TCP_cluster options:{GoMap:map[event:false hairpin_snat_ip:169.254.0.5 fd69::5 neighbor_responder:none reject:true skip_snat:false]} protocol:{GoSet:[tcp]} selection_fields:{GoSet:[]} vips:{GoMap:map[10.217.5.36:443:]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {7f9b8f25-db1a-4d02-a423-9afc5c2fb83c}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nF1203 14:10:49.548775 6841 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T14:10:49Z is after 2025-08-24T17:\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T14:10:48Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-psjpm_openshift-ovn-kubernetes(deeec30d-cad2-4dfd-adb6-64f6646581be)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wdlzx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ab98e365c1a4242a47b8cbe2dd85109d366d6002c4de1f18ddc950a46bea6c12\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wdlzx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2f875252fbe6a673a72fc9c56bb96899f62732305ad77b3108e3b244bfa91ecb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2f875252fbe6a673a72fc9c56bb96899f62732305ad77b3108e3b244bfa91ecb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T14:09:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T14:09:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wdlzx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T14:09:50Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-psjpm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T14:11:01Z is after 2025-08-24T17:21:41Z" Dec 03 14:11:01 crc kubenswrapper[4805]: I1203 14:11:01.179327 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-4gd9r" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6b36a3db-1047-487a-9d6c-4f51ffdd8cc1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:10:04Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:10:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:10:04Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:10:04Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4t6zl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4t6zl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T14:10:04Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-4gd9r\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T14:11:01Z is after 2025-08-24T17:21:41Z" Dec 03 14:11:01 crc kubenswrapper[4805]: I1203 14:11:01.197232 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7c462309-e492-44c8-923b-4e37fd6a98f1\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f3942cab663eda9769efb606b01dba99630de2e30b77a7888c8a57a2a844ee0d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0d86b4e22a0aec3b88cf3d60d37796252dc122c2c89dfa9f23123110c7b6e9be\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0d86b4e22a0aec3b88cf3d60d37796252dc122c2c89dfa9f23123110c7b6e9be\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T14:09:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T14:09:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T14:09:30Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T14:11:01Z is after 2025-08-24T17:21:41Z" Dec 03 14:11:01 crc kubenswrapper[4805]: I1203 14:11:01.209692 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-pdzgk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"860769de-6a5b-4cb9-a752-264e7e2e785f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5e2f0a0cc21505d609f81918365d5e4f5d21959ac12ab147d6a51601c9de4c14\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ggn4n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T14:09:50Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-pdzgk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T14:11:01Z is after 2025-08-24T17:21:41Z" Dec 03 14:11:01 crc kubenswrapper[4805]: I1203 14:11:01.221172 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 14:11:01 crc kubenswrapper[4805]: I1203 14:11:01.221220 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 14:11:01 crc kubenswrapper[4805]: I1203 14:11:01.221231 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 14:11:01 crc kubenswrapper[4805]: I1203 14:11:01.221249 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 14:11:01 crc kubenswrapper[4805]: I1203 14:11:01.221260 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T14:11:01Z","lastTransitionTime":"2025-12-03T14:11:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 14:11:01 crc kubenswrapper[4805]: I1203 14:11:01.225056 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-69h75" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5d1c694f-9847-4086-8b37-3573ddeb3ace\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:10:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:10:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:10:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:10:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://abc1d9b8582ca9fe4a51fb608fb39129a179ddcdfde0797a6369a55b2e0124cf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:10:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s578b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c7c4d434b035cac2a91a4f0846a2f4fb5bd68f881c28487860adb6db85e8ce75\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:10:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s578b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T14:10:03Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-69h75\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T14:11:01Z is after 2025-08-24T17:21:41Z" Dec 03 14:11:01 crc kubenswrapper[4805]: I1203 14:11:01.256042 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b6d36671-7173-4e9c-843f-005b4dcf576c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a837d7d491eaacbe4279b4014bb5fcbe9b514ba5cc82c708227bcd9b25a346e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6e5f04faf3b23197bb8ea11d3e574dc86f31e21d798734962001ec8c95de7c84\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2ff1a979387da41b00e95eb9cc8b2c9de966380a12fae687591df3c5b18a4497\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1e66e1cb18573a14e3c3ae16af4dbd9f8997a6ad38680c3ab4505646ebd48e7c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7d5b5c85e3de43e7e95111aa967248550607c26200c783251b68696f202f92a1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://474f72dac37a808fc4b9832306e76979b151d7221807ff3086013466ed1419c3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://474f72dac37a808fc4b9832306e76979b151d7221807ff3086013466ed1419c3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T14:09:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T14:09:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://74c8cea1a61c4775a0ad2f4ab4bb8a36437cf23d2ee953311d6ea967e297510a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://74c8cea1a61c4775a0ad2f4ab4bb8a36437cf23d2ee953311d6ea967e297510a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T14:09:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T14:09:31Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://a4f2f56a809adc01fba82eba0652d78f58770c132f77b1424f8b17068b1927e3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a4f2f56a809adc01fba82eba0652d78f58770c132f77b1424f8b17068b1927e3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T14:09:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T14:09:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T14:09:30Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T14:11:01Z is after 2025-08-24T17:21:41Z" Dec 03 14:11:01 crc kubenswrapper[4805]: I1203 14:11:01.270592 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ced8291d-0ce5-4e74-9eaf-2bf033774039\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:10:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:10:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3e5a34636f0a8c4d9179074596d28bc2eb7a504573398c596a637defd829907a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://29cd95be3f964415943c873e7914ffcea4abe50dfdf17f8a9cd03fc38df35981\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://951173b3d3d581275a6f6e43209b5e0611a701ca7a10f81504d723e24a2d21a3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e64730957bbbb2feb4d506e0ed9731a0594bf66fa1e8fcdb8ec944ded6308eed\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5da55f769e23ad52bac9c77da62c78e70fa34db8ed1a435e4eaf4dad3fd2b68a\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-03T14:09:48Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1203 14:09:42.992179 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1203 14:09:42.993233 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1830926909/tls.crt::/tmp/serving-cert-1830926909/tls.key\\\\\\\"\\\\nI1203 14:09:48.622403 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1203 14:09:48.626476 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1203 14:09:48.626527 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1203 14:09:48.626574 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1203 14:09:48.626582 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1203 14:09:48.633990 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1203 14:09:48.634024 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 14:09:48.634029 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 14:09:48.634034 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1203 14:09:48.634038 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1203 14:09:48.634042 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1203 14:09:48.634045 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1203 14:09:48.634344 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1203 14:09:48.644497 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T14:09:32Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bd92a160eb530c5687ebd7cd5b83b41820a2b808a0fd9514e5ace9d237e1bd26\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:32Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://af9c940135012c629447528c61743ed1e58708ab57ce76604c6da0479aa459cf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://af9c940135012c629447528c61743ed1e58708ab57ce76604c6da0479aa459cf\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T14:09:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T14:09:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T14:09:30Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T14:11:01Z is after 2025-08-24T17:21:41Z" Dec 03 14:11:01 crc kubenswrapper[4805]: I1203 14:11:01.284162 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6d48b4f1-5dfc-4f6b-a4e8-4e465c651c7b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:10:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:10:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5ea698a1e0de3b5201e2f9032d834d81d9aab9c2fb17aca582184c7090f1bd2f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://66aa257562f8e113fc1cd2936b8aa18e927f90018ad4c11a184c1c56444be648\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2fd6d50f839b603a90e5d00ab6e36e832eab2e299b7af1dbb84498c35bf1c08e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://77c26dae6ccd328a8a1ce4fe6ffe94c78061d8d3ac3727adb80a24c16f4ab533\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://77c26dae6ccd328a8a1ce4fe6ffe94c78061d8d3ac3727adb80a24c16f4ab533\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T14:09:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T14:09:31Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T14:09:30Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T14:11:01Z is after 2025-08-24T17:21:41Z" Dec 03 14:11:01 crc kubenswrapper[4805]: I1203 14:11:01.305328 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:49Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:49Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://38339331a3209d6787d1def81b72f4a8514bf91080a139cad6d8ac5cc6bef10e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T14:11:01Z is after 2025-08-24T17:21:41Z" Dec 03 14:11:01 crc kubenswrapper[4805]: I1203 14:11:01.323419 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:48Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:48Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T14:11:01Z is after 2025-08-24T17:21:41Z" Dec 03 14:11:01 crc kubenswrapper[4805]: I1203 14:11:01.324526 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 14:11:01 crc kubenswrapper[4805]: I1203 14:11:01.324617 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 14:11:01 crc kubenswrapper[4805]: I1203 14:11:01.324661 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 14:11:01 crc kubenswrapper[4805]: I1203 14:11:01.324689 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 14:11:01 crc kubenswrapper[4805]: I1203 14:11:01.324705 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T14:11:01Z","lastTransitionTime":"2025-12-03T14:11:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 14:11:01 crc kubenswrapper[4805]: I1203 14:11:01.336894 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:51Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:51Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b3123d3fa879097db88ed2f3cfb05631f440c24c5405503b2de3546f12e2e899\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T14:11:01Z is after 2025-08-24T17:21:41Z" Dec 03 14:11:01 crc kubenswrapper[4805]: I1203 14:11:01.349659 4805 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-bdtdg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"53a0e9e3-b6d8-44b7-a32b-aa3450442865\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T14:09:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f557e17ad105d6972d3c1480656c6073f67e28c0f5773aa365549951f4d161c7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T14:09:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9ww54\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T14:09:51Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-bdtdg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T14:11:01Z is after 2025-08-24T17:21:41Z" Dec 03 14:11:01 crc kubenswrapper[4805]: I1203 14:11:01.427148 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 14:11:01 crc kubenswrapper[4805]: I1203 14:11:01.427196 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 14:11:01 crc kubenswrapper[4805]: I1203 14:11:01.427207 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 14:11:01 crc kubenswrapper[4805]: I1203 14:11:01.427226 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 14:11:01 crc kubenswrapper[4805]: I1203 14:11:01.427237 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T14:11:01Z","lastTransitionTime":"2025-12-03T14:11:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 14:11:01 crc kubenswrapper[4805]: I1203 14:11:01.529427 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 14:11:01 crc kubenswrapper[4805]: I1203 14:11:01.529477 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 14:11:01 crc kubenswrapper[4805]: I1203 14:11:01.529489 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 14:11:01 crc kubenswrapper[4805]: I1203 14:11:01.529508 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 14:11:01 crc kubenswrapper[4805]: I1203 14:11:01.529540 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T14:11:01Z","lastTransitionTime":"2025-12-03T14:11:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 14:11:01 crc kubenswrapper[4805]: I1203 14:11:01.632589 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 14:11:01 crc kubenswrapper[4805]: I1203 14:11:01.632691 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 14:11:01 crc kubenswrapper[4805]: I1203 14:11:01.632754 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 14:11:01 crc kubenswrapper[4805]: I1203 14:11:01.632778 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 14:11:01 crc kubenswrapper[4805]: I1203 14:11:01.632868 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T14:11:01Z","lastTransitionTime":"2025-12-03T14:11:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 14:11:01 crc kubenswrapper[4805]: I1203 14:11:01.735403 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 14:11:01 crc kubenswrapper[4805]: I1203 14:11:01.735456 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 14:11:01 crc kubenswrapper[4805]: I1203 14:11:01.735477 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 14:11:01 crc kubenswrapper[4805]: I1203 14:11:01.735505 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 14:11:01 crc kubenswrapper[4805]: I1203 14:11:01.735525 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T14:11:01Z","lastTransitionTime":"2025-12-03T14:11:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 14:11:01 crc kubenswrapper[4805]: I1203 14:11:01.838666 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 14:11:01 crc kubenswrapper[4805]: I1203 14:11:01.838721 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 14:11:01 crc kubenswrapper[4805]: I1203 14:11:01.838739 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 14:11:01 crc kubenswrapper[4805]: I1203 14:11:01.838765 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 14:11:01 crc kubenswrapper[4805]: I1203 14:11:01.838786 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T14:11:01Z","lastTransitionTime":"2025-12-03T14:11:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 14:11:01 crc kubenswrapper[4805]: I1203 14:11:01.942142 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 14:11:01 crc kubenswrapper[4805]: I1203 14:11:01.942204 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 14:11:01 crc kubenswrapper[4805]: I1203 14:11:01.942221 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 14:11:01 crc kubenswrapper[4805]: I1203 14:11:01.942245 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 14:11:01 crc kubenswrapper[4805]: I1203 14:11:01.942263 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T14:11:01Z","lastTransitionTime":"2025-12-03T14:11:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 14:11:02 crc kubenswrapper[4805]: I1203 14:11:02.044699 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 14:11:02 crc kubenswrapper[4805]: I1203 14:11:02.044733 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 14:11:02 crc kubenswrapper[4805]: I1203 14:11:02.044743 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 14:11:02 crc kubenswrapper[4805]: I1203 14:11:02.044762 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 14:11:02 crc kubenswrapper[4805]: I1203 14:11:02.044771 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T14:11:02Z","lastTransitionTime":"2025-12-03T14:11:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 14:11:02 crc kubenswrapper[4805]: I1203 14:11:02.147687 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 14:11:02 crc kubenswrapper[4805]: I1203 14:11:02.147739 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 14:11:02 crc kubenswrapper[4805]: I1203 14:11:02.147756 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 14:11:02 crc kubenswrapper[4805]: I1203 14:11:02.147779 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 14:11:02 crc kubenswrapper[4805]: I1203 14:11:02.147799 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T14:11:02Z","lastTransitionTime":"2025-12-03T14:11:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 14:11:02 crc kubenswrapper[4805]: I1203 14:11:02.250622 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 14:11:02 crc kubenswrapper[4805]: I1203 14:11:02.250697 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 14:11:02 crc kubenswrapper[4805]: I1203 14:11:02.250706 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 14:11:02 crc kubenswrapper[4805]: I1203 14:11:02.250718 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 14:11:02 crc kubenswrapper[4805]: I1203 14:11:02.250727 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T14:11:02Z","lastTransitionTime":"2025-12-03T14:11:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 14:11:02 crc kubenswrapper[4805]: I1203 14:11:02.353871 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 14:11:02 crc kubenswrapper[4805]: I1203 14:11:02.353932 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 14:11:02 crc kubenswrapper[4805]: I1203 14:11:02.353956 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 14:11:02 crc kubenswrapper[4805]: I1203 14:11:02.353982 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 14:11:02 crc kubenswrapper[4805]: I1203 14:11:02.354001 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T14:11:02Z","lastTransitionTime":"2025-12-03T14:11:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 14:11:02 crc kubenswrapper[4805]: I1203 14:11:02.455676 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 14:11:02 crc kubenswrapper[4805]: I1203 14:11:02.455721 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 14:11:02 crc kubenswrapper[4805]: I1203 14:11:02.455730 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 14:11:02 crc kubenswrapper[4805]: I1203 14:11:02.455744 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 14:11:02 crc kubenswrapper[4805]: I1203 14:11:02.455754 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T14:11:02Z","lastTransitionTime":"2025-12-03T14:11:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 14:11:02 crc kubenswrapper[4805]: I1203 14:11:02.559128 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 14:11:02 crc kubenswrapper[4805]: I1203 14:11:02.559183 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 14:11:02 crc kubenswrapper[4805]: I1203 14:11:02.559199 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 14:11:02 crc kubenswrapper[4805]: I1203 14:11:02.559221 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 14:11:02 crc kubenswrapper[4805]: I1203 14:11:02.559237 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T14:11:02Z","lastTransitionTime":"2025-12-03T14:11:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 14:11:02 crc kubenswrapper[4805]: I1203 14:11:02.662026 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 14:11:02 crc kubenswrapper[4805]: I1203 14:11:02.662089 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 14:11:02 crc kubenswrapper[4805]: I1203 14:11:02.662113 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 14:11:02 crc kubenswrapper[4805]: I1203 14:11:02.662163 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 14:11:02 crc kubenswrapper[4805]: I1203 14:11:02.662186 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T14:11:02Z","lastTransitionTime":"2025-12-03T14:11:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 14:11:02 crc kubenswrapper[4805]: I1203 14:11:02.693417 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-4gd9r" Dec 03 14:11:02 crc kubenswrapper[4805]: I1203 14:11:02.693462 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 14:11:02 crc kubenswrapper[4805]: E1203 14:11:02.693589 4805 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-4gd9r" podUID="6b36a3db-1047-487a-9d6c-4f51ffdd8cc1" Dec 03 14:11:02 crc kubenswrapper[4805]: I1203 14:11:02.693645 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 14:11:02 crc kubenswrapper[4805]: I1203 14:11:02.693647 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 14:11:02 crc kubenswrapper[4805]: E1203 14:11:02.693768 4805 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 03 14:11:02 crc kubenswrapper[4805]: E1203 14:11:02.693938 4805 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 03 14:11:02 crc kubenswrapper[4805]: E1203 14:11:02.694070 4805 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 03 14:11:02 crc kubenswrapper[4805]: I1203 14:11:02.763801 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 14:11:02 crc kubenswrapper[4805]: I1203 14:11:02.763881 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 14:11:02 crc kubenswrapper[4805]: I1203 14:11:02.763899 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 14:11:02 crc kubenswrapper[4805]: I1203 14:11:02.763923 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 14:11:02 crc kubenswrapper[4805]: I1203 14:11:02.763940 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T14:11:02Z","lastTransitionTime":"2025-12-03T14:11:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 14:11:02 crc kubenswrapper[4805]: I1203 14:11:02.867777 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 14:11:02 crc kubenswrapper[4805]: I1203 14:11:02.867897 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 14:11:02 crc kubenswrapper[4805]: I1203 14:11:02.867922 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 14:11:02 crc kubenswrapper[4805]: I1203 14:11:02.867971 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 14:11:02 crc kubenswrapper[4805]: I1203 14:11:02.867990 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T14:11:02Z","lastTransitionTime":"2025-12-03T14:11:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 14:11:02 crc kubenswrapper[4805]: I1203 14:11:02.971700 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 14:11:02 crc kubenswrapper[4805]: I1203 14:11:02.971781 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 14:11:02 crc kubenswrapper[4805]: I1203 14:11:02.971807 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 14:11:02 crc kubenswrapper[4805]: I1203 14:11:02.971875 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 14:11:02 crc kubenswrapper[4805]: I1203 14:11:02.971901 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T14:11:02Z","lastTransitionTime":"2025-12-03T14:11:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 14:11:03 crc kubenswrapper[4805]: I1203 14:11:03.074940 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 14:11:03 crc kubenswrapper[4805]: I1203 14:11:03.074991 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 14:11:03 crc kubenswrapper[4805]: I1203 14:11:03.075008 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 14:11:03 crc kubenswrapper[4805]: I1203 14:11:03.075211 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 14:11:03 crc kubenswrapper[4805]: I1203 14:11:03.075228 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T14:11:03Z","lastTransitionTime":"2025-12-03T14:11:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 14:11:03 crc kubenswrapper[4805]: I1203 14:11:03.177829 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 14:11:03 crc kubenswrapper[4805]: I1203 14:11:03.177888 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 14:11:03 crc kubenswrapper[4805]: I1203 14:11:03.177896 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 14:11:03 crc kubenswrapper[4805]: I1203 14:11:03.177909 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 14:11:03 crc kubenswrapper[4805]: I1203 14:11:03.177944 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T14:11:03Z","lastTransitionTime":"2025-12-03T14:11:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 14:11:03 crc kubenswrapper[4805]: I1203 14:11:03.280440 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 14:11:03 crc kubenswrapper[4805]: I1203 14:11:03.280476 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 14:11:03 crc kubenswrapper[4805]: I1203 14:11:03.280485 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 14:11:03 crc kubenswrapper[4805]: I1203 14:11:03.280497 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 14:11:03 crc kubenswrapper[4805]: I1203 14:11:03.280506 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T14:11:03Z","lastTransitionTime":"2025-12-03T14:11:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 14:11:03 crc kubenswrapper[4805]: I1203 14:11:03.383546 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 14:11:03 crc kubenswrapper[4805]: I1203 14:11:03.383592 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 14:11:03 crc kubenswrapper[4805]: I1203 14:11:03.383604 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 14:11:03 crc kubenswrapper[4805]: I1203 14:11:03.383621 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 14:11:03 crc kubenswrapper[4805]: I1203 14:11:03.383636 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T14:11:03Z","lastTransitionTime":"2025-12-03T14:11:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 14:11:03 crc kubenswrapper[4805]: I1203 14:11:03.486167 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 14:11:03 crc kubenswrapper[4805]: I1203 14:11:03.486228 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 14:11:03 crc kubenswrapper[4805]: I1203 14:11:03.486246 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 14:11:03 crc kubenswrapper[4805]: I1203 14:11:03.486269 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 14:11:03 crc kubenswrapper[4805]: I1203 14:11:03.486302 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T14:11:03Z","lastTransitionTime":"2025-12-03T14:11:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 14:11:03 crc kubenswrapper[4805]: I1203 14:11:03.588978 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 14:11:03 crc kubenswrapper[4805]: I1203 14:11:03.589037 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 14:11:03 crc kubenswrapper[4805]: I1203 14:11:03.589059 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 14:11:03 crc kubenswrapper[4805]: I1203 14:11:03.589089 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 14:11:03 crc kubenswrapper[4805]: I1203 14:11:03.589111 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T14:11:03Z","lastTransitionTime":"2025-12-03T14:11:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 14:11:03 crc kubenswrapper[4805]: I1203 14:11:03.691445 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 14:11:03 crc kubenswrapper[4805]: I1203 14:11:03.691524 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 14:11:03 crc kubenswrapper[4805]: I1203 14:11:03.691547 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 14:11:03 crc kubenswrapper[4805]: I1203 14:11:03.691576 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 14:11:03 crc kubenswrapper[4805]: I1203 14:11:03.691599 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T14:11:03Z","lastTransitionTime":"2025-12-03T14:11:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 14:11:03 crc kubenswrapper[4805]: I1203 14:11:03.794247 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 14:11:03 crc kubenswrapper[4805]: I1203 14:11:03.794296 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 14:11:03 crc kubenswrapper[4805]: I1203 14:11:03.794313 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 14:11:03 crc kubenswrapper[4805]: I1203 14:11:03.794335 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 14:11:03 crc kubenswrapper[4805]: I1203 14:11:03.794352 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T14:11:03Z","lastTransitionTime":"2025-12-03T14:11:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 14:11:03 crc kubenswrapper[4805]: I1203 14:11:03.805755 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 14:11:03 crc kubenswrapper[4805]: I1203 14:11:03.805797 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 14:11:03 crc kubenswrapper[4805]: I1203 14:11:03.805809 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 14:11:03 crc kubenswrapper[4805]: I1203 14:11:03.805826 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 14:11:03 crc kubenswrapper[4805]: I1203 14:11:03.805860 4805 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T14:11:03Z","lastTransitionTime":"2025-12-03T14:11:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 14:11:03 crc kubenswrapper[4805]: I1203 14:11:03.866507 4805 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-cluster-version/cluster-version-operator-5c965bbfc6-6jshr"] Dec 03 14:11:03 crc kubenswrapper[4805]: I1203 14:11:03.866917 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-6jshr" Dec 03 14:11:03 crc kubenswrapper[4805]: I1203 14:11:03.869937 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-version"/"default-dockercfg-gxtc4" Dec 03 14:11:03 crc kubenswrapper[4805]: I1203 14:11:03.870442 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-version"/"cluster-version-operator-serving-cert" Dec 03 14:11:03 crc kubenswrapper[4805]: I1203 14:11:03.870536 4805 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-version"/"kube-root-ca.crt" Dec 03 14:11:03 crc kubenswrapper[4805]: I1203 14:11:03.871812 4805 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-version"/"openshift-service-ca.crt" Dec 03 14:11:03 crc kubenswrapper[4805]: I1203 14:11:03.946320 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/eba1b09b-1edc-4c16-bafd-bdaf840c2aae-serving-cert\") pod \"cluster-version-operator-5c965bbfc6-6jshr\" (UID: \"eba1b09b-1edc-4c16-bafd-bdaf840c2aae\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-6jshr" Dec 03 14:11:03 crc kubenswrapper[4805]: I1203 14:11:03.946370 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-cvo-updatepayloads\" (UniqueName: \"kubernetes.io/host-path/eba1b09b-1edc-4c16-bafd-bdaf840c2aae-etc-cvo-updatepayloads\") pod \"cluster-version-operator-5c965bbfc6-6jshr\" (UID: \"eba1b09b-1edc-4c16-bafd-bdaf840c2aae\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-6jshr" Dec 03 14:11:03 crc kubenswrapper[4805]: I1203 14:11:03.946421 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/eba1b09b-1edc-4c16-bafd-bdaf840c2aae-service-ca\") pod \"cluster-version-operator-5c965bbfc6-6jshr\" (UID: \"eba1b09b-1edc-4c16-bafd-bdaf840c2aae\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-6jshr" Dec 03 14:11:03 crc kubenswrapper[4805]: I1203 14:11:03.946449 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/eba1b09b-1edc-4c16-bafd-bdaf840c2aae-kube-api-access\") pod \"cluster-version-operator-5c965bbfc6-6jshr\" (UID: \"eba1b09b-1edc-4c16-bafd-bdaf840c2aae\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-6jshr" Dec 03 14:11:03 crc kubenswrapper[4805]: I1203 14:11:03.946510 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-ssl-certs\" (UniqueName: \"kubernetes.io/host-path/eba1b09b-1edc-4c16-bafd-bdaf840c2aae-etc-ssl-certs\") pod \"cluster-version-operator-5c965bbfc6-6jshr\" (UID: \"eba1b09b-1edc-4c16-bafd-bdaf840c2aae\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-6jshr" Dec 03 14:11:03 crc kubenswrapper[4805]: I1203 14:11:03.955451 4805 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-daemon-gskh4" podStartSLOduration=74.955426581 podStartE2EDuration="1m14.955426581s" podCreationTimestamp="2025-12-03 14:09:49 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 14:11:03.929022264 +0000 UTC m=+93.591939197" watchObservedRunningTime="2025-12-03 14:11:03.955426581 +0000 UTC m=+93.618343514" Dec 03 14:11:03 crc kubenswrapper[4805]: I1203 14:11:03.987419 4805 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" podStartSLOduration=10.987392753 podStartE2EDuration="10.987392753s" podCreationTimestamp="2025-12-03 14:10:53 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 14:11:03.976951031 +0000 UTC m=+93.639867964" watchObservedRunningTime="2025-12-03 14:11:03.987392753 +0000 UTC m=+93.650309696" Dec 03 14:11:03 crc kubenswrapper[4805]: I1203 14:11:03.988172 4805 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-dns/node-resolver-pdzgk" podStartSLOduration=74.988164075 podStartE2EDuration="1m14.988164075s" podCreationTimestamp="2025-12-03 14:09:49 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 14:11:03.987729602 +0000 UTC m=+93.650646535" watchObservedRunningTime="2025-12-03 14:11:03.988164075 +0000 UTC m=+93.651081008" Dec 03 14:11:04 crc kubenswrapper[4805]: I1203 14:11:04.012306 4805 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-69h75" podStartSLOduration=74.012283898 podStartE2EDuration="1m14.012283898s" podCreationTimestamp="2025-12-03 14:09:50 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 14:11:03.999058309 +0000 UTC m=+93.661975242" watchObservedRunningTime="2025-12-03 14:11:04.012283898 +0000 UTC m=+93.675200831" Dec 03 14:11:04 crc kubenswrapper[4805]: I1203 14:11:04.047887 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-ssl-certs\" (UniqueName: \"kubernetes.io/host-path/eba1b09b-1edc-4c16-bafd-bdaf840c2aae-etc-ssl-certs\") pod \"cluster-version-operator-5c965bbfc6-6jshr\" (UID: \"eba1b09b-1edc-4c16-bafd-bdaf840c2aae\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-6jshr" Dec 03 14:11:04 crc kubenswrapper[4805]: I1203 14:11:04.047960 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/eba1b09b-1edc-4c16-bafd-bdaf840c2aae-serving-cert\") pod \"cluster-version-operator-5c965bbfc6-6jshr\" (UID: \"eba1b09b-1edc-4c16-bafd-bdaf840c2aae\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-6jshr" Dec 03 14:11:04 crc kubenswrapper[4805]: I1203 14:11:04.047986 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-cvo-updatepayloads\" (UniqueName: \"kubernetes.io/host-path/eba1b09b-1edc-4c16-bafd-bdaf840c2aae-etc-cvo-updatepayloads\") pod \"cluster-version-operator-5c965bbfc6-6jshr\" (UID: \"eba1b09b-1edc-4c16-bafd-bdaf840c2aae\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-6jshr" Dec 03 14:11:04 crc kubenswrapper[4805]: I1203 14:11:04.048033 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/eba1b09b-1edc-4c16-bafd-bdaf840c2aae-service-ca\") pod \"cluster-version-operator-5c965bbfc6-6jshr\" (UID: \"eba1b09b-1edc-4c16-bafd-bdaf840c2aae\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-6jshr" Dec 03 14:11:04 crc kubenswrapper[4805]: I1203 14:11:04.048053 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/eba1b09b-1edc-4c16-bafd-bdaf840c2aae-kube-api-access\") pod \"cluster-version-operator-5c965bbfc6-6jshr\" (UID: \"eba1b09b-1edc-4c16-bafd-bdaf840c2aae\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-6jshr" Dec 03 14:11:04 crc kubenswrapper[4805]: I1203 14:11:04.048044 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-ssl-certs\" (UniqueName: \"kubernetes.io/host-path/eba1b09b-1edc-4c16-bafd-bdaf840c2aae-etc-ssl-certs\") pod \"cluster-version-operator-5c965bbfc6-6jshr\" (UID: \"eba1b09b-1edc-4c16-bafd-bdaf840c2aae\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-6jshr" Dec 03 14:11:04 crc kubenswrapper[4805]: I1203 14:11:04.048121 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-cvo-updatepayloads\" (UniqueName: \"kubernetes.io/host-path/eba1b09b-1edc-4c16-bafd-bdaf840c2aae-etc-cvo-updatepayloads\") pod \"cluster-version-operator-5c965bbfc6-6jshr\" (UID: \"eba1b09b-1edc-4c16-bafd-bdaf840c2aae\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-6jshr" Dec 03 14:11:04 crc kubenswrapper[4805]: I1203 14:11:04.049188 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/eba1b09b-1edc-4c16-bafd-bdaf840c2aae-service-ca\") pod \"cluster-version-operator-5c965bbfc6-6jshr\" (UID: \"eba1b09b-1edc-4c16-bafd-bdaf840c2aae\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-6jshr" Dec 03 14:11:04 crc kubenswrapper[4805]: I1203 14:11:04.060533 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/eba1b09b-1edc-4c16-bafd-bdaf840c2aae-serving-cert\") pod \"cluster-version-operator-5c965bbfc6-6jshr\" (UID: \"eba1b09b-1edc-4c16-bafd-bdaf840c2aae\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-6jshr" Dec 03 14:11:04 crc kubenswrapper[4805]: I1203 14:11:04.070424 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/eba1b09b-1edc-4c16-bafd-bdaf840c2aae-kube-api-access\") pod \"cluster-version-operator-5c965bbfc6-6jshr\" (UID: \"eba1b09b-1edc-4c16-bafd-bdaf840c2aae\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-6jshr" Dec 03 14:11:04 crc kubenswrapper[4805]: I1203 14:11:04.085605 4805 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/node-ca-bdtdg" podStartSLOduration=75.085587103 podStartE2EDuration="1m15.085587103s" podCreationTimestamp="2025-12-03 14:09:49 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 14:11:04.050883665 +0000 UTC m=+93.713800608" watchObservedRunningTime="2025-12-03 14:11:04.085587103 +0000 UTC m=+93.748504026" Dec 03 14:11:04 crc kubenswrapper[4805]: I1203 14:11:04.112089 4805 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver/kube-apiserver-crc" podStartSLOduration=76.112075083 podStartE2EDuration="1m16.112075083s" podCreationTimestamp="2025-12-03 14:09:48 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 14:11:04.111901688 +0000 UTC m=+93.774818601" watchObservedRunningTime="2025-12-03 14:11:04.112075083 +0000 UTC m=+93.774992006" Dec 03 14:11:04 crc kubenswrapper[4805]: I1203 14:11:04.112411 4805 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-etcd/etcd-crc" podStartSLOduration=71.112397122 podStartE2EDuration="1m11.112397122s" podCreationTimestamp="2025-12-03 14:09:53 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 14:11:04.085489641 +0000 UTC m=+93.748406574" watchObservedRunningTime="2025-12-03 14:11:04.112397122 +0000 UTC m=+93.775314045" Dec 03 14:11:04 crc kubenswrapper[4805]: I1203 14:11:04.144069 4805 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" podStartSLOduration=45.144049865 podStartE2EDuration="45.144049865s" podCreationTimestamp="2025-12-03 14:10:19 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 14:11:04.129822258 +0000 UTC m=+93.792739191" watchObservedRunningTime="2025-12-03 14:11:04.144049865 +0000 UTC m=+93.806966788" Dec 03 14:11:04 crc kubenswrapper[4805]: I1203 14:11:04.170532 4805 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podStartSLOduration=70.170514534 podStartE2EDuration="1m10.170514534s" podCreationTimestamp="2025-12-03 14:09:54 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 14:11:04.169691991 +0000 UTC m=+93.832608914" watchObservedRunningTime="2025-12-03 14:11:04.170514534 +0000 UTC m=+93.833431457" Dec 03 14:11:04 crc kubenswrapper[4805]: I1203 14:11:04.181080 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-6jshr" Dec 03 14:11:04 crc kubenswrapper[4805]: I1203 14:11:04.201432 4805 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/multus-k4mhs" podStartSLOduration=75.201412146 podStartE2EDuration="1m15.201412146s" podCreationTimestamp="2025-12-03 14:09:49 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 14:11:04.198937007 +0000 UTC m=+93.861853950" watchObservedRunningTime="2025-12-03 14:11:04.201412146 +0000 UTC m=+93.864329089" Dec 03 14:11:04 crc kubenswrapper[4805]: I1203 14:11:04.205152 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-6jshr" event={"ID":"eba1b09b-1edc-4c16-bafd-bdaf840c2aae","Type":"ContainerStarted","Data":"ab8b8fed81ab75cc6eae6d2e79295687a58d4464c28b613deba52eb592eb657e"} Dec 03 14:11:04 crc kubenswrapper[4805]: I1203 14:11:04.223448 4805 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/multus-additional-cni-plugins-bpk49" podStartSLOduration=75.22342891 podStartE2EDuration="1m15.22342891s" podCreationTimestamp="2025-12-03 14:09:49 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 14:11:04.222325239 +0000 UTC m=+93.885242162" watchObservedRunningTime="2025-12-03 14:11:04.22342891 +0000 UTC m=+93.886345853" Dec 03 14:11:04 crc kubenswrapper[4805]: I1203 14:11:04.694100 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 14:11:04 crc kubenswrapper[4805]: I1203 14:11:04.694200 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 14:11:04 crc kubenswrapper[4805]: E1203 14:11:04.694233 4805 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 03 14:11:04 crc kubenswrapper[4805]: I1203 14:11:04.694104 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 14:11:04 crc kubenswrapper[4805]: E1203 14:11:04.694348 4805 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 03 14:11:04 crc kubenswrapper[4805]: I1203 14:11:04.694119 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-4gd9r" Dec 03 14:11:04 crc kubenswrapper[4805]: E1203 14:11:04.694492 4805 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 03 14:11:04 crc kubenswrapper[4805]: E1203 14:11:04.694545 4805 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-4gd9r" podUID="6b36a3db-1047-487a-9d6c-4f51ffdd8cc1" Dec 03 14:11:05 crc kubenswrapper[4805]: I1203 14:11:05.209162 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-6jshr" event={"ID":"eba1b09b-1edc-4c16-bafd-bdaf840c2aae","Type":"ContainerStarted","Data":"d10e6e73899242cc3372708d9b53b34bbd5b93bf1c46db0e16261469818acdfc"} Dec 03 14:11:05 crc kubenswrapper[4805]: I1203 14:11:05.223557 4805 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-6jshr" podStartSLOduration=76.2235375 podStartE2EDuration="1m16.2235375s" podCreationTimestamp="2025-12-03 14:09:49 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 14:11:05.223215351 +0000 UTC m=+94.886132274" watchObservedRunningTime="2025-12-03 14:11:05.2235375 +0000 UTC m=+94.886454443" Dec 03 14:11:06 crc kubenswrapper[4805]: I1203 14:11:06.693998 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-4gd9r" Dec 03 14:11:06 crc kubenswrapper[4805]: I1203 14:11:06.694128 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 14:11:06 crc kubenswrapper[4805]: E1203 14:11:06.694724 4805 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-4gd9r" podUID="6b36a3db-1047-487a-9d6c-4f51ffdd8cc1" Dec 03 14:11:06 crc kubenswrapper[4805]: I1203 14:11:06.694345 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 14:11:06 crc kubenswrapper[4805]: E1203 14:11:06.694806 4805 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 03 14:11:06 crc kubenswrapper[4805]: E1203 14:11:06.694824 4805 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 03 14:11:06 crc kubenswrapper[4805]: I1203 14:11:06.694154 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 14:11:06 crc kubenswrapper[4805]: E1203 14:11:06.694920 4805 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 03 14:11:08 crc kubenswrapper[4805]: I1203 14:11:08.596066 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/6b36a3db-1047-487a-9d6c-4f51ffdd8cc1-metrics-certs\") pod \"network-metrics-daemon-4gd9r\" (UID: \"6b36a3db-1047-487a-9d6c-4f51ffdd8cc1\") " pod="openshift-multus/network-metrics-daemon-4gd9r" Dec 03 14:11:08 crc kubenswrapper[4805]: E1203 14:11:08.596245 4805 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Dec 03 14:11:08 crc kubenswrapper[4805]: E1203 14:11:08.596326 4805 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/6b36a3db-1047-487a-9d6c-4f51ffdd8cc1-metrics-certs podName:6b36a3db-1047-487a-9d6c-4f51ffdd8cc1 nodeName:}" failed. No retries permitted until 2025-12-03 14:12:12.596301026 +0000 UTC m=+162.259217989 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/6b36a3db-1047-487a-9d6c-4f51ffdd8cc1-metrics-certs") pod "network-metrics-daemon-4gd9r" (UID: "6b36a3db-1047-487a-9d6c-4f51ffdd8cc1") : object "openshift-multus"/"metrics-daemon-secret" not registered Dec 03 14:11:08 crc kubenswrapper[4805]: I1203 14:11:08.693804 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 14:11:08 crc kubenswrapper[4805]: I1203 14:11:08.693889 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 14:11:08 crc kubenswrapper[4805]: I1203 14:11:08.693811 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-4gd9r" Dec 03 14:11:08 crc kubenswrapper[4805]: E1203 14:11:08.694045 4805 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 03 14:11:08 crc kubenswrapper[4805]: I1203 14:11:08.694101 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 14:11:08 crc kubenswrapper[4805]: E1203 14:11:08.694289 4805 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-4gd9r" podUID="6b36a3db-1047-487a-9d6c-4f51ffdd8cc1" Dec 03 14:11:08 crc kubenswrapper[4805]: E1203 14:11:08.694361 4805 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 03 14:11:08 crc kubenswrapper[4805]: E1203 14:11:08.694438 4805 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 03 14:11:10 crc kubenswrapper[4805]: I1203 14:11:10.694201 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 14:11:10 crc kubenswrapper[4805]: I1203 14:11:10.694268 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-4gd9r" Dec 03 14:11:10 crc kubenswrapper[4805]: I1203 14:11:10.694330 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 14:11:10 crc kubenswrapper[4805]: I1203 14:11:10.694385 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 14:11:10 crc kubenswrapper[4805]: E1203 14:11:10.697476 4805 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 03 14:11:10 crc kubenswrapper[4805]: E1203 14:11:10.697581 4805 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 03 14:11:10 crc kubenswrapper[4805]: E1203 14:11:10.697683 4805 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-4gd9r" podUID="6b36a3db-1047-487a-9d6c-4f51ffdd8cc1" Dec 03 14:11:10 crc kubenswrapper[4805]: E1203 14:11:10.697787 4805 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 03 14:11:11 crc kubenswrapper[4805]: I1203 14:11:11.694334 4805 scope.go:117] "RemoveContainer" containerID="8f75988e25825e00abdb9e4b2292c5cc8ccab3dd028c882c2a38c796c0fd19c5" Dec 03 14:11:11 crc kubenswrapper[4805]: E1203 14:11:11.694514 4805 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-psjpm_openshift-ovn-kubernetes(deeec30d-cad2-4dfd-adb6-64f6646581be)\"" pod="openshift-ovn-kubernetes/ovnkube-node-psjpm" podUID="deeec30d-cad2-4dfd-adb6-64f6646581be" Dec 03 14:11:12 crc kubenswrapper[4805]: I1203 14:11:12.694182 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-4gd9r" Dec 03 14:11:12 crc kubenswrapper[4805]: I1203 14:11:12.694271 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 14:11:12 crc kubenswrapper[4805]: E1203 14:11:12.694327 4805 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-4gd9r" podUID="6b36a3db-1047-487a-9d6c-4f51ffdd8cc1" Dec 03 14:11:12 crc kubenswrapper[4805]: I1203 14:11:12.694355 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 14:11:12 crc kubenswrapper[4805]: E1203 14:11:12.694431 4805 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 03 14:11:12 crc kubenswrapper[4805]: I1203 14:11:12.694570 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 14:11:12 crc kubenswrapper[4805]: E1203 14:11:12.694684 4805 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 03 14:11:12 crc kubenswrapper[4805]: E1203 14:11:12.694875 4805 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 03 14:11:14 crc kubenswrapper[4805]: I1203 14:11:14.694275 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 14:11:14 crc kubenswrapper[4805]: I1203 14:11:14.694510 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-4gd9r" Dec 03 14:11:14 crc kubenswrapper[4805]: I1203 14:11:14.694526 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 14:11:14 crc kubenswrapper[4805]: E1203 14:11:14.694699 4805 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 03 14:11:14 crc kubenswrapper[4805]: I1203 14:11:14.694718 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 14:11:14 crc kubenswrapper[4805]: E1203 14:11:14.694833 4805 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 03 14:11:14 crc kubenswrapper[4805]: E1203 14:11:14.694902 4805 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 03 14:11:14 crc kubenswrapper[4805]: E1203 14:11:14.695008 4805 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-4gd9r" podUID="6b36a3db-1047-487a-9d6c-4f51ffdd8cc1" Dec 03 14:11:16 crc kubenswrapper[4805]: I1203 14:11:16.694259 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-4gd9r" Dec 03 14:11:16 crc kubenswrapper[4805]: I1203 14:11:16.694349 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 14:11:16 crc kubenswrapper[4805]: I1203 14:11:16.694349 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 14:11:16 crc kubenswrapper[4805]: E1203 14:11:16.694461 4805 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-4gd9r" podUID="6b36a3db-1047-487a-9d6c-4f51ffdd8cc1" Dec 03 14:11:16 crc kubenswrapper[4805]: I1203 14:11:16.694489 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 14:11:16 crc kubenswrapper[4805]: E1203 14:11:16.694584 4805 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 03 14:11:16 crc kubenswrapper[4805]: E1203 14:11:16.694766 4805 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 03 14:11:16 crc kubenswrapper[4805]: E1203 14:11:16.694923 4805 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 03 14:11:18 crc kubenswrapper[4805]: I1203 14:11:18.693934 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-4gd9r" Dec 03 14:11:18 crc kubenswrapper[4805]: E1203 14:11:18.694171 4805 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-4gd9r" podUID="6b36a3db-1047-487a-9d6c-4f51ffdd8cc1" Dec 03 14:11:18 crc kubenswrapper[4805]: I1203 14:11:18.694240 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 14:11:18 crc kubenswrapper[4805]: I1203 14:11:18.694238 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 14:11:18 crc kubenswrapper[4805]: I1203 14:11:18.694574 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 14:11:18 crc kubenswrapper[4805]: E1203 14:11:18.694636 4805 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 03 14:11:18 crc kubenswrapper[4805]: E1203 14:11:18.694756 4805 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 03 14:11:18 crc kubenswrapper[4805]: E1203 14:11:18.694806 4805 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 03 14:11:20 crc kubenswrapper[4805]: I1203 14:11:20.693996 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-4gd9r" Dec 03 14:11:20 crc kubenswrapper[4805]: I1203 14:11:20.694066 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 14:11:20 crc kubenswrapper[4805]: I1203 14:11:20.694101 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 14:11:20 crc kubenswrapper[4805]: I1203 14:11:20.694129 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 14:11:20 crc kubenswrapper[4805]: E1203 14:11:20.696173 4805 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 03 14:11:20 crc kubenswrapper[4805]: E1203 14:11:20.696307 4805 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-4gd9r" podUID="6b36a3db-1047-487a-9d6c-4f51ffdd8cc1" Dec 03 14:11:20 crc kubenswrapper[4805]: E1203 14:11:20.696371 4805 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 03 14:11:20 crc kubenswrapper[4805]: E1203 14:11:20.696817 4805 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 03 14:11:22 crc kubenswrapper[4805]: I1203 14:11:22.694187 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 14:11:22 crc kubenswrapper[4805]: E1203 14:11:22.694352 4805 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 03 14:11:22 crc kubenswrapper[4805]: I1203 14:11:22.694382 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 14:11:22 crc kubenswrapper[4805]: I1203 14:11:22.694414 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-4gd9r" Dec 03 14:11:22 crc kubenswrapper[4805]: I1203 14:11:22.694539 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 14:11:22 crc kubenswrapper[4805]: E1203 14:11:22.694558 4805 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 03 14:11:22 crc kubenswrapper[4805]: E1203 14:11:22.694695 4805 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-4gd9r" podUID="6b36a3db-1047-487a-9d6c-4f51ffdd8cc1" Dec 03 14:11:22 crc kubenswrapper[4805]: E1203 14:11:22.694787 4805 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 03 14:11:23 crc kubenswrapper[4805]: I1203 14:11:23.272401 4805 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-k4mhs_c09c1cdc-f461-45c9-8444-5b8764bccabc/kube-multus/1.log" Dec 03 14:11:23 crc kubenswrapper[4805]: I1203 14:11:23.273140 4805 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-k4mhs_c09c1cdc-f461-45c9-8444-5b8764bccabc/kube-multus/0.log" Dec 03 14:11:23 crc kubenswrapper[4805]: I1203 14:11:23.273189 4805 generic.go:334] "Generic (PLEG): container finished" podID="c09c1cdc-f461-45c9-8444-5b8764bccabc" containerID="7e077e19b6a5bf7d29b357982e12224d4b7058e0f9ac81b66b5a665c9bbf5014" exitCode=1 Dec 03 14:11:23 crc kubenswrapper[4805]: I1203 14:11:23.273253 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-k4mhs" event={"ID":"c09c1cdc-f461-45c9-8444-5b8764bccabc","Type":"ContainerDied","Data":"7e077e19b6a5bf7d29b357982e12224d4b7058e0f9ac81b66b5a665c9bbf5014"} Dec 03 14:11:23 crc kubenswrapper[4805]: I1203 14:11:23.273295 4805 scope.go:117] "RemoveContainer" containerID="3354c1a44f891182d04116fc76234a2b337d50567d032a3584f902a9516387cc" Dec 03 14:11:23 crc kubenswrapper[4805]: I1203 14:11:23.274579 4805 scope.go:117] "RemoveContainer" containerID="7e077e19b6a5bf7d29b357982e12224d4b7058e0f9ac81b66b5a665c9bbf5014" Dec 03 14:11:23 crc kubenswrapper[4805]: E1203 14:11:23.275012 4805 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-multus\" with CrashLoopBackOff: \"back-off 10s restarting failed container=kube-multus pod=multus-k4mhs_openshift-multus(c09c1cdc-f461-45c9-8444-5b8764bccabc)\"" pod="openshift-multus/multus-k4mhs" podUID="c09c1cdc-f461-45c9-8444-5b8764bccabc" Dec 03 14:11:24 crc kubenswrapper[4805]: I1203 14:11:24.277857 4805 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-k4mhs_c09c1cdc-f461-45c9-8444-5b8764bccabc/kube-multus/1.log" Dec 03 14:11:24 crc kubenswrapper[4805]: I1203 14:11:24.694008 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 14:11:24 crc kubenswrapper[4805]: I1203 14:11:24.694038 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 14:11:24 crc kubenswrapper[4805]: I1203 14:11:24.694047 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-4gd9r" Dec 03 14:11:24 crc kubenswrapper[4805]: E1203 14:11:24.694178 4805 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 03 14:11:24 crc kubenswrapper[4805]: I1203 14:11:24.694379 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 14:11:24 crc kubenswrapper[4805]: E1203 14:11:24.694452 4805 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 03 14:11:24 crc kubenswrapper[4805]: E1203 14:11:24.694625 4805 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-4gd9r" podUID="6b36a3db-1047-487a-9d6c-4f51ffdd8cc1" Dec 03 14:11:24 crc kubenswrapper[4805]: E1203 14:11:24.694771 4805 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 03 14:11:26 crc kubenswrapper[4805]: I1203 14:11:26.694287 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 14:11:26 crc kubenswrapper[4805]: I1203 14:11:26.694447 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-4gd9r" Dec 03 14:11:26 crc kubenswrapper[4805]: E1203 14:11:26.694586 4805 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 03 14:11:26 crc kubenswrapper[4805]: I1203 14:11:26.694631 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 14:11:26 crc kubenswrapper[4805]: I1203 14:11:26.694711 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 14:11:26 crc kubenswrapper[4805]: E1203 14:11:26.695039 4805 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-4gd9r" podUID="6b36a3db-1047-487a-9d6c-4f51ffdd8cc1" Dec 03 14:11:26 crc kubenswrapper[4805]: E1203 14:11:26.695207 4805 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 03 14:11:26 crc kubenswrapper[4805]: E1203 14:11:26.695592 4805 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 03 14:11:26 crc kubenswrapper[4805]: I1203 14:11:26.695880 4805 scope.go:117] "RemoveContainer" containerID="8f75988e25825e00abdb9e4b2292c5cc8ccab3dd028c882c2a38c796c0fd19c5" Dec 03 14:11:26 crc kubenswrapper[4805]: E1203 14:11:26.696026 4805 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-psjpm_openshift-ovn-kubernetes(deeec30d-cad2-4dfd-adb6-64f6646581be)\"" pod="openshift-ovn-kubernetes/ovnkube-node-psjpm" podUID="deeec30d-cad2-4dfd-adb6-64f6646581be" Dec 03 14:11:28 crc kubenswrapper[4805]: I1203 14:11:28.694099 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 14:11:28 crc kubenswrapper[4805]: I1203 14:11:28.694252 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 14:11:28 crc kubenswrapper[4805]: I1203 14:11:28.694322 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 14:11:28 crc kubenswrapper[4805]: E1203 14:11:28.694270 4805 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 03 14:11:28 crc kubenswrapper[4805]: I1203 14:11:28.694346 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-4gd9r" Dec 03 14:11:28 crc kubenswrapper[4805]: E1203 14:11:28.694569 4805 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 03 14:11:28 crc kubenswrapper[4805]: E1203 14:11:28.694712 4805 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 03 14:11:28 crc kubenswrapper[4805]: E1203 14:11:28.694894 4805 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-4gd9r" podUID="6b36a3db-1047-487a-9d6c-4f51ffdd8cc1" Dec 03 14:11:30 crc kubenswrapper[4805]: E1203 14:11:30.672636 4805 kubelet_node_status.go:497] "Node not becoming ready in time after startup" Dec 03 14:11:30 crc kubenswrapper[4805]: I1203 14:11:30.693641 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 14:11:30 crc kubenswrapper[4805]: I1203 14:11:30.693651 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-4gd9r" Dec 03 14:11:30 crc kubenswrapper[4805]: I1203 14:11:30.696029 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 14:11:30 crc kubenswrapper[4805]: E1203 14:11:30.696010 4805 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 03 14:11:30 crc kubenswrapper[4805]: I1203 14:11:30.696034 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 14:11:30 crc kubenswrapper[4805]: E1203 14:11:30.696180 4805 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-4gd9r" podUID="6b36a3db-1047-487a-9d6c-4f51ffdd8cc1" Dec 03 14:11:30 crc kubenswrapper[4805]: E1203 14:11:30.696294 4805 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 03 14:11:30 crc kubenswrapper[4805]: E1203 14:11:30.696401 4805 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 03 14:11:30 crc kubenswrapper[4805]: E1203 14:11:30.795426 4805 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Dec 03 14:11:32 crc kubenswrapper[4805]: I1203 14:11:32.693604 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 14:11:32 crc kubenswrapper[4805]: E1203 14:11:32.693905 4805 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 03 14:11:32 crc kubenswrapper[4805]: I1203 14:11:32.694134 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 14:11:32 crc kubenswrapper[4805]: I1203 14:11:32.694239 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-4gd9r" Dec 03 14:11:32 crc kubenswrapper[4805]: I1203 14:11:32.694322 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 14:11:32 crc kubenswrapper[4805]: E1203 14:11:32.694403 4805 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 03 14:11:32 crc kubenswrapper[4805]: E1203 14:11:32.694639 4805 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 03 14:11:32 crc kubenswrapper[4805]: E1203 14:11:32.694897 4805 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-4gd9r" podUID="6b36a3db-1047-487a-9d6c-4f51ffdd8cc1" Dec 03 14:11:34 crc kubenswrapper[4805]: I1203 14:11:34.694177 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 14:11:34 crc kubenswrapper[4805]: I1203 14:11:34.694170 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 14:11:34 crc kubenswrapper[4805]: E1203 14:11:34.694390 4805 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 03 14:11:34 crc kubenswrapper[4805]: E1203 14:11:34.694489 4805 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 03 14:11:34 crc kubenswrapper[4805]: I1203 14:11:34.694210 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 14:11:34 crc kubenswrapper[4805]: E1203 14:11:34.694643 4805 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 03 14:11:34 crc kubenswrapper[4805]: I1203 14:11:34.694744 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-4gd9r" Dec 03 14:11:34 crc kubenswrapper[4805]: E1203 14:11:34.694909 4805 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-4gd9r" podUID="6b36a3db-1047-487a-9d6c-4f51ffdd8cc1" Dec 03 14:11:35 crc kubenswrapper[4805]: E1203 14:11:35.796935 4805 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Dec 03 14:11:36 crc kubenswrapper[4805]: I1203 14:11:36.694503 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-4gd9r" Dec 03 14:11:36 crc kubenswrapper[4805]: I1203 14:11:36.694547 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 14:11:36 crc kubenswrapper[4805]: E1203 14:11:36.695318 4805 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-4gd9r" podUID="6b36a3db-1047-487a-9d6c-4f51ffdd8cc1" Dec 03 14:11:36 crc kubenswrapper[4805]: I1203 14:11:36.694540 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 14:11:36 crc kubenswrapper[4805]: I1203 14:11:36.694608 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 14:11:36 crc kubenswrapper[4805]: E1203 14:11:36.695467 4805 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 03 14:11:36 crc kubenswrapper[4805]: E1203 14:11:36.695663 4805 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 03 14:11:36 crc kubenswrapper[4805]: E1203 14:11:36.695788 4805 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 03 14:11:37 crc kubenswrapper[4805]: I1203 14:11:37.694663 4805 scope.go:117] "RemoveContainer" containerID="7e077e19b6a5bf7d29b357982e12224d4b7058e0f9ac81b66b5a665c9bbf5014" Dec 03 14:11:38 crc kubenswrapper[4805]: I1203 14:11:38.325550 4805 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-k4mhs_c09c1cdc-f461-45c9-8444-5b8764bccabc/kube-multus/1.log" Dec 03 14:11:38 crc kubenswrapper[4805]: I1203 14:11:38.325624 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-k4mhs" event={"ID":"c09c1cdc-f461-45c9-8444-5b8764bccabc","Type":"ContainerStarted","Data":"556324f2c4c09f251607fd41d095f4eae1aa27bcb9958f51449b88e6dedcaa86"} Dec 03 14:11:38 crc kubenswrapper[4805]: I1203 14:11:38.694349 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 14:11:38 crc kubenswrapper[4805]: I1203 14:11:38.694364 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-4gd9r" Dec 03 14:11:38 crc kubenswrapper[4805]: I1203 14:11:38.694513 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 14:11:38 crc kubenswrapper[4805]: E1203 14:11:38.694605 4805 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 03 14:11:38 crc kubenswrapper[4805]: I1203 14:11:38.694648 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 14:11:38 crc kubenswrapper[4805]: E1203 14:11:38.694866 4805 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-4gd9r" podUID="6b36a3db-1047-487a-9d6c-4f51ffdd8cc1" Dec 03 14:11:38 crc kubenswrapper[4805]: E1203 14:11:38.695517 4805 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 03 14:11:38 crc kubenswrapper[4805]: E1203 14:11:38.695694 4805 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 03 14:11:38 crc kubenswrapper[4805]: I1203 14:11:38.696298 4805 scope.go:117] "RemoveContainer" containerID="8f75988e25825e00abdb9e4b2292c5cc8ccab3dd028c882c2a38c796c0fd19c5" Dec 03 14:11:39 crc kubenswrapper[4805]: I1203 14:11:39.330872 4805 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-psjpm_deeec30d-cad2-4dfd-adb6-64f6646581be/ovnkube-controller/3.log" Dec 03 14:11:39 crc kubenswrapper[4805]: I1203 14:11:39.333357 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-psjpm" event={"ID":"deeec30d-cad2-4dfd-adb6-64f6646581be","Type":"ContainerStarted","Data":"cb029c1dbe665e82f006ab5f681796fa66f6ff814a3988c3c4b5863923b4a44f"} Dec 03 14:11:39 crc kubenswrapper[4805]: I1203 14:11:39.334228 4805 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-psjpm" Dec 03 14:11:39 crc kubenswrapper[4805]: I1203 14:11:39.358449 4805 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ovn-kubernetes/ovnkube-node-psjpm" podStartSLOduration=109.358417116 podStartE2EDuration="1m49.358417116s" podCreationTimestamp="2025-12-03 14:09:50 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 14:11:39.358382965 +0000 UTC m=+129.021299888" watchObservedRunningTime="2025-12-03 14:11:39.358417116 +0000 UTC m=+129.021334039" Dec 03 14:11:39 crc kubenswrapper[4805]: I1203 14:11:39.575754 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/network-metrics-daemon-4gd9r"] Dec 03 14:11:39 crc kubenswrapper[4805]: I1203 14:11:39.576212 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-4gd9r" Dec 03 14:11:39 crc kubenswrapper[4805]: E1203 14:11:39.576309 4805 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-4gd9r" podUID="6b36a3db-1047-487a-9d6c-4f51ffdd8cc1" Dec 03 14:11:40 crc kubenswrapper[4805]: I1203 14:11:40.694024 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 14:11:40 crc kubenswrapper[4805]: I1203 14:11:40.694028 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 14:11:40 crc kubenswrapper[4805]: E1203 14:11:40.696798 4805 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 03 14:11:40 crc kubenswrapper[4805]: I1203 14:11:40.696961 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 14:11:40 crc kubenswrapper[4805]: E1203 14:11:40.697096 4805 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 03 14:11:40 crc kubenswrapper[4805]: E1203 14:11:40.697227 4805 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 03 14:11:40 crc kubenswrapper[4805]: E1203 14:11:40.797899 4805 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Dec 03 14:11:41 crc kubenswrapper[4805]: I1203 14:11:41.693546 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-4gd9r" Dec 03 14:11:41 crc kubenswrapper[4805]: E1203 14:11:41.694422 4805 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-4gd9r" podUID="6b36a3db-1047-487a-9d6c-4f51ffdd8cc1" Dec 03 14:11:42 crc kubenswrapper[4805]: I1203 14:11:42.694464 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 14:11:42 crc kubenswrapper[4805]: I1203 14:11:42.694543 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 14:11:42 crc kubenswrapper[4805]: E1203 14:11:42.694617 4805 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 03 14:11:42 crc kubenswrapper[4805]: I1203 14:11:42.694684 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 14:11:42 crc kubenswrapper[4805]: E1203 14:11:42.694942 4805 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 03 14:11:42 crc kubenswrapper[4805]: E1203 14:11:42.695059 4805 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 03 14:11:43 crc kubenswrapper[4805]: I1203 14:11:43.693442 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-4gd9r" Dec 03 14:11:43 crc kubenswrapper[4805]: E1203 14:11:43.693581 4805 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-4gd9r" podUID="6b36a3db-1047-487a-9d6c-4f51ffdd8cc1" Dec 03 14:11:44 crc kubenswrapper[4805]: I1203 14:11:44.693647 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 14:11:44 crc kubenswrapper[4805]: I1203 14:11:44.693662 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 14:11:44 crc kubenswrapper[4805]: E1203 14:11:44.693799 4805 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 03 14:11:44 crc kubenswrapper[4805]: E1203 14:11:44.694017 4805 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 03 14:11:44 crc kubenswrapper[4805]: I1203 14:11:44.694248 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 14:11:44 crc kubenswrapper[4805]: E1203 14:11:44.694326 4805 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 03 14:11:45 crc kubenswrapper[4805]: I1203 14:11:45.693897 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-4gd9r" Dec 03 14:11:45 crc kubenswrapper[4805]: E1203 14:11:45.694084 4805 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-4gd9r" podUID="6b36a3db-1047-487a-9d6c-4f51ffdd8cc1" Dec 03 14:11:46 crc kubenswrapper[4805]: I1203 14:11:46.693551 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 14:11:46 crc kubenswrapper[4805]: I1203 14:11:46.693619 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 14:11:46 crc kubenswrapper[4805]: I1203 14:11:46.693669 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 14:11:46 crc kubenswrapper[4805]: I1203 14:11:46.696272 4805 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-diagnostics"/"openshift-service-ca.crt" Dec 03 14:11:46 crc kubenswrapper[4805]: I1203 14:11:46.696611 4805 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-console"/"networking-console-plugin" Dec 03 14:11:46 crc kubenswrapper[4805]: I1203 14:11:46.697754 4805 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-diagnostics"/"kube-root-ca.crt" Dec 03 14:11:46 crc kubenswrapper[4805]: I1203 14:11:46.698603 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-console"/"networking-console-plugin-cert" Dec 03 14:11:47 crc kubenswrapper[4805]: I1203 14:11:47.694363 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-4gd9r" Dec 03 14:11:47 crc kubenswrapper[4805]: I1203 14:11:47.696398 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"metrics-daemon-secret" Dec 03 14:11:47 crc kubenswrapper[4805]: I1203 14:11:47.696470 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"metrics-daemon-sa-dockercfg-d427c" Dec 03 14:11:49 crc kubenswrapper[4805]: I1203 14:11:49.282381 4805 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-psjpm" Dec 03 14:11:54 crc kubenswrapper[4805]: I1203 14:11:54.841981 4805 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeReady" Dec 03 14:11:54 crc kubenswrapper[4805]: I1203 14:11:54.892042 4805 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-apiserver/apiserver-76f77b778f-7nhvl"] Dec 03 14:11:54 crc kubenswrapper[4805]: I1203 14:11:54.892751 4805 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-oauth-apiserver/apiserver-7bbb656c7d-kfh2f"] Dec 03 14:11:54 crc kubenswrapper[4805]: I1203 14:11:54.893591 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver/apiserver-76f77b778f-7nhvl" Dec 03 14:11:54 crc kubenswrapper[4805]: I1203 14:11:54.894000 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-kfh2f" Dec 03 14:11:54 crc kubenswrapper[4805]: I1203 14:11:54.893600 4805 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-k7rgc"] Dec 03 14:11:54 crc kubenswrapper[4805]: I1203 14:11:54.894723 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-k7rgc" Dec 03 14:11:54 crc kubenswrapper[4805]: I1203 14:11:54.895902 4805 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-nsz7s"] Dec 03 14:11:54 crc kubenswrapper[4805]: I1203 14:11:54.896351 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-nsz7s" Dec 03 14:11:54 crc kubenswrapper[4805]: I1203 14:11:54.897882 4805 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-api/machine-api-operator-5694c8668f-jvspg"] Dec 03 14:11:54 crc kubenswrapper[4805]: I1203 14:11:54.898449 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/machine-api-operator-5694c8668f-jvspg" Dec 03 14:11:54 crc kubenswrapper[4805]: I1203 14:11:54.900256 4805 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-l6qxf"] Dec 03 14:11:54 crc kubenswrapper[4805]: I1203 14:11:54.901015 4805 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-vc9sn"] Dec 03 14:11:54 crc kubenswrapper[4805]: I1203 14:11:54.901111 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-l6qxf" Dec 03 14:11:54 crc kubenswrapper[4805]: I1203 14:11:54.901408 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"etcd-client" Dec 03 14:11:54 crc kubenswrapper[4805]: I1203 14:11:54.901567 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"machine-api-operator-dockercfg-mfbb7" Dec 03 14:11:54 crc kubenswrapper[4805]: I1203 14:11:54.901695 4805 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"kube-rbac-proxy" Dec 03 14:11:54 crc kubenswrapper[4805]: I1203 14:11:54.902171 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-vc9sn" Dec 03 14:11:54 crc kubenswrapper[4805]: I1203 14:11:54.904919 4805 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"openshift-service-ca.crt" Dec 03 14:11:54 crc kubenswrapper[4805]: I1203 14:11:54.904919 4805 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"audit-1" Dec 03 14:11:54 crc kubenswrapper[4805]: I1203 14:11:54.906380 4805 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-ntv6j"] Dec 03 14:11:54 crc kubenswrapper[4805]: I1203 14:11:54.906829 4805 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console/downloads-7954f5f757-s25zl"] Dec 03 14:11:54 crc kubenswrapper[4805]: I1203 14:11:54.906837 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-ntv6j" Dec 03 14:11:54 crc kubenswrapper[4805]: I1203 14:11:54.907506 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/downloads-7954f5f757-s25zl" Dec 03 14:11:54 crc kubenswrapper[4805]: I1203 14:11:54.910484 4805 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"machine-api-operator-images" Dec 03 14:11:54 crc kubenswrapper[4805]: I1203 14:11:54.912526 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"machine-api-operator-tls" Dec 03 14:11:54 crc kubenswrapper[4805]: I1203 14:11:54.913700 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"encryption-config-1" Dec 03 14:11:54 crc kubenswrapper[4805]: I1203 14:11:54.913910 4805 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"etcd-serving-ca" Dec 03 14:11:54 crc kubenswrapper[4805]: I1203 14:11:54.913941 4805 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"audit-1" Dec 03 14:11:54 crc kubenswrapper[4805]: I1203 14:11:54.913915 4805 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"config" Dec 03 14:11:54 crc kubenswrapper[4805]: I1203 14:11:54.914185 4805 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"openshift-service-ca.crt" Dec 03 14:11:54 crc kubenswrapper[4805]: I1203 14:11:54.919982 4805 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"image-import-ca" Dec 03 14:11:54 crc kubenswrapper[4805]: I1203 14:11:54.920759 4805 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"config" Dec 03 14:11:54 crc kubenswrapper[4805]: I1203 14:11:54.922804 4805 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"openshift-service-ca.crt" Dec 03 14:11:54 crc kubenswrapper[4805]: I1203 14:11:54.923560 4805 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"kube-root-ca.crt" Dec 03 14:11:54 crc kubenswrapper[4805]: I1203 14:11:54.923596 4805 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-service-ca.crt" Dec 03 14:11:54 crc kubenswrapper[4805]: I1203 14:11:54.923644 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"openshift-apiserver-sa-dockercfg-djjff" Dec 03 14:11:54 crc kubenswrapper[4805]: I1203 14:11:54.924487 4805 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"trusted-ca-bundle" Dec 03 14:11:54 crc kubenswrapper[4805]: I1203 14:11:54.924560 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"encryption-config-1" Dec 03 14:11:54 crc kubenswrapper[4805]: I1203 14:11:54.924505 4805 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"etcd-serving-ca" Dec 03 14:11:54 crc kubenswrapper[4805]: I1203 14:11:54.924893 4805 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"openshift-service-ca.crt" Dec 03 14:11:54 crc kubenswrapper[4805]: I1203 14:11:54.924973 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"etcd-client" Dec 03 14:11:54 crc kubenswrapper[4805]: I1203 14:11:54.925128 4805 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"config" Dec 03 14:11:54 crc kubenswrapper[4805]: I1203 14:11:54.925250 4805 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-cluster-machine-approver/machine-approver-56656f9798-nv2m9"] Dec 03 14:11:54 crc kubenswrapper[4805]: I1203 14:11:54.925453 4805 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"kube-root-ca.crt" Dec 03 14:11:54 crc kubenswrapper[4805]: I1203 14:11:54.925599 4805 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"kube-root-ca.crt" Dec 03 14:11:54 crc kubenswrapper[4805]: I1203 14:11:54.926003 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"serving-cert" Dec 03 14:11:54 crc kubenswrapper[4805]: I1203 14:11:54.926207 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-nv2m9" Dec 03 14:11:54 crc kubenswrapper[4805]: I1203 14:11:54.926567 4805 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"client-ca" Dec 03 14:11:54 crc kubenswrapper[4805]: I1203 14:11:54.926579 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"serving-cert" Dec 03 14:11:54 crc kubenswrapper[4805]: I1203 14:11:54.927044 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"route-controller-manager-sa-dockercfg-h2zr2" Dec 03 14:11:54 crc kubenswrapper[4805]: I1203 14:11:54.929096 4805 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console-operator/console-operator-58897d9998-46w2g"] Dec 03 14:11:54 crc kubenswrapper[4805]: I1203 14:11:54.940675 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console-operator/console-operator-58897d9998-46w2g" Dec 03 14:11:54 crc kubenswrapper[4805]: I1203 14:11:54.943980 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-idp-0-file-data" Dec 03 14:11:54 crc kubenswrapper[4805]: I1203 14:11:54.944164 4805 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"audit" Dec 03 14:11:54 crc kubenswrapper[4805]: I1203 14:11:54.944212 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"serving-cert" Dec 03 14:11:54 crc kubenswrapper[4805]: I1203 14:11:54.944249 4805 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-config" Dec 03 14:11:54 crc kubenswrapper[4805]: I1203 14:11:54.944311 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"serving-cert" Dec 03 14:11:54 crc kubenswrapper[4805]: I1203 14:11:54.943978 4805 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-samples-operator"/"openshift-service-ca.crt" Dec 03 14:11:54 crc kubenswrapper[4805]: I1203 14:11:54.943986 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-samples-operator"/"samples-operator-tls" Dec 03 14:11:54 crc kubenswrapper[4805]: I1203 14:11:54.944631 4805 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-service-ca" Dec 03 14:11:54 crc kubenswrapper[4805]: I1203 14:11:54.944043 4805 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"kube-root-ca.crt" Dec 03 14:11:54 crc kubenswrapper[4805]: I1203 14:11:54.944052 4805 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"openshift-service-ca.crt" Dec 03 14:11:54 crc kubenswrapper[4805]: I1203 14:11:54.944092 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-error" Dec 03 14:11:54 crc kubenswrapper[4805]: I1203 14:11:54.944135 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-samples-operator"/"cluster-samples-operator-dockercfg-xpp9w" Dec 03 14:11:54 crc kubenswrapper[4805]: I1203 14:11:54.944148 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-session" Dec 03 14:11:54 crc kubenswrapper[4805]: I1203 14:11:54.944146 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"openshift-controller-manager-sa-dockercfg-msq4c" Dec 03 14:11:54 crc kubenswrapper[4805]: I1203 14:11:54.945073 4805 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"client-ca" Dec 03 14:11:54 crc kubenswrapper[4805]: I1203 14:11:54.959384 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-dockercfg-xtcjv" Dec 03 14:11:54 crc kubenswrapper[4805]: I1203 14:11:54.959554 4805 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"kube-root-ca.crt" Dec 03 14:11:54 crc kubenswrapper[4805]: I1203 14:11:54.959877 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"oauth-apiserver-sa-dockercfg-6r2bq" Dec 03 14:11:54 crc kubenswrapper[4805]: I1203 14:11:54.960047 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-provider-selection" Dec 03 14:11:54 crc kubenswrapper[4805]: I1203 14:11:54.960141 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"oauth-openshift-dockercfg-znhcc" Dec 03 14:11:54 crc kubenswrapper[4805]: I1203 14:11:54.960431 4805 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"openshift-service-ca.crt" Dec 03 14:11:54 crc kubenswrapper[4805]: I1203 14:11:54.960513 4805 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"openshift-service-ca.crt" Dec 03 14:11:54 crc kubenswrapper[4805]: I1203 14:11:54.961112 4805 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"kube-root-ca.crt" Dec 03 14:11:54 crc kubenswrapper[4805]: I1203 14:11:54.962657 4805 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-cliconfig" Dec 03 14:11:54 crc kubenswrapper[4805]: I1203 14:11:54.962776 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-serving-cert" Dec 03 14:11:54 crc kubenswrapper[4805]: I1203 14:11:54.964081 4805 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"kube-rbac-proxy" Dec 03 14:11:54 crc kubenswrapper[4805]: I1203 14:11:54.964199 4805 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"machine-approver-config" Dec 03 14:11:54 crc kubenswrapper[4805]: I1203 14:11:54.964432 4805 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"openshift-service-ca.crt" Dec 03 14:11:54 crc kubenswrapper[4805]: I1203 14:11:54.964810 4805 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-samples-operator"/"kube-root-ca.crt" Dec 03 14:11:54 crc kubenswrapper[4805]: I1203 14:11:54.964981 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"default-dockercfg-chnjx" Dec 03 14:11:54 crc kubenswrapper[4805]: I1203 14:11:54.965062 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-serving-cert" Dec 03 14:11:54 crc kubenswrapper[4805]: I1203 14:11:54.965500 4805 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"openshift-service-ca.crt" Dec 03 14:11:54 crc kubenswrapper[4805]: I1203 14:11:54.967793 4805 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-config-operator/openshift-config-operator-7777fb866f-7wzgf"] Dec 03 14:11:54 crc kubenswrapper[4805]: I1203 14:11:54.967882 4805 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"trusted-ca-bundle" Dec 03 14:11:54 crc kubenswrapper[4805]: I1203 14:11:54.968227 4805 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-authentication-operator/authentication-operator-69f744f599-bs6tj"] Dec 03 14:11:54 crc kubenswrapper[4805]: I1203 14:11:54.968429 4805 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console/console-f9d7485db-pp9tv"] Dec 03 14:11:54 crc kubenswrapper[4805]: I1203 14:11:54.968675 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-f9d7485db-pp9tv" Dec 03 14:11:54 crc kubenswrapper[4805]: I1203 14:11:54.969084 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-config-operator/openshift-config-operator-7777fb866f-7wzgf" Dec 03 14:11:54 crc kubenswrapper[4805]: I1203 14:11:54.969280 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication-operator/authentication-operator-69f744f599-bs6tj" Dec 03 14:11:54 crc kubenswrapper[4805]: I1203 14:11:54.972965 4805 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"kube-root-ca.crt" Dec 03 14:11:54 crc kubenswrapper[4805]: I1203 14:11:54.973217 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console-operator"/"serving-cert" Dec 03 14:11:54 crc kubenswrapper[4805]: I1203 14:11:54.973257 4805 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"console-operator-config" Dec 03 14:11:54 crc kubenswrapper[4805]: I1203 14:11:54.973392 4805 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"oauth-serving-cert" Dec 03 14:11:54 crc kubenswrapper[4805]: I1203 14:11:54.974150 4805 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-config-operator"/"openshift-service-ca.crt" Dec 03 14:11:54 crc kubenswrapper[4805]: I1203 14:11:54.974351 4805 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"kube-root-ca.crt" Dec 03 14:11:54 crc kubenswrapper[4805]: I1203 14:11:54.974384 4805 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"console-config" Dec 03 14:11:54 crc kubenswrapper[4805]: I1203 14:11:54.974508 4805 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-operator-74547568cd-kvmpp"] Dec 03 14:11:54 crc kubenswrapper[4805]: I1203 14:11:54.974537 4805 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"kube-root-ca.crt" Dec 03 14:11:54 crc kubenswrapper[4805]: I1203 14:11:54.974624 4805 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"kube-root-ca.crt" Dec 03 14:11:54 crc kubenswrapper[4805]: I1203 14:11:54.974678 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console-operator"/"console-operator-dockercfg-4xjcr" Dec 03 14:11:54 crc kubenswrapper[4805]: I1203 14:11:54.975066 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-oauth-config" Dec 03 14:11:54 crc kubenswrapper[4805]: I1203 14:11:54.975281 4805 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/multus-admission-controller-857f4d67dd-tn8tx"] Dec 03 14:11:54 crc kubenswrapper[4805]: I1203 14:11:54.975613 4805 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-rr7fx"] Dec 03 14:11:54 crc kubenswrapper[4805]: I1203 14:11:54.975926 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-rr7fx" Dec 03 14:11:54 crc kubenswrapper[4805]: I1203 14:11:54.976460 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-serving-cert" Dec 03 14:11:54 crc kubenswrapper[4805]: I1203 14:11:54.976550 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-kvmpp" Dec 03 14:11:54 crc kubenswrapper[4805]: I1203 14:11:54.976707 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-admission-controller-857f4d67dd-tn8tx" Dec 03 14:11:54 crc kubenswrapper[4805]: I1203 14:11:54.977020 4805 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"service-ca" Dec 03 14:11:54 crc kubenswrapper[4805]: I1203 14:11:54.977534 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-machine-approver"/"machine-approver-sa-dockercfg-nl2j4" Dec 03 14:11:54 crc kubenswrapper[4805]: I1203 14:11:54.978222 4805 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-service-ca-operator/service-ca-operator-777779d784-z7lgj"] Dec 03 14:11:54 crc kubenswrapper[4805]: I1203 14:11:54.978634 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca-operator/service-ca-operator-777779d784-z7lgj" Dec 03 14:11:54 crc kubenswrapper[4805]: I1203 14:11:54.978967 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-router-certs" Dec 03 14:11:54 crc kubenswrapper[4805]: I1203 14:11:54.979028 4805 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-config-operator"/"kube-root-ca.crt" Dec 03 14:11:54 crc kubenswrapper[4805]: I1203 14:11:54.979193 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication-operator"/"authentication-operator-dockercfg-mz9bj" Dec 03 14:11:54 crc kubenswrapper[4805]: I1203 14:11:54.979293 4805 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"kube-root-ca.crt" Dec 03 14:11:54 crc kubenswrapper[4805]: I1203 14:11:54.979388 4805 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"service-ca-bundle" Dec 03 14:11:54 crc kubenswrapper[4805]: I1203 14:11:54.979461 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-machine-approver"/"machine-approver-tls" Dec 03 14:11:54 crc kubenswrapper[4805]: I1203 14:11:54.979478 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-config-operator"/"openshift-config-operator-dockercfg-7pc5z" Dec 03 14:11:54 crc kubenswrapper[4805]: I1203 14:11:54.979609 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-dockercfg-f62pw" Dec 03 14:11:54 crc kubenswrapper[4805]: I1203 14:11:54.979628 4805 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"openshift-service-ca.crt" Dec 03 14:11:54 crc kubenswrapper[4805]: I1203 14:11:54.979773 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-config-operator"/"config-operator-serving-cert" Dec 03 14:11:54 crc kubenswrapper[4805]: I1203 14:11:54.980469 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication-operator"/"serving-cert" Dec 03 14:11:54 crc kubenswrapper[4805]: I1203 14:11:54.983950 4805 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-global-ca" Dec 03 14:11:54 crc kubenswrapper[4805]: I1203 14:11:54.985295 4805 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-2vnrk"] Dec 03 14:11:54 crc kubenswrapper[4805]: I1203 14:11:54.985944 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-2vnrk" Dec 03 14:11:54 crc kubenswrapper[4805]: I1203 14:11:54.992747 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-login" Dec 03 14:11:54 crc kubenswrapper[4805]: I1203 14:11:54.994809 4805 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-trusted-ca-bundle" Dec 03 14:11:55 crc kubenswrapper[4805]: I1203 14:11:55.041228 4805 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29412840-gj7pr"] Dec 03 14:11:55 crc kubenswrapper[4805]: I1203 14:11:55.043171 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29412840-gj7pr" Dec 03 14:11:55 crc kubenswrapper[4805]: I1203 14:11:55.047811 4805 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-service-ca/service-ca-9c57cc56f-dbfrd"] Dec 03 14:11:55 crc kubenswrapper[4805]: I1203 14:11:55.048468 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/1658d8e1-ac63-40fe-98ce-900a9f9aeae5-audit-policies\") pod \"oauth-openshift-558db77b4-ntv6j\" (UID: \"1658d8e1-ac63-40fe-98ce-900a9f9aeae5\") " pod="openshift-authentication/oauth-openshift-558db77b4-ntv6j" Dec 03 14:11:55 crc kubenswrapper[4805]: I1203 14:11:55.048507 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b54908df-f04f-4515-ab36-842f0aef4a99-config\") pod \"controller-manager-879f6c89f-k7rgc\" (UID: \"b54908df-f04f-4515-ab36-842f0aef4a99\") " pod="openshift-controller-manager/controller-manager-879f6c89f-k7rgc" Dec 03 14:11:55 crc kubenswrapper[4805]: I1203 14:11:55.048534 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/b54908df-f04f-4515-ab36-842f0aef4a99-serving-cert\") pod \"controller-manager-879f6c89f-k7rgc\" (UID: \"b54908df-f04f-4515-ab36-842f0aef4a99\") " pod="openshift-controller-manager/controller-manager-879f6c89f-k7rgc" Dec 03 14:11:55 crc kubenswrapper[4805]: I1203 14:11:55.048563 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/e2034338-8a35-4fd2-9bbf-be733939fc25-oauth-serving-cert\") pod \"console-f9d7485db-pp9tv\" (UID: \"e2034338-8a35-4fd2-9bbf-be733939fc25\") " pod="openshift-console/console-f9d7485db-pp9tv" Dec 03 14:11:55 crc kubenswrapper[4805]: I1203 14:11:55.048594 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/1658d8e1-ac63-40fe-98ce-900a9f9aeae5-v4-0-config-system-serving-cert\") pod \"oauth-openshift-558db77b4-ntv6j\" (UID: \"1658d8e1-ac63-40fe-98ce-900a9f9aeae5\") " pod="openshift-authentication/oauth-openshift-558db77b4-ntv6j" Dec 03 14:11:55 crc kubenswrapper[4805]: I1203 14:11:55.048630 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/1658d8e1-ac63-40fe-98ce-900a9f9aeae5-v4-0-config-system-service-ca\") pod \"oauth-openshift-558db77b4-ntv6j\" (UID: \"1658d8e1-ac63-40fe-98ce-900a9f9aeae5\") " pod="openshift-authentication/oauth-openshift-558db77b4-ntv6j" Dec 03 14:11:55 crc kubenswrapper[4805]: I1203 14:11:55.048733 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dxgkw\" (UniqueName: \"kubernetes.io/projected/1658d8e1-ac63-40fe-98ce-900a9f9aeae5-kube-api-access-dxgkw\") pod \"oauth-openshift-558db77b4-ntv6j\" (UID: \"1658d8e1-ac63-40fe-98ce-900a9f9aeae5\") " pod="openshift-authentication/oauth-openshift-558db77b4-ntv6j" Dec 03 14:11:55 crc kubenswrapper[4805]: I1203 14:11:55.048761 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ca425595-aae1-44f2-9470-84ba52167748-config\") pod \"apiserver-76f77b778f-7nhvl\" (UID: \"ca425595-aae1-44f2-9470-84ba52167748\") " pod="openshift-apiserver/apiserver-76f77b778f-7nhvl" Dec 03 14:11:55 crc kubenswrapper[4805]: I1203 14:11:55.048788 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9b932fa5-f87b-46e1-8987-683cfd3b46fd-config\") pod \"service-ca-operator-777779d784-z7lgj\" (UID: \"9b932fa5-f87b-46e1-8987-683cfd3b46fd\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-z7lgj" Dec 03 14:11:55 crc kubenswrapper[4805]: I1203 14:11:55.048902 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/00941575-52e9-482d-b256-1b1c757d96e0-auth-proxy-config\") pod \"machine-approver-56656f9798-nv2m9\" (UID: \"00941575-52e9-482d-b256-1b1c757d96e0\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-nv2m9" Dec 03 14:11:55 crc kubenswrapper[4805]: I1203 14:11:55.048935 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pgzx9\" (UniqueName: \"kubernetes.io/projected/6c2d6fbc-771d-4aff-b1e8-3eda17204931-kube-api-access-pgzx9\") pod \"downloads-7954f5f757-s25zl\" (UID: \"6c2d6fbc-771d-4aff-b1e8-3eda17204931\") " pod="openshift-console/downloads-7954f5f757-s25zl" Dec 03 14:11:55 crc kubenswrapper[4805]: I1203 14:11:55.048966 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/9f53fe44-0d2c-4e3d-a682-df3a779a1c8a-trusted-ca\") pod \"console-operator-58897d9998-46w2g\" (UID: \"9f53fe44-0d2c-4e3d-a682-df3a779a1c8a\") " pod="openshift-console-operator/console-operator-58897d9998-46w2g" Dec 03 14:11:55 crc kubenswrapper[4805]: I1203 14:11:55.050493 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/ca425595-aae1-44f2-9470-84ba52167748-audit\") pod \"apiserver-76f77b778f-7nhvl\" (UID: \"ca425595-aae1-44f2-9470-84ba52167748\") " pod="openshift-apiserver/apiserver-76f77b778f-7nhvl" Dec 03 14:11:55 crc kubenswrapper[4805]: I1203 14:11:55.051694 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/1658d8e1-ac63-40fe-98ce-900a9f9aeae5-v4-0-config-system-cliconfig\") pod \"oauth-openshift-558db77b4-ntv6j\" (UID: \"1658d8e1-ac63-40fe-98ce-900a9f9aeae5\") " pod="openshift-authentication/oauth-openshift-558db77b4-ntv6j" Dec 03 14:11:55 crc kubenswrapper[4805]: I1203 14:11:55.052485 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/0a7d0b69-f9a6-4a40-9981-3dd999c400d5-audit-policies\") pod \"apiserver-7bbb656c7d-kfh2f\" (UID: \"0a7d0b69-f9a6-4a40-9981-3dd999c400d5\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-kfh2f" Dec 03 14:11:55 crc kubenswrapper[4805]: I1203 14:11:55.052518 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/dfa6979a-2473-469b-ac56-9086e35a63f4-config\") pod \"route-controller-manager-6576b87f9c-nsz7s\" (UID: \"dfa6979a-2473-469b-ac56-9086e35a63f4\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-nsz7s" Dec 03 14:11:55 crc kubenswrapper[4805]: I1203 14:11:55.052540 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4z5cq\" (UniqueName: \"kubernetes.io/projected/e5a33c32-3873-49b8-8a81-c1a0357d94fc-kube-api-access-4z5cq\") pod \"openshift-apiserver-operator-796bbdcf4f-l6qxf\" (UID: \"e5a33c32-3873-49b8-8a81-c1a0357d94fc\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-l6qxf" Dec 03 14:11:55 crc kubenswrapper[4805]: I1203 14:11:55.052560 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-46hn5\" (UniqueName: \"kubernetes.io/projected/e2034338-8a35-4fd2-9bbf-be733939fc25-kube-api-access-46hn5\") pod \"console-f9d7485db-pp9tv\" (UID: \"e2034338-8a35-4fd2-9bbf-be733939fc25\") " pod="openshift-console/console-f9d7485db-pp9tv" Dec 03 14:11:55 crc kubenswrapper[4805]: I1203 14:11:55.052596 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/0a7d0b69-f9a6-4a40-9981-3dd999c400d5-trusted-ca-bundle\") pod \"apiserver-7bbb656c7d-kfh2f\" (UID: \"0a7d0b69-f9a6-4a40-9981-3dd999c400d5\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-kfh2f" Dec 03 14:11:55 crc kubenswrapper[4805]: I1203 14:11:55.052617 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/c6b0a18a-5cf2-430a-a0ef-2c2776839f11-samples-operator-tls\") pod \"cluster-samples-operator-665b6dd947-vc9sn\" (UID: \"c6b0a18a-5cf2-430a-a0ef-2c2776839f11\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-vc9sn" Dec 03 14:11:55 crc kubenswrapper[4805]: I1203 14:11:55.052636 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/cdcb6b89-e405-479e-a60d-96681328634e-config\") pod \"authentication-operator-69f744f599-bs6tj\" (UID: \"cdcb6b89-e405-479e-a60d-96681328634e\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-bs6tj" Dec 03 14:11:55 crc kubenswrapper[4805]: I1203 14:11:55.052656 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/cdcb6b89-e405-479e-a60d-96681328634e-service-ca-bundle\") pod \"authentication-operator-69f744f599-bs6tj\" (UID: \"cdcb6b89-e405-479e-a60d-96681328634e\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-bs6tj" Dec 03 14:11:55 crc kubenswrapper[4805]: I1203 14:11:55.052676 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/b54908df-f04f-4515-ab36-842f0aef4a99-client-ca\") pod \"controller-manager-879f6c89f-k7rgc\" (UID: \"b54908df-f04f-4515-ab36-842f0aef4a99\") " pod="openshift-controller-manager/controller-manager-879f6c89f-k7rgc" Dec 03 14:11:55 crc kubenswrapper[4805]: I1203 14:11:55.052727 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/0a7d0b69-f9a6-4a40-9981-3dd999c400d5-etcd-client\") pod \"apiserver-7bbb656c7d-kfh2f\" (UID: \"0a7d0b69-f9a6-4a40-9981-3dd999c400d5\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-kfh2f" Dec 03 14:11:55 crc kubenswrapper[4805]: I1203 14:11:55.052935 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca/service-ca-9c57cc56f-dbfrd" Dec 03 14:11:55 crc kubenswrapper[4805]: I1203 14:11:55.052988 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fb99x\" (UniqueName: \"kubernetes.io/projected/33c718fb-affd-410b-9f53-ddb68ba9760c-kube-api-access-fb99x\") pod \"multus-admission-controller-857f4d67dd-tn8tx\" (UID: \"33c718fb-affd-410b-9f53-ddb68ba9760c\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-tn8tx" Dec 03 14:11:55 crc kubenswrapper[4805]: I1203 14:11:55.053155 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/ca425595-aae1-44f2-9470-84ba52167748-node-pullsecrets\") pod \"apiserver-76f77b778f-7nhvl\" (UID: \"ca425595-aae1-44f2-9470-84ba52167748\") " pod="openshift-apiserver/apiserver-76f77b778f-7nhvl" Dec 03 14:11:55 crc kubenswrapper[4805]: I1203 14:11:55.053247 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4v8xc\" (UniqueName: \"kubernetes.io/projected/ca425595-aae1-44f2-9470-84ba52167748-kube-api-access-4v8xc\") pod \"apiserver-76f77b778f-7nhvl\" (UID: \"ca425595-aae1-44f2-9470-84ba52167748\") " pod="openshift-apiserver/apiserver-76f77b778f-7nhvl" Dec 03 14:11:55 crc kubenswrapper[4805]: I1203 14:11:55.053268 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/04bd3442-dbeb-4cb4-bc9f-b07b9872231f-tmpfs\") pod \"packageserver-d55dfcdfc-rr7fx\" (UID: \"04bd3442-dbeb-4cb4-bc9f-b07b9872231f\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-rr7fx" Dec 03 14:11:55 crc kubenswrapper[4805]: I1203 14:11:55.053287 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/b54908df-f04f-4515-ab36-842f0aef4a99-proxy-ca-bundles\") pod \"controller-manager-879f6c89f-k7rgc\" (UID: \"b54908df-f04f-4515-ab36-842f0aef4a99\") " pod="openshift-controller-manager/controller-manager-879f6c89f-k7rgc" Dec 03 14:11:55 crc kubenswrapper[4805]: I1203 14:11:55.053306 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/33c718fb-affd-410b-9f53-ddb68ba9760c-webhook-certs\") pod \"multus-admission-controller-857f4d67dd-tn8tx\" (UID: \"33c718fb-affd-410b-9f53-ddb68ba9760c\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-tn8tx" Dec 03 14:11:55 crc kubenswrapper[4805]: I1203 14:11:55.053323 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/ca425595-aae1-44f2-9470-84ba52167748-image-import-ca\") pod \"apiserver-76f77b778f-7nhvl\" (UID: \"ca425595-aae1-44f2-9470-84ba52167748\") " pod="openshift-apiserver/apiserver-76f77b778f-7nhvl" Dec 03 14:11:55 crc kubenswrapper[4805]: I1203 14:11:55.053345 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-n7759\" (UniqueName: \"kubernetes.io/projected/dfa6979a-2473-469b-ac56-9086e35a63f4-kube-api-access-n7759\") pod \"route-controller-manager-6576b87f9c-nsz7s\" (UID: \"dfa6979a-2473-469b-ac56-9086e35a63f4\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-nsz7s" Dec 03 14:11:55 crc kubenswrapper[4805]: I1203 14:11:55.053375 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/e2034338-8a35-4fd2-9bbf-be733939fc25-console-oauth-config\") pod \"console-f9d7485db-pp9tv\" (UID: \"e2034338-8a35-4fd2-9bbf-be733939fc25\") " pod="openshift-console/console-f9d7485db-pp9tv" Dec 03 14:11:55 crc kubenswrapper[4805]: I1203 14:11:55.053454 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/f4b1faf2-bf8b-483c-b28f-5885a9e8fe44-serving-cert\") pod \"openshift-config-operator-7777fb866f-7wzgf\" (UID: \"f4b1faf2-bf8b-483c-b28f-5885a9e8fe44\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-7wzgf" Dec 03 14:11:55 crc kubenswrapper[4805]: I1203 14:11:55.053495 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/1658d8e1-ac63-40fe-98ce-900a9f9aeae5-audit-dir\") pod \"oauth-openshift-558db77b4-ntv6j\" (UID: \"1658d8e1-ac63-40fe-98ce-900a9f9aeae5\") " pod="openshift-authentication/oauth-openshift-558db77b4-ntv6j" Dec 03 14:11:55 crc kubenswrapper[4805]: I1203 14:11:55.053570 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/d5ea0764-4290-4eab-b4ce-dbd4909455c1-auth-proxy-config\") pod \"machine-config-operator-74547568cd-kvmpp\" (UID: \"d5ea0764-4290-4eab-b4ce-dbd4909455c1\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-kvmpp" Dec 03 14:11:55 crc kubenswrapper[4805]: I1203 14:11:55.053624 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/1658d8e1-ac63-40fe-98ce-900a9f9aeae5-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-558db77b4-ntv6j\" (UID: \"1658d8e1-ac63-40fe-98ce-900a9f9aeae5\") " pod="openshift-authentication/oauth-openshift-558db77b4-ntv6j" Dec 03 14:11:55 crc kubenswrapper[4805]: I1203 14:11:55.053644 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/e2034338-8a35-4fd2-9bbf-be733939fc25-console-config\") pod \"console-f9d7485db-pp9tv\" (UID: \"e2034338-8a35-4fd2-9bbf-be733939fc25\") " pod="openshift-console/console-f9d7485db-pp9tv" Dec 03 14:11:55 crc kubenswrapper[4805]: I1203 14:11:55.053662 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/0a7d0b69-f9a6-4a40-9981-3dd999c400d5-encryption-config\") pod \"apiserver-7bbb656c7d-kfh2f\" (UID: \"0a7d0b69-f9a6-4a40-9981-3dd999c400d5\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-kfh2f" Dec 03 14:11:55 crc kubenswrapper[4805]: I1203 14:11:55.053796 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sjpnv\" (UniqueName: \"kubernetes.io/projected/f4b1faf2-bf8b-483c-b28f-5885a9e8fe44-kube-api-access-sjpnv\") pod \"openshift-config-operator-7777fb866f-7wzgf\" (UID: \"f4b1faf2-bf8b-483c-b28f-5885a9e8fe44\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-7wzgf" Dec 03 14:11:55 crc kubenswrapper[4805]: I1203 14:11:55.053826 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/ca425595-aae1-44f2-9470-84ba52167748-serving-cert\") pod \"apiserver-76f77b778f-7nhvl\" (UID: \"ca425595-aae1-44f2-9470-84ba52167748\") " pod="openshift-apiserver/apiserver-76f77b778f-7nhvl" Dec 03 14:11:55 crc kubenswrapper[4805]: I1203 14:11:55.053919 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/0a7d0b69-f9a6-4a40-9981-3dd999c400d5-etcd-serving-ca\") pod \"apiserver-7bbb656c7d-kfh2f\" (UID: \"0a7d0b69-f9a6-4a40-9981-3dd999c400d5\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-kfh2f" Dec 03 14:11:55 crc kubenswrapper[4805]: I1203 14:11:55.053938 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/1658d8e1-ac63-40fe-98ce-900a9f9aeae5-v4-0-config-user-template-error\") pod \"oauth-openshift-558db77b4-ntv6j\" (UID: \"1658d8e1-ac63-40fe-98ce-900a9f9aeae5\") " pod="openshift-authentication/oauth-openshift-558db77b4-ntv6j" Dec 03 14:11:55 crc kubenswrapper[4805]: I1203 14:11:55.053956 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9b932fa5-f87b-46e1-8987-683cfd3b46fd-serving-cert\") pod \"service-ca-operator-777779d784-z7lgj\" (UID: \"9b932fa5-f87b-46e1-8987-683cfd3b46fd\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-z7lgj" Dec 03 14:11:55 crc kubenswrapper[4805]: I1203 14:11:55.053978 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/1658d8e1-ac63-40fe-98ce-900a9f9aeae5-v4-0-config-user-template-login\") pod \"oauth-openshift-558db77b4-ntv6j\" (UID: \"1658d8e1-ac63-40fe-98ce-900a9f9aeae5\") " pod="openshift-authentication/oauth-openshift-558db77b4-ntv6j" Dec 03 14:11:55 crc kubenswrapper[4805]: I1203 14:11:55.053996 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/dfd20bfc-5c11-4287-a4b3-df9e0ca4dcf1-images\") pod \"machine-api-operator-5694c8668f-jvspg\" (UID: \"dfd20bfc-5c11-4287-a4b3-df9e0ca4dcf1\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-jvspg" Dec 03 14:11:55 crc kubenswrapper[4805]: I1203 14:11:55.054013 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/1658d8e1-ac63-40fe-98ce-900a9f9aeae5-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-558db77b4-ntv6j\" (UID: \"1658d8e1-ac63-40fe-98ce-900a9f9aeae5\") " pod="openshift-authentication/oauth-openshift-558db77b4-ntv6j" Dec 03 14:11:55 crc kubenswrapper[4805]: I1203 14:11:55.054031 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/dfa6979a-2473-469b-ac56-9086e35a63f4-client-ca\") pod \"route-controller-manager-6576b87f9c-nsz7s\" (UID: \"dfa6979a-2473-469b-ac56-9086e35a63f4\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-nsz7s" Dec 03 14:11:55 crc kubenswrapper[4805]: I1203 14:11:55.054115 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-x6v8p\" (UniqueName: \"kubernetes.io/projected/9f53fe44-0d2c-4e3d-a682-df3a779a1c8a-kube-api-access-x6v8p\") pod \"console-operator-58897d9998-46w2g\" (UID: \"9f53fe44-0d2c-4e3d-a682-df3a779a1c8a\") " pod="openshift-console-operator/console-operator-58897d9998-46w2g" Dec 03 14:11:55 crc kubenswrapper[4805]: I1203 14:11:55.054136 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-89vsl\" (UniqueName: \"kubernetes.io/projected/b54908df-f04f-4515-ab36-842f0aef4a99-kube-api-access-89vsl\") pod \"controller-manager-879f6c89f-k7rgc\" (UID: \"b54908df-f04f-4515-ab36-842f0aef4a99\") " pod="openshift-controller-manager/controller-manager-879f6c89f-k7rgc" Dec 03 14:11:55 crc kubenswrapper[4805]: I1203 14:11:55.054163 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/dfd20bfc-5c11-4287-a4b3-df9e0ca4dcf1-config\") pod \"machine-api-operator-5694c8668f-jvspg\" (UID: \"dfd20bfc-5c11-4287-a4b3-df9e0ca4dcf1\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-jvspg" Dec 03 14:11:55 crc kubenswrapper[4805]: I1203 14:11:55.054228 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/dfd20bfc-5c11-4287-a4b3-df9e0ca4dcf1-machine-api-operator-tls\") pod \"machine-api-operator-5694c8668f-jvspg\" (UID: \"dfd20bfc-5c11-4287-a4b3-df9e0ca4dcf1\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-jvspg" Dec 03 14:11:55 crc kubenswrapper[4805]: I1203 14:11:55.054272 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/ca425595-aae1-44f2-9470-84ba52167748-etcd-serving-ca\") pod \"apiserver-76f77b778f-7nhvl\" (UID: \"ca425595-aae1-44f2-9470-84ba52167748\") " pod="openshift-apiserver/apiserver-76f77b778f-7nhvl" Dec 03 14:11:55 crc kubenswrapper[4805]: I1203 14:11:55.054321 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/ca425595-aae1-44f2-9470-84ba52167748-encryption-config\") pod \"apiserver-76f77b778f-7nhvl\" (UID: \"ca425595-aae1-44f2-9470-84ba52167748\") " pod="openshift-apiserver/apiserver-76f77b778f-7nhvl" Dec 03 14:11:55 crc kubenswrapper[4805]: I1203 14:11:55.054467 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e5a33c32-3873-49b8-8a81-c1a0357d94fc-config\") pod \"openshift-apiserver-operator-796bbdcf4f-l6qxf\" (UID: \"e5a33c32-3873-49b8-8a81-c1a0357d94fc\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-l6qxf" Dec 03 14:11:55 crc kubenswrapper[4805]: I1203 14:11:55.054492 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8gr5q\" (UniqueName: \"kubernetes.io/projected/d5ea0764-4290-4eab-b4ce-dbd4909455c1-kube-api-access-8gr5q\") pod \"machine-config-operator-74547568cd-kvmpp\" (UID: \"d5ea0764-4290-4eab-b4ce-dbd4909455c1\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-kvmpp" Dec 03 14:11:55 crc kubenswrapper[4805]: I1203 14:11:55.054513 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/d5ea0764-4290-4eab-b4ce-dbd4909455c1-proxy-tls\") pod \"machine-config-operator-74547568cd-kvmpp\" (UID: \"d5ea0764-4290-4eab-b4ce-dbd4909455c1\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-kvmpp" Dec 03 14:11:55 crc kubenswrapper[4805]: I1203 14:11:55.054561 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/cdcb6b89-e405-479e-a60d-96681328634e-serving-cert\") pod \"authentication-operator-69f744f599-bs6tj\" (UID: \"cdcb6b89-e405-479e-a60d-96681328634e\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-bs6tj" Dec 03 14:11:55 crc kubenswrapper[4805]: I1203 14:11:55.054581 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/04bd3442-dbeb-4cb4-bc9f-b07b9872231f-webhook-cert\") pod \"packageserver-d55dfcdfc-rr7fx\" (UID: \"04bd3442-dbeb-4cb4-bc9f-b07b9872231f\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-rr7fx" Dec 03 14:11:55 crc kubenswrapper[4805]: I1203 14:11:55.054596 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/cdcb6b89-e405-479e-a60d-96681328634e-trusted-ca-bundle\") pod \"authentication-operator-69f744f599-bs6tj\" (UID: \"cdcb6b89-e405-479e-a60d-96681328634e\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-bs6tj" Dec 03 14:11:55 crc kubenswrapper[4805]: I1203 14:11:55.054632 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/d5ea0764-4290-4eab-b4ce-dbd4909455c1-images\") pod \"machine-config-operator-74547568cd-kvmpp\" (UID: \"d5ea0764-4290-4eab-b4ce-dbd4909455c1\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-kvmpp" Dec 03 14:11:55 crc kubenswrapper[4805]: I1203 14:11:55.054650 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/1658d8e1-ac63-40fe-98ce-900a9f9aeae5-v4-0-config-system-router-certs\") pod \"oauth-openshift-558db77b4-ntv6j\" (UID: \"1658d8e1-ac63-40fe-98ce-900a9f9aeae5\") " pod="openshift-authentication/oauth-openshift-558db77b4-ntv6j" Dec 03 14:11:55 crc kubenswrapper[4805]: I1203 14:11:55.054670 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/1658d8e1-ac63-40fe-98ce-900a9f9aeae5-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-558db77b4-ntv6j\" (UID: \"1658d8e1-ac63-40fe-98ce-900a9f9aeae5\") " pod="openshift-authentication/oauth-openshift-558db77b4-ntv6j" Dec 03 14:11:55 crc kubenswrapper[4805]: I1203 14:11:55.054775 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/e2034338-8a35-4fd2-9bbf-be733939fc25-service-ca\") pod \"console-f9d7485db-pp9tv\" (UID: \"e2034338-8a35-4fd2-9bbf-be733939fc25\") " pod="openshift-console/console-f9d7485db-pp9tv" Dec 03 14:11:55 crc kubenswrapper[4805]: I1203 14:11:55.054802 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lwvs6\" (UniqueName: \"kubernetes.io/projected/00941575-52e9-482d-b256-1b1c757d96e0-kube-api-access-lwvs6\") pod \"machine-approver-56656f9798-nv2m9\" (UID: \"00941575-52e9-482d-b256-1b1c757d96e0\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-nv2m9" Dec 03 14:11:55 crc kubenswrapper[4805]: I1203 14:11:55.054826 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/0a7d0b69-f9a6-4a40-9981-3dd999c400d5-audit-dir\") pod \"apiserver-7bbb656c7d-kfh2f\" (UID: \"0a7d0b69-f9a6-4a40-9981-3dd999c400d5\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-kfh2f" Dec 03 14:11:55 crc kubenswrapper[4805]: I1203 14:11:55.054997 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/e2034338-8a35-4fd2-9bbf-be733939fc25-trusted-ca-bundle\") pod \"console-f9d7485db-pp9tv\" (UID: \"e2034338-8a35-4fd2-9bbf-be733939fc25\") " pod="openshift-console/console-f9d7485db-pp9tv" Dec 03 14:11:55 crc kubenswrapper[4805]: I1203 14:11:55.055000 4805 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"authentication-operator-config" Dec 03 14:11:55 crc kubenswrapper[4805]: I1203 14:11:55.055019 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/ca425595-aae1-44f2-9470-84ba52167748-audit-dir\") pod \"apiserver-76f77b778f-7nhvl\" (UID: \"ca425595-aae1-44f2-9470-84ba52167748\") " pod="openshift-apiserver/apiserver-76f77b778f-7nhvl" Dec 03 14:11:55 crc kubenswrapper[4805]: I1203 14:11:55.055155 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/e2034338-8a35-4fd2-9bbf-be733939fc25-console-serving-cert\") pod \"console-f9d7485db-pp9tv\" (UID: \"e2034338-8a35-4fd2-9bbf-be733939fc25\") " pod="openshift-console/console-f9d7485db-pp9tv" Dec 03 14:11:55 crc kubenswrapper[4805]: I1203 14:11:55.055176 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/ca425595-aae1-44f2-9470-84ba52167748-etcd-client\") pod \"apiserver-76f77b778f-7nhvl\" (UID: \"ca425595-aae1-44f2-9470-84ba52167748\") " pod="openshift-apiserver/apiserver-76f77b778f-7nhvl" Dec 03 14:11:55 crc kubenswrapper[4805]: I1203 14:11:55.055192 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"mco-proxy-tls" Dec 03 14:11:55 crc kubenswrapper[4805]: I1203 14:11:55.055358 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0a7d0b69-f9a6-4a40-9981-3dd999c400d5-serving-cert\") pod \"apiserver-7bbb656c7d-kfh2f\" (UID: \"0a7d0b69-f9a6-4a40-9981-3dd999c400d5\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-kfh2f" Dec 03 14:11:55 crc kubenswrapper[4805]: I1203 14:11:55.055422 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e5a33c32-3873-49b8-8a81-c1a0357d94fc-serving-cert\") pod \"openshift-apiserver-operator-796bbdcf4f-l6qxf\" (UID: \"e5a33c32-3873-49b8-8a81-c1a0357d94fc\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-l6qxf" Dec 03 14:11:55 crc kubenswrapper[4805]: I1203 14:11:55.055444 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/00941575-52e9-482d-b256-1b1c757d96e0-machine-approver-tls\") pod \"machine-approver-56656f9798-nv2m9\" (UID: \"00941575-52e9-482d-b256-1b1c757d96e0\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-nv2m9" Dec 03 14:11:55 crc kubenswrapper[4805]: I1203 14:11:55.055484 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/1658d8e1-ac63-40fe-98ce-900a9f9aeae5-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-558db77b4-ntv6j\" (UID: \"1658d8e1-ac63-40fe-98ce-900a9f9aeae5\") " pod="openshift-authentication/oauth-openshift-558db77b4-ntv6j" Dec 03 14:11:55 crc kubenswrapper[4805]: I1203 14:11:55.055501 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9f53fe44-0d2c-4e3d-a682-df3a779a1c8a-config\") pod \"console-operator-58897d9998-46w2g\" (UID: \"9f53fe44-0d2c-4e3d-a682-df3a779a1c8a\") " pod="openshift-console-operator/console-operator-58897d9998-46w2g" Dec 03 14:11:55 crc kubenswrapper[4805]: I1203 14:11:55.055520 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8zh8j\" (UniqueName: \"kubernetes.io/projected/cdcb6b89-e405-479e-a60d-96681328634e-kube-api-access-8zh8j\") pod \"authentication-operator-69f744f599-bs6tj\" (UID: \"cdcb6b89-e405-479e-a60d-96681328634e\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-bs6tj" Dec 03 14:11:55 crc kubenswrapper[4805]: I1203 14:11:55.055663 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6cs72\" (UniqueName: \"kubernetes.io/projected/04bd3442-dbeb-4cb4-bc9f-b07b9872231f-kube-api-access-6cs72\") pod \"packageserver-d55dfcdfc-rr7fx\" (UID: \"04bd3442-dbeb-4cb4-bc9f-b07b9872231f\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-rr7fx" Dec 03 14:11:55 crc kubenswrapper[4805]: I1203 14:11:55.063720 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/dfa6979a-2473-469b-ac56-9086e35a63f4-serving-cert\") pod \"route-controller-manager-6576b87f9c-nsz7s\" (UID: \"dfa6979a-2473-469b-ac56-9086e35a63f4\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-nsz7s" Dec 03 14:11:55 crc kubenswrapper[4805]: I1203 14:11:55.063741 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/00941575-52e9-482d-b256-1b1c757d96e0-config\") pod \"machine-approver-56656f9798-nv2m9\" (UID: \"00941575-52e9-482d-b256-1b1c757d96e0\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-nv2m9" Dec 03 14:11:55 crc kubenswrapper[4805]: I1203 14:11:55.063757 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/ca425595-aae1-44f2-9470-84ba52167748-trusted-ca-bundle\") pod \"apiserver-76f77b778f-7nhvl\" (UID: \"ca425595-aae1-44f2-9470-84ba52167748\") " pod="openshift-apiserver/apiserver-76f77b778f-7nhvl" Dec 03 14:11:55 crc kubenswrapper[4805]: I1203 14:11:55.063776 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9f53fe44-0d2c-4e3d-a682-df3a779a1c8a-serving-cert\") pod \"console-operator-58897d9998-46w2g\" (UID: \"9f53fe44-0d2c-4e3d-a682-df3a779a1c8a\") " pod="openshift-console-operator/console-operator-58897d9998-46w2g" Dec 03 14:11:55 crc kubenswrapper[4805]: I1203 14:11:55.063791 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rnn8j\" (UniqueName: \"kubernetes.io/projected/dfd20bfc-5c11-4287-a4b3-df9e0ca4dcf1-kube-api-access-rnn8j\") pod \"machine-api-operator-5694c8668f-jvspg\" (UID: \"dfd20bfc-5c11-4287-a4b3-df9e0ca4dcf1\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-jvspg" Dec 03 14:11:55 crc kubenswrapper[4805]: I1203 14:11:55.063806 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-p5sc5\" (UniqueName: \"kubernetes.io/projected/c6b0a18a-5cf2-430a-a0ef-2c2776839f11-kube-api-access-p5sc5\") pod \"cluster-samples-operator-665b6dd947-vc9sn\" (UID: \"c6b0a18a-5cf2-430a-a0ef-2c2776839f11\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-vc9sn" Dec 03 14:11:55 crc kubenswrapper[4805]: I1203 14:11:55.063820 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2cmpw\" (UniqueName: \"kubernetes.io/projected/0a7d0b69-f9a6-4a40-9981-3dd999c400d5-kube-api-access-2cmpw\") pod \"apiserver-7bbb656c7d-kfh2f\" (UID: \"0a7d0b69-f9a6-4a40-9981-3dd999c400d5\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-kfh2f" Dec 03 14:11:55 crc kubenswrapper[4805]: I1203 14:11:55.063935 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"packageserver-service-cert" Dec 03 14:11:55 crc kubenswrapper[4805]: I1203 14:11:55.064066 4805 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"openshift-service-ca.crt" Dec 03 14:11:55 crc kubenswrapper[4805]: I1203 14:11:55.064210 4805 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"machine-config-operator-images" Dec 03 14:11:55 crc kubenswrapper[4805]: I1203 14:11:55.064690 4805 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"trusted-ca-bundle" Dec 03 14:11:55 crc kubenswrapper[4805]: I1203 14:11:55.065395 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/f4b1faf2-bf8b-483c-b28f-5885a9e8fe44-available-featuregates\") pod \"openshift-config-operator-7777fb866f-7wzgf\" (UID: \"f4b1faf2-bf8b-483c-b28f-5885a9e8fe44\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-7wzgf" Dec 03 14:11:55 crc kubenswrapper[4805]: I1203 14:11:55.065434 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qzz6l\" (UniqueName: \"kubernetes.io/projected/9b932fa5-f87b-46e1-8987-683cfd3b46fd-kube-api-access-qzz6l\") pod \"service-ca-operator-777779d784-z7lgj\" (UID: \"9b932fa5-f87b-46e1-8987-683cfd3b46fd\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-z7lgj" Dec 03 14:11:55 crc kubenswrapper[4805]: I1203 14:11:55.065484 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/04bd3442-dbeb-4cb4-bc9f-b07b9872231f-apiservice-cert\") pod \"packageserver-d55dfcdfc-rr7fx\" (UID: \"04bd3442-dbeb-4cb4-bc9f-b07b9872231f\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-rr7fx" Dec 03 14:11:55 crc kubenswrapper[4805]: I1203 14:11:55.065503 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/1658d8e1-ac63-40fe-98ce-900a9f9aeae5-v4-0-config-system-session\") pod \"oauth-openshift-558db77b4-ntv6j\" (UID: \"1658d8e1-ac63-40fe-98ce-900a9f9aeae5\") " pod="openshift-authentication/oauth-openshift-558db77b4-ntv6j" Dec 03 14:11:55 crc kubenswrapper[4805]: I1203 14:11:55.066013 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-operator-dockercfg-98p87" Dec 03 14:11:55 crc kubenswrapper[4805]: I1203 14:11:55.066204 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-ac-dockercfg-9lkdf" Dec 03 14:11:55 crc kubenswrapper[4805]: I1203 14:11:55.066324 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-admission-controller-secret" Dec 03 14:11:55 crc kubenswrapper[4805]: I1203 14:11:55.069093 4805 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"trusted-ca" Dec 03 14:11:55 crc kubenswrapper[4805]: I1203 14:11:55.070557 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"olm-operator-serviceaccount-dockercfg-rq7zk" Dec 03 14:11:55 crc kubenswrapper[4805]: I1203 14:11:55.071017 4805 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"trusted-ca-bundle" Dec 03 14:11:55 crc kubenswrapper[4805]: I1203 14:11:55.071335 4805 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-etcd-operator/etcd-operator-b45778765-2x56c"] Dec 03 14:11:55 crc kubenswrapper[4805]: I1203 14:11:55.071533 4805 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"kube-root-ca.crt" Dec 03 14:11:55 crc kubenswrapper[4805]: I1203 14:11:55.071680 4805 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"service-ca-operator-config" Dec 03 14:11:55 crc kubenswrapper[4805]: I1203 14:11:55.071807 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd-operator/etcd-operator-b45778765-2x56c" Dec 03 14:11:55 crc kubenswrapper[4805]: I1203 14:11:55.072174 4805 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-dns-operator/dns-operator-744455d44c-rjf9m"] Dec 03 14:11:55 crc kubenswrapper[4805]: I1203 14:11:55.072583 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-ocp-branding-template" Dec 03 14:11:55 crc kubenswrapper[4805]: I1203 14:11:55.072863 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns-operator/dns-operator-744455d44c-rjf9m" Dec 03 14:11:55 crc kubenswrapper[4805]: I1203 14:11:55.073656 4805 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-spjlp"] Dec 03 14:11:55 crc kubenswrapper[4805]: I1203 14:11:55.074010 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-spjlp" Dec 03 14:11:55 crc kubenswrapper[4805]: I1203 14:11:55.075766 4805 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-controller-84d6567774-t4kl8"] Dec 03 14:11:55 crc kubenswrapper[4805]: I1203 14:11:55.076245 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-t4kl8" Dec 03 14:11:55 crc kubenswrapper[4805]: I1203 14:11:55.076494 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca-operator"/"service-ca-operator-dockercfg-rg9jl" Dec 03 14:11:55 crc kubenswrapper[4805]: I1203 14:11:55.078060 4805 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-jjzct"] Dec 03 14:11:55 crc kubenswrapper[4805]: I1203 14:11:55.078601 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-jjzct" Dec 03 14:11:55 crc kubenswrapper[4805]: I1203 14:11:55.079341 4805 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-jvsqf"] Dec 03 14:11:55 crc kubenswrapper[4805]: I1203 14:11:55.080003 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-jvsqf" Dec 03 14:11:55 crc kubenswrapper[4805]: I1203 14:11:55.080607 4805 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-storage-version-migrator/migrator-59844c95c7-h5tv7"] Dec 03 14:11:55 crc kubenswrapper[4805]: I1203 14:11:55.085053 4805 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ingress-operator/ingress-operator-5b745b69d9-j8pwh"] Dec 03 14:11:55 crc kubenswrapper[4805]: I1203 14:11:55.085798 4805 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-ss8kj"] Dec 03 14:11:55 crc kubenswrapper[4805]: I1203 14:11:55.086229 4805 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-g46rj"] Dec 03 14:11:55 crc kubenswrapper[4805]: I1203 14:11:55.086612 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-h5tv7" Dec 03 14:11:55 crc kubenswrapper[4805]: I1203 14:11:55.086612 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-j8pwh" Dec 03 14:11:55 crc kubenswrapper[4805]: I1203 14:11:55.087162 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-ss8kj" Dec 03 14:11:55 crc kubenswrapper[4805]: I1203 14:11:55.088822 4805 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-86589"] Dec 03 14:11:55 crc kubenswrapper[4805]: I1203 14:11:55.089257 4805 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/catalog-operator-68c6474976-rp9lf"] Dec 03 14:11:55 crc kubenswrapper[4805]: I1203 14:11:55.090309 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-g46rj" Dec 03 14:11:55 crc kubenswrapper[4805]: I1203 14:11:55.091076 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver/apiserver-76f77b778f-7nhvl"] Dec 03 14:11:55 crc kubenswrapper[4805]: I1203 14:11:55.091074 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-86589" Dec 03 14:11:55 crc kubenswrapper[4805]: I1203 14:11:55.091108 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-rp9lf" Dec 03 14:11:55 crc kubenswrapper[4805]: I1203 14:11:55.091298 4805 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-6n5xz"] Dec 03 14:11:55 crc kubenswrapper[4805]: I1203 14:11:55.091718 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-6n5xz" Dec 03 14:11:55 crc kubenswrapper[4805]: I1203 14:11:55.093799 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-k7rgc"] Dec 03 14:11:55 crc kubenswrapper[4805]: I1203 14:11:55.094741 4805 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ingress/router-default-5444994796-kxnqm"] Dec 03 14:11:55 crc kubenswrapper[4805]: I1203 14:11:55.095614 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress/router-default-5444994796-kxnqm" Dec 03 14:11:55 crc kubenswrapper[4805]: I1203 14:11:55.096048 4805 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-b7wrl"] Dec 03 14:11:55 crc kubenswrapper[4805]: I1203 14:11:55.096543 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-b7wrl" Dec 03 14:11:55 crc kubenswrapper[4805]: I1203 14:11:55.097280 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca-operator"/"serving-cert" Dec 03 14:11:55 crc kubenswrapper[4805]: I1203 14:11:55.099394 4805 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-25khc"] Dec 03 14:11:55 crc kubenswrapper[4805]: I1203 14:11:55.099876 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-25khc" Dec 03 14:11:55 crc kubenswrapper[4805]: I1203 14:11:55.101983 4805 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-94zfm"] Dec 03 14:11:55 crc kubenswrapper[4805]: I1203 14:11:55.102511 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-94zfm" Dec 03 14:11:55 crc kubenswrapper[4805]: I1203 14:11:55.103166 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-nsz7s"] Dec 03 14:11:55 crc kubenswrapper[4805]: I1203 14:11:55.104990 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-oauth-apiserver/apiserver-7bbb656c7d-kfh2f"] Dec 03 14:11:55 crc kubenswrapper[4805]: I1203 14:11:55.106231 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/downloads-7954f5f757-s25zl"] Dec 03 14:11:55 crc kubenswrapper[4805]: I1203 14:11:55.106981 4805 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ingress-canary/ingress-canary-cnlwr"] Dec 03 14:11:55 crc kubenswrapper[4805]: I1203 14:11:55.107784 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-canary/ingress-canary-cnlwr" Dec 03 14:11:55 crc kubenswrapper[4805]: I1203 14:11:55.109271 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-vc9sn"] Dec 03 14:11:55 crc kubenswrapper[4805]: I1203 14:11:55.111608 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns-operator/dns-operator-744455d44c-rjf9m"] Dec 03 14:11:55 crc kubenswrapper[4805]: I1203 14:11:55.115769 4805 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"kube-root-ca.crt" Dec 03 14:11:55 crc kubenswrapper[4805]: I1203 14:11:55.116007 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-config-operator/openshift-config-operator-7777fb866f-7wzgf"] Dec 03 14:11:55 crc kubenswrapper[4805]: I1203 14:11:55.116941 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console-operator/console-operator-58897d9998-46w2g"] Dec 03 14:11:55 crc kubenswrapper[4805]: I1203 14:11:55.118040 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-ntv6j"] Dec 03 14:11:55 crc kubenswrapper[4805]: I1203 14:11:55.119198 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-rr7fx"] Dec 03 14:11:55 crc kubenswrapper[4805]: I1203 14:11:55.120422 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29412840-gj7pr"] Dec 03 14:11:55 crc kubenswrapper[4805]: I1203 14:11:55.121464 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca-operator/service-ca-operator-777779d784-z7lgj"] Dec 03 14:11:55 crc kubenswrapper[4805]: I1203 14:11:55.126736 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-l6qxf"] Dec 03 14:11:55 crc kubenswrapper[4805]: I1203 14:11:55.128285 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-etcd-operator/etcd-operator-b45778765-2x56c"] Dec 03 14:11:55 crc kubenswrapper[4805]: I1203 14:11:55.133537 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/multus-admission-controller-857f4d67dd-tn8tx"] Dec 03 14:11:55 crc kubenswrapper[4805]: I1203 14:11:55.136057 4805 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"openshift-service-ca.crt" Dec 03 14:11:55 crc kubenswrapper[4805]: I1203 14:11:55.136194 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-controller-84d6567774-t4kl8"] Dec 03 14:11:55 crc kubenswrapper[4805]: I1203 14:11:55.137035 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-jvsqf"] Dec 03 14:11:55 crc kubenswrapper[4805]: I1203 14:11:55.138256 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/machine-api-operator-5694c8668f-jvspg"] Dec 03 14:11:55 crc kubenswrapper[4805]: I1203 14:11:55.139757 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-ss8kj"] Dec 03 14:11:55 crc kubenswrapper[4805]: I1203 14:11:55.141640 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-g46rj"] Dec 03 14:11:55 crc kubenswrapper[4805]: I1203 14:11:55.143044 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-spjlp"] Dec 03 14:11:55 crc kubenswrapper[4805]: I1203 14:11:55.144726 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-f9d7485db-pp9tv"] Dec 03 14:11:55 crc kubenswrapper[4805]: I1203 14:11:55.145858 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-94zfm"] Dec 03 14:11:55 crc kubenswrapper[4805]: I1203 14:11:55.147080 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication-operator/authentication-operator-69f744f599-bs6tj"] Dec 03 14:11:55 crc kubenswrapper[4805]: I1203 14:11:55.148139 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-2vnrk"] Dec 03 14:11:55 crc kubenswrapper[4805]: I1203 14:11:55.149148 4805 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-dns/dns-default-7wnwf"] Dec 03 14:11:55 crc kubenswrapper[4805]: I1203 14:11:55.149944 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/dns-default-7wnwf" Dec 03 14:11:55 crc kubenswrapper[4805]: I1203 14:11:55.150433 4805 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-server-d5pjl"] Dec 03 14:11:55 crc kubenswrapper[4805]: I1203 14:11:55.150830 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-server-d5pjl" Dec 03 14:11:55 crc kubenswrapper[4805]: I1203 14:11:55.151699 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-86589"] Dec 03 14:11:55 crc kubenswrapper[4805]: I1203 14:11:55.153088 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator/migrator-59844c95c7-h5tv7"] Dec 03 14:11:55 crc kubenswrapper[4805]: I1203 14:11:55.154558 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-operator/ingress-operator-5b745b69d9-j8pwh"] Dec 03 14:11:55 crc kubenswrapper[4805]: I1203 14:11:55.155575 4805 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-config" Dec 03 14:11:55 crc kubenswrapper[4805]: I1203 14:11:55.156933 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-6n5xz"] Dec 03 14:11:55 crc kubenswrapper[4805]: I1203 14:11:55.158048 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-canary/ingress-canary-cnlwr"] Dec 03 14:11:55 crc kubenswrapper[4805]: I1203 14:11:55.159495 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca/service-ca-9c57cc56f-dbfrd"] Dec 03 14:11:55 crc kubenswrapper[4805]: I1203 14:11:55.161238 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-jjzct"] Dec 03 14:11:55 crc kubenswrapper[4805]: I1203 14:11:55.162728 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/catalog-operator-68c6474976-rp9lf"] Dec 03 14:11:55 crc kubenswrapper[4805]: I1203 14:11:55.164390 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-operator-74547568cd-kvmpp"] Dec 03 14:11:55 crc kubenswrapper[4805]: I1203 14:11:55.165876 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns/dns-default-7wnwf"] Dec 03 14:11:55 crc kubenswrapper[4805]: I1203 14:11:55.166242 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/dfa6979a-2473-469b-ac56-9086e35a63f4-config\") pod \"route-controller-manager-6576b87f9c-nsz7s\" (UID: \"dfa6979a-2473-469b-ac56-9086e35a63f4\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-nsz7s" Dec 03 14:11:55 crc kubenswrapper[4805]: I1203 14:11:55.166268 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4z5cq\" (UniqueName: \"kubernetes.io/projected/e5a33c32-3873-49b8-8a81-c1a0357d94fc-kube-api-access-4z5cq\") pod \"openshift-apiserver-operator-796bbdcf4f-l6qxf\" (UID: \"e5a33c32-3873-49b8-8a81-c1a0357d94fc\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-l6qxf" Dec 03 14:11:55 crc kubenswrapper[4805]: I1203 14:11:55.166298 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-46hn5\" (UniqueName: \"kubernetes.io/projected/e2034338-8a35-4fd2-9bbf-be733939fc25-kube-api-access-46hn5\") pod \"console-f9d7485db-pp9tv\" (UID: \"e2034338-8a35-4fd2-9bbf-be733939fc25\") " pod="openshift-console/console-f9d7485db-pp9tv" Dec 03 14:11:55 crc kubenswrapper[4805]: I1203 14:11:55.166318 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/0a7d0b69-f9a6-4a40-9981-3dd999c400d5-trusted-ca-bundle\") pod \"apiserver-7bbb656c7d-kfh2f\" (UID: \"0a7d0b69-f9a6-4a40-9981-3dd999c400d5\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-kfh2f" Dec 03 14:11:55 crc kubenswrapper[4805]: I1203 14:11:55.166336 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/c6b0a18a-5cf2-430a-a0ef-2c2776839f11-samples-operator-tls\") pod \"cluster-samples-operator-665b6dd947-vc9sn\" (UID: \"c6b0a18a-5cf2-430a-a0ef-2c2776839f11\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-vc9sn" Dec 03 14:11:55 crc kubenswrapper[4805]: I1203 14:11:55.166351 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/cdcb6b89-e405-479e-a60d-96681328634e-config\") pod \"authentication-operator-69f744f599-bs6tj\" (UID: \"cdcb6b89-e405-479e-a60d-96681328634e\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-bs6tj" Dec 03 14:11:55 crc kubenswrapper[4805]: I1203 14:11:55.166367 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/cdcb6b89-e405-479e-a60d-96681328634e-service-ca-bundle\") pod \"authentication-operator-69f744f599-bs6tj\" (UID: \"cdcb6b89-e405-479e-a60d-96681328634e\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-bs6tj" Dec 03 14:11:55 crc kubenswrapper[4805]: I1203 14:11:55.166385 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/ae561e71-047a-4cf7-8378-d08adb458d9c-etcd-ca\") pod \"etcd-operator-b45778765-2x56c\" (UID: \"ae561e71-047a-4cf7-8378-d08adb458d9c\") " pod="openshift-etcd-operator/etcd-operator-b45778765-2x56c" Dec 03 14:11:55 crc kubenswrapper[4805]: I1203 14:11:55.166402 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/b54908df-f04f-4515-ab36-842f0aef4a99-client-ca\") pod \"controller-manager-879f6c89f-k7rgc\" (UID: \"b54908df-f04f-4515-ab36-842f0aef4a99\") " pod="openshift-controller-manager/controller-manager-879f6c89f-k7rgc" Dec 03 14:11:55 crc kubenswrapper[4805]: I1203 14:11:55.166417 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/0a7d0b69-f9a6-4a40-9981-3dd999c400d5-etcd-client\") pod \"apiserver-7bbb656c7d-kfh2f\" (UID: \"0a7d0b69-f9a6-4a40-9981-3dd999c400d5\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-kfh2f" Dec 03 14:11:55 crc kubenswrapper[4805]: I1203 14:11:55.166435 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/04bd3442-dbeb-4cb4-bc9f-b07b9872231f-tmpfs\") pod \"packageserver-d55dfcdfc-rr7fx\" (UID: \"04bd3442-dbeb-4cb4-bc9f-b07b9872231f\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-rr7fx" Dec 03 14:11:55 crc kubenswrapper[4805]: I1203 14:11:55.166451 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/b54908df-f04f-4515-ab36-842f0aef4a99-proxy-ca-bundles\") pod \"controller-manager-879f6c89f-k7rgc\" (UID: \"b54908df-f04f-4515-ab36-842f0aef4a99\") " pod="openshift-controller-manager/controller-manager-879f6c89f-k7rgc" Dec 03 14:11:55 crc kubenswrapper[4805]: I1203 14:11:55.166467 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/33c718fb-affd-410b-9f53-ddb68ba9760c-webhook-certs\") pod \"multus-admission-controller-857f4d67dd-tn8tx\" (UID: \"33c718fb-affd-410b-9f53-ddb68ba9760c\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-tn8tx" Dec 03 14:11:55 crc kubenswrapper[4805]: I1203 14:11:55.166484 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fb99x\" (UniqueName: \"kubernetes.io/projected/33c718fb-affd-410b-9f53-ddb68ba9760c-kube-api-access-fb99x\") pod \"multus-admission-controller-857f4d67dd-tn8tx\" (UID: \"33c718fb-affd-410b-9f53-ddb68ba9760c\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-tn8tx" Dec 03 14:11:55 crc kubenswrapper[4805]: I1203 14:11:55.166500 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/ca425595-aae1-44f2-9470-84ba52167748-node-pullsecrets\") pod \"apiserver-76f77b778f-7nhvl\" (UID: \"ca425595-aae1-44f2-9470-84ba52167748\") " pod="openshift-apiserver/apiserver-76f77b778f-7nhvl" Dec 03 14:11:55 crc kubenswrapper[4805]: I1203 14:11:55.166516 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4v8xc\" (UniqueName: \"kubernetes.io/projected/ca425595-aae1-44f2-9470-84ba52167748-kube-api-access-4v8xc\") pod \"apiserver-76f77b778f-7nhvl\" (UID: \"ca425595-aae1-44f2-9470-84ba52167748\") " pod="openshift-apiserver/apiserver-76f77b778f-7nhvl" Dec 03 14:11:55 crc kubenswrapper[4805]: I1203 14:11:55.166533 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-n7759\" (UniqueName: \"kubernetes.io/projected/dfa6979a-2473-469b-ac56-9086e35a63f4-kube-api-access-n7759\") pod \"route-controller-manager-6576b87f9c-nsz7s\" (UID: \"dfa6979a-2473-469b-ac56-9086e35a63f4\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-nsz7s" Dec 03 14:11:55 crc kubenswrapper[4805]: I1203 14:11:55.166550 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/e2034338-8a35-4fd2-9bbf-be733939fc25-console-oauth-config\") pod \"console-f9d7485db-pp9tv\" (UID: \"e2034338-8a35-4fd2-9bbf-be733939fc25\") " pod="openshift-console/console-f9d7485db-pp9tv" Dec 03 14:11:55 crc kubenswrapper[4805]: I1203 14:11:55.166565 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/ca425595-aae1-44f2-9470-84ba52167748-image-import-ca\") pod \"apiserver-76f77b778f-7nhvl\" (UID: \"ca425595-aae1-44f2-9470-84ba52167748\") " pod="openshift-apiserver/apiserver-76f77b778f-7nhvl" Dec 03 14:11:55 crc kubenswrapper[4805]: I1203 14:11:55.166583 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/e600aefb-0839-423f-8ac8-1916400cfa92-proxy-tls\") pod \"machine-config-controller-84d6567774-t4kl8\" (UID: \"e600aefb-0839-423f-8ac8-1916400cfa92\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-t4kl8" Dec 03 14:11:55 crc kubenswrapper[4805]: I1203 14:11:55.166604 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/1658d8e1-ac63-40fe-98ce-900a9f9aeae5-audit-dir\") pod \"oauth-openshift-558db77b4-ntv6j\" (UID: \"1658d8e1-ac63-40fe-98ce-900a9f9aeae5\") " pod="openshift-authentication/oauth-openshift-558db77b4-ntv6j" Dec 03 14:11:55 crc kubenswrapper[4805]: I1203 14:11:55.167238 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/f4b1faf2-bf8b-483c-b28f-5885a9e8fe44-serving-cert\") pod \"openshift-config-operator-7777fb866f-7wzgf\" (UID: \"f4b1faf2-bf8b-483c-b28f-5885a9e8fe44\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-7wzgf" Dec 03 14:11:55 crc kubenswrapper[4805]: I1203 14:11:55.168426 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e9022931-39f7-45cf-925f-fe19e1b870b8-serving-cert\") pod \"kube-apiserver-operator-766d6c64bb-2vnrk\" (UID: \"e9022931-39f7-45cf-925f-fe19e1b870b8\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-2vnrk" Dec 03 14:11:55 crc kubenswrapper[4805]: I1203 14:11:55.168459 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/d5ea0764-4290-4eab-b4ce-dbd4909455c1-auth-proxy-config\") pod \"machine-config-operator-74547568cd-kvmpp\" (UID: \"d5ea0764-4290-4eab-b4ce-dbd4909455c1\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-kvmpp" Dec 03 14:11:55 crc kubenswrapper[4805]: I1203 14:11:55.168483 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/1658d8e1-ac63-40fe-98ce-900a9f9aeae5-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-558db77b4-ntv6j\" (UID: \"1658d8e1-ac63-40fe-98ce-900a9f9aeae5\") " pod="openshift-authentication/oauth-openshift-558db77b4-ntv6j" Dec 03 14:11:55 crc kubenswrapper[4805]: I1203 14:11:55.168505 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/e2034338-8a35-4fd2-9bbf-be733939fc25-console-config\") pod \"console-f9d7485db-pp9tv\" (UID: \"e2034338-8a35-4fd2-9bbf-be733939fc25\") " pod="openshift-console/console-f9d7485db-pp9tv" Dec 03 14:11:55 crc kubenswrapper[4805]: I1203 14:11:55.168523 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/0a7d0b69-f9a6-4a40-9981-3dd999c400d5-encryption-config\") pod \"apiserver-7bbb656c7d-kfh2f\" (UID: \"0a7d0b69-f9a6-4a40-9981-3dd999c400d5\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-kfh2f" Dec 03 14:11:55 crc kubenswrapper[4805]: I1203 14:11:55.168540 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sjpnv\" (UniqueName: \"kubernetes.io/projected/f4b1faf2-bf8b-483c-b28f-5885a9e8fe44-kube-api-access-sjpnv\") pod \"openshift-config-operator-7777fb866f-7wzgf\" (UID: \"f4b1faf2-bf8b-483c-b28f-5885a9e8fe44\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-7wzgf" Dec 03 14:11:55 crc kubenswrapper[4805]: I1203 14:11:55.168556 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/1c55f37c-24b9-4800-9050-e0da883df2c0-signing-key\") pod \"service-ca-9c57cc56f-dbfrd\" (UID: \"1c55f37c-24b9-4800-9050-e0da883df2c0\") " pod="openshift-service-ca/service-ca-9c57cc56f-dbfrd" Dec 03 14:11:55 crc kubenswrapper[4805]: I1203 14:11:55.168575 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/1658d8e1-ac63-40fe-98ce-900a9f9aeae5-v4-0-config-user-template-error\") pod \"oauth-openshift-558db77b4-ntv6j\" (UID: \"1658d8e1-ac63-40fe-98ce-900a9f9aeae5\") " pod="openshift-authentication/oauth-openshift-558db77b4-ntv6j" Dec 03 14:11:55 crc kubenswrapper[4805]: I1203 14:11:55.168591 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/ca425595-aae1-44f2-9470-84ba52167748-serving-cert\") pod \"apiserver-76f77b778f-7nhvl\" (UID: \"ca425595-aae1-44f2-9470-84ba52167748\") " pod="openshift-apiserver/apiserver-76f77b778f-7nhvl" Dec 03 14:11:55 crc kubenswrapper[4805]: I1203 14:11:55.168607 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/0a7d0b69-f9a6-4a40-9981-3dd999c400d5-etcd-serving-ca\") pod \"apiserver-7bbb656c7d-kfh2f\" (UID: \"0a7d0b69-f9a6-4a40-9981-3dd999c400d5\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-kfh2f" Dec 03 14:11:55 crc kubenswrapper[4805]: I1203 14:11:55.168624 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-csvjg\" (UniqueName: \"kubernetes.io/projected/e600aefb-0839-423f-8ac8-1916400cfa92-kube-api-access-csvjg\") pod \"machine-config-controller-84d6567774-t4kl8\" (UID: \"e600aefb-0839-423f-8ac8-1916400cfa92\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-t4kl8" Dec 03 14:11:55 crc kubenswrapper[4805]: I1203 14:11:55.168642 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/1658d8e1-ac63-40fe-98ce-900a9f9aeae5-v4-0-config-user-template-login\") pod \"oauth-openshift-558db77b4-ntv6j\" (UID: \"1658d8e1-ac63-40fe-98ce-900a9f9aeae5\") " pod="openshift-authentication/oauth-openshift-558db77b4-ntv6j" Dec 03 14:11:55 crc kubenswrapper[4805]: I1203 14:11:55.168658 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/dfd20bfc-5c11-4287-a4b3-df9e0ca4dcf1-images\") pod \"machine-api-operator-5694c8668f-jvspg\" (UID: \"dfd20bfc-5c11-4287-a4b3-df9e0ca4dcf1\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-jvspg" Dec 03 14:11:55 crc kubenswrapper[4805]: I1203 14:11:55.168673 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9b932fa5-f87b-46e1-8987-683cfd3b46fd-serving-cert\") pod \"service-ca-operator-777779d784-z7lgj\" (UID: \"9b932fa5-f87b-46e1-8987-683cfd3b46fd\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-z7lgj" Dec 03 14:11:55 crc kubenswrapper[4805]: I1203 14:11:55.168689 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/1658d8e1-ac63-40fe-98ce-900a9f9aeae5-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-558db77b4-ntv6j\" (UID: \"1658d8e1-ac63-40fe-98ce-900a9f9aeae5\") " pod="openshift-authentication/oauth-openshift-558db77b4-ntv6j" Dec 03 14:11:55 crc kubenswrapper[4805]: I1203 14:11:55.168705 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/dfa6979a-2473-469b-ac56-9086e35a63f4-client-ca\") pod \"route-controller-manager-6576b87f9c-nsz7s\" (UID: \"dfa6979a-2473-469b-ac56-9086e35a63f4\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-nsz7s" Dec 03 14:11:55 crc kubenswrapper[4805]: I1203 14:11:55.168736 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-x6v8p\" (UniqueName: \"kubernetes.io/projected/9f53fe44-0d2c-4e3d-a682-df3a779a1c8a-kube-api-access-x6v8p\") pod \"console-operator-58897d9998-46w2g\" (UID: \"9f53fe44-0d2c-4e3d-a682-df3a779a1c8a\") " pod="openshift-console-operator/console-operator-58897d9998-46w2g" Dec 03 14:11:55 crc kubenswrapper[4805]: I1203 14:11:55.168756 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ddtf4\" (UniqueName: \"kubernetes.io/projected/ae561e71-047a-4cf7-8378-d08adb458d9c-kube-api-access-ddtf4\") pod \"etcd-operator-b45778765-2x56c\" (UID: \"ae561e71-047a-4cf7-8378-d08adb458d9c\") " pod="openshift-etcd-operator/etcd-operator-b45778765-2x56c" Dec 03 14:11:55 crc kubenswrapper[4805]: I1203 14:11:55.168773 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-89vsl\" (UniqueName: \"kubernetes.io/projected/b54908df-f04f-4515-ab36-842f0aef4a99-kube-api-access-89vsl\") pod \"controller-manager-879f6c89f-k7rgc\" (UID: \"b54908df-f04f-4515-ab36-842f0aef4a99\") " pod="openshift-controller-manager/controller-manager-879f6c89f-k7rgc" Dec 03 14:11:55 crc kubenswrapper[4805]: I1203 14:11:55.168789 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/dfd20bfc-5c11-4287-a4b3-df9e0ca4dcf1-config\") pod \"machine-api-operator-5694c8668f-jvspg\" (UID: \"dfd20bfc-5c11-4287-a4b3-df9e0ca4dcf1\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-jvspg" Dec 03 14:11:55 crc kubenswrapper[4805]: I1203 14:11:55.168811 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/dfd20bfc-5c11-4287-a4b3-df9e0ca4dcf1-machine-api-operator-tls\") pod \"machine-api-operator-5694c8668f-jvspg\" (UID: \"dfd20bfc-5c11-4287-a4b3-df9e0ca4dcf1\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-jvspg" Dec 03 14:11:55 crc kubenswrapper[4805]: I1203 14:11:55.168838 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7frlc\" (UniqueName: \"kubernetes.io/projected/8b954d96-a62c-434e-a340-8446d778c545-kube-api-access-7frlc\") pod \"collect-profiles-29412840-gj7pr\" (UID: \"8b954d96-a62c-434e-a340-8446d778c545\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29412840-gj7pr" Dec 03 14:11:55 crc kubenswrapper[4805]: I1203 14:11:55.168870 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/ae561e71-047a-4cf7-8378-d08adb458d9c-etcd-client\") pod \"etcd-operator-b45778765-2x56c\" (UID: \"ae561e71-047a-4cf7-8378-d08adb458d9c\") " pod="openshift-etcd-operator/etcd-operator-b45778765-2x56c" Dec 03 14:11:55 crc kubenswrapper[4805]: I1203 14:11:55.168888 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8gr5q\" (UniqueName: \"kubernetes.io/projected/d5ea0764-4290-4eab-b4ce-dbd4909455c1-kube-api-access-8gr5q\") pod \"machine-config-operator-74547568cd-kvmpp\" (UID: \"d5ea0764-4290-4eab-b4ce-dbd4909455c1\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-kvmpp" Dec 03 14:11:55 crc kubenswrapper[4805]: I1203 14:11:55.167990 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/cdcb6b89-e405-479e-a60d-96681328634e-config\") pod \"authentication-operator-69f744f599-bs6tj\" (UID: \"cdcb6b89-e405-479e-a60d-96681328634e\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-bs6tj" Dec 03 14:11:55 crc kubenswrapper[4805]: I1203 14:11:55.168955 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/ca425595-aae1-44f2-9470-84ba52167748-image-import-ca\") pod \"apiserver-76f77b778f-7nhvl\" (UID: \"ca425595-aae1-44f2-9470-84ba52167748\") " pod="openshift-apiserver/apiserver-76f77b778f-7nhvl" Dec 03 14:11:55 crc kubenswrapper[4805]: I1203 14:11:55.168904 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/ca425595-aae1-44f2-9470-84ba52167748-etcd-serving-ca\") pod \"apiserver-76f77b778f-7nhvl\" (UID: \"ca425595-aae1-44f2-9470-84ba52167748\") " pod="openshift-apiserver/apiserver-76f77b778f-7nhvl" Dec 03 14:11:55 crc kubenswrapper[4805]: I1203 14:11:55.169033 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/ca425595-aae1-44f2-9470-84ba52167748-encryption-config\") pod \"apiserver-76f77b778f-7nhvl\" (UID: \"ca425595-aae1-44f2-9470-84ba52167748\") " pod="openshift-apiserver/apiserver-76f77b778f-7nhvl" Dec 03 14:11:55 crc kubenswrapper[4805]: I1203 14:11:55.169065 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e5a33c32-3873-49b8-8a81-c1a0357d94fc-config\") pod \"openshift-apiserver-operator-796bbdcf4f-l6qxf\" (UID: \"e5a33c32-3873-49b8-8a81-c1a0357d94fc\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-l6qxf" Dec 03 14:11:55 crc kubenswrapper[4805]: I1203 14:11:55.169095 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/8b954d96-a62c-434e-a340-8446d778c545-config-volume\") pod \"collect-profiles-29412840-gj7pr\" (UID: \"8b954d96-a62c-434e-a340-8446d778c545\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29412840-gj7pr" Dec 03 14:11:55 crc kubenswrapper[4805]: I1203 14:11:55.169124 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/d5ea0764-4290-4eab-b4ce-dbd4909455c1-proxy-tls\") pod \"machine-config-operator-74547568cd-kvmpp\" (UID: \"d5ea0764-4290-4eab-b4ce-dbd4909455c1\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-kvmpp" Dec 03 14:11:55 crc kubenswrapper[4805]: I1203 14:11:55.169153 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/cdcb6b89-e405-479e-a60d-96681328634e-serving-cert\") pod \"authentication-operator-69f744f599-bs6tj\" (UID: \"cdcb6b89-e405-479e-a60d-96681328634e\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-bs6tj" Dec 03 14:11:55 crc kubenswrapper[4805]: I1203 14:11:55.169176 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/e9022931-39f7-45cf-925f-fe19e1b870b8-kube-api-access\") pod \"kube-apiserver-operator-766d6c64bb-2vnrk\" (UID: \"e9022931-39f7-45cf-925f-fe19e1b870b8\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-2vnrk" Dec 03 14:11:55 crc kubenswrapper[4805]: I1203 14:11:55.169205 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/04bd3442-dbeb-4cb4-bc9f-b07b9872231f-webhook-cert\") pod \"packageserver-d55dfcdfc-rr7fx\" (UID: \"04bd3442-dbeb-4cb4-bc9f-b07b9872231f\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-rr7fx" Dec 03 14:11:55 crc kubenswrapper[4805]: I1203 14:11:55.169228 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/cdcb6b89-e405-479e-a60d-96681328634e-trusted-ca-bundle\") pod \"authentication-operator-69f744f599-bs6tj\" (UID: \"cdcb6b89-e405-479e-a60d-96681328634e\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-bs6tj" Dec 03 14:11:55 crc kubenswrapper[4805]: I1203 14:11:55.169256 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lwvs6\" (UniqueName: \"kubernetes.io/projected/00941575-52e9-482d-b256-1b1c757d96e0-kube-api-access-lwvs6\") pod \"machine-approver-56656f9798-nv2m9\" (UID: \"00941575-52e9-482d-b256-1b1c757d96e0\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-nv2m9" Dec 03 14:11:55 crc kubenswrapper[4805]: I1203 14:11:55.169278 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/d5ea0764-4290-4eab-b4ce-dbd4909455c1-images\") pod \"machine-config-operator-74547568cd-kvmpp\" (UID: \"d5ea0764-4290-4eab-b4ce-dbd4909455c1\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-kvmpp" Dec 03 14:11:55 crc kubenswrapper[4805]: I1203 14:11:55.169303 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/1658d8e1-ac63-40fe-98ce-900a9f9aeae5-v4-0-config-system-router-certs\") pod \"oauth-openshift-558db77b4-ntv6j\" (UID: \"1658d8e1-ac63-40fe-98ce-900a9f9aeae5\") " pod="openshift-authentication/oauth-openshift-558db77b4-ntv6j" Dec 03 14:11:55 crc kubenswrapper[4805]: I1203 14:11:55.169327 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/1658d8e1-ac63-40fe-98ce-900a9f9aeae5-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-558db77b4-ntv6j\" (UID: \"1658d8e1-ac63-40fe-98ce-900a9f9aeae5\") " pod="openshift-authentication/oauth-openshift-558db77b4-ntv6j" Dec 03 14:11:55 crc kubenswrapper[4805]: I1203 14:11:55.169366 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/e2034338-8a35-4fd2-9bbf-be733939fc25-service-ca\") pod \"console-f9d7485db-pp9tv\" (UID: \"e2034338-8a35-4fd2-9bbf-be733939fc25\") " pod="openshift-console/console-f9d7485db-pp9tv" Dec 03 14:11:55 crc kubenswrapper[4805]: I1203 14:11:55.169392 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/e2034338-8a35-4fd2-9bbf-be733939fc25-trusted-ca-bundle\") pod \"console-f9d7485db-pp9tv\" (UID: \"e2034338-8a35-4fd2-9bbf-be733939fc25\") " pod="openshift-console/console-f9d7485db-pp9tv" Dec 03 14:11:55 crc kubenswrapper[4805]: I1203 14:11:55.169419 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/ca425595-aae1-44f2-9470-84ba52167748-audit-dir\") pod \"apiserver-76f77b778f-7nhvl\" (UID: \"ca425595-aae1-44f2-9470-84ba52167748\") " pod="openshift-apiserver/apiserver-76f77b778f-7nhvl" Dec 03 14:11:55 crc kubenswrapper[4805]: I1203 14:11:55.169442 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/0a7d0b69-f9a6-4a40-9981-3dd999c400d5-audit-dir\") pod \"apiserver-7bbb656c7d-kfh2f\" (UID: \"0a7d0b69-f9a6-4a40-9981-3dd999c400d5\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-kfh2f" Dec 03 14:11:55 crc kubenswrapper[4805]: I1203 14:11:55.169464 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/ca425595-aae1-44f2-9470-84ba52167748-etcd-serving-ca\") pod \"apiserver-76f77b778f-7nhvl\" (UID: \"ca425595-aae1-44f2-9470-84ba52167748\") " pod="openshift-apiserver/apiserver-76f77b778f-7nhvl" Dec 03 14:11:55 crc kubenswrapper[4805]: I1203 14:11:55.169468 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/e600aefb-0839-423f-8ac8-1916400cfa92-mcc-auth-proxy-config\") pod \"machine-config-controller-84d6567774-t4kl8\" (UID: \"e600aefb-0839-423f-8ac8-1916400cfa92\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-t4kl8" Dec 03 14:11:55 crc kubenswrapper[4805]: I1203 14:11:55.168159 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-25khc"] Dec 03 14:11:55 crc kubenswrapper[4805]: I1203 14:11:55.169495 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/1c55f37c-24b9-4800-9050-e0da883df2c0-signing-cabundle\") pod \"service-ca-9c57cc56f-dbfrd\" (UID: \"1c55f37c-24b9-4800-9050-e0da883df2c0\") " pod="openshift-service-ca/service-ca-9c57cc56f-dbfrd" Dec 03 14:11:55 crc kubenswrapper[4805]: I1203 14:11:55.169511 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-b7wrl"] Dec 03 14:11:55 crc kubenswrapper[4805]: I1203 14:11:55.169524 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/00941575-52e9-482d-b256-1b1c757d96e0-machine-approver-tls\") pod \"machine-approver-56656f9798-nv2m9\" (UID: \"00941575-52e9-482d-b256-1b1c757d96e0\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-nv2m9" Dec 03 14:11:55 crc kubenswrapper[4805]: I1203 14:11:55.168480 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/04bd3442-dbeb-4cb4-bc9f-b07b9872231f-tmpfs\") pod \"packageserver-d55dfcdfc-rr7fx\" (UID: \"04bd3442-dbeb-4cb4-bc9f-b07b9872231f\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-rr7fx" Dec 03 14:11:55 crc kubenswrapper[4805]: I1203 14:11:55.169549 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/1658d8e1-ac63-40fe-98ce-900a9f9aeae5-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-558db77b4-ntv6j\" (UID: \"1658d8e1-ac63-40fe-98ce-900a9f9aeae5\") " pod="openshift-authentication/oauth-openshift-558db77b4-ntv6j" Dec 03 14:11:55 crc kubenswrapper[4805]: I1203 14:11:55.169576 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9f53fe44-0d2c-4e3d-a682-df3a779a1c8a-config\") pod \"console-operator-58897d9998-46w2g\" (UID: \"9f53fe44-0d2c-4e3d-a682-df3a779a1c8a\") " pod="openshift-console-operator/console-operator-58897d9998-46w2g" Dec 03 14:11:55 crc kubenswrapper[4805]: I1203 14:11:55.169600 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/e2034338-8a35-4fd2-9bbf-be733939fc25-console-serving-cert\") pod \"console-f9d7485db-pp9tv\" (UID: \"e2034338-8a35-4fd2-9bbf-be733939fc25\") " pod="openshift-console/console-f9d7485db-pp9tv" Dec 03 14:11:55 crc kubenswrapper[4805]: I1203 14:11:55.169580 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/b54908df-f04f-4515-ab36-842f0aef4a99-proxy-ca-bundles\") pod \"controller-manager-879f6c89f-k7rgc\" (UID: \"b54908df-f04f-4515-ab36-842f0aef4a99\") " pod="openshift-controller-manager/controller-manager-879f6c89f-k7rgc" Dec 03 14:11:55 crc kubenswrapper[4805]: I1203 14:11:55.169805 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/e2034338-8a35-4fd2-9bbf-be733939fc25-console-config\") pod \"console-f9d7485db-pp9tv\" (UID: \"e2034338-8a35-4fd2-9bbf-be733939fc25\") " pod="openshift-console/console-f9d7485db-pp9tv" Dec 03 14:11:55 crc kubenswrapper[4805]: I1203 14:11:55.168134 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/cdcb6b89-e405-479e-a60d-96681328634e-service-ca-bundle\") pod \"authentication-operator-69f744f599-bs6tj\" (UID: \"cdcb6b89-e405-479e-a60d-96681328634e\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-bs6tj" Dec 03 14:11:55 crc kubenswrapper[4805]: I1203 14:11:55.168113 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/0a7d0b69-f9a6-4a40-9981-3dd999c400d5-trusted-ca-bundle\") pod \"apiserver-7bbb656c7d-kfh2f\" (UID: \"0a7d0b69-f9a6-4a40-9981-3dd999c400d5\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-kfh2f" Dec 03 14:11:55 crc kubenswrapper[4805]: I1203 14:11:55.169623 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/ca425595-aae1-44f2-9470-84ba52167748-etcd-client\") pod \"apiserver-76f77b778f-7nhvl\" (UID: \"ca425595-aae1-44f2-9470-84ba52167748\") " pod="openshift-apiserver/apiserver-76f77b778f-7nhvl" Dec 03 14:11:55 crc kubenswrapper[4805]: I1203 14:11:55.169915 4805 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["hostpath-provisioner/csi-hostpathplugin-krqsw"] Dec 03 14:11:55 crc kubenswrapper[4805]: I1203 14:11:55.169934 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0a7d0b69-f9a6-4a40-9981-3dd999c400d5-serving-cert\") pod \"apiserver-7bbb656c7d-kfh2f\" (UID: \"0a7d0b69-f9a6-4a40-9981-3dd999c400d5\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-kfh2f" Dec 03 14:11:55 crc kubenswrapper[4805]: I1203 14:11:55.169960 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e5a33c32-3873-49b8-8a81-c1a0357d94fc-serving-cert\") pod \"openshift-apiserver-operator-796bbdcf4f-l6qxf\" (UID: \"e5a33c32-3873-49b8-8a81-c1a0357d94fc\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-l6qxf" Dec 03 14:11:55 crc kubenswrapper[4805]: I1203 14:11:55.169988 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e9022931-39f7-45cf-925f-fe19e1b870b8-config\") pod \"kube-apiserver-operator-766d6c64bb-2vnrk\" (UID: \"e9022931-39f7-45cf-925f-fe19e1b870b8\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-2vnrk" Dec 03 14:11:55 crc kubenswrapper[4805]: I1203 14:11:55.170011 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/da68f128-1d8d-4256-a28d-a174766e1331-trusted-ca\") pod \"cluster-image-registry-operator-dc59b4c8b-jvsqf\" (UID: \"da68f128-1d8d-4256-a28d-a174766e1331\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-jvsqf" Dec 03 14:11:55 crc kubenswrapper[4805]: I1203 14:11:55.170040 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qpdfz\" (UniqueName: \"kubernetes.io/projected/da68f128-1d8d-4256-a28d-a174766e1331-kube-api-access-qpdfz\") pod \"cluster-image-registry-operator-dc59b4c8b-jvsqf\" (UID: \"da68f128-1d8d-4256-a28d-a174766e1331\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-jvsqf" Dec 03 14:11:55 crc kubenswrapper[4805]: I1203 14:11:55.170069 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6cs72\" (UniqueName: \"kubernetes.io/projected/04bd3442-dbeb-4cb4-bc9f-b07b9872231f-kube-api-access-6cs72\") pod \"packageserver-d55dfcdfc-rr7fx\" (UID: \"04bd3442-dbeb-4cb4-bc9f-b07b9872231f\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-rr7fx" Dec 03 14:11:55 crc kubenswrapper[4805]: I1203 14:11:55.170088 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/dfa6979a-2473-469b-ac56-9086e35a63f4-serving-cert\") pod \"route-controller-manager-6576b87f9c-nsz7s\" (UID: \"dfa6979a-2473-469b-ac56-9086e35a63f4\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-nsz7s" Dec 03 14:11:55 crc kubenswrapper[4805]: I1203 14:11:55.170105 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/00941575-52e9-482d-b256-1b1c757d96e0-config\") pod \"machine-approver-56656f9798-nv2m9\" (UID: \"00941575-52e9-482d-b256-1b1c757d96e0\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-nv2m9" Dec 03 14:11:55 crc kubenswrapper[4805]: I1203 14:11:55.170127 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8zh8j\" (UniqueName: \"kubernetes.io/projected/cdcb6b89-e405-479e-a60d-96681328634e-kube-api-access-8zh8j\") pod \"authentication-operator-69f744f599-bs6tj\" (UID: \"cdcb6b89-e405-479e-a60d-96681328634e\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-bs6tj" Dec 03 14:11:55 crc kubenswrapper[4805]: I1203 14:11:55.170145 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/8e437287-3b92-4c6a-abe9-84c786ba3220-metrics-tls\") pod \"dns-operator-744455d44c-rjf9m\" (UID: \"8e437287-3b92-4c6a-abe9-84c786ba3220\") " pod="openshift-dns-operator/dns-operator-744455d44c-rjf9m" Dec 03 14:11:55 crc kubenswrapper[4805]: I1203 14:11:55.170167 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9f53fe44-0d2c-4e3d-a682-df3a779a1c8a-serving-cert\") pod \"console-operator-58897d9998-46w2g\" (UID: \"9f53fe44-0d2c-4e3d-a682-df3a779a1c8a\") " pod="openshift-console-operator/console-operator-58897d9998-46w2g" Dec 03 14:11:55 crc kubenswrapper[4805]: I1203 14:11:55.170184 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/ca425595-aae1-44f2-9470-84ba52167748-trusted-ca-bundle\") pod \"apiserver-76f77b778f-7nhvl\" (UID: \"ca425595-aae1-44f2-9470-84ba52167748\") " pod="openshift-apiserver/apiserver-76f77b778f-7nhvl" Dec 03 14:11:55 crc kubenswrapper[4805]: I1203 14:11:55.170203 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/8b954d96-a62c-434e-a340-8446d778c545-secret-volume\") pod \"collect-profiles-29412840-gj7pr\" (UID: \"8b954d96-a62c-434e-a340-8446d778c545\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29412840-gj7pr" Dec 03 14:11:55 crc kubenswrapper[4805]: I1203 14:11:55.170222 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/da68f128-1d8d-4256-a28d-a174766e1331-bound-sa-token\") pod \"cluster-image-registry-operator-dc59b4c8b-jvsqf\" (UID: \"da68f128-1d8d-4256-a28d-a174766e1331\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-jvsqf" Dec 03 14:11:55 crc kubenswrapper[4805]: I1203 14:11:55.170238 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/ae561e71-047a-4cf7-8378-d08adb458d9c-serving-cert\") pod \"etcd-operator-b45778765-2x56c\" (UID: \"ae561e71-047a-4cf7-8378-d08adb458d9c\") " pod="openshift-etcd-operator/etcd-operator-b45778765-2x56c" Dec 03 14:11:55 crc kubenswrapper[4805]: I1203 14:11:55.170259 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rnn8j\" (UniqueName: \"kubernetes.io/projected/dfd20bfc-5c11-4287-a4b3-df9e0ca4dcf1-kube-api-access-rnn8j\") pod \"machine-api-operator-5694c8668f-jvspg\" (UID: \"dfd20bfc-5c11-4287-a4b3-df9e0ca4dcf1\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-jvspg" Dec 03 14:11:55 crc kubenswrapper[4805]: I1203 14:11:55.170278 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-p5sc5\" (UniqueName: \"kubernetes.io/projected/c6b0a18a-5cf2-430a-a0ef-2c2776839f11-kube-api-access-p5sc5\") pod \"cluster-samples-operator-665b6dd947-vc9sn\" (UID: \"c6b0a18a-5cf2-430a-a0ef-2c2776839f11\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-vc9sn" Dec 03 14:11:55 crc kubenswrapper[4805]: I1203 14:11:55.170297 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2cmpw\" (UniqueName: \"kubernetes.io/projected/0a7d0b69-f9a6-4a40-9981-3dd999c400d5-kube-api-access-2cmpw\") pod \"apiserver-7bbb656c7d-kfh2f\" (UID: \"0a7d0b69-f9a6-4a40-9981-3dd999c400d5\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-kfh2f" Dec 03 14:11:55 crc kubenswrapper[4805]: I1203 14:11:55.170327 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/04bd3442-dbeb-4cb4-bc9f-b07b9872231f-apiservice-cert\") pod \"packageserver-d55dfcdfc-rr7fx\" (UID: \"04bd3442-dbeb-4cb4-bc9f-b07b9872231f\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-rr7fx" Dec 03 14:11:55 crc kubenswrapper[4805]: I1203 14:11:55.170351 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/1658d8e1-ac63-40fe-98ce-900a9f9aeae5-v4-0-config-system-session\") pod \"oauth-openshift-558db77b4-ntv6j\" (UID: \"1658d8e1-ac63-40fe-98ce-900a9f9aeae5\") " pod="openshift-authentication/oauth-openshift-558db77b4-ntv6j" Dec 03 14:11:55 crc kubenswrapper[4805]: I1203 14:11:55.170413 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/f4b1faf2-bf8b-483c-b28f-5885a9e8fe44-available-featuregates\") pod \"openshift-config-operator-7777fb866f-7wzgf\" (UID: \"f4b1faf2-bf8b-483c-b28f-5885a9e8fe44\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-7wzgf" Dec 03 14:11:55 crc kubenswrapper[4805]: I1203 14:11:55.170437 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qzz6l\" (UniqueName: \"kubernetes.io/projected/9b932fa5-f87b-46e1-8987-683cfd3b46fd-kube-api-access-qzz6l\") pod \"service-ca-operator-777779d784-z7lgj\" (UID: \"9b932fa5-f87b-46e1-8987-683cfd3b46fd\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-z7lgj" Dec 03 14:11:55 crc kubenswrapper[4805]: I1203 14:11:55.170466 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/1658d8e1-ac63-40fe-98ce-900a9f9aeae5-audit-policies\") pod \"oauth-openshift-558db77b4-ntv6j\" (UID: \"1658d8e1-ac63-40fe-98ce-900a9f9aeae5\") " pod="openshift-authentication/oauth-openshift-558db77b4-ntv6j" Dec 03 14:11:55 crc kubenswrapper[4805]: I1203 14:11:55.170487 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b54908df-f04f-4515-ab36-842f0aef4a99-config\") pod \"controller-manager-879f6c89f-k7rgc\" (UID: \"b54908df-f04f-4515-ab36-842f0aef4a99\") " pod="openshift-controller-manager/controller-manager-879f6c89f-k7rgc" Dec 03 14:11:55 crc kubenswrapper[4805]: I1203 14:11:55.170526 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/b54908df-f04f-4515-ab36-842f0aef4a99-serving-cert\") pod \"controller-manager-879f6c89f-k7rgc\" (UID: \"b54908df-f04f-4515-ab36-842f0aef4a99\") " pod="openshift-controller-manager/controller-manager-879f6c89f-k7rgc" Dec 03 14:11:55 crc kubenswrapper[4805]: I1203 14:11:55.170558 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/e2034338-8a35-4fd2-9bbf-be733939fc25-oauth-serving-cert\") pod \"console-f9d7485db-pp9tv\" (UID: \"e2034338-8a35-4fd2-9bbf-be733939fc25\") " pod="openshift-console/console-f9d7485db-pp9tv" Dec 03 14:11:55 crc kubenswrapper[4805]: I1203 14:11:55.170576 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/da68f128-1d8d-4256-a28d-a174766e1331-image-registry-operator-tls\") pod \"cluster-image-registry-operator-dc59b4c8b-jvsqf\" (UID: \"da68f128-1d8d-4256-a28d-a174766e1331\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-jvsqf" Dec 03 14:11:55 crc kubenswrapper[4805]: I1203 14:11:55.170599 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ae561e71-047a-4cf7-8378-d08adb458d9c-config\") pod \"etcd-operator-b45778765-2x56c\" (UID: \"ae561e71-047a-4cf7-8378-d08adb458d9c\") " pod="openshift-etcd-operator/etcd-operator-b45778765-2x56c" Dec 03 14:11:55 crc kubenswrapper[4805]: I1203 14:11:55.170603 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/e2034338-8a35-4fd2-9bbf-be733939fc25-service-ca\") pod \"console-f9d7485db-pp9tv\" (UID: \"e2034338-8a35-4fd2-9bbf-be733939fc25\") " pod="openshift-console/console-f9d7485db-pp9tv" Dec 03 14:11:55 crc kubenswrapper[4805]: I1203 14:11:55.170628 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/ae561e71-047a-4cf7-8378-d08adb458d9c-etcd-service-ca\") pod \"etcd-operator-b45778765-2x56c\" (UID: \"ae561e71-047a-4cf7-8378-d08adb458d9c\") " pod="openshift-etcd-operator/etcd-operator-b45778765-2x56c" Dec 03 14:11:55 crc kubenswrapper[4805]: I1203 14:11:55.170657 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4djzx\" (UniqueName: \"kubernetes.io/projected/1c55f37c-24b9-4800-9050-e0da883df2c0-kube-api-access-4djzx\") pod \"service-ca-9c57cc56f-dbfrd\" (UID: \"1c55f37c-24b9-4800-9050-e0da883df2c0\") " pod="openshift-service-ca/service-ca-9c57cc56f-dbfrd" Dec 03 14:11:55 crc kubenswrapper[4805]: I1203 14:11:55.170688 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/1658d8e1-ac63-40fe-98ce-900a9f9aeae5-v4-0-config-system-serving-cert\") pod \"oauth-openshift-558db77b4-ntv6j\" (UID: \"1658d8e1-ac63-40fe-98ce-900a9f9aeae5\") " pod="openshift-authentication/oauth-openshift-558db77b4-ntv6j" Dec 03 14:11:55 crc kubenswrapper[4805]: I1203 14:11:55.170720 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/1658d8e1-ac63-40fe-98ce-900a9f9aeae5-v4-0-config-system-service-ca\") pod \"oauth-openshift-558db77b4-ntv6j\" (UID: \"1658d8e1-ac63-40fe-98ce-900a9f9aeae5\") " pod="openshift-authentication/oauth-openshift-558db77b4-ntv6j" Dec 03 14:11:55 crc kubenswrapper[4805]: I1203 14:11:55.170745 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dxgkw\" (UniqueName: \"kubernetes.io/projected/1658d8e1-ac63-40fe-98ce-900a9f9aeae5-kube-api-access-dxgkw\") pod \"oauth-openshift-558db77b4-ntv6j\" (UID: \"1658d8e1-ac63-40fe-98ce-900a9f9aeae5\") " pod="openshift-authentication/oauth-openshift-558db77b4-ntv6j" Dec 03 14:11:55 crc kubenswrapper[4805]: I1203 14:11:55.170772 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ca425595-aae1-44f2-9470-84ba52167748-config\") pod \"apiserver-76f77b778f-7nhvl\" (UID: \"ca425595-aae1-44f2-9470-84ba52167748\") " pod="openshift-apiserver/apiserver-76f77b778f-7nhvl" Dec 03 14:11:55 crc kubenswrapper[4805]: I1203 14:11:55.170810 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9b932fa5-f87b-46e1-8987-683cfd3b46fd-config\") pod \"service-ca-operator-777779d784-z7lgj\" (UID: \"9b932fa5-f87b-46e1-8987-683cfd3b46fd\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-z7lgj" Dec 03 14:11:55 crc kubenswrapper[4805]: I1203 14:11:55.170891 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/00941575-52e9-482d-b256-1b1c757d96e0-auth-proxy-config\") pod \"machine-approver-56656f9798-nv2m9\" (UID: \"00941575-52e9-482d-b256-1b1c757d96e0\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-nv2m9" Dec 03 14:11:55 crc kubenswrapper[4805]: I1203 14:11:55.170962 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pgzx9\" (UniqueName: \"kubernetes.io/projected/6c2d6fbc-771d-4aff-b1e8-3eda17204931-kube-api-access-pgzx9\") pod \"downloads-7954f5f757-s25zl\" (UID: \"6c2d6fbc-771d-4aff-b1e8-3eda17204931\") " pod="openshift-console/downloads-7954f5f757-s25zl" Dec 03 14:11:55 crc kubenswrapper[4805]: I1203 14:11:55.170989 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/9f53fe44-0d2c-4e3d-a682-df3a779a1c8a-trusted-ca\") pod \"console-operator-58897d9998-46w2g\" (UID: \"9f53fe44-0d2c-4e3d-a682-df3a779a1c8a\") " pod="openshift-console-operator/console-operator-58897d9998-46w2g" Dec 03 14:11:55 crc kubenswrapper[4805]: I1203 14:11:55.171016 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="hostpath-provisioner/csi-hostpathplugin-krqsw" Dec 03 14:11:55 crc kubenswrapper[4805]: I1203 14:11:55.172529 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["hostpath-provisioner/csi-hostpathplugin-krqsw"] Dec 03 14:11:55 crc kubenswrapper[4805]: I1203 14:11:55.171022 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/ca425595-aae1-44f2-9470-84ba52167748-audit\") pod \"apiserver-76f77b778f-7nhvl\" (UID: \"ca425595-aae1-44f2-9470-84ba52167748\") " pod="openshift-apiserver/apiserver-76f77b778f-7nhvl" Dec 03 14:11:55 crc kubenswrapper[4805]: I1203 14:11:55.172620 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/1658d8e1-ac63-40fe-98ce-900a9f9aeae5-v4-0-config-system-cliconfig\") pod \"oauth-openshift-558db77b4-ntv6j\" (UID: \"1658d8e1-ac63-40fe-98ce-900a9f9aeae5\") " pod="openshift-authentication/oauth-openshift-558db77b4-ntv6j" Dec 03 14:11:55 crc kubenswrapper[4805]: I1203 14:11:55.172644 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/0a7d0b69-f9a6-4a40-9981-3dd999c400d5-audit-policies\") pod \"apiserver-7bbb656c7d-kfh2f\" (UID: \"0a7d0b69-f9a6-4a40-9981-3dd999c400d5\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-kfh2f" Dec 03 14:11:55 crc kubenswrapper[4805]: I1203 14:11:55.172678 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-v8t8r\" (UniqueName: \"kubernetes.io/projected/8e437287-3b92-4c6a-abe9-84c786ba3220-kube-api-access-v8t8r\") pod \"dns-operator-744455d44c-rjf9m\" (UID: \"8e437287-3b92-4c6a-abe9-84c786ba3220\") " pod="openshift-dns-operator/dns-operator-744455d44c-rjf9m" Dec 03 14:11:55 crc kubenswrapper[4805]: I1203 14:11:55.173708 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/ca425595-aae1-44f2-9470-84ba52167748-audit-dir\") pod \"apiserver-76f77b778f-7nhvl\" (UID: \"ca425595-aae1-44f2-9470-84ba52167748\") " pod="openshift-apiserver/apiserver-76f77b778f-7nhvl" Dec 03 14:11:55 crc kubenswrapper[4805]: I1203 14:11:55.167689 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/dfa6979a-2473-469b-ac56-9086e35a63f4-config\") pod \"route-controller-manager-6576b87f9c-nsz7s\" (UID: \"dfa6979a-2473-469b-ac56-9086e35a63f4\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-nsz7s" Dec 03 14:11:55 crc kubenswrapper[4805]: I1203 14:11:55.167177 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/ca425595-aae1-44f2-9470-84ba52167748-node-pullsecrets\") pod \"apiserver-76f77b778f-7nhvl\" (UID: \"ca425595-aae1-44f2-9470-84ba52167748\") " pod="openshift-apiserver/apiserver-76f77b778f-7nhvl" Dec 03 14:11:55 crc kubenswrapper[4805]: I1203 14:11:55.174071 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/e2034338-8a35-4fd2-9bbf-be733939fc25-trusted-ca-bundle\") pod \"console-f9d7485db-pp9tv\" (UID: \"e2034338-8a35-4fd2-9bbf-be733939fc25\") " pod="openshift-console/console-f9d7485db-pp9tv" Dec 03 14:11:55 crc kubenswrapper[4805]: I1203 14:11:55.174314 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/0a7d0b69-f9a6-4a40-9981-3dd999c400d5-etcd-serving-ca\") pod \"apiserver-7bbb656c7d-kfh2f\" (UID: \"0a7d0b69-f9a6-4a40-9981-3dd999c400d5\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-kfh2f" Dec 03 14:11:55 crc kubenswrapper[4805]: I1203 14:11:55.174372 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/1658d8e1-ac63-40fe-98ce-900a9f9aeae5-v4-0-config-system-cliconfig\") pod \"oauth-openshift-558db77b4-ntv6j\" (UID: \"1658d8e1-ac63-40fe-98ce-900a9f9aeae5\") " pod="openshift-authentication/oauth-openshift-558db77b4-ntv6j" Dec 03 14:11:55 crc kubenswrapper[4805]: I1203 14:11:55.174418 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e5a33c32-3873-49b8-8a81-c1a0357d94fc-config\") pod \"openshift-apiserver-operator-796bbdcf4f-l6qxf\" (UID: \"e5a33c32-3873-49b8-8a81-c1a0357d94fc\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-l6qxf" Dec 03 14:11:55 crc kubenswrapper[4805]: I1203 14:11:55.166920 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/1658d8e1-ac63-40fe-98ce-900a9f9aeae5-audit-dir\") pod \"oauth-openshift-558db77b4-ntv6j\" (UID: \"1658d8e1-ac63-40fe-98ce-900a9f9aeae5\") " pod="openshift-authentication/oauth-openshift-558db77b4-ntv6j" Dec 03 14:11:55 crc kubenswrapper[4805]: I1203 14:11:55.174798 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/0a7d0b69-f9a6-4a40-9981-3dd999c400d5-audit-policies\") pod \"apiserver-7bbb656c7d-kfh2f\" (UID: \"0a7d0b69-f9a6-4a40-9981-3dd999c400d5\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-kfh2f" Dec 03 14:11:55 crc kubenswrapper[4805]: I1203 14:11:55.175127 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/0a7d0b69-f9a6-4a40-9981-3dd999c400d5-audit-dir\") pod \"apiserver-7bbb656c7d-kfh2f\" (UID: \"0a7d0b69-f9a6-4a40-9981-3dd999c400d5\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-kfh2f" Dec 03 14:11:55 crc kubenswrapper[4805]: I1203 14:11:55.176297 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/ca425595-aae1-44f2-9470-84ba52167748-audit\") pod \"apiserver-76f77b778f-7nhvl\" (UID: \"ca425595-aae1-44f2-9470-84ba52167748\") " pod="openshift-apiserver/apiserver-76f77b778f-7nhvl" Dec 03 14:11:55 crc kubenswrapper[4805]: I1203 14:11:55.176520 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/1658d8e1-ac63-40fe-98ce-900a9f9aeae5-audit-policies\") pod \"oauth-openshift-558db77b4-ntv6j\" (UID: \"1658d8e1-ac63-40fe-98ce-900a9f9aeae5\") " pod="openshift-authentication/oauth-openshift-558db77b4-ntv6j" Dec 03 14:11:55 crc kubenswrapper[4805]: I1203 14:11:55.176670 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/b54908df-f04f-4515-ab36-842f0aef4a99-client-ca\") pod \"controller-manager-879f6c89f-k7rgc\" (UID: \"b54908df-f04f-4515-ab36-842f0aef4a99\") " pod="openshift-controller-manager/controller-manager-879f6c89f-k7rgc" Dec 03 14:11:55 crc kubenswrapper[4805]: I1203 14:11:55.176800 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/c6b0a18a-5cf2-430a-a0ef-2c2776839f11-samples-operator-tls\") pod \"cluster-samples-operator-665b6dd947-vc9sn\" (UID: \"c6b0a18a-5cf2-430a-a0ef-2c2776839f11\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-vc9sn" Dec 03 14:11:55 crc kubenswrapper[4805]: I1203 14:11:55.177026 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-dockercfg-x57mr" Dec 03 14:11:55 crc kubenswrapper[4805]: I1203 14:11:55.177698 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/1658d8e1-ac63-40fe-98ce-900a9f9aeae5-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-558db77b4-ntv6j\" (UID: \"1658d8e1-ac63-40fe-98ce-900a9f9aeae5\") " pod="openshift-authentication/oauth-openshift-558db77b4-ntv6j" Dec 03 14:11:55 crc kubenswrapper[4805]: I1203 14:11:55.178403 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/f4b1faf2-bf8b-483c-b28f-5885a9e8fe44-serving-cert\") pod \"openshift-config-operator-7777fb866f-7wzgf\" (UID: \"f4b1faf2-bf8b-483c-b28f-5885a9e8fe44\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-7wzgf" Dec 03 14:11:55 crc kubenswrapper[4805]: I1203 14:11:55.178606 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/04bd3442-dbeb-4cb4-bc9f-b07b9872231f-webhook-cert\") pod \"packageserver-d55dfcdfc-rr7fx\" (UID: \"04bd3442-dbeb-4cb4-bc9f-b07b9872231f\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-rr7fx" Dec 03 14:11:55 crc kubenswrapper[4805]: I1203 14:11:55.178683 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/0a7d0b69-f9a6-4a40-9981-3dd999c400d5-etcd-client\") pod \"apiserver-7bbb656c7d-kfh2f\" (UID: \"0a7d0b69-f9a6-4a40-9981-3dd999c400d5\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-kfh2f" Dec 03 14:11:55 crc kubenswrapper[4805]: I1203 14:11:55.178689 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9f53fe44-0d2c-4e3d-a682-df3a779a1c8a-config\") pod \"console-operator-58897d9998-46w2g\" (UID: \"9f53fe44-0d2c-4e3d-a682-df3a779a1c8a\") " pod="openshift-console-operator/console-operator-58897d9998-46w2g" Dec 03 14:11:55 crc kubenswrapper[4805]: I1203 14:11:55.179110 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ca425595-aae1-44f2-9470-84ba52167748-config\") pod \"apiserver-76f77b778f-7nhvl\" (UID: \"ca425595-aae1-44f2-9470-84ba52167748\") " pod="openshift-apiserver/apiserver-76f77b778f-7nhvl" Dec 03 14:11:55 crc kubenswrapper[4805]: I1203 14:11:55.179191 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/1658d8e1-ac63-40fe-98ce-900a9f9aeae5-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-558db77b4-ntv6j\" (UID: \"1658d8e1-ac63-40fe-98ce-900a9f9aeae5\") " pod="openshift-authentication/oauth-openshift-558db77b4-ntv6j" Dec 03 14:11:55 crc kubenswrapper[4805]: I1203 14:11:55.179820 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/33c718fb-affd-410b-9f53-ddb68ba9760c-webhook-certs\") pod \"multus-admission-controller-857f4d67dd-tn8tx\" (UID: \"33c718fb-affd-410b-9f53-ddb68ba9760c\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-tn8tx" Dec 03 14:11:55 crc kubenswrapper[4805]: I1203 14:11:55.180000 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"images\" (UniqueName: \"kubernetes.io/configmap/d5ea0764-4290-4eab-b4ce-dbd4909455c1-images\") pod \"machine-config-operator-74547568cd-kvmpp\" (UID: \"d5ea0764-4290-4eab-b4ce-dbd4909455c1\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-kvmpp" Dec 03 14:11:55 crc kubenswrapper[4805]: I1203 14:11:55.180068 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/0a7d0b69-f9a6-4a40-9981-3dd999c400d5-encryption-config\") pod \"apiserver-7bbb656c7d-kfh2f\" (UID: \"0a7d0b69-f9a6-4a40-9981-3dd999c400d5\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-kfh2f" Dec 03 14:11:55 crc kubenswrapper[4805]: I1203 14:11:55.180188 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/04bd3442-dbeb-4cb4-bc9f-b07b9872231f-apiservice-cert\") pod \"packageserver-d55dfcdfc-rr7fx\" (UID: \"04bd3442-dbeb-4cb4-bc9f-b07b9872231f\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-rr7fx" Dec 03 14:11:55 crc kubenswrapper[4805]: I1203 14:11:55.180347 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/1658d8e1-ac63-40fe-98ce-900a9f9aeae5-v4-0-config-user-template-error\") pod \"oauth-openshift-558db77b4-ntv6j\" (UID: \"1658d8e1-ac63-40fe-98ce-900a9f9aeae5\") " pod="openshift-authentication/oauth-openshift-558db77b4-ntv6j" Dec 03 14:11:55 crc kubenswrapper[4805]: I1203 14:11:55.180519 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b54908df-f04f-4515-ab36-842f0aef4a99-config\") pod \"controller-manager-879f6c89f-k7rgc\" (UID: \"b54908df-f04f-4515-ab36-842f0aef4a99\") " pod="openshift-controller-manager/controller-manager-879f6c89f-k7rgc" Dec 03 14:11:55 crc kubenswrapper[4805]: I1203 14:11:55.180513 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/1658d8e1-ac63-40fe-98ce-900a9f9aeae5-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-558db77b4-ntv6j\" (UID: \"1658d8e1-ac63-40fe-98ce-900a9f9aeae5\") " pod="openshift-authentication/oauth-openshift-558db77b4-ntv6j" Dec 03 14:11:55 crc kubenswrapper[4805]: I1203 14:11:55.180594 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/ca425595-aae1-44f2-9470-84ba52167748-encryption-config\") pod \"apiserver-76f77b778f-7nhvl\" (UID: \"ca425595-aae1-44f2-9470-84ba52167748\") " pod="openshift-apiserver/apiserver-76f77b778f-7nhvl" Dec 03 14:11:55 crc kubenswrapper[4805]: I1203 14:11:55.180706 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/e2034338-8a35-4fd2-9bbf-be733939fc25-oauth-serving-cert\") pod \"console-f9d7485db-pp9tv\" (UID: \"e2034338-8a35-4fd2-9bbf-be733939fc25\") " pod="openshift-console/console-f9d7485db-pp9tv" Dec 03 14:11:55 crc kubenswrapper[4805]: I1203 14:11:55.181076 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/cdcb6b89-e405-479e-a60d-96681328634e-trusted-ca-bundle\") pod \"authentication-operator-69f744f599-bs6tj\" (UID: \"cdcb6b89-e405-479e-a60d-96681328634e\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-bs6tj" Dec 03 14:11:55 crc kubenswrapper[4805]: I1203 14:11:55.181162 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/00941575-52e9-482d-b256-1b1c757d96e0-config\") pod \"machine-approver-56656f9798-nv2m9\" (UID: \"00941575-52e9-482d-b256-1b1c757d96e0\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-nv2m9" Dec 03 14:11:55 crc kubenswrapper[4805]: I1203 14:11:55.181502 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/1658d8e1-ac63-40fe-98ce-900a9f9aeae5-v4-0-config-system-session\") pod \"oauth-openshift-558db77b4-ntv6j\" (UID: \"1658d8e1-ac63-40fe-98ce-900a9f9aeae5\") " pod="openshift-authentication/oauth-openshift-558db77b4-ntv6j" Dec 03 14:11:55 crc kubenswrapper[4805]: I1203 14:11:55.181672 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0a7d0b69-f9a6-4a40-9981-3dd999c400d5-serving-cert\") pod \"apiserver-7bbb656c7d-kfh2f\" (UID: \"0a7d0b69-f9a6-4a40-9981-3dd999c400d5\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-kfh2f" Dec 03 14:11:55 crc kubenswrapper[4805]: I1203 14:11:55.184416 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9b932fa5-f87b-46e1-8987-683cfd3b46fd-config\") pod \"service-ca-operator-777779d784-z7lgj\" (UID: \"9b932fa5-f87b-46e1-8987-683cfd3b46fd\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-z7lgj" Dec 03 14:11:55 crc kubenswrapper[4805]: I1203 14:11:55.182206 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"images\" (UniqueName: \"kubernetes.io/configmap/dfd20bfc-5c11-4287-a4b3-df9e0ca4dcf1-images\") pod \"machine-api-operator-5694c8668f-jvspg\" (UID: \"dfd20bfc-5c11-4287-a4b3-df9e0ca4dcf1\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-jvspg" Dec 03 14:11:55 crc kubenswrapper[4805]: I1203 14:11:55.184445 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/b54908df-f04f-4515-ab36-842f0aef4a99-serving-cert\") pod \"controller-manager-879f6c89f-k7rgc\" (UID: \"b54908df-f04f-4515-ab36-842f0aef4a99\") " pod="openshift-controller-manager/controller-manager-879f6c89f-k7rgc" Dec 03 14:11:55 crc kubenswrapper[4805]: I1203 14:11:55.182552 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/cdcb6b89-e405-479e-a60d-96681328634e-serving-cert\") pod \"authentication-operator-69f744f599-bs6tj\" (UID: \"cdcb6b89-e405-479e-a60d-96681328634e\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-bs6tj" Dec 03 14:11:55 crc kubenswrapper[4805]: I1203 14:11:55.182566 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/00941575-52e9-482d-b256-1b1c757d96e0-machine-approver-tls\") pod \"machine-approver-56656f9798-nv2m9\" (UID: \"00941575-52e9-482d-b256-1b1c757d96e0\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-nv2m9" Dec 03 14:11:55 crc kubenswrapper[4805]: I1203 14:11:55.181224 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/d5ea0764-4290-4eab-b4ce-dbd4909455c1-auth-proxy-config\") pod \"machine-config-operator-74547568cd-kvmpp\" (UID: \"d5ea0764-4290-4eab-b4ce-dbd4909455c1\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-kvmpp" Dec 03 14:11:55 crc kubenswrapper[4805]: I1203 14:11:55.182887 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/ca425595-aae1-44f2-9470-84ba52167748-trusted-ca-bundle\") pod \"apiserver-76f77b778f-7nhvl\" (UID: \"ca425595-aae1-44f2-9470-84ba52167748\") " pod="openshift-apiserver/apiserver-76f77b778f-7nhvl" Dec 03 14:11:55 crc kubenswrapper[4805]: I1203 14:11:55.183072 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/1658d8e1-ac63-40fe-98ce-900a9f9aeae5-v4-0-config-system-service-ca\") pod \"oauth-openshift-558db77b4-ntv6j\" (UID: \"1658d8e1-ac63-40fe-98ce-900a9f9aeae5\") " pod="openshift-authentication/oauth-openshift-558db77b4-ntv6j" Dec 03 14:11:55 crc kubenswrapper[4805]: I1203 14:11:55.183312 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/dfa6979a-2473-469b-ac56-9086e35a63f4-client-ca\") pod \"route-controller-manager-6576b87f9c-nsz7s\" (UID: \"dfa6979a-2473-469b-ac56-9086e35a63f4\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-nsz7s" Dec 03 14:11:55 crc kubenswrapper[4805]: I1203 14:11:55.183450 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/00941575-52e9-482d-b256-1b1c757d96e0-auth-proxy-config\") pod \"machine-approver-56656f9798-nv2m9\" (UID: \"00941575-52e9-482d-b256-1b1c757d96e0\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-nv2m9" Dec 03 14:11:55 crc kubenswrapper[4805]: I1203 14:11:55.183721 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/dfd20bfc-5c11-4287-a4b3-df9e0ca4dcf1-config\") pod \"machine-api-operator-5694c8668f-jvspg\" (UID: \"dfd20bfc-5c11-4287-a4b3-df9e0ca4dcf1\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-jvspg" Dec 03 14:11:55 crc kubenswrapper[4805]: I1203 14:11:55.183786 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9f53fe44-0d2c-4e3d-a682-df3a779a1c8a-serving-cert\") pod \"console-operator-58897d9998-46w2g\" (UID: \"9f53fe44-0d2c-4e3d-a682-df3a779a1c8a\") " pod="openshift-console-operator/console-operator-58897d9998-46w2g" Dec 03 14:11:55 crc kubenswrapper[4805]: I1203 14:11:55.182032 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/f4b1faf2-bf8b-483c-b28f-5885a9e8fe44-available-featuregates\") pod \"openshift-config-operator-7777fb866f-7wzgf\" (UID: \"f4b1faf2-bf8b-483c-b28f-5885a9e8fe44\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-7wzgf" Dec 03 14:11:55 crc kubenswrapper[4805]: I1203 14:11:55.182265 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/d5ea0764-4290-4eab-b4ce-dbd4909455c1-proxy-tls\") pod \"machine-config-operator-74547568cd-kvmpp\" (UID: \"d5ea0764-4290-4eab-b4ce-dbd4909455c1\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-kvmpp" Dec 03 14:11:55 crc kubenswrapper[4805]: I1203 14:11:55.184613 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/9f53fe44-0d2c-4e3d-a682-df3a779a1c8a-trusted-ca\") pod \"console-operator-58897d9998-46w2g\" (UID: \"9f53fe44-0d2c-4e3d-a682-df3a779a1c8a\") " pod="openshift-console-operator/console-operator-58897d9998-46w2g" Dec 03 14:11:55 crc kubenswrapper[4805]: I1203 14:11:55.185210 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e5a33c32-3873-49b8-8a81-c1a0357d94fc-serving-cert\") pod \"openshift-apiserver-operator-796bbdcf4f-l6qxf\" (UID: \"e5a33c32-3873-49b8-8a81-c1a0357d94fc\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-l6qxf" Dec 03 14:11:55 crc kubenswrapper[4805]: I1203 14:11:55.185318 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/1658d8e1-ac63-40fe-98ce-900a9f9aeae5-v4-0-config-user-template-login\") pod \"oauth-openshift-558db77b4-ntv6j\" (UID: \"1658d8e1-ac63-40fe-98ce-900a9f9aeae5\") " pod="openshift-authentication/oauth-openshift-558db77b4-ntv6j" Dec 03 14:11:55 crc kubenswrapper[4805]: I1203 14:11:55.185447 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/ca425595-aae1-44f2-9470-84ba52167748-serving-cert\") pod \"apiserver-76f77b778f-7nhvl\" (UID: \"ca425595-aae1-44f2-9470-84ba52167748\") " pod="openshift-apiserver/apiserver-76f77b778f-7nhvl" Dec 03 14:11:55 crc kubenswrapper[4805]: I1203 14:11:55.185963 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/ca425595-aae1-44f2-9470-84ba52167748-etcd-client\") pod \"apiserver-76f77b778f-7nhvl\" (UID: \"ca425595-aae1-44f2-9470-84ba52167748\") " pod="openshift-apiserver/apiserver-76f77b778f-7nhvl" Dec 03 14:11:55 crc kubenswrapper[4805]: I1203 14:11:55.186105 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/dfd20bfc-5c11-4287-a4b3-df9e0ca4dcf1-machine-api-operator-tls\") pod \"machine-api-operator-5694c8668f-jvspg\" (UID: \"dfd20bfc-5c11-4287-a4b3-df9e0ca4dcf1\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-jvspg" Dec 03 14:11:55 crc kubenswrapper[4805]: I1203 14:11:55.189384 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/1658d8e1-ac63-40fe-98ce-900a9f9aeae5-v4-0-config-system-serving-cert\") pod \"oauth-openshift-558db77b4-ntv6j\" (UID: \"1658d8e1-ac63-40fe-98ce-900a9f9aeae5\") " pod="openshift-authentication/oauth-openshift-558db77b4-ntv6j" Dec 03 14:11:55 crc kubenswrapper[4805]: I1203 14:11:55.189755 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/dfa6979a-2473-469b-ac56-9086e35a63f4-serving-cert\") pod \"route-controller-manager-6576b87f9c-nsz7s\" (UID: \"dfa6979a-2473-469b-ac56-9086e35a63f4\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-nsz7s" Dec 03 14:11:55 crc kubenswrapper[4805]: I1203 14:11:55.189757 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/1658d8e1-ac63-40fe-98ce-900a9f9aeae5-v4-0-config-system-router-certs\") pod \"oauth-openshift-558db77b4-ntv6j\" (UID: \"1658d8e1-ac63-40fe-98ce-900a9f9aeae5\") " pod="openshift-authentication/oauth-openshift-558db77b4-ntv6j" Dec 03 14:11:55 crc kubenswrapper[4805]: I1203 14:11:55.189897 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/e2034338-8a35-4fd2-9bbf-be733939fc25-console-oauth-config\") pod \"console-f9d7485db-pp9tv\" (UID: \"e2034338-8a35-4fd2-9bbf-be733939fc25\") " pod="openshift-console/console-f9d7485db-pp9tv" Dec 03 14:11:55 crc kubenswrapper[4805]: I1203 14:11:55.190061 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/1658d8e1-ac63-40fe-98ce-900a9f9aeae5-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-558db77b4-ntv6j\" (UID: \"1658d8e1-ac63-40fe-98ce-900a9f9aeae5\") " pod="openshift-authentication/oauth-openshift-558db77b4-ntv6j" Dec 03 14:11:55 crc kubenswrapper[4805]: I1203 14:11:55.190191 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9b932fa5-f87b-46e1-8987-683cfd3b46fd-serving-cert\") pod \"service-ca-operator-777779d784-z7lgj\" (UID: \"9b932fa5-f87b-46e1-8987-683cfd3b46fd\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-z7lgj" Dec 03 14:11:55 crc kubenswrapper[4805]: I1203 14:11:55.190973 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/e2034338-8a35-4fd2-9bbf-be733939fc25-console-serving-cert\") pod \"console-f9d7485db-pp9tv\" (UID: \"e2034338-8a35-4fd2-9bbf-be733939fc25\") " pod="openshift-console/console-f9d7485db-pp9tv" Dec 03 14:11:55 crc kubenswrapper[4805]: I1203 14:11:55.196065 4805 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver-operator"/"kube-root-ca.crt" Dec 03 14:11:55 crc kubenswrapper[4805]: I1203 14:11:55.216081 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-serving-cert" Dec 03 14:11:55 crc kubenswrapper[4805]: I1203 14:11:55.237050 4805 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Dec 03 14:11:55 crc kubenswrapper[4805]: I1203 14:11:55.255670 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"pprof-cert" Dec 03 14:11:55 crc kubenswrapper[4805]: I1203 14:11:55.273633 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/e600aefb-0839-423f-8ac8-1916400cfa92-mcc-auth-proxy-config\") pod \"machine-config-controller-84d6567774-t4kl8\" (UID: \"e600aefb-0839-423f-8ac8-1916400cfa92\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-t4kl8" Dec 03 14:11:55 crc kubenswrapper[4805]: I1203 14:11:55.273676 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/1c55f37c-24b9-4800-9050-e0da883df2c0-signing-cabundle\") pod \"service-ca-9c57cc56f-dbfrd\" (UID: \"1c55f37c-24b9-4800-9050-e0da883df2c0\") " pod="openshift-service-ca/service-ca-9c57cc56f-dbfrd" Dec 03 14:11:55 crc kubenswrapper[4805]: I1203 14:11:55.273694 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/da68f128-1d8d-4256-a28d-a174766e1331-trusted-ca\") pod \"cluster-image-registry-operator-dc59b4c8b-jvsqf\" (UID: \"da68f128-1d8d-4256-a28d-a174766e1331\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-jvsqf" Dec 03 14:11:55 crc kubenswrapper[4805]: I1203 14:11:55.273711 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qpdfz\" (UniqueName: \"kubernetes.io/projected/da68f128-1d8d-4256-a28d-a174766e1331-kube-api-access-qpdfz\") pod \"cluster-image-registry-operator-dc59b4c8b-jvsqf\" (UID: \"da68f128-1d8d-4256-a28d-a174766e1331\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-jvsqf" Dec 03 14:11:55 crc kubenswrapper[4805]: I1203 14:11:55.273727 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e9022931-39f7-45cf-925f-fe19e1b870b8-config\") pod \"kube-apiserver-operator-766d6c64bb-2vnrk\" (UID: \"e9022931-39f7-45cf-925f-fe19e1b870b8\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-2vnrk" Dec 03 14:11:55 crc kubenswrapper[4805]: I1203 14:11:55.273752 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/8e437287-3b92-4c6a-abe9-84c786ba3220-metrics-tls\") pod \"dns-operator-744455d44c-rjf9m\" (UID: \"8e437287-3b92-4c6a-abe9-84c786ba3220\") " pod="openshift-dns-operator/dns-operator-744455d44c-rjf9m" Dec 03 14:11:55 crc kubenswrapper[4805]: I1203 14:11:55.273768 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/8b954d96-a62c-434e-a340-8446d778c545-secret-volume\") pod \"collect-profiles-29412840-gj7pr\" (UID: \"8b954d96-a62c-434e-a340-8446d778c545\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29412840-gj7pr" Dec 03 14:11:55 crc kubenswrapper[4805]: I1203 14:11:55.273783 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/da68f128-1d8d-4256-a28d-a174766e1331-bound-sa-token\") pod \"cluster-image-registry-operator-dc59b4c8b-jvsqf\" (UID: \"da68f128-1d8d-4256-a28d-a174766e1331\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-jvsqf" Dec 03 14:11:55 crc kubenswrapper[4805]: I1203 14:11:55.273797 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/ae561e71-047a-4cf7-8378-d08adb458d9c-serving-cert\") pod \"etcd-operator-b45778765-2x56c\" (UID: \"ae561e71-047a-4cf7-8378-d08adb458d9c\") " pod="openshift-etcd-operator/etcd-operator-b45778765-2x56c" Dec 03 14:11:55 crc kubenswrapper[4805]: I1203 14:11:55.273872 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ae561e71-047a-4cf7-8378-d08adb458d9c-config\") pod \"etcd-operator-b45778765-2x56c\" (UID: \"ae561e71-047a-4cf7-8378-d08adb458d9c\") " pod="openshift-etcd-operator/etcd-operator-b45778765-2x56c" Dec 03 14:11:55 crc kubenswrapper[4805]: I1203 14:11:55.273895 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/ae561e71-047a-4cf7-8378-d08adb458d9c-etcd-service-ca\") pod \"etcd-operator-b45778765-2x56c\" (UID: \"ae561e71-047a-4cf7-8378-d08adb458d9c\") " pod="openshift-etcd-operator/etcd-operator-b45778765-2x56c" Dec 03 14:11:55 crc kubenswrapper[4805]: I1203 14:11:55.273917 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4djzx\" (UniqueName: \"kubernetes.io/projected/1c55f37c-24b9-4800-9050-e0da883df2c0-kube-api-access-4djzx\") pod \"service-ca-9c57cc56f-dbfrd\" (UID: \"1c55f37c-24b9-4800-9050-e0da883df2c0\") " pod="openshift-service-ca/service-ca-9c57cc56f-dbfrd" Dec 03 14:11:55 crc kubenswrapper[4805]: I1203 14:11:55.273932 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/da68f128-1d8d-4256-a28d-a174766e1331-image-registry-operator-tls\") pod \"cluster-image-registry-operator-dc59b4c8b-jvsqf\" (UID: \"da68f128-1d8d-4256-a28d-a174766e1331\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-jvsqf" Dec 03 14:11:55 crc kubenswrapper[4805]: I1203 14:11:55.273974 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-v8t8r\" (UniqueName: \"kubernetes.io/projected/8e437287-3b92-4c6a-abe9-84c786ba3220-kube-api-access-v8t8r\") pod \"dns-operator-744455d44c-rjf9m\" (UID: \"8e437287-3b92-4c6a-abe9-84c786ba3220\") " pod="openshift-dns-operator/dns-operator-744455d44c-rjf9m" Dec 03 14:11:55 crc kubenswrapper[4805]: I1203 14:11:55.274005 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/ae561e71-047a-4cf7-8378-d08adb458d9c-etcd-ca\") pod \"etcd-operator-b45778765-2x56c\" (UID: \"ae561e71-047a-4cf7-8378-d08adb458d9c\") " pod="openshift-etcd-operator/etcd-operator-b45778765-2x56c" Dec 03 14:11:55 crc kubenswrapper[4805]: I1203 14:11:55.274036 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/e600aefb-0839-423f-8ac8-1916400cfa92-proxy-tls\") pod \"machine-config-controller-84d6567774-t4kl8\" (UID: \"e600aefb-0839-423f-8ac8-1916400cfa92\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-t4kl8" Dec 03 14:11:55 crc kubenswrapper[4805]: I1203 14:11:55.274050 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e9022931-39f7-45cf-925f-fe19e1b870b8-serving-cert\") pod \"kube-apiserver-operator-766d6c64bb-2vnrk\" (UID: \"e9022931-39f7-45cf-925f-fe19e1b870b8\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-2vnrk" Dec 03 14:11:55 crc kubenswrapper[4805]: I1203 14:11:55.274074 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/1c55f37c-24b9-4800-9050-e0da883df2c0-signing-key\") pod \"service-ca-9c57cc56f-dbfrd\" (UID: \"1c55f37c-24b9-4800-9050-e0da883df2c0\") " pod="openshift-service-ca/service-ca-9c57cc56f-dbfrd" Dec 03 14:11:55 crc kubenswrapper[4805]: I1203 14:11:55.274089 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-csvjg\" (UniqueName: \"kubernetes.io/projected/e600aefb-0839-423f-8ac8-1916400cfa92-kube-api-access-csvjg\") pod \"machine-config-controller-84d6567774-t4kl8\" (UID: \"e600aefb-0839-423f-8ac8-1916400cfa92\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-t4kl8" Dec 03 14:11:55 crc kubenswrapper[4805]: I1203 14:11:55.274111 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ddtf4\" (UniqueName: \"kubernetes.io/projected/ae561e71-047a-4cf7-8378-d08adb458d9c-kube-api-access-ddtf4\") pod \"etcd-operator-b45778765-2x56c\" (UID: \"ae561e71-047a-4cf7-8378-d08adb458d9c\") " pod="openshift-etcd-operator/etcd-operator-b45778765-2x56c" Dec 03 14:11:55 crc kubenswrapper[4805]: I1203 14:11:55.274135 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7frlc\" (UniqueName: \"kubernetes.io/projected/8b954d96-a62c-434e-a340-8446d778c545-kube-api-access-7frlc\") pod \"collect-profiles-29412840-gj7pr\" (UID: \"8b954d96-a62c-434e-a340-8446d778c545\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29412840-gj7pr" Dec 03 14:11:55 crc kubenswrapper[4805]: I1203 14:11:55.274149 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/ae561e71-047a-4cf7-8378-d08adb458d9c-etcd-client\") pod \"etcd-operator-b45778765-2x56c\" (UID: \"ae561e71-047a-4cf7-8378-d08adb458d9c\") " pod="openshift-etcd-operator/etcd-operator-b45778765-2x56c" Dec 03 14:11:55 crc kubenswrapper[4805]: I1203 14:11:55.274174 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/8b954d96-a62c-434e-a340-8446d778c545-config-volume\") pod \"collect-profiles-29412840-gj7pr\" (UID: \"8b954d96-a62c-434e-a340-8446d778c545\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29412840-gj7pr" Dec 03 14:11:55 crc kubenswrapper[4805]: I1203 14:11:55.274189 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/e9022931-39f7-45cf-925f-fe19e1b870b8-kube-api-access\") pod \"kube-apiserver-operator-766d6c64bb-2vnrk\" (UID: \"e9022931-39f7-45cf-925f-fe19e1b870b8\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-2vnrk" Dec 03 14:11:55 crc kubenswrapper[4805]: I1203 14:11:55.275785 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/e600aefb-0839-423f-8ac8-1916400cfa92-mcc-auth-proxy-config\") pod \"machine-config-controller-84d6567774-t4kl8\" (UID: \"e600aefb-0839-423f-8ac8-1916400cfa92\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-t4kl8" Dec 03 14:11:55 crc kubenswrapper[4805]: I1203 14:11:55.276414 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e9022931-39f7-45cf-925f-fe19e1b870b8-config\") pod \"kube-apiserver-operator-766d6c64bb-2vnrk\" (UID: \"e9022931-39f7-45cf-925f-fe19e1b870b8\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-2vnrk" Dec 03 14:11:55 crc kubenswrapper[4805]: I1203 14:11:55.277584 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Dec 03 14:11:55 crc kubenswrapper[4805]: I1203 14:11:55.278960 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/8b954d96-a62c-434e-a340-8446d778c545-config-volume\") pod \"collect-profiles-29412840-gj7pr\" (UID: \"8b954d96-a62c-434e-a340-8446d778c545\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29412840-gj7pr" Dec 03 14:11:55 crc kubenswrapper[4805]: I1203 14:11:55.280467 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/8b954d96-a62c-434e-a340-8446d778c545-secret-volume\") pod \"collect-profiles-29412840-gj7pr\" (UID: \"8b954d96-a62c-434e-a340-8446d778c545\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29412840-gj7pr" Dec 03 14:11:55 crc kubenswrapper[4805]: I1203 14:11:55.280890 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e9022931-39f7-45cf-925f-fe19e1b870b8-serving-cert\") pod \"kube-apiserver-operator-766d6c64bb-2vnrk\" (UID: \"e9022931-39f7-45cf-925f-fe19e1b870b8\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-2vnrk" Dec 03 14:11:55 crc kubenswrapper[4805]: I1203 14:11:55.298356 4805 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"openshift-service-ca.crt" Dec 03 14:11:55 crc kubenswrapper[4805]: I1203 14:11:55.316820 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca"/"service-ca-dockercfg-pn86c" Dec 03 14:11:55 crc kubenswrapper[4805]: I1203 14:11:55.337011 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca"/"signing-key" Dec 03 14:11:55 crc kubenswrapper[4805]: I1203 14:11:55.356106 4805 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"kube-root-ca.crt" Dec 03 14:11:55 crc kubenswrapper[4805]: I1203 14:11:55.366249 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/1c55f37c-24b9-4800-9050-e0da883df2c0-signing-key\") pod \"service-ca-9c57cc56f-dbfrd\" (UID: \"1c55f37c-24b9-4800-9050-e0da883df2c0\") " pod="openshift-service-ca/service-ca-9c57cc56f-dbfrd" Dec 03 14:11:55 crc kubenswrapper[4805]: I1203 14:11:55.377016 4805 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"signing-cabundle" Dec 03 14:11:55 crc kubenswrapper[4805]: I1203 14:11:55.386643 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/1c55f37c-24b9-4800-9050-e0da883df2c0-signing-cabundle\") pod \"service-ca-9c57cc56f-dbfrd\" (UID: \"1c55f37c-24b9-4800-9050-e0da883df2c0\") " pod="openshift-service-ca/service-ca-9c57cc56f-dbfrd" Dec 03 14:11:55 crc kubenswrapper[4805]: I1203 14:11:55.397327 4805 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"openshift-service-ca.crt" Dec 03 14:11:55 crc kubenswrapper[4805]: I1203 14:11:55.416357 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-operator-serving-cert" Dec 03 14:11:55 crc kubenswrapper[4805]: I1203 14:11:55.421877 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/ae561e71-047a-4cf7-8378-d08adb458d9c-serving-cert\") pod \"etcd-operator-b45778765-2x56c\" (UID: \"ae561e71-047a-4cf7-8378-d08adb458d9c\") " pod="openshift-etcd-operator/etcd-operator-b45778765-2x56c" Dec 03 14:11:55 crc kubenswrapper[4805]: I1203 14:11:55.436269 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-client" Dec 03 14:11:55 crc kubenswrapper[4805]: I1203 14:11:55.442321 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/ae561e71-047a-4cf7-8378-d08adb458d9c-etcd-client\") pod \"etcd-operator-b45778765-2x56c\" (UID: \"ae561e71-047a-4cf7-8378-d08adb458d9c\") " pod="openshift-etcd-operator/etcd-operator-b45778765-2x56c" Dec 03 14:11:55 crc kubenswrapper[4805]: I1203 14:11:55.456346 4805 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-service-ca-bundle" Dec 03 14:11:55 crc kubenswrapper[4805]: I1203 14:11:55.458930 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/ae561e71-047a-4cf7-8378-d08adb458d9c-etcd-service-ca\") pod \"etcd-operator-b45778765-2x56c\" (UID: \"ae561e71-047a-4cf7-8378-d08adb458d9c\") " pod="openshift-etcd-operator/etcd-operator-b45778765-2x56c" Dec 03 14:11:55 crc kubenswrapper[4805]: I1203 14:11:55.476712 4805 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"kube-root-ca.crt" Dec 03 14:11:55 crc kubenswrapper[4805]: I1203 14:11:55.496009 4805 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-ca-bundle" Dec 03 14:11:55 crc kubenswrapper[4805]: I1203 14:11:55.498716 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/ae561e71-047a-4cf7-8378-d08adb458d9c-etcd-ca\") pod \"etcd-operator-b45778765-2x56c\" (UID: \"ae561e71-047a-4cf7-8378-d08adb458d9c\") " pod="openshift-etcd-operator/etcd-operator-b45778765-2x56c" Dec 03 14:11:55 crc kubenswrapper[4805]: I1203 14:11:55.516259 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-operator-dockercfg-r9srn" Dec 03 14:11:55 crc kubenswrapper[4805]: I1203 14:11:55.536517 4805 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-operator-config" Dec 03 14:11:55 crc kubenswrapper[4805]: I1203 14:11:55.539474 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ae561e71-047a-4cf7-8378-d08adb458d9c-config\") pod \"etcd-operator-b45778765-2x56c\" (UID: \"ae561e71-047a-4cf7-8378-d08adb458d9c\") " pod="openshift-etcd-operator/etcd-operator-b45778765-2x56c" Dec 03 14:11:55 crc kubenswrapper[4805]: I1203 14:11:55.576118 4805 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns-operator"/"openshift-service-ca.crt" Dec 03 14:11:55 crc kubenswrapper[4805]: I1203 14:11:55.596185 4805 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns-operator"/"kube-root-ca.crt" Dec 03 14:11:55 crc kubenswrapper[4805]: I1203 14:11:55.617588 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns-operator"/"dns-operator-dockercfg-9mqw5" Dec 03 14:11:55 crc kubenswrapper[4805]: I1203 14:11:55.637224 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns-operator"/"metrics-tls" Dec 03 14:11:55 crc kubenswrapper[4805]: I1203 14:11:55.650369 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/8e437287-3b92-4c6a-abe9-84c786ba3220-metrics-tls\") pod \"dns-operator-744455d44c-rjf9m\" (UID: \"8e437287-3b92-4c6a-abe9-84c786ba3220\") " pod="openshift-dns-operator/dns-operator-744455d44c-rjf9m" Dec 03 14:11:55 crc kubenswrapper[4805]: I1203 14:11:55.656568 4805 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"openshift-service-ca.crt" Dec 03 14:11:55 crc kubenswrapper[4805]: I1203 14:11:55.678133 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"marketplace-operator-dockercfg-5nsgg" Dec 03 14:11:55 crc kubenswrapper[4805]: I1203 14:11:55.697322 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"marketplace-operator-metrics" Dec 03 14:11:55 crc kubenswrapper[4805]: I1203 14:11:55.727483 4805 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"marketplace-trusted-ca" Dec 03 14:11:55 crc kubenswrapper[4805]: I1203 14:11:55.736658 4805 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"kube-root-ca.crt" Dec 03 14:11:55 crc kubenswrapper[4805]: I1203 14:11:55.756366 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"mcc-proxy-tls" Dec 03 14:11:55 crc kubenswrapper[4805]: I1203 14:11:55.761342 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/e600aefb-0839-423f-8ac8-1916400cfa92-proxy-tls\") pod \"machine-config-controller-84d6567774-t4kl8\" (UID: \"e600aefb-0839-423f-8ac8-1916400cfa92\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-t4kl8" Dec 03 14:11:55 crc kubenswrapper[4805]: I1203 14:11:55.776042 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-controller-dockercfg-c2lfx" Dec 03 14:11:55 crc kubenswrapper[4805]: I1203 14:11:55.806059 4805 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"trusted-ca" Dec 03 14:11:55 crc kubenswrapper[4805]: I1203 14:11:55.808037 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/da68f128-1d8d-4256-a28d-a174766e1331-trusted-ca\") pod \"cluster-image-registry-operator-dc59b4c8b-jvsqf\" (UID: \"da68f128-1d8d-4256-a28d-a174766e1331\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-jvsqf" Dec 03 14:11:55 crc kubenswrapper[4805]: I1203 14:11:55.816586 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"registry-dockercfg-kzzsd" Dec 03 14:11:55 crc kubenswrapper[4805]: I1203 14:11:55.836618 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"image-registry-tls" Dec 03 14:11:55 crc kubenswrapper[4805]: I1203 14:11:55.855949 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"installation-pull-secrets" Dec 03 14:11:55 crc kubenswrapper[4805]: I1203 14:11:55.877359 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"image-registry-operator-tls" Dec 03 14:11:55 crc kubenswrapper[4805]: I1203 14:11:55.881862 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/da68f128-1d8d-4256-a28d-a174766e1331-image-registry-operator-tls\") pod \"cluster-image-registry-operator-dc59b4c8b-jvsqf\" (UID: \"da68f128-1d8d-4256-a28d-a174766e1331\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-jvsqf" Dec 03 14:11:55 crc kubenswrapper[4805]: I1203 14:11:55.896314 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"cluster-image-registry-operator-dockercfg-m4qtx" Dec 03 14:11:55 crc kubenswrapper[4805]: I1203 14:11:55.916687 4805 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator"/"openshift-service-ca.crt" Dec 03 14:11:55 crc kubenswrapper[4805]: I1203 14:11:55.937291 4805 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"openshift-service-ca.crt" Dec 03 14:11:55 crc kubenswrapper[4805]: I1203 14:11:55.956632 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator"/"kube-storage-version-migrator-sa-dockercfg-5xfcg" Dec 03 14:11:55 crc kubenswrapper[4805]: I1203 14:11:55.977160 4805 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator"/"kube-root-ca.crt" Dec 03 14:11:55 crc kubenswrapper[4805]: I1203 14:11:55.996934 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-operator"/"ingress-operator-dockercfg-7lnqk" Dec 03 14:11:56 crc kubenswrapper[4805]: I1203 14:11:56.016120 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-operator"/"metrics-tls" Dec 03 14:11:56 crc kubenswrapper[4805]: I1203 14:11:56.043701 4805 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"trusted-ca" Dec 03 14:11:56 crc kubenswrapper[4805]: I1203 14:11:56.057076 4805 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"kube-root-ca.crt" Dec 03 14:11:56 crc kubenswrapper[4805]: I1203 14:11:56.076574 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"olm-operator-serving-cert" Dec 03 14:11:56 crc kubenswrapper[4805]: I1203 14:11:56.095082 4805 request.go:700] Waited for 1.00629222s due to client-side throttling, not priority and fairness, request: GET:https://api-int.crc.testing:6443/api/v1/namespaces/openshift-image-registry/persistentvolumeclaims/crc-image-registry-storage Dec 03 14:11:56 crc kubenswrapper[4805]: I1203 14:11:56.118122 4805 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"openshift-service-ca.crt" Dec 03 14:11:56 crc kubenswrapper[4805]: I1203 14:11:56.142069 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator-operator"/"kube-storage-version-migrator-operator-dockercfg-2bh8d" Dec 03 14:11:56 crc kubenswrapper[4805]: I1203 14:11:56.157303 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator-operator"/"serving-cert" Dec 03 14:11:56 crc kubenswrapper[4805]: I1203 14:11:56.176015 4805 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"config" Dec 03 14:11:56 crc kubenswrapper[4805]: I1203 14:11:56.196266 4805 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"kube-root-ca.crt" Dec 03 14:11:56 crc kubenswrapper[4805]: I1203 14:11:56.216260 4805 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-scheduler-operator"/"kube-root-ca.crt" Dec 03 14:11:56 crc kubenswrapper[4805]: I1203 14:11:56.236993 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-scheduler-operator"/"openshift-kube-scheduler-operator-dockercfg-qt55r" Dec 03 14:11:56 crc kubenswrapper[4805]: I1203 14:11:56.257025 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-scheduler-operator"/"kube-scheduler-operator-serving-cert" Dec 03 14:11:56 crc kubenswrapper[4805]: I1203 14:11:56.276546 4805 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-scheduler-operator"/"openshift-kube-scheduler-operator-config" Dec 03 14:11:56 crc kubenswrapper[4805]: I1203 14:11:56.297203 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"catalog-operator-serving-cert" Dec 03 14:11:56 crc kubenswrapper[4805]: I1203 14:11:56.316199 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-dockercfg-gkqpw" Dec 03 14:11:56 crc kubenswrapper[4805]: I1203 14:11:56.336550 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-serving-cert" Dec 03 14:11:56 crc kubenswrapper[4805]: I1203 14:11:56.355926 4805 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-config" Dec 03 14:11:56 crc kubenswrapper[4805]: I1203 14:11:56.376448 4805 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager-operator"/"kube-root-ca.crt" Dec 03 14:11:56 crc kubenswrapper[4805]: I1203 14:11:56.396645 4805 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"service-ca-bundle" Dec 03 14:11:56 crc kubenswrapper[4805]: I1203 14:11:56.417508 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-metrics-certs-default" Dec 03 14:11:56 crc kubenswrapper[4805]: I1203 14:11:56.437166 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-dockercfg-zdk86" Dec 03 14:11:56 crc kubenswrapper[4805]: I1203 14:11:56.456354 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-certs-default" Dec 03 14:11:56 crc kubenswrapper[4805]: I1203 14:11:56.476507 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-stats-default" Dec 03 14:11:56 crc kubenswrapper[4805]: I1203 14:11:56.497334 4805 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"kube-root-ca.crt" Dec 03 14:11:56 crc kubenswrapper[4805]: I1203 14:11:56.516828 4805 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"openshift-service-ca.crt" Dec 03 14:11:56 crc kubenswrapper[4805]: I1203 14:11:56.538694 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"package-server-manager-serving-cert" Dec 03 14:11:56 crc kubenswrapper[4805]: I1203 14:11:56.556496 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"control-plane-machine-set-operator-tls" Dec 03 14:11:56 crc kubenswrapper[4805]: I1203 14:11:56.577128 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"control-plane-machine-set-operator-dockercfg-k9rxt" Dec 03 14:11:56 crc kubenswrapper[4805]: I1203 14:11:56.595746 4805 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"openshift-service-ca.crt" Dec 03 14:11:56 crc kubenswrapper[4805]: I1203 14:11:56.616147 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-dockercfg-vw8fw" Dec 03 14:11:56 crc kubenswrapper[4805]: I1203 14:11:56.636961 4805 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-config" Dec 03 14:11:56 crc kubenswrapper[4805]: I1203 14:11:56.655779 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-serving-cert" Dec 03 14:11:56 crc kubenswrapper[4805]: I1203 14:11:56.676451 4805 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"kube-root-ca.crt" Dec 03 14:11:56 crc kubenswrapper[4805]: I1203 14:11:56.697339 4805 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-canary"/"kube-root-ca.crt" Dec 03 14:11:56 crc kubenswrapper[4805]: I1203 14:11:56.698635 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 14:11:56 crc kubenswrapper[4805]: E1203 14:11:56.698905 4805 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-03 14:13:58.698883852 +0000 UTC m=+268.361800835 (durationBeforeRetry 2m2s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 14:11:56 crc kubenswrapper[4805]: I1203 14:11:56.717517 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-canary"/"canary-serving-cert" Dec 03 14:11:56 crc kubenswrapper[4805]: I1203 14:11:56.736329 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-canary"/"default-dockercfg-2llfx" Dec 03 14:11:56 crc kubenswrapper[4805]: I1203 14:11:56.756273 4805 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-canary"/"openshift-service-ca.crt" Dec 03 14:11:56 crc kubenswrapper[4805]: I1203 14:11:56.777647 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"dns-dockercfg-jwfmh" Dec 03 14:11:56 crc kubenswrapper[4805]: I1203 14:11:56.796087 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"dns-default-metrics-tls" Dec 03 14:11:56 crc kubenswrapper[4805]: I1203 14:11:56.800197 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 14:11:56 crc kubenswrapper[4805]: I1203 14:11:56.800291 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 14:11:56 crc kubenswrapper[4805]: I1203 14:11:56.800388 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 14:11:56 crc kubenswrapper[4805]: I1203 14:11:56.800510 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 14:11:56 crc kubenswrapper[4805]: I1203 14:11:56.801500 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 14:11:56 crc kubenswrapper[4805]: I1203 14:11:56.804424 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 14:11:56 crc kubenswrapper[4805]: I1203 14:11:56.805290 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 14:11:56 crc kubenswrapper[4805]: I1203 14:11:56.807612 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 14:11:56 crc kubenswrapper[4805]: I1203 14:11:56.817211 4805 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"dns-default" Dec 03 14:11:56 crc kubenswrapper[4805]: I1203 14:11:56.836627 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-server-tls" Dec 03 14:11:56 crc kubenswrapper[4805]: I1203 14:11:56.856228 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-server-dockercfg-qx5rd" Dec 03 14:11:56 crc kubenswrapper[4805]: I1203 14:11:56.876240 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"node-bootstrapper-token" Dec 03 14:11:56 crc kubenswrapper[4805]: I1203 14:11:56.911504 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4z5cq\" (UniqueName: \"kubernetes.io/projected/e5a33c32-3873-49b8-8a81-c1a0357d94fc-kube-api-access-4z5cq\") pod \"openshift-apiserver-operator-796bbdcf4f-l6qxf\" (UID: \"e5a33c32-3873-49b8-8a81-c1a0357d94fc\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-l6qxf" Dec 03 14:11:56 crc kubenswrapper[4805]: I1203 14:11:56.917539 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 14:11:56 crc kubenswrapper[4805]: I1203 14:11:56.932761 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 14:11:56 crc kubenswrapper[4805]: I1203 14:11:56.935627 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-46hn5\" (UniqueName: \"kubernetes.io/projected/e2034338-8a35-4fd2-9bbf-be733939fc25-kube-api-access-46hn5\") pod \"console-f9d7485db-pp9tv\" (UID: \"e2034338-8a35-4fd2-9bbf-be733939fc25\") " pod="openshift-console/console-f9d7485db-pp9tv" Dec 03 14:11:56 crc kubenswrapper[4805]: I1203 14:11:56.945008 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 14:11:56 crc kubenswrapper[4805]: I1203 14:11:56.954820 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fb99x\" (UniqueName: \"kubernetes.io/projected/33c718fb-affd-410b-9f53-ddb68ba9760c-kube-api-access-fb99x\") pod \"multus-admission-controller-857f4d67dd-tn8tx\" (UID: \"33c718fb-affd-410b-9f53-ddb68ba9760c\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-tn8tx" Dec 03 14:11:56 crc kubenswrapper[4805]: I1203 14:11:56.972545 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4v8xc\" (UniqueName: \"kubernetes.io/projected/ca425595-aae1-44f2-9470-84ba52167748-kube-api-access-4v8xc\") pod \"apiserver-76f77b778f-7nhvl\" (UID: \"ca425595-aae1-44f2-9470-84ba52167748\") " pod="openshift-apiserver/apiserver-76f77b778f-7nhvl" Dec 03 14:11:56 crc kubenswrapper[4805]: I1203 14:11:56.993751 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-n7759\" (UniqueName: \"kubernetes.io/projected/dfa6979a-2473-469b-ac56-9086e35a63f4-kube-api-access-n7759\") pod \"route-controller-manager-6576b87f9c-nsz7s\" (UID: \"dfa6979a-2473-469b-ac56-9086e35a63f4\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-nsz7s" Dec 03 14:11:57 crc kubenswrapper[4805]: I1203 14:11:57.015348 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver/apiserver-76f77b778f-7nhvl" Dec 03 14:11:57 crc kubenswrapper[4805]: I1203 14:11:57.018134 4805 reflector.go:368] Caches populated for *v1.ConfigMap from object-"hostpath-provisioner"/"openshift-service-ca.crt" Dec 03 14:11:57 crc kubenswrapper[4805]: I1203 14:11:57.020503 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sjpnv\" (UniqueName: \"kubernetes.io/projected/f4b1faf2-bf8b-483c-b28f-5885a9e8fe44-kube-api-access-sjpnv\") pod \"openshift-config-operator-7777fb866f-7wzgf\" (UID: \"f4b1faf2-bf8b-483c-b28f-5885a9e8fe44\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-7wzgf" Dec 03 14:11:57 crc kubenswrapper[4805]: I1203 14:11:57.020937 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-admission-controller-857f4d67dd-tn8tx" Dec 03 14:11:57 crc kubenswrapper[4805]: I1203 14:11:57.025051 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-l6qxf" Dec 03 14:11:57 crc kubenswrapper[4805]: I1203 14:11:57.038358 4805 reflector.go:368] Caches populated for *v1.Secret from object-"hostpath-provisioner"/"csi-hostpath-provisioner-sa-dockercfg-qd74k" Dec 03 14:11:57 crc kubenswrapper[4805]: I1203 14:11:57.056402 4805 reflector.go:368] Caches populated for *v1.ConfigMap from object-"hostpath-provisioner"/"kube-root-ca.crt" Dec 03 14:11:57 crc kubenswrapper[4805]: I1203 14:11:57.095650 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2cmpw\" (UniqueName: \"kubernetes.io/projected/0a7d0b69-f9a6-4a40-9981-3dd999c400d5-kube-api-access-2cmpw\") pod \"apiserver-7bbb656c7d-kfh2f\" (UID: \"0a7d0b69-f9a6-4a40-9981-3dd999c400d5\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-kfh2f" Dec 03 14:11:57 crc kubenswrapper[4805]: I1203 14:11:57.112370 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lwvs6\" (UniqueName: \"kubernetes.io/projected/00941575-52e9-482d-b256-1b1c757d96e0-kube-api-access-lwvs6\") pod \"machine-approver-56656f9798-nv2m9\" (UID: \"00941575-52e9-482d-b256-1b1c757d96e0\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-nv2m9" Dec 03 14:11:57 crc kubenswrapper[4805]: I1203 14:11:57.114932 4805 request.go:700] Waited for 1.933994619s due to client-side throttling, not priority and fairness, request: POST:https://api-int.crc.testing:6443/api/v1/namespaces/openshift-service-ca-operator/serviceaccounts/service-ca-operator/token Dec 03 14:11:57 crc kubenswrapper[4805]: I1203 14:11:57.131626 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qzz6l\" (UniqueName: \"kubernetes.io/projected/9b932fa5-f87b-46e1-8987-683cfd3b46fd-kube-api-access-qzz6l\") pod \"service-ca-operator-777779d784-z7lgj\" (UID: \"9b932fa5-f87b-46e1-8987-683cfd3b46fd\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-z7lgj" Dec 03 14:11:57 crc kubenswrapper[4805]: I1203 14:11:57.145410 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-nsz7s" Dec 03 14:11:57 crc kubenswrapper[4805]: I1203 14:11:57.152256 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8zh8j\" (UniqueName: \"kubernetes.io/projected/cdcb6b89-e405-479e-a60d-96681328634e-kube-api-access-8zh8j\") pod \"authentication-operator-69f744f599-bs6tj\" (UID: \"cdcb6b89-e405-479e-a60d-96681328634e\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-bs6tj" Dec 03 14:11:57 crc kubenswrapper[4805]: I1203 14:11:57.174520 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rnn8j\" (UniqueName: \"kubernetes.io/projected/dfd20bfc-5c11-4287-a4b3-df9e0ca4dcf1-kube-api-access-rnn8j\") pod \"machine-api-operator-5694c8668f-jvspg\" (UID: \"dfd20bfc-5c11-4287-a4b3-df9e0ca4dcf1\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-jvspg" Dec 03 14:11:57 crc kubenswrapper[4805]: I1203 14:11:57.190692 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-p5sc5\" (UniqueName: \"kubernetes.io/projected/c6b0a18a-5cf2-430a-a0ef-2c2776839f11-kube-api-access-p5sc5\") pod \"cluster-samples-operator-665b6dd947-vc9sn\" (UID: \"c6b0a18a-5cf2-430a-a0ef-2c2776839f11\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-vc9sn" Dec 03 14:11:57 crc kubenswrapper[4805]: I1203 14:11:57.207127 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-nv2m9" Dec 03 14:11:57 crc kubenswrapper[4805]: I1203 14:11:57.224153 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-x6v8p\" (UniqueName: \"kubernetes.io/projected/9f53fe44-0d2c-4e3d-a682-df3a779a1c8a-kube-api-access-x6v8p\") pod \"console-operator-58897d9998-46w2g\" (UID: \"9f53fe44-0d2c-4e3d-a682-df3a779a1c8a\") " pod="openshift-console-operator/console-operator-58897d9998-46w2g" Dec 03 14:11:57 crc kubenswrapper[4805]: I1203 14:11:57.232145 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-f9d7485db-pp9tv" Dec 03 14:11:57 crc kubenswrapper[4805]: I1203 14:11:57.241994 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6cs72\" (UniqueName: \"kubernetes.io/projected/04bd3442-dbeb-4cb4-bc9f-b07b9872231f-kube-api-access-6cs72\") pod \"packageserver-d55dfcdfc-rr7fx\" (UID: \"04bd3442-dbeb-4cb4-bc9f-b07b9872231f\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-rr7fx" Dec 03 14:11:57 crc kubenswrapper[4805]: I1203 14:11:57.242705 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-config-operator/openshift-config-operator-7777fb866f-7wzgf" Dec 03 14:11:57 crc kubenswrapper[4805]: I1203 14:11:57.250121 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication-operator/authentication-operator-69f744f599-bs6tj" Dec 03 14:11:57 crc kubenswrapper[4805]: I1203 14:11:57.261822 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-rr7fx" Dec 03 14:11:57 crc kubenswrapper[4805]: I1203 14:11:57.263870 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pgzx9\" (UniqueName: \"kubernetes.io/projected/6c2d6fbc-771d-4aff-b1e8-3eda17204931-kube-api-access-pgzx9\") pod \"downloads-7954f5f757-s25zl\" (UID: \"6c2d6fbc-771d-4aff-b1e8-3eda17204931\") " pod="openshift-console/downloads-7954f5f757-s25zl" Dec 03 14:11:57 crc kubenswrapper[4805]: I1203 14:11:57.278303 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-89vsl\" (UniqueName: \"kubernetes.io/projected/b54908df-f04f-4515-ab36-842f0aef4a99-kube-api-access-89vsl\") pod \"controller-manager-879f6c89f-k7rgc\" (UID: \"b54908df-f04f-4515-ab36-842f0aef4a99\") " pod="openshift-controller-manager/controller-manager-879f6c89f-k7rgc" Dec 03 14:11:57 crc kubenswrapper[4805]: I1203 14:11:57.292476 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8gr5q\" (UniqueName: \"kubernetes.io/projected/d5ea0764-4290-4eab-b4ce-dbd4909455c1-kube-api-access-8gr5q\") pod \"machine-config-operator-74547568cd-kvmpp\" (UID: \"d5ea0764-4290-4eab-b4ce-dbd4909455c1\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-kvmpp" Dec 03 14:11:57 crc kubenswrapper[4805]: I1203 14:11:57.314311 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dxgkw\" (UniqueName: \"kubernetes.io/projected/1658d8e1-ac63-40fe-98ce-900a9f9aeae5-kube-api-access-dxgkw\") pod \"oauth-openshift-558db77b4-ntv6j\" (UID: \"1658d8e1-ac63-40fe-98ce-900a9f9aeae5\") " pod="openshift-authentication/oauth-openshift-558db77b4-ntv6j" Dec 03 14:11:57 crc kubenswrapper[4805]: I1203 14:11:57.330772 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca-operator/service-ca-operator-777779d784-z7lgj" Dec 03 14:11:57 crc kubenswrapper[4805]: I1203 14:11:57.336636 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qpdfz\" (UniqueName: \"kubernetes.io/projected/da68f128-1d8d-4256-a28d-a174766e1331-kube-api-access-qpdfz\") pod \"cluster-image-registry-operator-dc59b4c8b-jvsqf\" (UID: \"da68f128-1d8d-4256-a28d-a174766e1331\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-jvsqf" Dec 03 14:11:57 crc kubenswrapper[4805]: I1203 14:11:57.353688 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/da68f128-1d8d-4256-a28d-a174766e1331-bound-sa-token\") pod \"cluster-image-registry-operator-dc59b4c8b-jvsqf\" (UID: \"da68f128-1d8d-4256-a28d-a174766e1331\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-jvsqf" Dec 03 14:11:57 crc kubenswrapper[4805]: I1203 14:11:57.364106 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-k7rgc" Dec 03 14:11:57 crc kubenswrapper[4805]: I1203 14:11:57.364918 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-kfh2f" Dec 03 14:11:57 crc kubenswrapper[4805]: I1203 14:11:57.373306 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-nsz7s"] Dec 03 14:11:57 crc kubenswrapper[4805]: I1203 14:11:57.376510 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ddtf4\" (UniqueName: \"kubernetes.io/projected/ae561e71-047a-4cf7-8378-d08adb458d9c-kube-api-access-ddtf4\") pod \"etcd-operator-b45778765-2x56c\" (UID: \"ae561e71-047a-4cf7-8378-d08adb458d9c\") " pod="openshift-etcd-operator/etcd-operator-b45778765-2x56c" Dec 03 14:11:57 crc kubenswrapper[4805]: I1203 14:11:57.394299 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4djzx\" (UniqueName: \"kubernetes.io/projected/1c55f37c-24b9-4800-9050-e0da883df2c0-kube-api-access-4djzx\") pod \"service-ca-9c57cc56f-dbfrd\" (UID: \"1c55f37c-24b9-4800-9050-e0da883df2c0\") " pod="openshift-service-ca/service-ca-9c57cc56f-dbfrd" Dec 03 14:11:57 crc kubenswrapper[4805]: I1203 14:11:57.413489 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-v8t8r\" (UniqueName: \"kubernetes.io/projected/8e437287-3b92-4c6a-abe9-84c786ba3220-kube-api-access-v8t8r\") pod \"dns-operator-744455d44c-rjf9m\" (UID: \"8e437287-3b92-4c6a-abe9-84c786ba3220\") " pod="openshift-dns-operator/dns-operator-744455d44c-rjf9m" Dec 03 14:11:57 crc kubenswrapper[4805]: I1203 14:11:57.416655 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-jvsqf" Dec 03 14:11:57 crc kubenswrapper[4805]: I1203 14:11:57.419183 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" event={"ID":"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8","Type":"ContainerStarted","Data":"36ba8ee06981b907aa3e6a2e2933f0292afea76d35c230e5ca36421c70545baf"} Dec 03 14:11:57 crc kubenswrapper[4805]: I1203 14:11:57.427620 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" event={"ID":"3b6479f0-333b-4a96-9adf-2099afdc2447","Type":"ContainerStarted","Data":"b6bfa93e30fa35f2dfc2bdfa2933df357bfd1d9471715f0fe1c124a2e7cd19c6"} Dec 03 14:11:57 crc kubenswrapper[4805]: I1203 14:11:57.431462 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-csvjg\" (UniqueName: \"kubernetes.io/projected/e600aefb-0839-423f-8ac8-1916400cfa92-kube-api-access-csvjg\") pod \"machine-config-controller-84d6567774-t4kl8\" (UID: \"e600aefb-0839-423f-8ac8-1916400cfa92\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-t4kl8" Dec 03 14:11:57 crc kubenswrapper[4805]: I1203 14:11:57.431954 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-nv2m9" event={"ID":"00941575-52e9-482d-b256-1b1c757d96e0","Type":"ContainerStarted","Data":"a7a5b1903f15cc4b6efc88cea0ec4da61ea75a45de3b587e146b0d1795ee8e56"} Dec 03 14:11:57 crc kubenswrapper[4805]: I1203 14:11:57.441947 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" event={"ID":"9d751cbb-f2e2-430d-9754-c882a5e924a5","Type":"ContainerStarted","Data":"51ce4ea1461cac22545ee40744737fe7dab293ee1dd335495c24373e660b8dbe"} Dec 03 14:11:57 crc kubenswrapper[4805]: I1203 14:11:57.453178 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/machine-api-operator-5694c8668f-jvspg" Dec 03 14:11:57 crc kubenswrapper[4805]: I1203 14:11:57.454628 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7frlc\" (UniqueName: \"kubernetes.io/projected/8b954d96-a62c-434e-a340-8446d778c545-kube-api-access-7frlc\") pod \"collect-profiles-29412840-gj7pr\" (UID: \"8b954d96-a62c-434e-a340-8446d778c545\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29412840-gj7pr" Dec 03 14:11:57 crc kubenswrapper[4805]: I1203 14:11:57.477827 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/e9022931-39f7-45cf-925f-fe19e1b870b8-kube-api-access\") pod \"kube-apiserver-operator-766d6c64bb-2vnrk\" (UID: \"e9022931-39f7-45cf-925f-fe19e1b870b8\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-2vnrk" Dec 03 14:11:57 crc kubenswrapper[4805]: I1203 14:11:57.478010 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-vc9sn" Dec 03 14:11:57 crc kubenswrapper[4805]: I1203 14:11:57.490779 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-ntv6j" Dec 03 14:11:57 crc kubenswrapper[4805]: I1203 14:11:57.500774 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/downloads-7954f5f757-s25zl" Dec 03 14:11:57 crc kubenswrapper[4805]: I1203 14:11:57.507761 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver/apiserver-76f77b778f-7nhvl"] Dec 03 14:11:57 crc kubenswrapper[4805]: I1203 14:11:57.507926 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/66812e7e-aa85-4045-8e7a-8967b84f8849-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-spjlp\" (UID: \"66812e7e-aa85-4045-8e7a-8967b84f8849\") " pod="openshift-marketplace/marketplace-operator-79b997595-spjlp" Dec 03 14:11:57 crc kubenswrapper[4805]: I1203 14:11:57.508007 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/4655a83e-4980-4a5c-9f61-e305ae418822-ca-trust-extracted\") pod \"image-registry-697d97f7c8-jjzct\" (UID: \"4655a83e-4980-4a5c-9f61-e305ae418822\") " pod="openshift-image-registry/image-registry-697d97f7c8-jjzct" Dec 03 14:11:57 crc kubenswrapper[4805]: I1203 14:11:57.508097 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/4655a83e-4980-4a5c-9f61-e305ae418822-registry-tls\") pod \"image-registry-697d97f7c8-jjzct\" (UID: \"4655a83e-4980-4a5c-9f61-e305ae418822\") " pod="openshift-image-registry/image-registry-697d97f7c8-jjzct" Dec 03 14:11:57 crc kubenswrapper[4805]: I1203 14:11:57.508160 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-jjzct\" (UID: \"4655a83e-4980-4a5c-9f61-e305ae418822\") " pod="openshift-image-registry/image-registry-697d97f7c8-jjzct" Dec 03 14:11:57 crc kubenswrapper[4805]: I1203 14:11:57.508206 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/66812e7e-aa85-4045-8e7a-8967b84f8849-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-spjlp\" (UID: \"66812e7e-aa85-4045-8e7a-8967b84f8849\") " pod="openshift-marketplace/marketplace-operator-79b997595-spjlp" Dec 03 14:11:57 crc kubenswrapper[4805]: I1203 14:11:57.508237 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/4655a83e-4980-4a5c-9f61-e305ae418822-installation-pull-secrets\") pod \"image-registry-697d97f7c8-jjzct\" (UID: \"4655a83e-4980-4a5c-9f61-e305ae418822\") " pod="openshift-image-registry/image-registry-697d97f7c8-jjzct" Dec 03 14:11:57 crc kubenswrapper[4805]: I1203 14:11:57.508262 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-n5f7x\" (UniqueName: \"kubernetes.io/projected/66812e7e-aa85-4045-8e7a-8967b84f8849-kube-api-access-n5f7x\") pod \"marketplace-operator-79b997595-spjlp\" (UID: \"66812e7e-aa85-4045-8e7a-8967b84f8849\") " pod="openshift-marketplace/marketplace-operator-79b997595-spjlp" Dec 03 14:11:57 crc kubenswrapper[4805]: I1203 14:11:57.508324 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/4655a83e-4980-4a5c-9f61-e305ae418822-trusted-ca\") pod \"image-registry-697d97f7c8-jjzct\" (UID: \"4655a83e-4980-4a5c-9f61-e305ae418822\") " pod="openshift-image-registry/image-registry-697d97f7c8-jjzct" Dec 03 14:11:57 crc kubenswrapper[4805]: I1203 14:11:57.508359 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-m7vtp\" (UniqueName: \"kubernetes.io/projected/4655a83e-4980-4a5c-9f61-e305ae418822-kube-api-access-m7vtp\") pod \"image-registry-697d97f7c8-jjzct\" (UID: \"4655a83e-4980-4a5c-9f61-e305ae418822\") " pod="openshift-image-registry/image-registry-697d97f7c8-jjzct" Dec 03 14:11:57 crc kubenswrapper[4805]: I1203 14:11:57.508373 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/4655a83e-4980-4a5c-9f61-e305ae418822-registry-certificates\") pod \"image-registry-697d97f7c8-jjzct\" (UID: \"4655a83e-4980-4a5c-9f61-e305ae418822\") " pod="openshift-image-registry/image-registry-697d97f7c8-jjzct" Dec 03 14:11:57 crc kubenswrapper[4805]: I1203 14:11:57.508413 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/4655a83e-4980-4a5c-9f61-e305ae418822-bound-sa-token\") pod \"image-registry-697d97f7c8-jjzct\" (UID: \"4655a83e-4980-4a5c-9f61-e305ae418822\") " pod="openshift-image-registry/image-registry-697d97f7c8-jjzct" Dec 03 14:11:57 crc kubenswrapper[4805]: E1203 14:11:57.513860 4805 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-03 14:11:58.013824412 +0000 UTC m=+147.676741335 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-jjzct" (UID: "4655a83e-4980-4a5c-9f61-e305ae418822") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 14:11:57 crc kubenswrapper[4805]: I1203 14:11:57.525235 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-l6qxf"] Dec 03 14:11:57 crc kubenswrapper[4805]: I1203 14:11:57.535748 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console-operator/console-operator-58897d9998-46w2g" Dec 03 14:11:57 crc kubenswrapper[4805]: I1203 14:11:57.557341 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/multus-admission-controller-857f4d67dd-tn8tx"] Dec 03 14:11:57 crc kubenswrapper[4805]: I1203 14:11:57.566012 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-kvmpp" Dec 03 14:11:57 crc kubenswrapper[4805]: W1203 14:11:57.568633 4805 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podca425595_aae1_44f2_9470_84ba52167748.slice/crio-0b4ecae8312cfa07d802b90c3709b70903f090fa27bffa48228a136f3698db9b WatchSource:0}: Error finding container 0b4ecae8312cfa07d802b90c3709b70903f090fa27bffa48228a136f3698db9b: Status 404 returned error can't find the container with id 0b4ecae8312cfa07d802b90c3709b70903f090fa27bffa48228a136f3698db9b Dec 03 14:11:57 crc kubenswrapper[4805]: I1203 14:11:57.608863 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 14:11:57 crc kubenswrapper[4805]: I1203 14:11:57.609600 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/b22cbc60-5975-42a4-85d3-6add6917fd88-service-ca-bundle\") pod \"router-default-5444994796-kxnqm\" (UID: \"b22cbc60-5975-42a4-85d3-6add6917fd88\") " pod="openshift-ingress/router-default-5444994796-kxnqm" Dec 03 14:11:57 crc kubenswrapper[4805]: I1203 14:11:57.610186 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registration-dir\" (UniqueName: \"kubernetes.io/host-path/8801ff25-d273-4354-ab39-16e979d0b73d-registration-dir\") pod \"csi-hostpathplugin-krqsw\" (UID: \"8801ff25-d273-4354-ab39-16e979d0b73d\") " pod="hostpath-provisioner/csi-hostpathplugin-krqsw" Dec 03 14:11:57 crc kubenswrapper[4805]: I1203 14:11:57.610331 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/66812e7e-aa85-4045-8e7a-8967b84f8849-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-spjlp\" (UID: \"66812e7e-aa85-4045-8e7a-8967b84f8849\") " pod="openshift-marketplace/marketplace-operator-79b997595-spjlp" Dec 03 14:11:57 crc kubenswrapper[4805]: E1203 14:11:57.610412 4805 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-03 14:11:58.110389027 +0000 UTC m=+147.773306040 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 14:11:57 crc kubenswrapper[4805]: I1203 14:11:57.611100 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/4655a83e-4980-4a5c-9f61-e305ae418822-installation-pull-secrets\") pod \"image-registry-697d97f7c8-jjzct\" (UID: \"4655a83e-4980-4a5c-9f61-e305ae418822\") " pod="openshift-image-registry/image-registry-697d97f7c8-jjzct" Dec 03 14:11:57 crc kubenswrapper[4805]: I1203 14:11:57.612409 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tlt9k\" (UniqueName: \"kubernetes.io/projected/6be05973-dc60-437f-959c-a9309cd16d5b-kube-api-access-tlt9k\") pod \"ingress-canary-cnlwr\" (UID: \"6be05973-dc60-437f-959c-a9309cd16d5b\") " pod="openshift-ingress-canary/ingress-canary-cnlwr" Dec 03 14:11:57 crc kubenswrapper[4805]: I1203 14:11:57.612478 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/6be05973-dc60-437f-959c-a9309cd16d5b-cert\") pod \"ingress-canary-cnlwr\" (UID: \"6be05973-dc60-437f-959c-a9309cd16d5b\") " pod="openshift-ingress-canary/ingress-canary-cnlwr" Dec 03 14:11:57 crc kubenswrapper[4805]: I1203 14:11:57.612582 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-n5f7x\" (UniqueName: \"kubernetes.io/projected/66812e7e-aa85-4045-8e7a-8967b84f8849-kube-api-access-n5f7x\") pod \"marketplace-operator-79b997595-spjlp\" (UID: \"66812e7e-aa85-4045-8e7a-8967b84f8849\") " pod="openshift-marketplace/marketplace-operator-79b997595-spjlp" Dec 03 14:11:57 crc kubenswrapper[4805]: I1203 14:11:57.613190 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/776b097e-ad39-49a4-aa4f-a326a04bc30b-profile-collector-cert\") pod \"olm-operator-6b444d44fb-ss8kj\" (UID: \"776b097e-ad39-49a4-aa4f-a326a04bc30b\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-ss8kj" Dec 03 14:11:57 crc kubenswrapper[4805]: I1203 14:11:57.613230 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-j686r\" (UniqueName: \"kubernetes.io/projected/776b097e-ad39-49a4-aa4f-a326a04bc30b-kube-api-access-j686r\") pod \"olm-operator-6b444d44fb-ss8kj\" (UID: \"776b097e-ad39-49a4-aa4f-a326a04bc30b\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-ss8kj" Dec 03 14:11:57 crc kubenswrapper[4805]: I1203 14:11:57.613259 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/4655a83e-4980-4a5c-9f61-e305ae418822-trusted-ca\") pod \"image-registry-697d97f7c8-jjzct\" (UID: \"4655a83e-4980-4a5c-9f61-e305ae418822\") " pod="openshift-image-registry/image-registry-697d97f7c8-jjzct" Dec 03 14:11:57 crc kubenswrapper[4805]: I1203 14:11:57.613300 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/b22cbc60-5975-42a4-85d3-6add6917fd88-stats-auth\") pod \"router-default-5444994796-kxnqm\" (UID: \"b22cbc60-5975-42a4-85d3-6add6917fd88\") " pod="openshift-ingress/router-default-5444994796-kxnqm" Dec 03 14:11:57 crc kubenswrapper[4805]: I1203 14:11:57.613357 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-m7vtp\" (UniqueName: \"kubernetes.io/projected/4655a83e-4980-4a5c-9f61-e305ae418822-kube-api-access-m7vtp\") pod \"image-registry-697d97f7c8-jjzct\" (UID: \"4655a83e-4980-4a5c-9f61-e305ae418822\") " pod="openshift-image-registry/image-registry-697d97f7c8-jjzct" Dec 03 14:11:57 crc kubenswrapper[4805]: I1203 14:11:57.613396 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/4655a83e-4980-4a5c-9f61-e305ae418822-registry-certificates\") pod \"image-registry-697d97f7c8-jjzct\" (UID: \"4655a83e-4980-4a5c-9f61-e305ae418822\") " pod="openshift-image-registry/image-registry-697d97f7c8-jjzct" Dec 03 14:11:57 crc kubenswrapper[4805]: I1203 14:11:57.613523 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/4655a83e-4980-4a5c-9f61-e305ae418822-bound-sa-token\") pod \"image-registry-697d97f7c8-jjzct\" (UID: \"4655a83e-4980-4a5c-9f61-e305ae418822\") " pod="openshift-image-registry/image-registry-697d97f7c8-jjzct" Dec 03 14:11:57 crc kubenswrapper[4805]: I1203 14:11:57.613555 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/66812e7e-aa85-4045-8e7a-8967b84f8849-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-spjlp\" (UID: \"66812e7e-aa85-4045-8e7a-8967b84f8849\") " pod="openshift-marketplace/marketplace-operator-79b997595-spjlp" Dec 03 14:11:57 crc kubenswrapper[4805]: I1203 14:11:57.613626 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"csi-data-dir\" (UniqueName: \"kubernetes.io/host-path/8801ff25-d273-4354-ab39-16e979d0b73d-csi-data-dir\") pod \"csi-hostpathplugin-krqsw\" (UID: \"8801ff25-d273-4354-ab39-16e979d0b73d\") " pod="hostpath-provisioner/csi-hostpathplugin-krqsw" Dec 03 14:11:57 crc kubenswrapper[4805]: I1203 14:11:57.613684 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/776b097e-ad39-49a4-aa4f-a326a04bc30b-srv-cert\") pod \"olm-operator-6b444d44fb-ss8kj\" (UID: \"776b097e-ad39-49a4-aa4f-a326a04bc30b\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-ss8kj" Dec 03 14:11:57 crc kubenswrapper[4805]: I1203 14:11:57.613719 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/b22cbc60-5975-42a4-85d3-6add6917fd88-default-certificate\") pod \"router-default-5444994796-kxnqm\" (UID: \"b22cbc60-5975-42a4-85d3-6add6917fd88\") " pod="openshift-ingress/router-default-5444994796-kxnqm" Dec 03 14:11:57 crc kubenswrapper[4805]: I1203 14:11:57.613744 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/4655a83e-4980-4a5c-9f61-e305ae418822-ca-trust-extracted\") pod \"image-registry-697d97f7c8-jjzct\" (UID: \"4655a83e-4980-4a5c-9f61-e305ae418822\") " pod="openshift-image-registry/image-registry-697d97f7c8-jjzct" Dec 03 14:11:57 crc kubenswrapper[4805]: I1203 14:11:57.613761 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/b22cbc60-5975-42a4-85d3-6add6917fd88-metrics-certs\") pod \"router-default-5444994796-kxnqm\" (UID: \"b22cbc60-5975-42a4-85d3-6add6917fd88\") " pod="openshift-ingress/router-default-5444994796-kxnqm" Dec 03 14:11:57 crc kubenswrapper[4805]: I1203 14:11:57.613787 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"mountpoint-dir\" (UniqueName: \"kubernetes.io/host-path/8801ff25-d273-4354-ab39-16e979d0b73d-mountpoint-dir\") pod \"csi-hostpathplugin-krqsw\" (UID: \"8801ff25-d273-4354-ab39-16e979d0b73d\") " pod="hostpath-provisioner/csi-hostpathplugin-krqsw" Dec 03 14:11:57 crc kubenswrapper[4805]: I1203 14:11:57.613817 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"socket-dir\" (UniqueName: \"kubernetes.io/host-path/8801ff25-d273-4354-ab39-16e979d0b73d-socket-dir\") pod \"csi-hostpathplugin-krqsw\" (UID: \"8801ff25-d273-4354-ab39-16e979d0b73d\") " pod="hostpath-provisioner/csi-hostpathplugin-krqsw" Dec 03 14:11:57 crc kubenswrapper[4805]: I1203 14:11:57.613956 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5cec10a3-ed60-4193-be30-a78cb80bd51d-config\") pod \"openshift-controller-manager-operator-756b6f6bc6-94zfm\" (UID: \"5cec10a3-ed60-4193-be30-a78cb80bd51d\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-94zfm" Dec 03 14:11:57 crc kubenswrapper[4805]: I1203 14:11:57.614070 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gl58d\" (UniqueName: \"kubernetes.io/projected/8801ff25-d273-4354-ab39-16e979d0b73d-kube-api-access-gl58d\") pod \"csi-hostpathplugin-krqsw\" (UID: \"8801ff25-d273-4354-ab39-16e979d0b73d\") " pod="hostpath-provisioner/csi-hostpathplugin-krqsw" Dec 03 14:11:57 crc kubenswrapper[4805]: I1203 14:11:57.614113 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-dir\" (UniqueName: \"kubernetes.io/host-path/8801ff25-d273-4354-ab39-16e979d0b73d-plugins-dir\") pod \"csi-hostpathplugin-krqsw\" (UID: \"8801ff25-d273-4354-ab39-16e979d0b73d\") " pod="hostpath-provisioner/csi-hostpathplugin-krqsw" Dec 03 14:11:57 crc kubenswrapper[4805]: I1203 14:11:57.614133 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/4655a83e-4980-4a5c-9f61-e305ae418822-registry-tls\") pod \"image-registry-697d97f7c8-jjzct\" (UID: \"4655a83e-4980-4a5c-9f61-e305ae418822\") " pod="openshift-image-registry/image-registry-697d97f7c8-jjzct" Dec 03 14:11:57 crc kubenswrapper[4805]: I1203 14:11:57.614171 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vwppd\" (UniqueName: \"kubernetes.io/projected/b22cbc60-5975-42a4-85d3-6add6917fd88-kube-api-access-vwppd\") pod \"router-default-5444994796-kxnqm\" (UID: \"b22cbc60-5975-42a4-85d3-6add6917fd88\") " pod="openshift-ingress/router-default-5444994796-kxnqm" Dec 03 14:11:57 crc kubenswrapper[4805]: I1203 14:11:57.614344 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-jjzct\" (UID: \"4655a83e-4980-4a5c-9f61-e305ae418822\") " pod="openshift-image-registry/image-registry-697d97f7c8-jjzct" Dec 03 14:11:57 crc kubenswrapper[4805]: I1203 14:11:57.614471 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5cec10a3-ed60-4193-be30-a78cb80bd51d-serving-cert\") pod \"openshift-controller-manager-operator-756b6f6bc6-94zfm\" (UID: \"5cec10a3-ed60-4193-be30-a78cb80bd51d\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-94zfm" Dec 03 14:11:57 crc kubenswrapper[4805]: I1203 14:11:57.614624 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fz449\" (UniqueName: \"kubernetes.io/projected/5cec10a3-ed60-4193-be30-a78cb80bd51d-kube-api-access-fz449\") pod \"openshift-controller-manager-operator-756b6f6bc6-94zfm\" (UID: \"5cec10a3-ed60-4193-be30-a78cb80bd51d\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-94zfm" Dec 03 14:11:57 crc kubenswrapper[4805]: I1203 14:11:57.617914 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/4655a83e-4980-4a5c-9f61-e305ae418822-trusted-ca\") pod \"image-registry-697d97f7c8-jjzct\" (UID: \"4655a83e-4980-4a5c-9f61-e305ae418822\") " pod="openshift-image-registry/image-registry-697d97f7c8-jjzct" Dec 03 14:11:57 crc kubenswrapper[4805]: I1203 14:11:57.618712 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/4655a83e-4980-4a5c-9f61-e305ae418822-ca-trust-extracted\") pod \"image-registry-697d97f7c8-jjzct\" (UID: \"4655a83e-4980-4a5c-9f61-e305ae418822\") " pod="openshift-image-registry/image-registry-697d97f7c8-jjzct" Dec 03 14:11:57 crc kubenswrapper[4805]: E1203 14:11:57.618817 4805 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-03 14:11:58.118803887 +0000 UTC m=+147.781720810 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-jjzct" (UID: "4655a83e-4980-4a5c-9f61-e305ae418822") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 14:11:57 crc kubenswrapper[4805]: I1203 14:11:57.619323 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/66812e7e-aa85-4045-8e7a-8967b84f8849-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-spjlp\" (UID: \"66812e7e-aa85-4045-8e7a-8967b84f8849\") " pod="openshift-marketplace/marketplace-operator-79b997595-spjlp" Dec 03 14:11:57 crc kubenswrapper[4805]: I1203 14:11:57.620096 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/4655a83e-4980-4a5c-9f61-e305ae418822-registry-certificates\") pod \"image-registry-697d97f7c8-jjzct\" (UID: \"4655a83e-4980-4a5c-9f61-e305ae418822\") " pod="openshift-image-registry/image-registry-697d97f7c8-jjzct" Dec 03 14:11:57 crc kubenswrapper[4805]: W1203 14:11:57.620735 4805 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pode5a33c32_3873_49b8_8a81_c1a0357d94fc.slice/crio-899c06fc674e60b644ad74b97902e673391dad8df3ff3dfe810bd3c7b43168f3 WatchSource:0}: Error finding container 899c06fc674e60b644ad74b97902e673391dad8df3ff3dfe810bd3c7b43168f3: Status 404 returned error can't find the container with id 899c06fc674e60b644ad74b97902e673391dad8df3ff3dfe810bd3c7b43168f3 Dec 03 14:11:57 crc kubenswrapper[4805]: I1203 14:11:57.622795 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/66812e7e-aa85-4045-8e7a-8967b84f8849-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-spjlp\" (UID: \"66812e7e-aa85-4045-8e7a-8967b84f8849\") " pod="openshift-marketplace/marketplace-operator-79b997595-spjlp" Dec 03 14:11:57 crc kubenswrapper[4805]: I1203 14:11:57.624249 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/4655a83e-4980-4a5c-9f61-e305ae418822-installation-pull-secrets\") pod \"image-registry-697d97f7c8-jjzct\" (UID: \"4655a83e-4980-4a5c-9f61-e305ae418822\") " pod="openshift-image-registry/image-registry-697d97f7c8-jjzct" Dec 03 14:11:57 crc kubenswrapper[4805]: I1203 14:11:57.635364 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/4655a83e-4980-4a5c-9f61-e305ae418822-registry-tls\") pod \"image-registry-697d97f7c8-jjzct\" (UID: \"4655a83e-4980-4a5c-9f61-e305ae418822\") " pod="openshift-image-registry/image-registry-697d97f7c8-jjzct" Dec 03 14:11:57 crc kubenswrapper[4805]: I1203 14:11:57.640063 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-2vnrk" Dec 03 14:11:57 crc kubenswrapper[4805]: I1203 14:11:57.645533 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29412840-gj7pr" Dec 03 14:11:57 crc kubenswrapper[4805]: I1203 14:11:57.651898 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca/service-ca-9c57cc56f-dbfrd" Dec 03 14:11:57 crc kubenswrapper[4805]: I1203 14:11:57.655594 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-n5f7x\" (UniqueName: \"kubernetes.io/projected/66812e7e-aa85-4045-8e7a-8967b84f8849-kube-api-access-n5f7x\") pod \"marketplace-operator-79b997595-spjlp\" (UID: \"66812e7e-aa85-4045-8e7a-8967b84f8849\") " pod="openshift-marketplace/marketplace-operator-79b997595-spjlp" Dec 03 14:11:57 crc kubenswrapper[4805]: I1203 14:11:57.658975 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd-operator/etcd-operator-b45778765-2x56c" Dec 03 14:11:57 crc kubenswrapper[4805]: I1203 14:11:57.667573 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns-operator/dns-operator-744455d44c-rjf9m" Dec 03 14:11:57 crc kubenswrapper[4805]: I1203 14:11:57.674196 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-spjlp" Dec 03 14:11:57 crc kubenswrapper[4805]: I1203 14:11:57.676105 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/4655a83e-4980-4a5c-9f61-e305ae418822-bound-sa-token\") pod \"image-registry-697d97f7c8-jjzct\" (UID: \"4655a83e-4980-4a5c-9f61-e305ae418822\") " pod="openshift-image-registry/image-registry-697d97f7c8-jjzct" Dec 03 14:11:57 crc kubenswrapper[4805]: I1203 14:11:57.696595 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-t4kl8" Dec 03 14:11:57 crc kubenswrapper[4805]: I1203 14:11:57.699994 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-m7vtp\" (UniqueName: \"kubernetes.io/projected/4655a83e-4980-4a5c-9f61-e305ae418822-kube-api-access-m7vtp\") pod \"image-registry-697d97f7c8-jjzct\" (UID: \"4655a83e-4980-4a5c-9f61-e305ae418822\") " pod="openshift-image-registry/image-registry-697d97f7c8-jjzct" Dec 03 14:11:57 crc kubenswrapper[4805]: I1203 14:11:57.718629 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 14:11:57 crc kubenswrapper[4805]: I1203 14:11:57.718738 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"socket-dir\" (UniqueName: \"kubernetes.io/host-path/8801ff25-d273-4354-ab39-16e979d0b73d-socket-dir\") pod \"csi-hostpathplugin-krqsw\" (UID: \"8801ff25-d273-4354-ab39-16e979d0b73d\") " pod="hostpath-provisioner/csi-hostpathplugin-krqsw" Dec 03 14:11:57 crc kubenswrapper[4805]: I1203 14:11:57.718763 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6107a436-ce5e-4fde-890b-a5ac9f2ada10-config\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-86589\" (UID: \"6107a436-ce5e-4fde-890b-a5ac9f2ada10\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-86589" Dec 03 14:11:57 crc kubenswrapper[4805]: I1203 14:11:57.718779 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-22kbd\" (UniqueName: \"kubernetes.io/projected/b44fb2fa-6155-40fc-b5f7-eb5d1fa391bb-kube-api-access-22kbd\") pod \"package-server-manager-789f6589d5-b7wrl\" (UID: \"b44fb2fa-6155-40fc-b5f7-eb5d1fa391bb\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-b7wrl" Dec 03 14:11:57 crc kubenswrapper[4805]: I1203 14:11:57.718793 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/b1313b26-f0b6-462b-8210-c19b0d4065a4-config-volume\") pod \"dns-default-7wnwf\" (UID: \"b1313b26-f0b6-462b-8210-c19b0d4065a4\") " pod="openshift-dns/dns-default-7wnwf" Dec 03 14:11:57 crc kubenswrapper[4805]: I1203 14:11:57.718813 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5cec10a3-ed60-4193-be30-a78cb80bd51d-config\") pod \"openshift-controller-manager-operator-756b6f6bc6-94zfm\" (UID: \"5cec10a3-ed60-4193-be30-a78cb80bd51d\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-94zfm" Dec 03 14:11:57 crc kubenswrapper[4805]: I1203 14:11:57.718871 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gl58d\" (UniqueName: \"kubernetes.io/projected/8801ff25-d273-4354-ab39-16e979d0b73d-kube-api-access-gl58d\") pod \"csi-hostpathplugin-krqsw\" (UID: \"8801ff25-d273-4354-ab39-16e979d0b73d\") " pod="hostpath-provisioner/csi-hostpathplugin-krqsw" Dec 03 14:11:57 crc kubenswrapper[4805]: I1203 14:11:57.718888 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/b1313b26-f0b6-462b-8210-c19b0d4065a4-metrics-tls\") pod \"dns-default-7wnwf\" (UID: \"b1313b26-f0b6-462b-8210-c19b0d4065a4\") " pod="openshift-dns/dns-default-7wnwf" Dec 03 14:11:57 crc kubenswrapper[4805]: I1203 14:11:57.718905 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-r7pr5\" (UniqueName: \"kubernetes.io/projected/9bd2f61c-2b92-488f-b632-1e627b48a518-kube-api-access-r7pr5\") pod \"control-plane-machine-set-operator-78cbb6b69f-25khc\" (UID: \"9bd2f61c-2b92-488f-b632-1e627b48a518\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-25khc" Dec 03 14:11:57 crc kubenswrapper[4805]: I1203 14:11:57.718922 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f71c6d49-bfbf-4c57-a4f8-5cd22d1038cb-config\") pod \"kube-storage-version-migrator-operator-b67b599dd-g46rj\" (UID: \"f71c6d49-bfbf-4c57-a4f8-5cd22d1038cb\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-g46rj" Dec 03 14:11:57 crc kubenswrapper[4805]: I1203 14:11:57.718942 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-dir\" (UniqueName: \"kubernetes.io/host-path/8801ff25-d273-4354-ab39-16e979d0b73d-plugins-dir\") pod \"csi-hostpathplugin-krqsw\" (UID: \"8801ff25-d273-4354-ab39-16e979d0b73d\") " pod="hostpath-provisioner/csi-hostpathplugin-krqsw" Dec 03 14:11:57 crc kubenswrapper[4805]: I1203 14:11:57.718958 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/f71c6d49-bfbf-4c57-a4f8-5cd22d1038cb-serving-cert\") pod \"kube-storage-version-migrator-operator-b67b599dd-g46rj\" (UID: \"f71c6d49-bfbf-4c57-a4f8-5cd22d1038cb\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-g46rj" Dec 03 14:11:57 crc kubenswrapper[4805]: I1203 14:11:57.718980 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/220a8b3d-e5fa-486f-ad11-4fb9a43ab092-trusted-ca\") pod \"ingress-operator-5b745b69d9-j8pwh\" (UID: \"220a8b3d-e5fa-486f-ad11-4fb9a43ab092\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-j8pwh" Dec 03 14:11:57 crc kubenswrapper[4805]: I1203 14:11:57.718994 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vwppd\" (UniqueName: \"kubernetes.io/projected/b22cbc60-5975-42a4-85d3-6add6917fd88-kube-api-access-vwppd\") pod \"router-default-5444994796-kxnqm\" (UID: \"b22cbc60-5975-42a4-85d3-6add6917fd88\") " pod="openshift-ingress/router-default-5444994796-kxnqm" Dec 03 14:11:57 crc kubenswrapper[4805]: I1203 14:11:57.719010 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/6107a436-ce5e-4fde-890b-a5ac9f2ada10-kube-api-access\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-86589\" (UID: \"6107a436-ce5e-4fde-890b-a5ac9f2ada10\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-86589" Dec 03 14:11:57 crc kubenswrapper[4805]: I1203 14:11:57.719034 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/27ffe4cd-8ab3-49dd-8e5d-91b7d2439a27-node-bootstrap-token\") pod \"machine-config-server-d5pjl\" (UID: \"27ffe4cd-8ab3-49dd-8e5d-91b7d2439a27\") " pod="openshift-machine-config-operator/machine-config-server-d5pjl" Dec 03 14:11:57 crc kubenswrapper[4805]: I1203 14:11:57.719050 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5cec10a3-ed60-4193-be30-a78cb80bd51d-serving-cert\") pod \"openshift-controller-manager-operator-756b6f6bc6-94zfm\" (UID: \"5cec10a3-ed60-4193-be30-a78cb80bd51d\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-94zfm" Dec 03 14:11:57 crc kubenswrapper[4805]: I1203 14:11:57.719064 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-k648b\" (UniqueName: \"kubernetes.io/projected/4aa1a2c6-8700-4f70-b804-f57371df7025-kube-api-access-k648b\") pod \"migrator-59844c95c7-h5tv7\" (UID: \"4aa1a2c6-8700-4f70-b804-f57371df7025\") " pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-h5tv7" Dec 03 14:11:57 crc kubenswrapper[4805]: I1203 14:11:57.719082 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fz449\" (UniqueName: \"kubernetes.io/projected/5cec10a3-ed60-4193-be30-a78cb80bd51d-kube-api-access-fz449\") pod \"openshift-controller-manager-operator-756b6f6bc6-94zfm\" (UID: \"5cec10a3-ed60-4193-be30-a78cb80bd51d\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-94zfm" Dec 03 14:11:57 crc kubenswrapper[4805]: I1203 14:11:57.719106 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/b22cbc60-5975-42a4-85d3-6add6917fd88-service-ca-bundle\") pod \"router-default-5444994796-kxnqm\" (UID: \"b22cbc60-5975-42a4-85d3-6add6917fd88\") " pod="openshift-ingress/router-default-5444994796-kxnqm" Dec 03 14:11:57 crc kubenswrapper[4805]: I1203 14:11:57.719120 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6107a436-ce5e-4fde-890b-a5ac9f2ada10-serving-cert\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-86589\" (UID: \"6107a436-ce5e-4fde-890b-a5ac9f2ada10\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-86589" Dec 03 14:11:57 crc kubenswrapper[4805]: I1203 14:11:57.719139 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/secret/27ffe4cd-8ab3-49dd-8e5d-91b7d2439a27-certs\") pod \"machine-config-server-d5pjl\" (UID: \"27ffe4cd-8ab3-49dd-8e5d-91b7d2439a27\") " pod="openshift-machine-config-operator/machine-config-server-d5pjl" Dec 03 14:11:57 crc kubenswrapper[4805]: I1203 14:11:57.719160 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registration-dir\" (UniqueName: \"kubernetes.io/host-path/8801ff25-d273-4354-ab39-16e979d0b73d-registration-dir\") pod \"csi-hostpathplugin-krqsw\" (UID: \"8801ff25-d273-4354-ab39-16e979d0b73d\") " pod="hostpath-provisioner/csi-hostpathplugin-krqsw" Dec 03 14:11:57 crc kubenswrapper[4805]: I1203 14:11:57.719177 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/220a8b3d-e5fa-486f-ad11-4fb9a43ab092-metrics-tls\") pod \"ingress-operator-5b745b69d9-j8pwh\" (UID: \"220a8b3d-e5fa-486f-ad11-4fb9a43ab092\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-j8pwh" Dec 03 14:11:57 crc kubenswrapper[4805]: I1203 14:11:57.719202 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pgvvg\" (UniqueName: \"kubernetes.io/projected/f71c6d49-bfbf-4c57-a4f8-5cd22d1038cb-kube-api-access-pgvvg\") pod \"kube-storage-version-migrator-operator-b67b599dd-g46rj\" (UID: \"f71c6d49-bfbf-4c57-a4f8-5cd22d1038cb\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-g46rj" Dec 03 14:11:57 crc kubenswrapper[4805]: I1203 14:11:57.719225 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/220a8b3d-e5fa-486f-ad11-4fb9a43ab092-bound-sa-token\") pod \"ingress-operator-5b745b69d9-j8pwh\" (UID: \"220a8b3d-e5fa-486f-ad11-4fb9a43ab092\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-j8pwh" Dec 03 14:11:57 crc kubenswrapper[4805]: I1203 14:11:57.719243 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tlt9k\" (UniqueName: \"kubernetes.io/projected/6be05973-dc60-437f-959c-a9309cd16d5b-kube-api-access-tlt9k\") pod \"ingress-canary-cnlwr\" (UID: \"6be05973-dc60-437f-959c-a9309cd16d5b\") " pod="openshift-ingress-canary/ingress-canary-cnlwr" Dec 03 14:11:57 crc kubenswrapper[4805]: I1203 14:11:57.719262 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/9bd2f61c-2b92-488f-b632-1e627b48a518-control-plane-machine-set-operator-tls\") pod \"control-plane-machine-set-operator-78cbb6b69f-25khc\" (UID: \"9bd2f61c-2b92-488f-b632-1e627b48a518\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-25khc" Dec 03 14:11:57 crc kubenswrapper[4805]: I1203 14:11:57.719280 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/6be05973-dc60-437f-959c-a9309cd16d5b-cert\") pod \"ingress-canary-cnlwr\" (UID: \"6be05973-dc60-437f-959c-a9309cd16d5b\") " pod="openshift-ingress-canary/ingress-canary-cnlwr" Dec 03 14:11:57 crc kubenswrapper[4805]: I1203 14:11:57.719297 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/a8d74d87-86ee-4723-91bc-1d25db30909f-srv-cert\") pod \"catalog-operator-68c6474976-rp9lf\" (UID: \"a8d74d87-86ee-4723-91bc-1d25db30909f\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-rp9lf" Dec 03 14:11:57 crc kubenswrapper[4805]: I1203 14:11:57.719319 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/776b097e-ad39-49a4-aa4f-a326a04bc30b-profile-collector-cert\") pod \"olm-operator-6b444d44fb-ss8kj\" (UID: \"776b097e-ad39-49a4-aa4f-a326a04bc30b\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-ss8kj" Dec 03 14:11:57 crc kubenswrapper[4805]: I1203 14:11:57.719336 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-j686r\" (UniqueName: \"kubernetes.io/projected/776b097e-ad39-49a4-aa4f-a326a04bc30b-kube-api-access-j686r\") pod \"olm-operator-6b444d44fb-ss8kj\" (UID: \"776b097e-ad39-49a4-aa4f-a326a04bc30b\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-ss8kj" Dec 03 14:11:57 crc kubenswrapper[4805]: I1203 14:11:57.719353 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/b22cbc60-5975-42a4-85d3-6add6917fd88-stats-auth\") pod \"router-default-5444994796-kxnqm\" (UID: \"b22cbc60-5975-42a4-85d3-6add6917fd88\") " pod="openshift-ingress/router-default-5444994796-kxnqm" Dec 03 14:11:57 crc kubenswrapper[4805]: I1203 14:11:57.719371 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/470bfa35-69f4-4865-ac37-0aa43d229b05-serving-cert\") pod \"kube-controller-manager-operator-78b949d7b-6n5xz\" (UID: \"470bfa35-69f4-4865-ac37-0aa43d229b05\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-6n5xz" Dec 03 14:11:57 crc kubenswrapper[4805]: I1203 14:11:57.719401 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-47lhk\" (UniqueName: \"kubernetes.io/projected/b1313b26-f0b6-462b-8210-c19b0d4065a4-kube-api-access-47lhk\") pod \"dns-default-7wnwf\" (UID: \"b1313b26-f0b6-462b-8210-c19b0d4065a4\") " pod="openshift-dns/dns-default-7wnwf" Dec 03 14:11:57 crc kubenswrapper[4805]: I1203 14:11:57.719417 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/470bfa35-69f4-4865-ac37-0aa43d229b05-config\") pod \"kube-controller-manager-operator-78b949d7b-6n5xz\" (UID: \"470bfa35-69f4-4865-ac37-0aa43d229b05\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-6n5xz" Dec 03 14:11:57 crc kubenswrapper[4805]: I1203 14:11:57.719436 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/b44fb2fa-6155-40fc-b5f7-eb5d1fa391bb-package-server-manager-serving-cert\") pod \"package-server-manager-789f6589d5-b7wrl\" (UID: \"b44fb2fa-6155-40fc-b5f7-eb5d1fa391bb\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-b7wrl" Dec 03 14:11:57 crc kubenswrapper[4805]: I1203 14:11:57.719452 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/470bfa35-69f4-4865-ac37-0aa43d229b05-kube-api-access\") pod \"kube-controller-manager-operator-78b949d7b-6n5xz\" (UID: \"470bfa35-69f4-4865-ac37-0aa43d229b05\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-6n5xz" Dec 03 14:11:57 crc kubenswrapper[4805]: I1203 14:11:57.719467 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"csi-data-dir\" (UniqueName: \"kubernetes.io/host-path/8801ff25-d273-4354-ab39-16e979d0b73d-csi-data-dir\") pod \"csi-hostpathplugin-krqsw\" (UID: \"8801ff25-d273-4354-ab39-16e979d0b73d\") " pod="hostpath-provisioner/csi-hostpathplugin-krqsw" Dec 03 14:11:57 crc kubenswrapper[4805]: I1203 14:11:57.719483 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/a8d74d87-86ee-4723-91bc-1d25db30909f-profile-collector-cert\") pod \"catalog-operator-68c6474976-rp9lf\" (UID: \"a8d74d87-86ee-4723-91bc-1d25db30909f\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-rp9lf" Dec 03 14:11:57 crc kubenswrapper[4805]: I1203 14:11:57.719499 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rfhkx\" (UniqueName: \"kubernetes.io/projected/a8d74d87-86ee-4723-91bc-1d25db30909f-kube-api-access-rfhkx\") pod \"catalog-operator-68c6474976-rp9lf\" (UID: \"a8d74d87-86ee-4723-91bc-1d25db30909f\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-rp9lf" Dec 03 14:11:57 crc kubenswrapper[4805]: I1203 14:11:57.719515 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/776b097e-ad39-49a4-aa4f-a326a04bc30b-srv-cert\") pod \"olm-operator-6b444d44fb-ss8kj\" (UID: \"776b097e-ad39-49a4-aa4f-a326a04bc30b\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-ss8kj" Dec 03 14:11:57 crc kubenswrapper[4805]: I1203 14:11:57.719531 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nmkbv\" (UniqueName: \"kubernetes.io/projected/220a8b3d-e5fa-486f-ad11-4fb9a43ab092-kube-api-access-nmkbv\") pod \"ingress-operator-5b745b69d9-j8pwh\" (UID: \"220a8b3d-e5fa-486f-ad11-4fb9a43ab092\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-j8pwh" Dec 03 14:11:57 crc kubenswrapper[4805]: I1203 14:11:57.719547 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/b22cbc60-5975-42a4-85d3-6add6917fd88-default-certificate\") pod \"router-default-5444994796-kxnqm\" (UID: \"b22cbc60-5975-42a4-85d3-6add6917fd88\") " pod="openshift-ingress/router-default-5444994796-kxnqm" Dec 03 14:11:57 crc kubenswrapper[4805]: I1203 14:11:57.719566 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/b22cbc60-5975-42a4-85d3-6add6917fd88-metrics-certs\") pod \"router-default-5444994796-kxnqm\" (UID: \"b22cbc60-5975-42a4-85d3-6add6917fd88\") " pod="openshift-ingress/router-default-5444994796-kxnqm" Dec 03 14:11:57 crc kubenswrapper[4805]: I1203 14:11:57.719581 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vtfj2\" (UniqueName: \"kubernetes.io/projected/27ffe4cd-8ab3-49dd-8e5d-91b7d2439a27-kube-api-access-vtfj2\") pod \"machine-config-server-d5pjl\" (UID: \"27ffe4cd-8ab3-49dd-8e5d-91b7d2439a27\") " pod="openshift-machine-config-operator/machine-config-server-d5pjl" Dec 03 14:11:57 crc kubenswrapper[4805]: I1203 14:11:57.719596 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"mountpoint-dir\" (UniqueName: \"kubernetes.io/host-path/8801ff25-d273-4354-ab39-16e979d0b73d-mountpoint-dir\") pod \"csi-hostpathplugin-krqsw\" (UID: \"8801ff25-d273-4354-ab39-16e979d0b73d\") " pod="hostpath-provisioner/csi-hostpathplugin-krqsw" Dec 03 14:11:57 crc kubenswrapper[4805]: I1203 14:11:57.719665 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"mountpoint-dir\" (UniqueName: \"kubernetes.io/host-path/8801ff25-d273-4354-ab39-16e979d0b73d-mountpoint-dir\") pod \"csi-hostpathplugin-krqsw\" (UID: \"8801ff25-d273-4354-ab39-16e979d0b73d\") " pod="hostpath-provisioner/csi-hostpathplugin-krqsw" Dec 03 14:11:57 crc kubenswrapper[4805]: E1203 14:11:57.719724 4805 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-03 14:11:58.219711335 +0000 UTC m=+147.882628258 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 14:11:57 crc kubenswrapper[4805]: I1203 14:11:57.719916 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"socket-dir\" (UniqueName: \"kubernetes.io/host-path/8801ff25-d273-4354-ab39-16e979d0b73d-socket-dir\") pod \"csi-hostpathplugin-krqsw\" (UID: \"8801ff25-d273-4354-ab39-16e979d0b73d\") " pod="hostpath-provisioner/csi-hostpathplugin-krqsw" Dec 03 14:11:57 crc kubenswrapper[4805]: I1203 14:11:57.720407 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5cec10a3-ed60-4193-be30-a78cb80bd51d-config\") pod \"openshift-controller-manager-operator-756b6f6bc6-94zfm\" (UID: \"5cec10a3-ed60-4193-be30-a78cb80bd51d\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-94zfm" Dec 03 14:11:57 crc kubenswrapper[4805]: I1203 14:11:57.721190 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-dir\" (UniqueName: \"kubernetes.io/host-path/8801ff25-d273-4354-ab39-16e979d0b73d-plugins-dir\") pod \"csi-hostpathplugin-krqsw\" (UID: \"8801ff25-d273-4354-ab39-16e979d0b73d\") " pod="hostpath-provisioner/csi-hostpathplugin-krqsw" Dec 03 14:11:57 crc kubenswrapper[4805]: I1203 14:11:57.721752 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/b22cbc60-5975-42a4-85d3-6add6917fd88-service-ca-bundle\") pod \"router-default-5444994796-kxnqm\" (UID: \"b22cbc60-5975-42a4-85d3-6add6917fd88\") " pod="openshift-ingress/router-default-5444994796-kxnqm" Dec 03 14:11:57 crc kubenswrapper[4805]: I1203 14:11:57.723765 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/6be05973-dc60-437f-959c-a9309cd16d5b-cert\") pod \"ingress-canary-cnlwr\" (UID: \"6be05973-dc60-437f-959c-a9309cd16d5b\") " pod="openshift-ingress-canary/ingress-canary-cnlwr" Dec 03 14:11:57 crc kubenswrapper[4805]: I1203 14:11:57.724179 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5cec10a3-ed60-4193-be30-a78cb80bd51d-serving-cert\") pod \"openshift-controller-manager-operator-756b6f6bc6-94zfm\" (UID: \"5cec10a3-ed60-4193-be30-a78cb80bd51d\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-94zfm" Dec 03 14:11:57 crc kubenswrapper[4805]: I1203 14:11:57.724202 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registration-dir\" (UniqueName: \"kubernetes.io/host-path/8801ff25-d273-4354-ab39-16e979d0b73d-registration-dir\") pod \"csi-hostpathplugin-krqsw\" (UID: \"8801ff25-d273-4354-ab39-16e979d0b73d\") " pod="hostpath-provisioner/csi-hostpathplugin-krqsw" Dec 03 14:11:57 crc kubenswrapper[4805]: I1203 14:11:57.724263 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"csi-data-dir\" (UniqueName: \"kubernetes.io/host-path/8801ff25-d273-4354-ab39-16e979d0b73d-csi-data-dir\") pod \"csi-hostpathplugin-krqsw\" (UID: \"8801ff25-d273-4354-ab39-16e979d0b73d\") " pod="hostpath-provisioner/csi-hostpathplugin-krqsw" Dec 03 14:11:57 crc kubenswrapper[4805]: I1203 14:11:57.724962 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/776b097e-ad39-49a4-aa4f-a326a04bc30b-srv-cert\") pod \"olm-operator-6b444d44fb-ss8kj\" (UID: \"776b097e-ad39-49a4-aa4f-a326a04bc30b\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-ss8kj" Dec 03 14:11:57 crc kubenswrapper[4805]: I1203 14:11:57.725393 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/b22cbc60-5975-42a4-85d3-6add6917fd88-metrics-certs\") pod \"router-default-5444994796-kxnqm\" (UID: \"b22cbc60-5975-42a4-85d3-6add6917fd88\") " pod="openshift-ingress/router-default-5444994796-kxnqm" Dec 03 14:11:57 crc kubenswrapper[4805]: I1203 14:11:57.726685 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/776b097e-ad39-49a4-aa4f-a326a04bc30b-profile-collector-cert\") pod \"olm-operator-6b444d44fb-ss8kj\" (UID: \"776b097e-ad39-49a4-aa4f-a326a04bc30b\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-ss8kj" Dec 03 14:11:57 crc kubenswrapper[4805]: I1203 14:11:57.727539 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/b22cbc60-5975-42a4-85d3-6add6917fd88-default-certificate\") pod \"router-default-5444994796-kxnqm\" (UID: \"b22cbc60-5975-42a4-85d3-6add6917fd88\") " pod="openshift-ingress/router-default-5444994796-kxnqm" Dec 03 14:11:57 crc kubenswrapper[4805]: I1203 14:11:57.728240 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/b22cbc60-5975-42a4-85d3-6add6917fd88-stats-auth\") pod \"router-default-5444994796-kxnqm\" (UID: \"b22cbc60-5975-42a4-85d3-6add6917fd88\") " pod="openshift-ingress/router-default-5444994796-kxnqm" Dec 03 14:11:57 crc kubenswrapper[4805]: I1203 14:11:57.760011 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tlt9k\" (UniqueName: \"kubernetes.io/projected/6be05973-dc60-437f-959c-a9309cd16d5b-kube-api-access-tlt9k\") pod \"ingress-canary-cnlwr\" (UID: \"6be05973-dc60-437f-959c-a9309cd16d5b\") " pod="openshift-ingress-canary/ingress-canary-cnlwr" Dec 03 14:11:57 crc kubenswrapper[4805]: I1203 14:11:57.772387 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-config-operator/openshift-config-operator-7777fb866f-7wzgf"] Dec 03 14:11:57 crc kubenswrapper[4805]: I1203 14:11:57.781204 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication-operator/authentication-operator-69f744f599-bs6tj"] Dec 03 14:11:57 crc kubenswrapper[4805]: I1203 14:11:57.783375 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gl58d\" (UniqueName: \"kubernetes.io/projected/8801ff25-d273-4354-ab39-16e979d0b73d-kube-api-access-gl58d\") pod \"csi-hostpathplugin-krqsw\" (UID: \"8801ff25-d273-4354-ab39-16e979d0b73d\") " pod="hostpath-provisioner/csi-hostpathplugin-krqsw" Dec 03 14:11:57 crc kubenswrapper[4805]: I1203 14:11:57.807636 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vwppd\" (UniqueName: \"kubernetes.io/projected/b22cbc60-5975-42a4-85d3-6add6917fd88-kube-api-access-vwppd\") pod \"router-default-5444994796-kxnqm\" (UID: \"b22cbc60-5975-42a4-85d3-6add6917fd88\") " pod="openshift-ingress/router-default-5444994796-kxnqm" Dec 03 14:11:57 crc kubenswrapper[4805]: I1203 14:11:57.812572 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fz449\" (UniqueName: \"kubernetes.io/projected/5cec10a3-ed60-4193-be30-a78cb80bd51d-kube-api-access-fz449\") pod \"openshift-controller-manager-operator-756b6f6bc6-94zfm\" (UID: \"5cec10a3-ed60-4193-be30-a78cb80bd51d\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-94zfm" Dec 03 14:11:57 crc kubenswrapper[4805]: I1203 14:11:57.815233 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress/router-default-5444994796-kxnqm" Dec 03 14:11:57 crc kubenswrapper[4805]: I1203 14:11:57.821618 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/b1313b26-f0b6-462b-8210-c19b0d4065a4-metrics-tls\") pod \"dns-default-7wnwf\" (UID: \"b1313b26-f0b6-462b-8210-c19b0d4065a4\") " pod="openshift-dns/dns-default-7wnwf" Dec 03 14:11:57 crc kubenswrapper[4805]: I1203 14:11:57.821652 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-r7pr5\" (UniqueName: \"kubernetes.io/projected/9bd2f61c-2b92-488f-b632-1e627b48a518-kube-api-access-r7pr5\") pod \"control-plane-machine-set-operator-78cbb6b69f-25khc\" (UID: \"9bd2f61c-2b92-488f-b632-1e627b48a518\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-25khc" Dec 03 14:11:57 crc kubenswrapper[4805]: I1203 14:11:57.821673 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f71c6d49-bfbf-4c57-a4f8-5cd22d1038cb-config\") pod \"kube-storage-version-migrator-operator-b67b599dd-g46rj\" (UID: \"f71c6d49-bfbf-4c57-a4f8-5cd22d1038cb\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-g46rj" Dec 03 14:11:57 crc kubenswrapper[4805]: I1203 14:11:57.821696 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/f71c6d49-bfbf-4c57-a4f8-5cd22d1038cb-serving-cert\") pod \"kube-storage-version-migrator-operator-b67b599dd-g46rj\" (UID: \"f71c6d49-bfbf-4c57-a4f8-5cd22d1038cb\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-g46rj" Dec 03 14:11:57 crc kubenswrapper[4805]: I1203 14:11:57.821715 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/220a8b3d-e5fa-486f-ad11-4fb9a43ab092-trusted-ca\") pod \"ingress-operator-5b745b69d9-j8pwh\" (UID: \"220a8b3d-e5fa-486f-ad11-4fb9a43ab092\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-j8pwh" Dec 03 14:11:57 crc kubenswrapper[4805]: I1203 14:11:57.821740 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/6107a436-ce5e-4fde-890b-a5ac9f2ada10-kube-api-access\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-86589\" (UID: \"6107a436-ce5e-4fde-890b-a5ac9f2ada10\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-86589" Dec 03 14:11:57 crc kubenswrapper[4805]: I1203 14:11:57.821773 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-jjzct\" (UID: \"4655a83e-4980-4a5c-9f61-e305ae418822\") " pod="openshift-image-registry/image-registry-697d97f7c8-jjzct" Dec 03 14:11:57 crc kubenswrapper[4805]: I1203 14:11:57.821797 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/27ffe4cd-8ab3-49dd-8e5d-91b7d2439a27-node-bootstrap-token\") pod \"machine-config-server-d5pjl\" (UID: \"27ffe4cd-8ab3-49dd-8e5d-91b7d2439a27\") " pod="openshift-machine-config-operator/machine-config-server-d5pjl" Dec 03 14:11:57 crc kubenswrapper[4805]: I1203 14:11:57.821818 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-k648b\" (UniqueName: \"kubernetes.io/projected/4aa1a2c6-8700-4f70-b804-f57371df7025-kube-api-access-k648b\") pod \"migrator-59844c95c7-h5tv7\" (UID: \"4aa1a2c6-8700-4f70-b804-f57371df7025\") " pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-h5tv7" Dec 03 14:11:57 crc kubenswrapper[4805]: I1203 14:11:57.822902 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6107a436-ce5e-4fde-890b-a5ac9f2ada10-serving-cert\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-86589\" (UID: \"6107a436-ce5e-4fde-890b-a5ac9f2ada10\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-86589" Dec 03 14:11:57 crc kubenswrapper[4805]: I1203 14:11:57.822945 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/secret/27ffe4cd-8ab3-49dd-8e5d-91b7d2439a27-certs\") pod \"machine-config-server-d5pjl\" (UID: \"27ffe4cd-8ab3-49dd-8e5d-91b7d2439a27\") " pod="openshift-machine-config-operator/machine-config-server-d5pjl" Dec 03 14:11:57 crc kubenswrapper[4805]: I1203 14:11:57.822973 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/220a8b3d-e5fa-486f-ad11-4fb9a43ab092-metrics-tls\") pod \"ingress-operator-5b745b69d9-j8pwh\" (UID: \"220a8b3d-e5fa-486f-ad11-4fb9a43ab092\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-j8pwh" Dec 03 14:11:57 crc kubenswrapper[4805]: I1203 14:11:57.822991 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pgvvg\" (UniqueName: \"kubernetes.io/projected/f71c6d49-bfbf-4c57-a4f8-5cd22d1038cb-kube-api-access-pgvvg\") pod \"kube-storage-version-migrator-operator-b67b599dd-g46rj\" (UID: \"f71c6d49-bfbf-4c57-a4f8-5cd22d1038cb\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-g46rj" Dec 03 14:11:57 crc kubenswrapper[4805]: I1203 14:11:57.823024 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/220a8b3d-e5fa-486f-ad11-4fb9a43ab092-bound-sa-token\") pod \"ingress-operator-5b745b69d9-j8pwh\" (UID: \"220a8b3d-e5fa-486f-ad11-4fb9a43ab092\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-j8pwh" Dec 03 14:11:57 crc kubenswrapper[4805]: I1203 14:11:57.823050 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/9bd2f61c-2b92-488f-b632-1e627b48a518-control-plane-machine-set-operator-tls\") pod \"control-plane-machine-set-operator-78cbb6b69f-25khc\" (UID: \"9bd2f61c-2b92-488f-b632-1e627b48a518\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-25khc" Dec 03 14:11:57 crc kubenswrapper[4805]: I1203 14:11:57.823069 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/a8d74d87-86ee-4723-91bc-1d25db30909f-srv-cert\") pod \"catalog-operator-68c6474976-rp9lf\" (UID: \"a8d74d87-86ee-4723-91bc-1d25db30909f\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-rp9lf" Dec 03 14:11:57 crc kubenswrapper[4805]: I1203 14:11:57.823123 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/470bfa35-69f4-4865-ac37-0aa43d229b05-serving-cert\") pod \"kube-controller-manager-operator-78b949d7b-6n5xz\" (UID: \"470bfa35-69f4-4865-ac37-0aa43d229b05\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-6n5xz" Dec 03 14:11:57 crc kubenswrapper[4805]: I1203 14:11:57.823145 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-47lhk\" (UniqueName: \"kubernetes.io/projected/b1313b26-f0b6-462b-8210-c19b0d4065a4-kube-api-access-47lhk\") pod \"dns-default-7wnwf\" (UID: \"b1313b26-f0b6-462b-8210-c19b0d4065a4\") " pod="openshift-dns/dns-default-7wnwf" Dec 03 14:11:57 crc kubenswrapper[4805]: I1203 14:11:57.823165 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/470bfa35-69f4-4865-ac37-0aa43d229b05-config\") pod \"kube-controller-manager-operator-78b949d7b-6n5xz\" (UID: \"470bfa35-69f4-4865-ac37-0aa43d229b05\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-6n5xz" Dec 03 14:11:57 crc kubenswrapper[4805]: I1203 14:11:57.823183 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/b44fb2fa-6155-40fc-b5f7-eb5d1fa391bb-package-server-manager-serving-cert\") pod \"package-server-manager-789f6589d5-b7wrl\" (UID: \"b44fb2fa-6155-40fc-b5f7-eb5d1fa391bb\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-b7wrl" Dec 03 14:11:57 crc kubenswrapper[4805]: I1203 14:11:57.823199 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/470bfa35-69f4-4865-ac37-0aa43d229b05-kube-api-access\") pod \"kube-controller-manager-operator-78b949d7b-6n5xz\" (UID: \"470bfa35-69f4-4865-ac37-0aa43d229b05\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-6n5xz" Dec 03 14:11:57 crc kubenswrapper[4805]: I1203 14:11:57.823221 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/a8d74d87-86ee-4723-91bc-1d25db30909f-profile-collector-cert\") pod \"catalog-operator-68c6474976-rp9lf\" (UID: \"a8d74d87-86ee-4723-91bc-1d25db30909f\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-rp9lf" Dec 03 14:11:57 crc kubenswrapper[4805]: I1203 14:11:57.823238 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rfhkx\" (UniqueName: \"kubernetes.io/projected/a8d74d87-86ee-4723-91bc-1d25db30909f-kube-api-access-rfhkx\") pod \"catalog-operator-68c6474976-rp9lf\" (UID: \"a8d74d87-86ee-4723-91bc-1d25db30909f\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-rp9lf" Dec 03 14:11:57 crc kubenswrapper[4805]: I1203 14:11:57.823255 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nmkbv\" (UniqueName: \"kubernetes.io/projected/220a8b3d-e5fa-486f-ad11-4fb9a43ab092-kube-api-access-nmkbv\") pod \"ingress-operator-5b745b69d9-j8pwh\" (UID: \"220a8b3d-e5fa-486f-ad11-4fb9a43ab092\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-j8pwh" Dec 03 14:11:57 crc kubenswrapper[4805]: I1203 14:11:57.823288 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vtfj2\" (UniqueName: \"kubernetes.io/projected/27ffe4cd-8ab3-49dd-8e5d-91b7d2439a27-kube-api-access-vtfj2\") pod \"machine-config-server-d5pjl\" (UID: \"27ffe4cd-8ab3-49dd-8e5d-91b7d2439a27\") " pod="openshift-machine-config-operator/machine-config-server-d5pjl" Dec 03 14:11:57 crc kubenswrapper[4805]: I1203 14:11:57.823312 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6107a436-ce5e-4fde-890b-a5ac9f2ada10-config\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-86589\" (UID: \"6107a436-ce5e-4fde-890b-a5ac9f2ada10\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-86589" Dec 03 14:11:57 crc kubenswrapper[4805]: I1203 14:11:57.823333 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-22kbd\" (UniqueName: \"kubernetes.io/projected/b44fb2fa-6155-40fc-b5f7-eb5d1fa391bb-kube-api-access-22kbd\") pod \"package-server-manager-789f6589d5-b7wrl\" (UID: \"b44fb2fa-6155-40fc-b5f7-eb5d1fa391bb\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-b7wrl" Dec 03 14:11:57 crc kubenswrapper[4805]: I1203 14:11:57.823346 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/b1313b26-f0b6-462b-8210-c19b0d4065a4-config-volume\") pod \"dns-default-7wnwf\" (UID: \"b1313b26-f0b6-462b-8210-c19b0d4065a4\") " pod="openshift-dns/dns-default-7wnwf" Dec 03 14:11:57 crc kubenswrapper[4805]: I1203 14:11:57.823943 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/b1313b26-f0b6-462b-8210-c19b0d4065a4-config-volume\") pod \"dns-default-7wnwf\" (UID: \"b1313b26-f0b6-462b-8210-c19b0d4065a4\") " pod="openshift-dns/dns-default-7wnwf" Dec 03 14:11:57 crc kubenswrapper[4805]: I1203 14:11:57.824584 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/470bfa35-69f4-4865-ac37-0aa43d229b05-config\") pod \"kube-controller-manager-operator-78b949d7b-6n5xz\" (UID: \"470bfa35-69f4-4865-ac37-0aa43d229b05\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-6n5xz" Dec 03 14:11:57 crc kubenswrapper[4805]: I1203 14:11:57.825906 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f71c6d49-bfbf-4c57-a4f8-5cd22d1038cb-config\") pod \"kube-storage-version-migrator-operator-b67b599dd-g46rj\" (UID: \"f71c6d49-bfbf-4c57-a4f8-5cd22d1038cb\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-g46rj" Dec 03 14:11:57 crc kubenswrapper[4805]: E1203 14:11:57.826360 4805 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-03 14:11:58.326344199 +0000 UTC m=+147.989261112 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-jjzct" (UID: "4655a83e-4980-4a5c-9f61-e305ae418822") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 14:11:57 crc kubenswrapper[4805]: I1203 14:11:57.828017 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6107a436-ce5e-4fde-890b-a5ac9f2ada10-config\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-86589\" (UID: \"6107a436-ce5e-4fde-890b-a5ac9f2ada10\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-86589" Dec 03 14:11:57 crc kubenswrapper[4805]: I1203 14:11:57.828325 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/220a8b3d-e5fa-486f-ad11-4fb9a43ab092-trusted-ca\") pod \"ingress-operator-5b745b69d9-j8pwh\" (UID: \"220a8b3d-e5fa-486f-ad11-4fb9a43ab092\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-j8pwh" Dec 03 14:11:57 crc kubenswrapper[4805]: I1203 14:11:57.833018 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/a8d74d87-86ee-4723-91bc-1d25db30909f-srv-cert\") pod \"catalog-operator-68c6474976-rp9lf\" (UID: \"a8d74d87-86ee-4723-91bc-1d25db30909f\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-rp9lf" Dec 03 14:11:57 crc kubenswrapper[4805]: I1203 14:11:57.834457 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/a8d74d87-86ee-4723-91bc-1d25db30909f-profile-collector-cert\") pod \"catalog-operator-68c6474976-rp9lf\" (UID: \"a8d74d87-86ee-4723-91bc-1d25db30909f\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-rp9lf" Dec 03 14:11:57 crc kubenswrapper[4805]: I1203 14:11:57.836475 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/b44fb2fa-6155-40fc-b5f7-eb5d1fa391bb-package-server-manager-serving-cert\") pod \"package-server-manager-789f6589d5-b7wrl\" (UID: \"b44fb2fa-6155-40fc-b5f7-eb5d1fa391bb\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-b7wrl" Dec 03 14:11:57 crc kubenswrapper[4805]: I1203 14:11:57.842268 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-94zfm" Dec 03 14:11:57 crc kubenswrapper[4805]: I1203 14:11:57.847152 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/470bfa35-69f4-4865-ac37-0aa43d229b05-serving-cert\") pod \"kube-controller-manager-operator-78b949d7b-6n5xz\" (UID: \"470bfa35-69f4-4865-ac37-0aa43d229b05\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-6n5xz" Dec 03 14:11:57 crc kubenswrapper[4805]: I1203 14:11:57.842415 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca-operator/service-ca-operator-777779d784-z7lgj"] Dec 03 14:11:57 crc kubenswrapper[4805]: I1203 14:11:57.848442 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-j686r\" (UniqueName: \"kubernetes.io/projected/776b097e-ad39-49a4-aa4f-a326a04bc30b-kube-api-access-j686r\") pod \"olm-operator-6b444d44fb-ss8kj\" (UID: \"776b097e-ad39-49a4-aa4f-a326a04bc30b\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-ss8kj" Dec 03 14:11:57 crc kubenswrapper[4805]: I1203 14:11:57.849207 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/220a8b3d-e5fa-486f-ad11-4fb9a43ab092-metrics-tls\") pod \"ingress-operator-5b745b69d9-j8pwh\" (UID: \"220a8b3d-e5fa-486f-ad11-4fb9a43ab092\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-j8pwh" Dec 03 14:11:57 crc kubenswrapper[4805]: I1203 14:11:57.850625 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-canary/ingress-canary-cnlwr" Dec 03 14:11:57 crc kubenswrapper[4805]: I1203 14:11:57.850746 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6107a436-ce5e-4fde-890b-a5ac9f2ada10-serving-cert\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-86589\" (UID: \"6107a436-ce5e-4fde-890b-a5ac9f2ada10\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-86589" Dec 03 14:11:57 crc kubenswrapper[4805]: I1203 14:11:57.851192 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-rr7fx"] Dec 03 14:11:57 crc kubenswrapper[4805]: I1203 14:11:57.851394 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/f71c6d49-bfbf-4c57-a4f8-5cd22d1038cb-serving-cert\") pod \"kube-storage-version-migrator-operator-b67b599dd-g46rj\" (UID: \"f71c6d49-bfbf-4c57-a4f8-5cd22d1038cb\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-g46rj" Dec 03 14:11:57 crc kubenswrapper[4805]: I1203 14:11:57.852400 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/b1313b26-f0b6-462b-8210-c19b0d4065a4-metrics-tls\") pod \"dns-default-7wnwf\" (UID: \"b1313b26-f0b6-462b-8210-c19b0d4065a4\") " pod="openshift-dns/dns-default-7wnwf" Dec 03 14:11:57 crc kubenswrapper[4805]: I1203 14:11:57.853513 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/9bd2f61c-2b92-488f-b632-1e627b48a518-control-plane-machine-set-operator-tls\") pod \"control-plane-machine-set-operator-78cbb6b69f-25khc\" (UID: \"9bd2f61c-2b92-488f-b632-1e627b48a518\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-25khc" Dec 03 14:11:57 crc kubenswrapper[4805]: I1203 14:11:57.854072 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"certs\" (UniqueName: \"kubernetes.io/secret/27ffe4cd-8ab3-49dd-8e5d-91b7d2439a27-certs\") pod \"machine-config-server-d5pjl\" (UID: \"27ffe4cd-8ab3-49dd-8e5d-91b7d2439a27\") " pod="openshift-machine-config-operator/machine-config-server-d5pjl" Dec 03 14:11:57 crc kubenswrapper[4805]: I1203 14:11:57.855880 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-f9d7485db-pp9tv"] Dec 03 14:11:57 crc kubenswrapper[4805]: I1203 14:11:57.857029 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-jvsqf"] Dec 03 14:11:57 crc kubenswrapper[4805]: I1203 14:11:57.859268 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/27ffe4cd-8ab3-49dd-8e5d-91b7d2439a27-node-bootstrap-token\") pod \"machine-config-server-d5pjl\" (UID: \"27ffe4cd-8ab3-49dd-8e5d-91b7d2439a27\") " pod="openshift-machine-config-operator/machine-config-server-d5pjl" Dec 03 14:11:57 crc kubenswrapper[4805]: I1203 14:11:57.882799 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/220a8b3d-e5fa-486f-ad11-4fb9a43ab092-bound-sa-token\") pod \"ingress-operator-5b745b69d9-j8pwh\" (UID: \"220a8b3d-e5fa-486f-ad11-4fb9a43ab092\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-j8pwh" Dec 03 14:11:57 crc kubenswrapper[4805]: I1203 14:11:57.887827 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-vc9sn"] Dec 03 14:11:57 crc kubenswrapper[4805]: I1203 14:11:57.888297 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="hostpath-provisioner/csi-hostpathplugin-krqsw" Dec 03 14:11:57 crc kubenswrapper[4805]: I1203 14:11:57.899896 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console-operator/console-operator-58897d9998-46w2g"] Dec 03 14:11:57 crc kubenswrapper[4805]: I1203 14:11:57.903584 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rfhkx\" (UniqueName: \"kubernetes.io/projected/a8d74d87-86ee-4723-91bc-1d25db30909f-kube-api-access-rfhkx\") pod \"catalog-operator-68c6474976-rp9lf\" (UID: \"a8d74d87-86ee-4723-91bc-1d25db30909f\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-rp9lf" Dec 03 14:11:57 crc kubenswrapper[4805]: W1203 14:11:57.913787 4805 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pode2034338_8a35_4fd2_9bbf_be733939fc25.slice/crio-b956cee4baf3251b5e35ca263db8e1fbff64366ec9922a8a0edb94f3320e271f WatchSource:0}: Error finding container b956cee4baf3251b5e35ca263db8e1fbff64366ec9922a8a0edb94f3320e271f: Status 404 returned error can't find the container with id b956cee4baf3251b5e35ca263db8e1fbff64366ec9922a8a0edb94f3320e271f Dec 03 14:11:57 crc kubenswrapper[4805]: W1203 14:11:57.915585 4805 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podda68f128_1d8d_4256_a28d_a174766e1331.slice/crio-e3c12f009e423435140f819a7bf2cb58eba47296fda5be47e3eb32450f191c02 WatchSource:0}: Error finding container e3c12f009e423435140f819a7bf2cb58eba47296fda5be47e3eb32450f191c02: Status 404 returned error can't find the container with id e3c12f009e423435140f819a7bf2cb58eba47296fda5be47e3eb32450f191c02 Dec 03 14:11:57 crc kubenswrapper[4805]: W1203 14:11:57.916419 4805 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod9b932fa5_f87b_46e1_8987_683cfd3b46fd.slice/crio-d326226b86633862d3a8f197c279aea7712e56fec09eec20ebd223a28208a291 WatchSource:0}: Error finding container d326226b86633862d3a8f197c279aea7712e56fec09eec20ebd223a28208a291: Status 404 returned error can't find the container with id d326226b86633862d3a8f197c279aea7712e56fec09eec20ebd223a28208a291 Dec 03 14:11:57 crc kubenswrapper[4805]: I1203 14:11:57.917458 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-r7pr5\" (UniqueName: \"kubernetes.io/projected/9bd2f61c-2b92-488f-b632-1e627b48a518-kube-api-access-r7pr5\") pod \"control-plane-machine-set-operator-78cbb6b69f-25khc\" (UID: \"9bd2f61c-2b92-488f-b632-1e627b48a518\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-25khc" Dec 03 14:11:57 crc kubenswrapper[4805]: I1203 14:11:57.924039 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 14:11:57 crc kubenswrapper[4805]: E1203 14:11:57.924447 4805 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-03 14:11:58.424429171 +0000 UTC m=+148.087346094 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 14:11:57 crc kubenswrapper[4805]: I1203 14:11:57.939664 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-k648b\" (UniqueName: \"kubernetes.io/projected/4aa1a2c6-8700-4f70-b804-f57371df7025-kube-api-access-k648b\") pod \"migrator-59844c95c7-h5tv7\" (UID: \"4aa1a2c6-8700-4f70-b804-f57371df7025\") " pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-h5tv7" Dec 03 14:11:57 crc kubenswrapper[4805]: I1203 14:11:57.956030 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/machine-api-operator-5694c8668f-jvspg"] Dec 03 14:11:57 crc kubenswrapper[4805]: I1203 14:11:57.960482 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-k7rgc"] Dec 03 14:11:57 crc kubenswrapper[4805]: I1203 14:11:57.972327 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-oauth-apiserver/apiserver-7bbb656c7d-kfh2f"] Dec 03 14:11:57 crc kubenswrapper[4805]: I1203 14:11:57.985552 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vtfj2\" (UniqueName: \"kubernetes.io/projected/27ffe4cd-8ab3-49dd-8e5d-91b7d2439a27-kube-api-access-vtfj2\") pod \"machine-config-server-d5pjl\" (UID: \"27ffe4cd-8ab3-49dd-8e5d-91b7d2439a27\") " pod="openshift-machine-config-operator/machine-config-server-d5pjl" Dec 03 14:11:58 crc kubenswrapper[4805]: I1203 14:11:58.007588 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nmkbv\" (UniqueName: \"kubernetes.io/projected/220a8b3d-e5fa-486f-ad11-4fb9a43ab092-kube-api-access-nmkbv\") pod \"ingress-operator-5b745b69d9-j8pwh\" (UID: \"220a8b3d-e5fa-486f-ad11-4fb9a43ab092\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-j8pwh" Dec 03 14:11:58 crc kubenswrapper[4805]: I1203 14:11:58.008133 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/6107a436-ce5e-4fde-890b-a5ac9f2ada10-kube-api-access\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-86589\" (UID: \"6107a436-ce5e-4fde-890b-a5ac9f2ada10\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-86589" Dec 03 14:11:58 crc kubenswrapper[4805]: W1203 14:11:58.016551 4805 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod9f53fe44_0d2c_4e3d_a682_df3a779a1c8a.slice/crio-2fa8dc0d3b9582a0e761f47f8cc7bc420e22bb088375dee7f23e5568c389f3ab WatchSource:0}: Error finding container 2fa8dc0d3b9582a0e761f47f8cc7bc420e22bb088375dee7f23e5568c389f3ab: Status 404 returned error can't find the container with id 2fa8dc0d3b9582a0e761f47f8cc7bc420e22bb088375dee7f23e5568c389f3ab Dec 03 14:11:58 crc kubenswrapper[4805]: I1203 14:11:58.020281 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/470bfa35-69f4-4865-ac37-0aa43d229b05-kube-api-access\") pod \"kube-controller-manager-operator-78b949d7b-6n5xz\" (UID: \"470bfa35-69f4-4865-ac37-0aa43d229b05\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-6n5xz" Dec 03 14:11:58 crc kubenswrapper[4805]: I1203 14:11:58.023258 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-h5tv7" Dec 03 14:11:58 crc kubenswrapper[4805]: I1203 14:11:58.025444 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-jjzct\" (UID: \"4655a83e-4980-4a5c-9f61-e305ae418822\") " pod="openshift-image-registry/image-registry-697d97f7c8-jjzct" Dec 03 14:11:58 crc kubenswrapper[4805]: E1203 14:11:58.025746 4805 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-03 14:11:58.525733599 +0000 UTC m=+148.188650522 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-jjzct" (UID: "4655a83e-4980-4a5c-9f61-e305ae418822") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 14:11:58 crc kubenswrapper[4805]: W1203 14:11:58.029345 4805 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podb22cbc60_5975_42a4_85d3_6add6917fd88.slice/crio-8afb328b6eec9f0fb5510c2c0268688250e99470e4b23c184cf70f7e2a4d2ccc WatchSource:0}: Error finding container 8afb328b6eec9f0fb5510c2c0268688250e99470e4b23c184cf70f7e2a4d2ccc: Status 404 returned error can't find the container with id 8afb328b6eec9f0fb5510c2c0268688250e99470e4b23c184cf70f7e2a4d2ccc Dec 03 14:11:58 crc kubenswrapper[4805]: I1203 14:11:58.029554 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-j8pwh" Dec 03 14:11:58 crc kubenswrapper[4805]: I1203 14:11:58.032783 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-47lhk\" (UniqueName: \"kubernetes.io/projected/b1313b26-f0b6-462b-8210-c19b0d4065a4-kube-api-access-47lhk\") pod \"dns-default-7wnwf\" (UID: \"b1313b26-f0b6-462b-8210-c19b0d4065a4\") " pod="openshift-dns/dns-default-7wnwf" Dec 03 14:11:58 crc kubenswrapper[4805]: I1203 14:11:58.059129 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-22kbd\" (UniqueName: \"kubernetes.io/projected/b44fb2fa-6155-40fc-b5f7-eb5d1fa391bb-kube-api-access-22kbd\") pod \"package-server-manager-789f6589d5-b7wrl\" (UID: \"b44fb2fa-6155-40fc-b5f7-eb5d1fa391bb\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-b7wrl" Dec 03 14:11:58 crc kubenswrapper[4805]: I1203 14:11:58.067031 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-ss8kj" Dec 03 14:11:58 crc kubenswrapper[4805]: W1203 14:11:58.070412 4805 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podb54908df_f04f_4515_ab36_842f0aef4a99.slice/crio-1aebe087880b686ed928502ca004a539fe527c336faf9330357b68468b25841c WatchSource:0}: Error finding container 1aebe087880b686ed928502ca004a539fe527c336faf9330357b68468b25841c: Status 404 returned error can't find the container with id 1aebe087880b686ed928502ca004a539fe527c336faf9330357b68468b25841c Dec 03 14:11:58 crc kubenswrapper[4805]: W1203 14:11:58.071598 4805 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-poddfd20bfc_5c11_4287_a4b3_df9e0ca4dcf1.slice/crio-88f9e67ebf81e77891b5bf0f53d48db2d210295dc030c9e52abc201c500e5061 WatchSource:0}: Error finding container 88f9e67ebf81e77891b5bf0f53d48db2d210295dc030c9e52abc201c500e5061: Status 404 returned error can't find the container with id 88f9e67ebf81e77891b5bf0f53d48db2d210295dc030c9e52abc201c500e5061 Dec 03 14:11:58 crc kubenswrapper[4805]: I1203 14:11:58.075171 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pgvvg\" (UniqueName: \"kubernetes.io/projected/f71c6d49-bfbf-4c57-a4f8-5cd22d1038cb-kube-api-access-pgvvg\") pod \"kube-storage-version-migrator-operator-b67b599dd-g46rj\" (UID: \"f71c6d49-bfbf-4c57-a4f8-5cd22d1038cb\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-g46rj" Dec 03 14:11:58 crc kubenswrapper[4805]: I1203 14:11:58.079614 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-g46rj" Dec 03 14:11:58 crc kubenswrapper[4805]: I1203 14:11:58.087744 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-86589" Dec 03 14:11:58 crc kubenswrapper[4805]: I1203 14:11:58.095995 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-rp9lf" Dec 03 14:11:58 crc kubenswrapper[4805]: I1203 14:11:58.105263 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-6n5xz" Dec 03 14:11:58 crc kubenswrapper[4805]: I1203 14:11:58.125134 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-b7wrl" Dec 03 14:11:58 crc kubenswrapper[4805]: I1203 14:11:58.126295 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 14:11:58 crc kubenswrapper[4805]: E1203 14:11:58.133564 4805 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-03 14:11:58.63353341 +0000 UTC m=+148.296450343 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 14:11:58 crc kubenswrapper[4805]: I1203 14:11:58.135606 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-jjzct\" (UID: \"4655a83e-4980-4a5c-9f61-e305ae418822\") " pod="openshift-image-registry/image-registry-697d97f7c8-jjzct" Dec 03 14:11:58 crc kubenswrapper[4805]: I1203 14:11:58.135561 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-25khc" Dec 03 14:11:58 crc kubenswrapper[4805]: E1203 14:11:58.136182 4805 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-03 14:11:58.636168023 +0000 UTC m=+148.299084946 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-jjzct" (UID: "4655a83e-4980-4a5c-9f61-e305ae418822") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 14:11:58 crc kubenswrapper[4805]: I1203 14:11:58.159486 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/dns-default-7wnwf" Dec 03 14:11:58 crc kubenswrapper[4805]: I1203 14:11:58.166175 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-server-d5pjl" Dec 03 14:11:58 crc kubenswrapper[4805]: I1203 14:11:58.197306 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/downloads-7954f5f757-s25zl"] Dec 03 14:11:58 crc kubenswrapper[4805]: I1203 14:11:58.199056 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-ntv6j"] Dec 03 14:11:58 crc kubenswrapper[4805]: I1203 14:11:58.240569 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 14:11:58 crc kubenswrapper[4805]: E1203 14:11:58.241249 4805 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-03 14:11:58.74122938 +0000 UTC m=+148.404146303 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 14:11:58 crc kubenswrapper[4805]: I1203 14:11:58.241290 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-operator-74547568cd-kvmpp"] Dec 03 14:11:58 crc kubenswrapper[4805]: I1203 14:11:58.244255 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-2vnrk"] Dec 03 14:11:58 crc kubenswrapper[4805]: I1203 14:11:58.343253 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-jjzct\" (UID: \"4655a83e-4980-4a5c-9f61-e305ae418822\") " pod="openshift-image-registry/image-registry-697d97f7c8-jjzct" Dec 03 14:11:58 crc kubenswrapper[4805]: E1203 14:11:58.343653 4805 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-03 14:11:58.843638194 +0000 UTC m=+148.506555117 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-jjzct" (UID: "4655a83e-4980-4a5c-9f61-e305ae418822") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 14:11:58 crc kubenswrapper[4805]: I1203 14:11:58.446059 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 14:11:58 crc kubenswrapper[4805]: E1203 14:11:58.446564 4805 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-03 14:11:58.94654915 +0000 UTC m=+148.609466073 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 14:11:58 crc kubenswrapper[4805]: I1203 14:11:58.469797 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca-operator/service-ca-operator-777779d784-z7lgj" event={"ID":"9b932fa5-f87b-46e1-8987-683cfd3b46fd","Type":"ContainerStarted","Data":"d326226b86633862d3a8f197c279aea7712e56fec09eec20ebd223a28208a291"} Dec 03 14:11:58 crc kubenswrapper[4805]: I1203 14:11:58.483850 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-nv2m9" event={"ID":"00941575-52e9-482d-b256-1b1c757d96e0","Type":"ContainerStarted","Data":"e32a60804d392b35acc6ac735c94ebed20dfe42e28374204ec00e7c476c0b62e"} Dec 03 14:11:58 crc kubenswrapper[4805]: I1203 14:11:58.485103 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress/router-default-5444994796-kxnqm" event={"ID":"b22cbc60-5975-42a4-85d3-6add6917fd88","Type":"ContainerStarted","Data":"8afb328b6eec9f0fb5510c2c0268688250e99470e4b23c184cf70f7e2a4d2ccc"} Dec 03 14:11:58 crc kubenswrapper[4805]: I1203 14:11:58.487630 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-f9d7485db-pp9tv" event={"ID":"e2034338-8a35-4fd2-9bbf-be733939fc25","Type":"ContainerStarted","Data":"b956cee4baf3251b5e35ca263db8e1fbff64366ec9922a8a0edb94f3320e271f"} Dec 03 14:11:58 crc kubenswrapper[4805]: I1203 14:11:58.488321 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/downloads-7954f5f757-s25zl" event={"ID":"6c2d6fbc-771d-4aff-b1e8-3eda17204931","Type":"ContainerStarted","Data":"9cc5ec6b001c36b7855168407f944ec3282b36640eb727b407ad7318eeb6e6d3"} Dec 03 14:11:58 crc kubenswrapper[4805]: I1203 14:11:58.489348 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-admission-controller-857f4d67dd-tn8tx" event={"ID":"33c718fb-affd-410b-9f53-ddb68ba9760c","Type":"ContainerStarted","Data":"2199fca40c53f02994dd719c93b7005df41afcae10091e4cd82223a501cf8eeb"} Dec 03 14:11:58 crc kubenswrapper[4805]: I1203 14:11:58.489364 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-admission-controller-857f4d67dd-tn8tx" event={"ID":"33c718fb-affd-410b-9f53-ddb68ba9760c","Type":"ContainerStarted","Data":"bb70c8276d867da4d5374169c3f0457397ce4169de497a4d3ff29ae52237736a"} Dec 03 14:11:58 crc kubenswrapper[4805]: I1203 14:11:58.492053 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-kfh2f" event={"ID":"0a7d0b69-f9a6-4a40-9981-3dd999c400d5","Type":"ContainerStarted","Data":"115428ee38a88a720993342081e641758088db30061a751b8d88934bd91084ce"} Dec 03 14:11:58 crc kubenswrapper[4805]: I1203 14:11:58.498490 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" event={"ID":"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8","Type":"ContainerStarted","Data":"38e7232f657a794e36c167b266171939e43a5bd070e918b73cd5ff5e8f947d6e"} Dec 03 14:11:58 crc kubenswrapper[4805]: I1203 14:11:58.509139 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-kvmpp" event={"ID":"d5ea0764-4290-4eab-b4ce-dbd4909455c1","Type":"ContainerStarted","Data":"43ae16609984b8647013cd23760a886a13783f401a0f867666a27d9c5a458a95"} Dec 03 14:11:58 crc kubenswrapper[4805]: I1203 14:11:58.532980 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" event={"ID":"9d751cbb-f2e2-430d-9754-c882a5e924a5","Type":"ContainerStarted","Data":"f4a0af1e0fd69602423e2c0196e59dc9f4cafdc8f258d10624951c6d943ac3ae"} Dec 03 14:11:58 crc kubenswrapper[4805]: I1203 14:11:58.538481 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29412840-gj7pr"] Dec 03 14:11:58 crc kubenswrapper[4805]: I1203 14:11:58.545908 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-spjlp"] Dec 03 14:11:58 crc kubenswrapper[4805]: I1203 14:11:58.548144 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-jjzct\" (UID: \"4655a83e-4980-4a5c-9f61-e305ae418822\") " pod="openshift-image-registry/image-registry-697d97f7c8-jjzct" Dec 03 14:11:58 crc kubenswrapper[4805]: E1203 14:11:58.549427 4805 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-03 14:11:59.049412214 +0000 UTC m=+148.712329137 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-jjzct" (UID: "4655a83e-4980-4a5c-9f61-e305ae418822") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 14:11:58 crc kubenswrapper[4805]: I1203 14:11:58.556756 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca/service-ca-9c57cc56f-dbfrd"] Dec 03 14:11:58 crc kubenswrapper[4805]: I1203 14:11:58.560690 4805 generic.go:334] "Generic (PLEG): container finished" podID="ca425595-aae1-44f2-9470-84ba52167748" containerID="b9b15ca18a9452cf0e2757873b1f381d4e6302d52308b8ac2e7470b3e102e2b9" exitCode=0 Dec 03 14:11:58 crc kubenswrapper[4805]: I1203 14:11:58.560787 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-7nhvl" event={"ID":"ca425595-aae1-44f2-9470-84ba52167748","Type":"ContainerDied","Data":"b9b15ca18a9452cf0e2757873b1f381d4e6302d52308b8ac2e7470b3e102e2b9"} Dec 03 14:11:58 crc kubenswrapper[4805]: I1203 14:11:58.560829 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-7nhvl" event={"ID":"ca425595-aae1-44f2-9470-84ba52167748","Type":"ContainerStarted","Data":"0b4ecae8312cfa07d802b90c3709b70903f090fa27bffa48228a136f3698db9b"} Dec 03 14:11:58 crc kubenswrapper[4805]: I1203 14:11:58.574487 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-ntv6j" event={"ID":"1658d8e1-ac63-40fe-98ce-900a9f9aeae5","Type":"ContainerStarted","Data":"b58229b467e31512c2984b233d11b240a14366ce5c3263aeab9af7a21498a9a6"} Dec 03 14:11:58 crc kubenswrapper[4805]: I1203 14:11:58.574968 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-controller-84d6567774-t4kl8"] Dec 03 14:11:58 crc kubenswrapper[4805]: I1203 14:11:58.581883 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-l6qxf" event={"ID":"e5a33c32-3873-49b8-8a81-c1a0357d94fc","Type":"ContainerStarted","Data":"956563191f647ab3c361b96fbb02387ff874fc84f5e4c4799725ccccfc1bc882"} Dec 03 14:11:58 crc kubenswrapper[4805]: I1203 14:11:58.582032 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-l6qxf" event={"ID":"e5a33c32-3873-49b8-8a81-c1a0357d94fc","Type":"ContainerStarted","Data":"899c06fc674e60b644ad74b97902e673391dad8df3ff3dfe810bd3c7b43168f3"} Dec 03 14:11:58 crc kubenswrapper[4805]: I1203 14:11:58.589197 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-k7rgc" event={"ID":"b54908df-f04f-4515-ab36-842f0aef4a99","Type":"ContainerStarted","Data":"1aebe087880b686ed928502ca004a539fe527c336faf9330357b68468b25841c"} Dec 03 14:11:58 crc kubenswrapper[4805]: I1203 14:11:58.593125 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" event={"ID":"3b6479f0-333b-4a96-9adf-2099afdc2447","Type":"ContainerStarted","Data":"d43f5cf98cd8c7b941cf88c5bec7b7bd39bc201c4e431f244dfcad5c2a964408"} Dec 03 14:11:58 crc kubenswrapper[4805]: I1203 14:11:58.593501 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-etcd-operator/etcd-operator-b45778765-2x56c"] Dec 03 14:11:58 crc kubenswrapper[4805]: I1203 14:11:58.593532 4805 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 14:11:58 crc kubenswrapper[4805]: I1203 14:11:58.595984 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication-operator/authentication-operator-69f744f599-bs6tj" event={"ID":"cdcb6b89-e405-479e-a60d-96681328634e","Type":"ContainerStarted","Data":"c51511f39836e706c4a4ab60cf3b6c5e89e74e92b6fce7f145142fd594c18b2d"} Dec 03 14:11:58 crc kubenswrapper[4805]: I1203 14:11:58.599073 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-nsz7s" event={"ID":"dfa6979a-2473-469b-ac56-9086e35a63f4","Type":"ContainerStarted","Data":"6093b96c1fb59901b9b95e4d8fc91991e09c62d3131912bd96f23954fdca62c2"} Dec 03 14:11:58 crc kubenswrapper[4805]: I1203 14:11:58.599104 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-nsz7s" event={"ID":"dfa6979a-2473-469b-ac56-9086e35a63f4","Type":"ContainerStarted","Data":"ff8b6532555301b7ddcd5390cb0fd97c161d7b3fd3aa87e3f6e8d0d6a140f457"} Dec 03 14:11:58 crc kubenswrapper[4805]: I1203 14:11:58.602764 4805 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-nsz7s" Dec 03 14:11:58 crc kubenswrapper[4805]: I1203 14:11:58.611680 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-canary/ingress-canary-cnlwr"] Dec 03 14:11:58 crc kubenswrapper[4805]: I1203 14:11:58.615406 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns-operator/dns-operator-744455d44c-rjf9m"] Dec 03 14:11:58 crc kubenswrapper[4805]: I1203 14:11:58.616610 4805 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-nsz7s" Dec 03 14:11:58 crc kubenswrapper[4805]: I1203 14:11:58.619660 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-jvsqf" event={"ID":"da68f128-1d8d-4256-a28d-a174766e1331","Type":"ContainerStarted","Data":"e3c12f009e423435140f819a7bf2cb58eba47296fda5be47e3eb32450f191c02"} Dec 03 14:11:58 crc kubenswrapper[4805]: I1203 14:11:58.623707 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-config-operator/openshift-config-operator-7777fb866f-7wzgf" event={"ID":"f4b1faf2-bf8b-483c-b28f-5885a9e8fe44","Type":"ContainerStarted","Data":"32bd5b8fb2158da83a9b964df4325dcac9356c33ad962060b23f31b4caeb5523"} Dec 03 14:11:58 crc kubenswrapper[4805]: I1203 14:11:58.626601 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/machine-api-operator-5694c8668f-jvspg" event={"ID":"dfd20bfc-5c11-4287-a4b3-df9e0ca4dcf1","Type":"ContainerStarted","Data":"88f9e67ebf81e77891b5bf0f53d48db2d210295dc030c9e52abc201c500e5061"} Dec 03 14:11:58 crc kubenswrapper[4805]: I1203 14:11:58.632269 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-vc9sn" event={"ID":"c6b0a18a-5cf2-430a-a0ef-2c2776839f11","Type":"ContainerStarted","Data":"95c7b49c35f34f4b154c020c9d75a9d34acd6eb5d88f3afb376facc7d64b5268"} Dec 03 14:11:58 crc kubenswrapper[4805]: I1203 14:11:58.635804 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-rr7fx" event={"ID":"04bd3442-dbeb-4cb4-bc9f-b07b9872231f","Type":"ContainerStarted","Data":"c30ffa4fdcc0c39382d6f8bcb6c3e134fbd786825b3b31a700c0ad37e69ab5f3"} Dec 03 14:11:58 crc kubenswrapper[4805]: I1203 14:11:58.636624 4805 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-rr7fx" Dec 03 14:11:58 crc kubenswrapper[4805]: I1203 14:11:58.638087 4805 patch_prober.go:28] interesting pod/packageserver-d55dfcdfc-rr7fx container/packageserver namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.15:5443/healthz\": dial tcp 10.217.0.15:5443: connect: connection refused" start-of-body= Dec 03 14:11:58 crc kubenswrapper[4805]: I1203 14:11:58.638116 4805 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-rr7fx" podUID="04bd3442-dbeb-4cb4-bc9f-b07b9872231f" containerName="packageserver" probeResult="failure" output="Get \"https://10.217.0.15:5443/healthz\": dial tcp 10.217.0.15:5443: connect: connection refused" Dec 03 14:11:58 crc kubenswrapper[4805]: I1203 14:11:58.638786 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-2vnrk" event={"ID":"e9022931-39f7-45cf-925f-fe19e1b870b8","Type":"ContainerStarted","Data":"3753f6e3135fc2ec58b838a54157c6f6cbbc67c4cb3d3a8547a8e9e60871c40c"} Dec 03 14:11:58 crc kubenswrapper[4805]: I1203 14:11:58.640070 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console-operator/console-operator-58897d9998-46w2g" event={"ID":"9f53fe44-0d2c-4e3d-a682-df3a779a1c8a","Type":"ContainerStarted","Data":"2fa8dc0d3b9582a0e761f47f8cc7bc420e22bb088375dee7f23e5568c389f3ab"} Dec 03 14:11:58 crc kubenswrapper[4805]: I1203 14:11:58.650270 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 14:11:58 crc kubenswrapper[4805]: E1203 14:11:58.651738 4805 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-03 14:11:59.151721665 +0000 UTC m=+148.814638588 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 14:11:58 crc kubenswrapper[4805]: I1203 14:11:58.720299 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["hostpath-provisioner/csi-hostpathplugin-krqsw"] Dec 03 14:11:58 crc kubenswrapper[4805]: I1203 14:11:58.753891 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-jjzct\" (UID: \"4655a83e-4980-4a5c-9f61-e305ae418822\") " pod="openshift-image-registry/image-registry-697d97f7c8-jjzct" Dec 03 14:11:58 crc kubenswrapper[4805]: E1203 14:11:58.755093 4805 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-03 14:11:59.255079872 +0000 UTC m=+148.917996795 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-jjzct" (UID: "4655a83e-4980-4a5c-9f61-e305ae418822") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 14:11:58 crc kubenswrapper[4805]: I1203 14:11:58.764150 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-94zfm"] Dec 03 14:11:58 crc kubenswrapper[4805]: I1203 14:11:58.854740 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 14:11:58 crc kubenswrapper[4805]: E1203 14:11:58.855161 4805 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-03 14:11:59.35513262 +0000 UTC m=+149.018049553 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 14:11:58 crc kubenswrapper[4805]: I1203 14:11:58.956738 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-jjzct\" (UID: \"4655a83e-4980-4a5c-9f61-e305ae418822\") " pod="openshift-image-registry/image-registry-697d97f7c8-jjzct" Dec 03 14:11:58 crc kubenswrapper[4805]: E1203 14:11:58.957343 4805 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-03 14:11:59.457307798 +0000 UTC m=+149.120224801 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-jjzct" (UID: "4655a83e-4980-4a5c-9f61-e305ae418822") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 14:11:59 crc kubenswrapper[4805]: I1203 14:11:59.059283 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 14:11:59 crc kubenswrapper[4805]: E1203 14:11:59.059708 4805 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-03 14:11:59.559676661 +0000 UTC m=+149.222593634 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 14:11:59 crc kubenswrapper[4805]: I1203 14:11:59.161636 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-jjzct\" (UID: \"4655a83e-4980-4a5c-9f61-e305ae418822\") " pod="openshift-image-registry/image-registry-697d97f7c8-jjzct" Dec 03 14:11:59 crc kubenswrapper[4805]: E1203 14:11:59.162218 4805 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-03 14:11:59.662206098 +0000 UTC m=+149.325123021 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-jjzct" (UID: "4655a83e-4980-4a5c-9f61-e305ae418822") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 14:11:59 crc kubenswrapper[4805]: I1203 14:11:59.278318 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 14:11:59 crc kubenswrapper[4805]: E1203 14:11:59.278621 4805 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-03 14:11:59.778588663 +0000 UTC m=+149.441505596 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 14:11:59 crc kubenswrapper[4805]: I1203 14:11:59.288669 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator/migrator-59844c95c7-h5tv7"] Dec 03 14:11:59 crc kubenswrapper[4805]: I1203 14:11:59.312908 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/catalog-operator-68c6474976-rp9lf"] Dec 03 14:11:59 crc kubenswrapper[4805]: I1203 14:11:59.386626 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-jjzct\" (UID: \"4655a83e-4980-4a5c-9f61-e305ae418822\") " pod="openshift-image-registry/image-registry-697d97f7c8-jjzct" Dec 03 14:11:59 crc kubenswrapper[4805]: E1203 14:11:59.387292 4805 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-03 14:11:59.887272296 +0000 UTC m=+149.550189219 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-jjzct" (UID: "4655a83e-4980-4a5c-9f61-e305ae418822") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 14:11:59 crc kubenswrapper[4805]: I1203 14:11:59.487537 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-ss8kj"] Dec 03 14:11:59 crc kubenswrapper[4805]: I1203 14:11:59.488249 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 14:11:59 crc kubenswrapper[4805]: E1203 14:11:59.488545 4805 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-03 14:11:59.988528702 +0000 UTC m=+149.651445625 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 14:11:59 crc kubenswrapper[4805]: I1203 14:11:59.507470 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-operator/ingress-operator-5b745b69d9-j8pwh"] Dec 03 14:11:59 crc kubenswrapper[4805]: I1203 14:11:59.530156 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns/dns-default-7wnwf"] Dec 03 14:11:59 crc kubenswrapper[4805]: W1203 14:11:59.585208 4805 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podb1313b26_f0b6_462b_8210_c19b0d4065a4.slice/crio-0cf35206d71dfa09dd5cf68e62bd8527ccff85c6dc6850a8ccc2f5f628c038a5 WatchSource:0}: Error finding container 0cf35206d71dfa09dd5cf68e62bd8527ccff85c6dc6850a8ccc2f5f628c038a5: Status 404 returned error can't find the container with id 0cf35206d71dfa09dd5cf68e62bd8527ccff85c6dc6850a8ccc2f5f628c038a5 Dec 03 14:11:59 crc kubenswrapper[4805]: I1203 14:11:59.593555 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-jjzct\" (UID: \"4655a83e-4980-4a5c-9f61-e305ae418822\") " pod="openshift-image-registry/image-registry-697d97f7c8-jjzct" Dec 03 14:11:59 crc kubenswrapper[4805]: E1203 14:11:59.593911 4805 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-03 14:12:00.093899527 +0000 UTC m=+149.756816450 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-jjzct" (UID: "4655a83e-4980-4a5c-9f61-e305ae418822") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 14:11:59 crc kubenswrapper[4805]: I1203 14:11:59.607717 4805 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-nsz7s" podStartSLOduration=129.607698303 podStartE2EDuration="2m9.607698303s" podCreationTimestamp="2025-12-03 14:09:50 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 14:11:59.569363336 +0000 UTC m=+149.232280269" watchObservedRunningTime="2025-12-03 14:11:59.607698303 +0000 UTC m=+149.270615226" Dec 03 14:11:59 crc kubenswrapper[4805]: W1203 14:11:59.610382 4805 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-poda8d74d87_86ee_4723_91bc_1d25db30909f.slice/crio-d48f2123e0e17923dfad87b9287be618af2bf0f11d6067d4cb299fcdb1a24684 WatchSource:0}: Error finding container d48f2123e0e17923dfad87b9287be618af2bf0f11d6067d4cb299fcdb1a24684: Status 404 returned error can't find the container with id d48f2123e0e17923dfad87b9287be618af2bf0f11d6067d4cb299fcdb1a24684 Dec 03 14:11:59 crc kubenswrapper[4805]: I1203 14:11:59.640484 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-b7wrl"] Dec 03 14:11:59 crc kubenswrapper[4805]: I1203 14:11:59.689179 4805 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-l6qxf" podStartSLOduration=130.689162291 podStartE2EDuration="2m10.689162291s" podCreationTimestamp="2025-12-03 14:09:49 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 14:11:59.688549657 +0000 UTC m=+149.351466580" watchObservedRunningTime="2025-12-03 14:11:59.689162291 +0000 UTC m=+149.352079214" Dec 03 14:11:59 crc kubenswrapper[4805]: I1203 14:11:59.691256 4805 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-jvsqf" podStartSLOduration=130.69121334 podStartE2EDuration="2m10.69121334s" podCreationTimestamp="2025-12-03 14:09:49 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 14:11:59.659414927 +0000 UTC m=+149.322331850" watchObservedRunningTime="2025-12-03 14:11:59.69121334 +0000 UTC m=+149.354130263" Dec 03 14:11:59 crc kubenswrapper[4805]: I1203 14:11:59.696810 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 14:11:59 crc kubenswrapper[4805]: E1203 14:11:59.697203 4805 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-03 14:12:00.197189101 +0000 UTC m=+149.860106024 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 14:11:59 crc kubenswrapper[4805]: W1203 14:11:59.697262 4805 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod220a8b3d_e5fa_486f_ad11_4fb9a43ab092.slice/crio-3e808e95bbfda6ef80d8cd544cb1483a8a840f93f77bb2363295e8f4cd92da58 WatchSource:0}: Error finding container 3e808e95bbfda6ef80d8cd544cb1483a8a840f93f77bb2363295e8f4cd92da58: Status 404 returned error can't find the container with id 3e808e95bbfda6ef80d8cd544cb1483a8a840f93f77bb2363295e8f4cd92da58 Dec 03 14:11:59 crc kubenswrapper[4805]: I1203 14:11:59.720727 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress/router-default-5444994796-kxnqm" event={"ID":"b22cbc60-5975-42a4-85d3-6add6917fd88","Type":"ContainerStarted","Data":"2335f2d3668d5ef559f3dc39da977ace5eb317b28b45628960fe2068fc8776b4"} Dec 03 14:11:59 crc kubenswrapper[4805]: I1203 14:11:59.734707 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-server-d5pjl" event={"ID":"27ffe4cd-8ab3-49dd-8e5d-91b7d2439a27","Type":"ContainerStarted","Data":"107104c76692006c85aa43c5bd97905467205ce3b2e67259c6823c1e0e06aa9f"} Dec 03 14:11:59 crc kubenswrapper[4805]: I1203 14:11:59.750724 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-k7rgc" event={"ID":"b54908df-f04f-4515-ab36-842f0aef4a99","Type":"ContainerStarted","Data":"4129f75f5d57f0e656ccb2de4e65efc62a7b3c8132a246f5fb03c630d83ee6ab"} Dec 03 14:11:59 crc kubenswrapper[4805]: I1203 14:11:59.751584 4805 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-controller-manager/controller-manager-879f6c89f-k7rgc" Dec 03 14:11:59 crc kubenswrapper[4805]: I1203 14:11:59.751813 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-86589"] Dec 03 14:11:59 crc kubenswrapper[4805]: I1203 14:11:59.753510 4805 patch_prober.go:28] interesting pod/controller-manager-879f6c89f-k7rgc container/controller-manager namespace/openshift-controller-manager: Readiness probe status=failure output="Get \"https://10.217.0.7:8443/healthz\": dial tcp 10.217.0.7:8443: connect: connection refused" start-of-body= Dec 03 14:11:59 crc kubenswrapper[4805]: I1203 14:11:59.753545 4805 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-controller-manager/controller-manager-879f6c89f-k7rgc" podUID="b54908df-f04f-4515-ab36-842f0aef4a99" containerName="controller-manager" probeResult="failure" output="Get \"https://10.217.0.7:8443/healthz\": dial tcp 10.217.0.7:8443: connect: connection refused" Dec 03 14:11:59 crc kubenswrapper[4805]: I1203 14:11:59.753691 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-jvsqf" event={"ID":"da68f128-1d8d-4256-a28d-a174766e1331","Type":"ContainerStarted","Data":"5ab18f7ea154290fdafe167a48293ee7c7f39099f0a3bc1438f755d827d47ffd"} Dec 03 14:11:59 crc kubenswrapper[4805]: I1203 14:11:59.798441 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-jjzct\" (UID: \"4655a83e-4980-4a5c-9f61-e305ae418822\") " pod="openshift-image-registry/image-registry-697d97f7c8-jjzct" Dec 03 14:11:59 crc kubenswrapper[4805]: I1203 14:11:59.798726 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29412840-gj7pr" event={"ID":"8b954d96-a62c-434e-a340-8446d778c545","Type":"ContainerStarted","Data":"8f4f233b53a4e215b29e4cb9baef3ccf262fa81a4f6bddc1a28d3a846ab83332"} Dec 03 14:11:59 crc kubenswrapper[4805]: I1203 14:11:59.798820 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29412840-gj7pr" event={"ID":"8b954d96-a62c-434e-a340-8446d778c545","Type":"ContainerStarted","Data":"3d3293741e2741cb9b99650b37e0faa09813046145b469055b1661244883e262"} Dec 03 14:11:59 crc kubenswrapper[4805]: E1203 14:11:59.799810 4805 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-03 14:12:00.2997897 +0000 UTC m=+149.962706683 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-jjzct" (UID: "4655a83e-4980-4a5c-9f61-e305ae418822") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 14:11:59 crc kubenswrapper[4805]: I1203 14:11:59.819989 4805 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-ingress/router-default-5444994796-kxnqm" Dec 03 14:11:59 crc kubenswrapper[4805]: I1203 14:11:59.827222 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication-operator/authentication-operator-69f744f599-bs6tj" event={"ID":"cdcb6b89-e405-479e-a60d-96681328634e","Type":"ContainerStarted","Data":"1c90349f83c2bbeec3ff5e0c34fd8a0ef97c25e781886bd7cc8be59c7481c2fb"} Dec 03 14:11:59 crc kubenswrapper[4805]: I1203 14:11:59.830177 4805 patch_prober.go:28] interesting pod/router-default-5444994796-kxnqm container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 03 14:11:59 crc kubenswrapper[4805]: [-]has-synced failed: reason withheld Dec 03 14:11:59 crc kubenswrapper[4805]: [+]process-running ok Dec 03 14:11:59 crc kubenswrapper[4805]: healthz check failed Dec 03 14:11:59 crc kubenswrapper[4805]: I1203 14:11:59.830220 4805 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-kxnqm" podUID="b22cbc60-5975-42a4-85d3-6add6917fd88" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 03 14:11:59 crc kubenswrapper[4805]: I1203 14:11:59.837433 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-krqsw" event={"ID":"8801ff25-d273-4354-ab39-16e979d0b73d","Type":"ContainerStarted","Data":"1889b7ffa519324b4444ac16456a78fbf2663488b440e124a52c9c7daf8c368b"} Dec 03 14:11:59 crc kubenswrapper[4805]: I1203 14:11:59.863226 4805 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-authentication-operator/authentication-operator-69f744f599-bs6tj" podStartSLOduration=130.863200621 podStartE2EDuration="2m10.863200621s" podCreationTimestamp="2025-12-03 14:09:49 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 14:11:59.819165209 +0000 UTC m=+149.482082132" watchObservedRunningTime="2025-12-03 14:11:59.863200621 +0000 UTC m=+149.526117544" Dec 03 14:11:59 crc kubenswrapper[4805]: I1203 14:11:59.880056 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-spjlp" event={"ID":"66812e7e-aa85-4045-8e7a-8967b84f8849","Type":"ContainerStarted","Data":"ab75814bb815ba66757f696e18a982e87da1a1e3fa65e8f535b0363cc08e85b8"} Dec 03 14:11:59 crc kubenswrapper[4805]: I1203 14:11:59.882470 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-rr7fx" event={"ID":"04bd3442-dbeb-4cb4-bc9f-b07b9872231f","Type":"ContainerStarted","Data":"03b72a21f379f9ede07a434dd0ecfeabac852138853e3e2c8a0054db5b6692f4"} Dec 03 14:11:59 crc kubenswrapper[4805]: I1203 14:11:59.886153 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-t4kl8" event={"ID":"e600aefb-0839-423f-8ac8-1916400cfa92","Type":"ContainerStarted","Data":"fd4de0f93d9519ea804c3fa09afa89d29ddce67f9175b7a0f1885168c3861f11"} Dec 03 14:11:59 crc kubenswrapper[4805]: I1203 14:11:59.891080 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-25khc"] Dec 03 14:11:59 crc kubenswrapper[4805]: I1203 14:11:59.891136 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-6n5xz"] Dec 03 14:11:59 crc kubenswrapper[4805]: I1203 14:11:59.899456 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 14:11:59 crc kubenswrapper[4805]: E1203 14:11:59.900547 4805 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-03 14:12:00.400530654 +0000 UTC m=+150.063447577 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 14:11:59 crc kubenswrapper[4805]: I1203 14:11:59.901237 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-vc9sn" event={"ID":"c6b0a18a-5cf2-430a-a0ef-2c2776839f11","Type":"ContainerStarted","Data":"42d3b0d2d070be057104cc04d6c80cbbe30f071dbcb0d01b791a0851735477e0"} Dec 03 14:11:59 crc kubenswrapper[4805]: I1203 14:11:59.903267 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-g46rj"] Dec 03 14:11:59 crc kubenswrapper[4805]: I1203 14:11:59.905960 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns-operator/dns-operator-744455d44c-rjf9m" event={"ID":"8e437287-3b92-4c6a-abe9-84c786ba3220","Type":"ContainerStarted","Data":"b9bdff669b800184cd29490e5c9521dd6ad484541472aa5d069e63090dec700a"} Dec 03 14:11:59 crc kubenswrapper[4805]: I1203 14:11:59.913579 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-kvmpp" event={"ID":"d5ea0764-4290-4eab-b4ce-dbd4909455c1","Type":"ContainerStarted","Data":"139a5c679766a5bb66604143748d1c1d4788add2b74c9db316c16f04beb28044"} Dec 03 14:11:59 crc kubenswrapper[4805]: I1203 14:11:59.916807 4805 generic.go:334] "Generic (PLEG): container finished" podID="f4b1faf2-bf8b-483c-b28f-5885a9e8fe44" containerID="347ffc4398cf93310590b9bf02cc720bb723bc77ecfd4f8262cd0b08711e7a60" exitCode=0 Dec 03 14:11:59 crc kubenswrapper[4805]: I1203 14:11:59.916919 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-config-operator/openshift-config-operator-7777fb866f-7wzgf" event={"ID":"f4b1faf2-bf8b-483c-b28f-5885a9e8fe44","Type":"ContainerDied","Data":"347ffc4398cf93310590b9bf02cc720bb723bc77ecfd4f8262cd0b08711e7a60"} Dec 03 14:11:59 crc kubenswrapper[4805]: I1203 14:11:59.920062 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/downloads-7954f5f757-s25zl" event={"ID":"6c2d6fbc-771d-4aff-b1e8-3eda17204931","Type":"ContainerStarted","Data":"2a4f988987b094af15b577199c7b9e968d63dd68fb871eaf29fae499db22c947"} Dec 03 14:11:59 crc kubenswrapper[4805]: I1203 14:11:59.920765 4805 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console/downloads-7954f5f757-s25zl" Dec 03 14:11:59 crc kubenswrapper[4805]: I1203 14:11:59.922942 4805 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-rr7fx" podStartSLOduration=129.922930535 podStartE2EDuration="2m9.922930535s" podCreationTimestamp="2025-12-03 14:09:50 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 14:11:59.919244628 +0000 UTC m=+149.582161551" watchObservedRunningTime="2025-12-03 14:11:59.922930535 +0000 UTC m=+149.585847448" Dec 03 14:11:59 crc kubenswrapper[4805]: I1203 14:11:59.925114 4805 patch_prober.go:28] interesting pod/downloads-7954f5f757-s25zl container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.11:8080/\": dial tcp 10.217.0.11:8080: connect: connection refused" start-of-body= Dec 03 14:11:59 crc kubenswrapper[4805]: I1203 14:11:59.925161 4805 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-s25zl" podUID="6c2d6fbc-771d-4aff-b1e8-3eda17204931" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.11:8080/\": dial tcp 10.217.0.11:8080: connect: connection refused" Dec 03 14:11:59 crc kubenswrapper[4805]: I1203 14:11:59.956200 4805 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-rr7fx" Dec 03 14:11:59 crc kubenswrapper[4805]: I1203 14:11:59.970523 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console-operator/console-operator-58897d9998-46w2g" event={"ID":"9f53fe44-0d2c-4e3d-a682-df3a779a1c8a","Type":"ContainerStarted","Data":"1428ce93164ad4dc1bfd2e67e6ce37bb09a33cc2972bc1a38d30584e83875966"} Dec 03 14:11:59 crc kubenswrapper[4805]: I1203 14:11:59.970772 4805 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console-operator/console-operator-58897d9998-46w2g" Dec 03 14:11:59 crc kubenswrapper[4805]: I1203 14:11:59.977251 4805 patch_prober.go:28] interesting pod/console-operator-58897d9998-46w2g container/console-operator namespace/openshift-console-operator: Readiness probe status=failure output="Get \"https://10.217.0.13:8443/readyz\": dial tcp 10.217.0.13:8443: connect: connection refused" start-of-body= Dec 03 14:11:59 crc kubenswrapper[4805]: I1203 14:11:59.977313 4805 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console-operator/console-operator-58897d9998-46w2g" podUID="9f53fe44-0d2c-4e3d-a682-df3a779a1c8a" containerName="console-operator" probeResult="failure" output="Get \"https://10.217.0.13:8443/readyz\": dial tcp 10.217.0.13:8443: connect: connection refused" Dec 03 14:11:59 crc kubenswrapper[4805]: I1203 14:11:59.992105 4805 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-2vnrk" podStartSLOduration=129.992080522 podStartE2EDuration="2m9.992080522s" podCreationTimestamp="2025-12-03 14:09:50 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 14:11:59.959622473 +0000 UTC m=+149.622539396" watchObservedRunningTime="2025-12-03 14:11:59.992080522 +0000 UTC m=+149.654997445" Dec 03 14:11:59 crc kubenswrapper[4805]: I1203 14:11:59.998330 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-94zfm" event={"ID":"5cec10a3-ed60-4193-be30-a78cb80bd51d","Type":"ContainerStarted","Data":"8c96098011243113dc16911e505366bef1f2516fdb19fbc758e700367e19a820"} Dec 03 14:12:00 crc kubenswrapper[4805]: I1203 14:12:00.000870 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-jjzct\" (UID: \"4655a83e-4980-4a5c-9f61-e305ae418822\") " pod="openshift-image-registry/image-registry-697d97f7c8-jjzct" Dec 03 14:12:00 crc kubenswrapper[4805]: E1203 14:12:00.004817 4805 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-03 14:12:00.504803503 +0000 UTC m=+150.167720426 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-jjzct" (UID: "4655a83e-4980-4a5c-9f61-e305ae418822") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 14:12:00 crc kubenswrapper[4805]: I1203 14:12:00.019875 4805 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ingress/router-default-5444994796-kxnqm" podStartSLOduration=130.019855439 podStartE2EDuration="2m10.019855439s" podCreationTimestamp="2025-12-03 14:09:50 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 14:12:00.01779537 +0000 UTC m=+149.680712293" watchObservedRunningTime="2025-12-03 14:12:00.019855439 +0000 UTC m=+149.682772352" Dec 03 14:12:00 crc kubenswrapper[4805]: I1203 14:12:00.031781 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-7nhvl" event={"ID":"ca425595-aae1-44f2-9470-84ba52167748","Type":"ContainerStarted","Data":"2aaeb949f9bcfe32c5be0564a05918aebaa074069c27ab939c5a1c55d6e8a256"} Dec 03 14:12:00 crc kubenswrapper[4805]: I1203 14:12:00.037257 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-canary/ingress-canary-cnlwr" event={"ID":"6be05973-dc60-437f-959c-a9309cd16d5b","Type":"ContainerStarted","Data":"44dea830099691f3b6109bd5e3497529945f520cd41038f102e2fc557c37bad9"} Dec 03 14:12:00 crc kubenswrapper[4805]: I1203 14:12:00.050628 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca/service-ca-9c57cc56f-dbfrd" event={"ID":"1c55f37c-24b9-4800-9050-e0da883df2c0","Type":"ContainerStarted","Data":"278af814aa6fe80b47fd0a90cfe3779e6eab8c62875ffc864219465aa7a3dfee"} Dec 03 14:12:00 crc kubenswrapper[4805]: I1203 14:12:00.053057 4805 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager/controller-manager-879f6c89f-k7rgc" podStartSLOduration=131.053044765 podStartE2EDuration="2m11.053044765s" podCreationTimestamp="2025-12-03 14:09:49 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 14:12:00.051933848 +0000 UTC m=+149.714850771" watchObservedRunningTime="2025-12-03 14:12:00.053044765 +0000 UTC m=+149.715961688" Dec 03 14:12:00 crc kubenswrapper[4805]: I1203 14:12:00.072988 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-ss8kj" event={"ID":"776b097e-ad39-49a4-aa4f-a326a04bc30b","Type":"ContainerStarted","Data":"7e81cea2bfda1d2cd999ab69f7141f6f40eae93d7c9d4ad17c4bca0a9f67a77f"} Dec 03 14:12:00 crc kubenswrapper[4805]: I1203 14:12:00.086218 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-f9d7485db-pp9tv" event={"ID":"e2034338-8a35-4fd2-9bbf-be733939fc25","Type":"ContainerStarted","Data":"2a23c6b4702b54967793e4d45488577f06641ef8826624873c784933f74405a3"} Dec 03 14:12:00 crc kubenswrapper[4805]: I1203 14:12:00.110420 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 14:12:00 crc kubenswrapper[4805]: E1203 14:12:00.110677 4805 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-03 14:12:00.610656688 +0000 UTC m=+150.273573611 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 14:12:00 crc kubenswrapper[4805]: I1203 14:12:00.110933 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-jjzct\" (UID: \"4655a83e-4980-4a5c-9f61-e305ae418822\") " pod="openshift-image-registry/image-registry-697d97f7c8-jjzct" Dec 03 14:12:00 crc kubenswrapper[4805]: E1203 14:12:00.112657 4805 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-03 14:12:00.612649395 +0000 UTC m=+150.275566318 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-jjzct" (UID: "4655a83e-4980-4a5c-9f61-e305ae418822") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 14:12:00 crc kubenswrapper[4805]: I1203 14:12:00.143939 4805 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/collect-profiles-29412840-gj7pr" podStartSLOduration=131.143919786 podStartE2EDuration="2m11.143919786s" podCreationTimestamp="2025-12-03 14:09:49 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 14:12:00.103618202 +0000 UTC m=+149.766535125" watchObservedRunningTime="2025-12-03 14:12:00.143919786 +0000 UTC m=+149.806836709" Dec 03 14:12:00 crc kubenswrapper[4805]: I1203 14:12:00.145781 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-rp9lf" event={"ID":"a8d74d87-86ee-4723-91bc-1d25db30909f","Type":"ContainerStarted","Data":"d48f2123e0e17923dfad87b9287be618af2bf0f11d6067d4cb299fcdb1a24684"} Dec 03 14:12:00 crc kubenswrapper[4805]: I1203 14:12:00.198523 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca-operator/service-ca-operator-777779d784-z7lgj" event={"ID":"9b932fa5-f87b-46e1-8987-683cfd3b46fd","Type":"ContainerStarted","Data":"8892098b4b4bfe58379bb7f8f97902a5481645e5f09208f27032836ff266b41f"} Dec 03 14:12:00 crc kubenswrapper[4805]: I1203 14:12:00.201126 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/machine-api-operator-5694c8668f-jvspg" event={"ID":"dfd20bfc-5c11-4287-a4b3-df9e0ca4dcf1","Type":"ContainerStarted","Data":"3c4f5d33a626dd0dce3eecdacb44fea6e460a9619afd4b8edfbd50fda2b7d7af"} Dec 03 14:12:00 crc kubenswrapper[4805]: I1203 14:12:00.215336 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 14:12:00 crc kubenswrapper[4805]: E1203 14:12:00.216453 4805 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-03 14:12:00.716435402 +0000 UTC m=+150.379352335 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 14:12:00 crc kubenswrapper[4805]: I1203 14:12:00.225170 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-nv2m9" event={"ID":"00941575-52e9-482d-b256-1b1c757d96e0","Type":"ContainerStarted","Data":"184d48be47301e02bf34a60dcdf73aa214b03f765b698f759fff9a34f3f49e3a"} Dec 03 14:12:00 crc kubenswrapper[4805]: I1203 14:12:00.256016 4805 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console-operator/console-operator-58897d9998-46w2g" podStartSLOduration=131.255998379 podStartE2EDuration="2m11.255998379s" podCreationTimestamp="2025-12-03 14:09:49 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 14:12:00.255298882 +0000 UTC m=+149.918215805" watchObservedRunningTime="2025-12-03 14:12:00.255998379 +0000 UTC m=+149.918915302" Dec 03 14:12:00 crc kubenswrapper[4805]: I1203 14:12:00.276120 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-admission-controller-857f4d67dd-tn8tx" event={"ID":"33c718fb-affd-410b-9f53-ddb68ba9760c","Type":"ContainerStarted","Data":"bdf87c596cc8815a016cb3930d8d8f2f26cf218c5f15fd2f4100a952c4f2169a"} Dec 03 14:12:00 crc kubenswrapper[4805]: I1203 14:12:00.279629 4805 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console/console-f9d7485db-pp9tv" podStartSLOduration=131.279617828 podStartE2EDuration="2m11.279617828s" podCreationTimestamp="2025-12-03 14:09:49 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 14:12:00.278126792 +0000 UTC m=+149.941043705" watchObservedRunningTime="2025-12-03 14:12:00.279617828 +0000 UTC m=+149.942534751" Dec 03 14:12:00 crc kubenswrapper[4805]: I1203 14:12:00.289240 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-h5tv7" event={"ID":"4aa1a2c6-8700-4f70-b804-f57371df7025","Type":"ContainerStarted","Data":"91dd97a68232abe0895d6009c059111279c9c01ba35a70f72dd5a8c76db2d5e6"} Dec 03 14:12:00 crc kubenswrapper[4805]: I1203 14:12:00.301479 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/dns-default-7wnwf" event={"ID":"b1313b26-f0b6-462b-8210-c19b0d4065a4","Type":"ContainerStarted","Data":"0cf35206d71dfa09dd5cf68e62bd8527ccff85c6dc6850a8ccc2f5f628c038a5"} Dec 03 14:12:00 crc kubenswrapper[4805]: I1203 14:12:00.308417 4805 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console/downloads-7954f5f757-s25zl" podStartSLOduration=131.308400439 podStartE2EDuration="2m11.308400439s" podCreationTimestamp="2025-12-03 14:09:49 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 14:12:00.306866433 +0000 UTC m=+149.969783356" watchObservedRunningTime="2025-12-03 14:12:00.308400439 +0000 UTC m=+149.971317362" Dec 03 14:12:00 crc kubenswrapper[4805]: I1203 14:12:00.316438 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd-operator/etcd-operator-b45778765-2x56c" event={"ID":"ae561e71-047a-4cf7-8378-d08adb458d9c","Type":"ContainerStarted","Data":"7c9defffdfa8a56ce16a38acd3bd3a98cca55488438d1e1d0dfac356083f612c"} Dec 03 14:12:00 crc kubenswrapper[4805]: I1203 14:12:00.318057 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-jjzct\" (UID: \"4655a83e-4980-4a5c-9f61-e305ae418822\") " pod="openshift-image-registry/image-registry-697d97f7c8-jjzct" Dec 03 14:12:00 crc kubenswrapper[4805]: E1203 14:12:00.319316 4805 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-03 14:12:00.819299757 +0000 UTC m=+150.482216670 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-jjzct" (UID: "4655a83e-4980-4a5c-9f61-e305ae418822") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 14:12:00 crc kubenswrapper[4805]: I1203 14:12:00.338934 4805 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-nv2m9" podStartSLOduration=131.338915081 podStartE2EDuration="2m11.338915081s" podCreationTimestamp="2025-12-03 14:09:49 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 14:12:00.338832499 +0000 UTC m=+150.001749422" watchObservedRunningTime="2025-12-03 14:12:00.338915081 +0000 UTC m=+150.001832004" Dec 03 14:12:00 crc kubenswrapper[4805]: I1203 14:12:00.370442 4805 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-service-ca-operator/service-ca-operator-777779d784-z7lgj" podStartSLOduration=130.370426137 podStartE2EDuration="2m10.370426137s" podCreationTimestamp="2025-12-03 14:09:50 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 14:12:00.369053625 +0000 UTC m=+150.031970548" watchObservedRunningTime="2025-12-03 14:12:00.370426137 +0000 UTC m=+150.033343060" Dec 03 14:12:00 crc kubenswrapper[4805]: I1203 14:12:00.419155 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 14:12:00 crc kubenswrapper[4805]: E1203 14:12:00.420185 4805 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-03 14:12:00.920169214 +0000 UTC m=+150.583086137 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 14:12:00 crc kubenswrapper[4805]: I1203 14:12:00.455641 4805 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/multus-admission-controller-857f4d67dd-tn8tx" podStartSLOduration=130.455623844 podStartE2EDuration="2m10.455623844s" podCreationTimestamp="2025-12-03 14:09:50 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 14:12:00.415061224 +0000 UTC m=+150.077978147" watchObservedRunningTime="2025-12-03 14:12:00.455623844 +0000 UTC m=+150.118540767" Dec 03 14:12:00 crc kubenswrapper[4805]: I1203 14:12:00.521321 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-jjzct\" (UID: \"4655a83e-4980-4a5c-9f61-e305ae418822\") " pod="openshift-image-registry/image-registry-697d97f7c8-jjzct" Dec 03 14:12:00 crc kubenswrapper[4805]: E1203 14:12:00.521624 4805 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-03 14:12:01.021613166 +0000 UTC m=+150.684530089 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-jjzct" (UID: "4655a83e-4980-4a5c-9f61-e305ae418822") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 14:12:00 crc kubenswrapper[4805]: I1203 14:12:00.623024 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 14:12:00 crc kubenswrapper[4805]: E1203 14:12:00.623593 4805 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-03 14:12:01.123578619 +0000 UTC m=+150.786495542 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 14:12:00 crc kubenswrapper[4805]: I1203 14:12:00.724454 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-jjzct\" (UID: \"4655a83e-4980-4a5c-9f61-e305ae418822\") " pod="openshift-image-registry/image-registry-697d97f7c8-jjzct" Dec 03 14:12:00 crc kubenswrapper[4805]: E1203 14:12:00.724790 4805 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-03 14:12:01.224777915 +0000 UTC m=+150.887694838 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-jjzct" (UID: "4655a83e-4980-4a5c-9f61-e305ae418822") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 14:12:00 crc kubenswrapper[4805]: I1203 14:12:00.746952 4805 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-etcd-operator/etcd-operator-b45778765-2x56c" podStartSLOduration=130.746817386 podStartE2EDuration="2m10.746817386s" podCreationTimestamp="2025-12-03 14:09:50 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 14:12:00.457131269 +0000 UTC m=+150.120048192" watchObservedRunningTime="2025-12-03 14:12:00.746817386 +0000 UTC m=+150.409734309" Dec 03 14:12:00 crc kubenswrapper[4805]: I1203 14:12:00.821195 4805 patch_prober.go:28] interesting pod/router-default-5444994796-kxnqm container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 03 14:12:00 crc kubenswrapper[4805]: [-]has-synced failed: reason withheld Dec 03 14:12:00 crc kubenswrapper[4805]: [+]process-running ok Dec 03 14:12:00 crc kubenswrapper[4805]: healthz check failed Dec 03 14:12:00 crc kubenswrapper[4805]: I1203 14:12:00.821243 4805 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-kxnqm" podUID="b22cbc60-5975-42a4-85d3-6add6917fd88" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 03 14:12:00 crc kubenswrapper[4805]: I1203 14:12:00.826314 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 14:12:00 crc kubenswrapper[4805]: E1203 14:12:00.826751 4805 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-03 14:12:01.326734418 +0000 UTC m=+150.989651341 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 14:12:00 crc kubenswrapper[4805]: I1203 14:12:00.927460 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-jjzct\" (UID: \"4655a83e-4980-4a5c-9f61-e305ae418822\") " pod="openshift-image-registry/image-registry-697d97f7c8-jjzct" Dec 03 14:12:00 crc kubenswrapper[4805]: E1203 14:12:00.928168 4805 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-03 14:12:01.428157339 +0000 UTC m=+151.091074262 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-jjzct" (UID: "4655a83e-4980-4a5c-9f61-e305ae418822") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 14:12:01 crc kubenswrapper[4805]: I1203 14:12:01.029234 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 14:12:01 crc kubenswrapper[4805]: E1203 14:12:01.029597 4805 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-03 14:12:01.5295831 +0000 UTC m=+151.192500023 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 14:12:01 crc kubenswrapper[4805]: I1203 14:12:01.130406 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-jjzct\" (UID: \"4655a83e-4980-4a5c-9f61-e305ae418822\") " pod="openshift-image-registry/image-registry-697d97f7c8-jjzct" Dec 03 14:12:01 crc kubenswrapper[4805]: E1203 14:12:01.131065 4805 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-03 14:12:01.631053661 +0000 UTC m=+151.293970584 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-jjzct" (UID: "4655a83e-4980-4a5c-9f61-e305ae418822") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 14:12:01 crc kubenswrapper[4805]: I1203 14:12:01.231425 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 14:12:01 crc kubenswrapper[4805]: E1203 14:12:01.231738 4805 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-03 14:12:01.731722574 +0000 UTC m=+151.394639497 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 14:12:01 crc kubenswrapper[4805]: I1203 14:12:01.342383 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-jjzct\" (UID: \"4655a83e-4980-4a5c-9f61-e305ae418822\") " pod="openshift-image-registry/image-registry-697d97f7c8-jjzct" Dec 03 14:12:01 crc kubenswrapper[4805]: E1203 14:12:01.343030 4805 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-03 14:12:01.843015758 +0000 UTC m=+151.505932681 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-jjzct" (UID: "4655a83e-4980-4a5c-9f61-e305ae418822") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 14:12:01 crc kubenswrapper[4805]: I1203 14:12:01.393903 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-25khc" event={"ID":"9bd2f61c-2b92-488f-b632-1e627b48a518","Type":"ContainerStarted","Data":"ba786c5132c0ac9a4dd5e7f522874746ef5a24fdf4abcfde52463f35828c80ea"} Dec 03 14:12:01 crc kubenswrapper[4805]: I1203 14:12:01.393952 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-25khc" event={"ID":"9bd2f61c-2b92-488f-b632-1e627b48a518","Type":"ContainerStarted","Data":"9554cffe9a8e506ae8fd1dbcfc6ae1ad9b34d50a4fc7ac0a7ffc0cdddf95674c"} Dec 03 14:12:01 crc kubenswrapper[4805]: I1203 14:12:01.404759 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns-operator/dns-operator-744455d44c-rjf9m" event={"ID":"8e437287-3b92-4c6a-abe9-84c786ba3220","Type":"ContainerStarted","Data":"2d1b97a9907a2867f9780265aad7b51f59e8eaef44ee38c1464556c9ea5de397"} Dec 03 14:12:01 crc kubenswrapper[4805]: I1203 14:12:01.447436 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 14:12:01 crc kubenswrapper[4805]: E1203 14:12:01.447694 4805 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-03 14:12:01.947675166 +0000 UTC m=+151.610592089 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 14:12:01 crc kubenswrapper[4805]: I1203 14:12:01.449169 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/machine-api-operator-5694c8668f-jvspg" event={"ID":"dfd20bfc-5c11-4287-a4b3-df9e0ca4dcf1","Type":"ContainerStarted","Data":"1e7bda6f96e1aba6703afe2817fea3ad34fb235902bfb7aeaa5d047eb9a4fa0d"} Dec 03 14:12:01 crc kubenswrapper[4805]: I1203 14:12:01.490288 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca/service-ca-9c57cc56f-dbfrd" event={"ID":"1c55f37c-24b9-4800-9050-e0da883df2c0","Type":"ContainerStarted","Data":"ef1b7511d30e8a587ef377ffdef26ff544c2bae5fdc41cdb2e5d0a3078e9fbe7"} Dec 03 14:12:01 crc kubenswrapper[4805]: I1203 14:12:01.494162 4805 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-25khc" podStartSLOduration=131.494141256 podStartE2EDuration="2m11.494141256s" podCreationTimestamp="2025-12-03 14:09:50 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 14:12:01.439659596 +0000 UTC m=+151.102576519" watchObservedRunningTime="2025-12-03 14:12:01.494141256 +0000 UTC m=+151.157058189" Dec 03 14:12:01 crc kubenswrapper[4805]: I1203 14:12:01.495432 4805 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-api/machine-api-operator-5694c8668f-jvspg" podStartSLOduration=131.495426716 podStartE2EDuration="2m11.495426716s" podCreationTimestamp="2025-12-03 14:09:50 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 14:12:01.491208316 +0000 UTC m=+151.154125249" watchObservedRunningTime="2025-12-03 14:12:01.495426716 +0000 UTC m=+151.158343649" Dec 03 14:12:01 crc kubenswrapper[4805]: I1203 14:12:01.517215 4805 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-service-ca/service-ca-9c57cc56f-dbfrd" podStartSLOduration=131.517191461 podStartE2EDuration="2m11.517191461s" podCreationTimestamp="2025-12-03 14:09:50 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 14:12:01.511719122 +0000 UTC m=+151.174636045" watchObservedRunningTime="2025-12-03 14:12:01.517191461 +0000 UTC m=+151.180108384" Dec 03 14:12:01 crc kubenswrapper[4805]: I1203 14:12:01.548421 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-jjzct\" (UID: \"4655a83e-4980-4a5c-9f61-e305ae418822\") " pod="openshift-image-registry/image-registry-697d97f7c8-jjzct" Dec 03 14:12:01 crc kubenswrapper[4805]: E1203 14:12:01.548717 4805 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-03 14:12:02.048704417 +0000 UTC m=+151.711621340 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-jjzct" (UID: "4655a83e-4980-4a5c-9f61-e305ae418822") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 14:12:01 crc kubenswrapper[4805]: I1203 14:12:01.561559 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-config-operator/openshift-config-operator-7777fb866f-7wzgf" event={"ID":"f4b1faf2-bf8b-483c-b28f-5885a9e8fe44","Type":"ContainerStarted","Data":"f13ecff1f7b341d8b2a0e49f1cdd482c9bc65e9823aac9aba37709b0ebce02f5"} Dec 03 14:12:01 crc kubenswrapper[4805]: I1203 14:12:01.561774 4805 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-config-operator/openshift-config-operator-7777fb866f-7wzgf" Dec 03 14:12:01 crc kubenswrapper[4805]: I1203 14:12:01.586896 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-t4kl8" event={"ID":"e600aefb-0839-423f-8ac8-1916400cfa92","Type":"ContainerStarted","Data":"f2079938383ad75ff7bf9ecdf18a161542ff1945498f88884592fe975c017336"} Dec 03 14:12:01 crc kubenswrapper[4805]: I1203 14:12:01.633340 4805 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-config-operator/openshift-config-operator-7777fb866f-7wzgf" podStartSLOduration=132.63332385 podStartE2EDuration="2m12.63332385s" podCreationTimestamp="2025-12-03 14:09:49 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 14:12:01.631365484 +0000 UTC m=+151.294282407" watchObservedRunningTime="2025-12-03 14:12:01.63332385 +0000 UTC m=+151.296240773" Dec 03 14:12:01 crc kubenswrapper[4805]: I1203 14:12:01.647687 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-j8pwh" event={"ID":"220a8b3d-e5fa-486f-ad11-4fb9a43ab092","Type":"ContainerStarted","Data":"d182d68e4d44d1e4c1a020a132124b918aef3c44399bd2eee5382babe9c90601"} Dec 03 14:12:01 crc kubenswrapper[4805]: I1203 14:12:01.647726 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-j8pwh" event={"ID":"220a8b3d-e5fa-486f-ad11-4fb9a43ab092","Type":"ContainerStarted","Data":"3e808e95bbfda6ef80d8cd544cb1483a8a840f93f77bb2363295e8f4cd92da58"} Dec 03 14:12:01 crc kubenswrapper[4805]: I1203 14:12:01.649725 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 14:12:01 crc kubenswrapper[4805]: E1203 14:12:01.650525 4805 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-03 14:12:02.150494647 +0000 UTC m=+151.813411580 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 14:12:01 crc kubenswrapper[4805]: I1203 14:12:01.659802 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-krqsw" event={"ID":"8801ff25-d273-4354-ab39-16e979d0b73d","Type":"ContainerStarted","Data":"d953cc0b2ec94513a37aeb4c55db612d570e5bc1c34cb01b41c6c5c549d4a77e"} Dec 03 14:12:01 crc kubenswrapper[4805]: I1203 14:12:01.708502 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-2vnrk" event={"ID":"e9022931-39f7-45cf-925f-fe19e1b870b8","Type":"ContainerStarted","Data":"4ccbf3ce66da2a9d735c913676bf3ea7c25d11e4d78d486b90b0b112ccba767c"} Dec 03 14:12:01 crc kubenswrapper[4805]: I1203 14:12:01.739007 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-server-d5pjl" event={"ID":"27ffe4cd-8ab3-49dd-8e5d-91b7d2439a27","Type":"ContainerStarted","Data":"41fb941dd9eb2e76c743c6a4eb1bb9dc53786fae14bc41211da8bd12cac91f4b"} Dec 03 14:12:01 crc kubenswrapper[4805]: I1203 14:12:01.751226 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-jjzct\" (UID: \"4655a83e-4980-4a5c-9f61-e305ae418822\") " pod="openshift-image-registry/image-registry-697d97f7c8-jjzct" Dec 03 14:12:01 crc kubenswrapper[4805]: E1203 14:12:01.752734 4805 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-03 14:12:02.252719406 +0000 UTC m=+151.915636329 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-jjzct" (UID: "4655a83e-4980-4a5c-9f61-e305ae418822") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 14:12:01 crc kubenswrapper[4805]: I1203 14:12:01.771037 4805 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-server-d5pjl" podStartSLOduration=7.7710166990000005 podStartE2EDuration="7.771016699s" podCreationTimestamp="2025-12-03 14:11:54 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 14:12:01.769294879 +0000 UTC m=+151.432211822" watchObservedRunningTime="2025-12-03 14:12:01.771016699 +0000 UTC m=+151.433933622" Dec 03 14:12:01 crc kubenswrapper[4805]: I1203 14:12:01.782566 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-ss8kj" event={"ID":"776b097e-ad39-49a4-aa4f-a326a04bc30b","Type":"ContainerStarted","Data":"998125fbb2ced587edb865da9b4501e1aac5eb0453cbd4844454b5add36ecd1c"} Dec 03 14:12:01 crc kubenswrapper[4805]: I1203 14:12:01.784506 4805 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-ss8kj" Dec 03 14:12:01 crc kubenswrapper[4805]: I1203 14:12:01.785424 4805 patch_prober.go:28] interesting pod/olm-operator-6b444d44fb-ss8kj container/olm-operator namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.23:8443/healthz\": dial tcp 10.217.0.23:8443: connect: connection refused" start-of-body= Dec 03 14:12:01 crc kubenswrapper[4805]: I1203 14:12:01.785532 4805 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-ss8kj" podUID="776b097e-ad39-49a4-aa4f-a326a04bc30b" containerName="olm-operator" probeResult="failure" output="Get \"https://10.217.0.23:8443/healthz\": dial tcp 10.217.0.23:8443: connect: connection refused" Dec 03 14:12:01 crc kubenswrapper[4805]: I1203 14:12:01.796801 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-b7wrl" event={"ID":"b44fb2fa-6155-40fc-b5f7-eb5d1fa391bb","Type":"ContainerStarted","Data":"fad91d8628c612766239dab976180a049af721cbbc46658f2e9c99a1212a8b1d"} Dec 03 14:12:01 crc kubenswrapper[4805]: I1203 14:12:01.796873 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-b7wrl" event={"ID":"b44fb2fa-6155-40fc-b5f7-eb5d1fa391bb","Type":"ContainerStarted","Data":"eb0ff9cbd7c68e53b3f3d95a0f014a30d9fd77c43227a4dd6c189c29c15b17be"} Dec 03 14:12:01 crc kubenswrapper[4805]: I1203 14:12:01.806414 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-rp9lf" event={"ID":"a8d74d87-86ee-4723-91bc-1d25db30909f","Type":"ContainerStarted","Data":"4312abec25c1e4cfb61dd11cec2827bb5faec7b0798065371dd582ad41ef793a"} Dec 03 14:12:01 crc kubenswrapper[4805]: I1203 14:12:01.808094 4805 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-rp9lf" Dec 03 14:12:01 crc kubenswrapper[4805]: I1203 14:12:01.835267 4805 patch_prober.go:28] interesting pod/router-default-5444994796-kxnqm container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 03 14:12:01 crc kubenswrapper[4805]: [-]has-synced failed: reason withheld Dec 03 14:12:01 crc kubenswrapper[4805]: [+]process-running ok Dec 03 14:12:01 crc kubenswrapper[4805]: healthz check failed Dec 03 14:12:01 crc kubenswrapper[4805]: I1203 14:12:01.835588 4805 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-kxnqm" podUID="b22cbc60-5975-42a4-85d3-6add6917fd88" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 03 14:12:01 crc kubenswrapper[4805]: I1203 14:12:01.849333 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-spjlp" event={"ID":"66812e7e-aa85-4045-8e7a-8967b84f8849","Type":"ContainerStarted","Data":"a9f2e6afd451ccf4c9cf049256fc598e0cbb0ee7e78cd6ed913ff8afdff9b2bf"} Dec 03 14:12:01 crc kubenswrapper[4805]: I1203 14:12:01.850247 4805 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/marketplace-operator-79b997595-spjlp" Dec 03 14:12:01 crc kubenswrapper[4805]: I1203 14:12:01.850595 4805 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-ss8kj" podStartSLOduration=131.850573142 podStartE2EDuration="2m11.850573142s" podCreationTimestamp="2025-12-03 14:09:50 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 14:12:01.835433794 +0000 UTC m=+151.498350717" watchObservedRunningTime="2025-12-03 14:12:01.850573142 +0000 UTC m=+151.513490055" Dec 03 14:12:01 crc kubenswrapper[4805]: I1203 14:12:01.852300 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 14:12:01 crc kubenswrapper[4805]: E1203 14:12:01.853728 4805 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-03 14:12:02.353703377 +0000 UTC m=+152.016620360 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 14:12:01 crc kubenswrapper[4805]: I1203 14:12:01.856114 4805 patch_prober.go:28] interesting pod/marketplace-operator-79b997595-spjlp container/marketplace-operator namespace/openshift-marketplace: Readiness probe status=failure output="Get \"http://10.217.0.18:8080/healthz\": dial tcp 10.217.0.18:8080: connect: connection refused" start-of-body= Dec 03 14:12:01 crc kubenswrapper[4805]: I1203 14:12:01.856164 4805 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-marketplace/marketplace-operator-79b997595-spjlp" podUID="66812e7e-aa85-4045-8e7a-8967b84f8849" containerName="marketplace-operator" probeResult="failure" output="Get \"http://10.217.0.18:8080/healthz\": dial tcp 10.217.0.18:8080: connect: connection refused" Dec 03 14:12:01 crc kubenswrapper[4805]: I1203 14:12:01.856365 4805 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-rp9lf" Dec 03 14:12:01 crc kubenswrapper[4805]: I1203 14:12:01.878192 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-ntv6j" event={"ID":"1658d8e1-ac63-40fe-98ce-900a9f9aeae5","Type":"ContainerStarted","Data":"bf5f45a82b82b4c46cfce651ba0e5f2273e655f4722c7485eed466b0c536c77f"} Dec 03 14:12:01 crc kubenswrapper[4805]: I1203 14:12:01.879247 4805 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-authentication/oauth-openshift-558db77b4-ntv6j" Dec 03 14:12:01 crc kubenswrapper[4805]: I1203 14:12:01.900936 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-vc9sn" event={"ID":"c6b0a18a-5cf2-430a-a0ef-2c2776839f11","Type":"ContainerStarted","Data":"b1cfc0249069b436fd8ae6b349892121b6068f595139fae579f28eb7f29f3522"} Dec 03 14:12:01 crc kubenswrapper[4805]: I1203 14:12:01.924721 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-canary/ingress-canary-cnlwr" event={"ID":"6be05973-dc60-437f-959c-a9309cd16d5b","Type":"ContainerStarted","Data":"fcdaa278558dd2702fd0d8f202ea90dbff3756a7676d809351e6990ed9bab8c2"} Dec 03 14:12:01 crc kubenswrapper[4805]: I1203 14:12:01.925175 4805 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-authentication/oauth-openshift-558db77b4-ntv6j" Dec 03 14:12:01 crc kubenswrapper[4805]: I1203 14:12:01.940870 4805 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/marketplace-operator-79b997595-spjlp" podStartSLOduration=131.940832579 podStartE2EDuration="2m11.940832579s" podCreationTimestamp="2025-12-03 14:09:50 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 14:12:01.940272106 +0000 UTC m=+151.603189019" watchObservedRunningTime="2025-12-03 14:12:01.940832579 +0000 UTC m=+151.603749492" Dec 03 14:12:01 crc kubenswrapper[4805]: I1203 14:12:01.943142 4805 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-rp9lf" podStartSLOduration=131.943134163 podStartE2EDuration="2m11.943134163s" podCreationTimestamp="2025-12-03 14:09:50 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 14:12:01.885672213 +0000 UTC m=+151.548589136" watchObservedRunningTime="2025-12-03 14:12:01.943134163 +0000 UTC m=+151.606051086" Dec 03 14:12:01 crc kubenswrapper[4805]: I1203 14:12:01.952609 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-86589" event={"ID":"6107a436-ce5e-4fde-890b-a5ac9f2ada10","Type":"ContainerStarted","Data":"2b42361ef13a9ebd29e4871e0792da5f198fd60aeaa245f080eb5e2d01ece8f9"} Dec 03 14:12:01 crc kubenswrapper[4805]: I1203 14:12:01.952655 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-86589" event={"ID":"6107a436-ce5e-4fde-890b-a5ac9f2ada10","Type":"ContainerStarted","Data":"ee82066e2d5ca5bb8c67dadd9bc94070200f2bd3ebb9d298200aa9b2f9592138"} Dec 03 14:12:01 crc kubenswrapper[4805]: I1203 14:12:01.955314 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-jjzct\" (UID: \"4655a83e-4980-4a5c-9f61-e305ae418822\") " pod="openshift-image-registry/image-registry-697d97f7c8-jjzct" Dec 03 14:12:01 crc kubenswrapper[4805]: E1203 14:12:01.959330 4805 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-03 14:12:02.459316087 +0000 UTC m=+152.122233010 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-jjzct" (UID: "4655a83e-4980-4a5c-9f61-e305ae418822") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 14:12:01 crc kubenswrapper[4805]: I1203 14:12:01.973262 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd-operator/etcd-operator-b45778765-2x56c" event={"ID":"ae561e71-047a-4cf7-8378-d08adb458d9c","Type":"ContainerStarted","Data":"ee6717e3a8be7201b7b8d972fae6204ece20a425c866fc8f028dfb04a831b0ca"} Dec 03 14:12:01 crc kubenswrapper[4805]: I1203 14:12:01.987680 4805 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-authentication/oauth-openshift-558db77b4-ntv6j" podStartSLOduration=132.987662537 podStartE2EDuration="2m12.987662537s" podCreationTimestamp="2025-12-03 14:09:49 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 14:12:01.979476044 +0000 UTC m=+151.642392997" watchObservedRunningTime="2025-12-03 14:12:01.987662537 +0000 UTC m=+151.650579460" Dec 03 14:12:02 crc kubenswrapper[4805]: I1203 14:12:02.002086 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-kvmpp" event={"ID":"d5ea0764-4290-4eab-b4ce-dbd4909455c1","Type":"ContainerStarted","Data":"645d276a385ec88ca7e7e5d6dc83acb6390856ec4833e534907a3f77091b85f9"} Dec 03 14:12:02 crc kubenswrapper[4805]: I1203 14:12:02.018935 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-h5tv7" event={"ID":"4aa1a2c6-8700-4f70-b804-f57371df7025","Type":"ContainerStarted","Data":"62644b594e5a11be41894ea9c2c7052052f604c4b28880a4c0de16fed921bb2a"} Dec 03 14:12:02 crc kubenswrapper[4805]: I1203 14:12:02.032675 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/dns-default-7wnwf" event={"ID":"b1313b26-f0b6-462b-8210-c19b0d4065a4","Type":"ContainerStarted","Data":"dcce4754d7107d47ab0f6867cf5437def0127c90a03b64391860378c24730334"} Dec 03 14:12:02 crc kubenswrapper[4805]: I1203 14:12:02.032714 4805 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-dns/dns-default-7wnwf" Dec 03 14:12:02 crc kubenswrapper[4805]: I1203 14:12:02.038586 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-6n5xz" event={"ID":"470bfa35-69f4-4865-ac37-0aa43d229b05","Type":"ContainerStarted","Data":"ae0bb671ec094cfac67f3d7cda5419d2964cea7e135ba4ed534a61fb61d2ec36"} Dec 03 14:12:02 crc kubenswrapper[4805]: I1203 14:12:02.056625 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 14:12:02 crc kubenswrapper[4805]: E1203 14:12:02.057280 4805 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-03 14:12:02.557265605 +0000 UTC m=+152.220182528 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 14:12:02 crc kubenswrapper[4805]: I1203 14:12:02.057485 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-jjzct\" (UID: \"4655a83e-4980-4a5c-9f61-e305ae418822\") " pod="openshift-image-registry/image-registry-697d97f7c8-jjzct" Dec 03 14:12:02 crc kubenswrapper[4805]: E1203 14:12:02.061891 4805 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-03 14:12:02.561873614 +0000 UTC m=+152.224790537 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-jjzct" (UID: "4655a83e-4980-4a5c-9f61-e305ae418822") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 14:12:02 crc kubenswrapper[4805]: I1203 14:12:02.069070 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-g46rj" event={"ID":"f71c6d49-bfbf-4c57-a4f8-5cd22d1038cb","Type":"ContainerStarted","Data":"f65b7512aa7bfe777d1860b33f40e4fa7f561fd82fe1a41c437c482e9f5568ad"} Dec 03 14:12:02 crc kubenswrapper[4805]: I1203 14:12:02.090406 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-94zfm" event={"ID":"5cec10a3-ed60-4193-be30-a78cb80bd51d","Type":"ContainerStarted","Data":"0fe8bae902d07492d7425621fed3c83901b2157a62a5f94afaffe329f4249a8d"} Dec 03 14:12:02 crc kubenswrapper[4805]: I1203 14:12:02.099818 4805 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-86589" podStartSLOduration=132.099803422 podStartE2EDuration="2m12.099803422s" podCreationTimestamp="2025-12-03 14:09:50 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 14:12:02.098570063 +0000 UTC m=+151.761486986" watchObservedRunningTime="2025-12-03 14:12:02.099803422 +0000 UTC m=+151.762720345" Dec 03 14:12:02 crc kubenswrapper[4805]: I1203 14:12:02.110130 4805 generic.go:334] "Generic (PLEG): container finished" podID="0a7d0b69-f9a6-4a40-9981-3dd999c400d5" containerID="3decc11b46609c728dfadde986bff692a4e80603719c9a3a8e62ed9951a958f0" exitCode=0 Dec 03 14:12:02 crc kubenswrapper[4805]: I1203 14:12:02.111005 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-kfh2f" event={"ID":"0a7d0b69-f9a6-4a40-9981-3dd999c400d5","Type":"ContainerDied","Data":"3decc11b46609c728dfadde986bff692a4e80603719c9a3a8e62ed9951a958f0"} Dec 03 14:12:02 crc kubenswrapper[4805]: I1203 14:12:02.111444 4805 patch_prober.go:28] interesting pod/downloads-7954f5f757-s25zl container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.11:8080/\": dial tcp 10.217.0.11:8080: connect: connection refused" start-of-body= Dec 03 14:12:02 crc kubenswrapper[4805]: I1203 14:12:02.111484 4805 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-s25zl" podUID="6c2d6fbc-771d-4aff-b1e8-3eda17204931" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.11:8080/\": dial tcp 10.217.0.11:8080: connect: connection refused" Dec 03 14:12:02 crc kubenswrapper[4805]: I1203 14:12:02.131916 4805 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console-operator/console-operator-58897d9998-46w2g" Dec 03 14:12:02 crc kubenswrapper[4805]: I1203 14:12:02.135780 4805 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-controller-manager/controller-manager-879f6c89f-k7rgc" Dec 03 14:12:02 crc kubenswrapper[4805]: I1203 14:12:02.161614 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 14:12:02 crc kubenswrapper[4805]: E1203 14:12:02.162041 4805 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-03 14:12:02.662022524 +0000 UTC m=+152.324939447 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 14:12:02 crc kubenswrapper[4805]: I1203 14:12:02.162720 4805 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-kvmpp" podStartSLOduration=132.16270103 podStartE2EDuration="2m12.16270103s" podCreationTimestamp="2025-12-03 14:09:50 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 14:12:02.131935162 +0000 UTC m=+151.794852085" watchObservedRunningTime="2025-12-03 14:12:02.16270103 +0000 UTC m=+151.825617953" Dec 03 14:12:02 crc kubenswrapper[4805]: I1203 14:12:02.199578 4805 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ingress-canary/ingress-canary-cnlwr" podStartSLOduration=8.199562892 podStartE2EDuration="8.199562892s" podCreationTimestamp="2025-12-03 14:11:54 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 14:12:02.163453057 +0000 UTC m=+151.826369980" watchObservedRunningTime="2025-12-03 14:12:02.199562892 +0000 UTC m=+151.862479815" Dec 03 14:12:02 crc kubenswrapper[4805]: I1203 14:12:02.199817 4805 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-vc9sn" podStartSLOduration=133.199813908 podStartE2EDuration="2m13.199813908s" podCreationTimestamp="2025-12-03 14:09:49 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 14:12:02.197196696 +0000 UTC m=+151.860113619" watchObservedRunningTime="2025-12-03 14:12:02.199813908 +0000 UTC m=+151.862730831" Dec 03 14:12:02 crc kubenswrapper[4805]: I1203 14:12:02.264400 4805 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-g46rj" podStartSLOduration=132.264381406 podStartE2EDuration="2m12.264381406s" podCreationTimestamp="2025-12-03 14:09:50 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 14:12:02.256401678 +0000 UTC m=+151.919318601" watchObservedRunningTime="2025-12-03 14:12:02.264381406 +0000 UTC m=+151.927298329" Dec 03 14:12:02 crc kubenswrapper[4805]: I1203 14:12:02.267037 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-jjzct\" (UID: \"4655a83e-4980-4a5c-9f61-e305ae418822\") " pod="openshift-image-registry/image-registry-697d97f7c8-jjzct" Dec 03 14:12:02 crc kubenswrapper[4805]: E1203 14:12:02.268190 4805 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-03 14:12:02.768176656 +0000 UTC m=+152.431093579 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-jjzct" (UID: "4655a83e-4980-4a5c-9f61-e305ae418822") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 14:12:02 crc kubenswrapper[4805]: I1203 14:12:02.369901 4805 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-kfh2f" Dec 03 14:12:02 crc kubenswrapper[4805]: I1203 14:12:02.370225 4805 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-kfh2f" Dec 03 14:12:02 crc kubenswrapper[4805]: I1203 14:12:02.370797 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 14:12:02 crc kubenswrapper[4805]: E1203 14:12:02.371183 4805 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-03 14:12:02.871149874 +0000 UTC m=+152.534066797 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 14:12:02 crc kubenswrapper[4805]: I1203 14:12:02.374275 4805 patch_prober.go:28] interesting pod/apiserver-7bbb656c7d-kfh2f container/oauth-apiserver namespace/openshift-oauth-apiserver: Startup probe status=failure output="Get \"https://10.217.0.6:8443/livez\": dial tcp 10.217.0.6:8443: connect: connection refused" start-of-body= Dec 03 14:12:02 crc kubenswrapper[4805]: I1203 14:12:02.374340 4805 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-kfh2f" podUID="0a7d0b69-f9a6-4a40-9981-3dd999c400d5" containerName="oauth-apiserver" probeResult="failure" output="Get \"https://10.217.0.6:8443/livez\": dial tcp 10.217.0.6:8443: connect: connection refused" Dec 03 14:12:02 crc kubenswrapper[4805]: I1203 14:12:02.473688 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-jjzct\" (UID: \"4655a83e-4980-4a5c-9f61-e305ae418822\") " pod="openshift-image-registry/image-registry-697d97f7c8-jjzct" Dec 03 14:12:02 crc kubenswrapper[4805]: E1203 14:12:02.474408 4805 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-03 14:12:02.974391277 +0000 UTC m=+152.637308200 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-jjzct" (UID: "4655a83e-4980-4a5c-9f61-e305ae418822") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 14:12:02 crc kubenswrapper[4805]: I1203 14:12:02.558187 4805 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-6n5xz" podStartSLOduration=132.558171191 podStartE2EDuration="2m12.558171191s" podCreationTimestamp="2025-12-03 14:09:50 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 14:12:02.427975039 +0000 UTC m=+152.090891962" watchObservedRunningTime="2025-12-03 14:12:02.558171191 +0000 UTC m=+152.221088114" Dec 03 14:12:02 crc kubenswrapper[4805]: I1203 14:12:02.574445 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 14:12:02 crc kubenswrapper[4805]: E1203 14:12:02.574878 4805 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-03 14:12:03.074861946 +0000 UTC m=+152.737778869 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 14:12:02 crc kubenswrapper[4805]: I1203 14:12:02.620199 4805 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-dns/dns-default-7wnwf" podStartSLOduration=8.620184628 podStartE2EDuration="8.620184628s" podCreationTimestamp="2025-12-03 14:11:54 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 14:12:02.619587614 +0000 UTC m=+152.282504537" watchObservedRunningTime="2025-12-03 14:12:02.620184628 +0000 UTC m=+152.283101551" Dec 03 14:12:02 crc kubenswrapper[4805]: I1203 14:12:02.620693 4805 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-h5tv7" podStartSLOduration=132.62068947 podStartE2EDuration="2m12.62068947s" podCreationTimestamp="2025-12-03 14:09:50 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 14:12:02.55813062 +0000 UTC m=+152.221047543" watchObservedRunningTime="2025-12-03 14:12:02.62068947 +0000 UTC m=+152.283606393" Dec 03 14:12:02 crc kubenswrapper[4805]: I1203 14:12:02.642163 4805 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-94zfm" podStartSLOduration=133.642130408 podStartE2EDuration="2m13.642130408s" podCreationTimestamp="2025-12-03 14:09:49 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 14:12:02.641459122 +0000 UTC m=+152.304376045" watchObservedRunningTime="2025-12-03 14:12:02.642130408 +0000 UTC m=+152.305047331" Dec 03 14:12:02 crc kubenswrapper[4805]: I1203 14:12:02.679279 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-jjzct\" (UID: \"4655a83e-4980-4a5c-9f61-e305ae418822\") " pod="openshift-image-registry/image-registry-697d97f7c8-jjzct" Dec 03 14:12:02 crc kubenswrapper[4805]: E1203 14:12:02.679590 4805 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-03 14:12:03.179578234 +0000 UTC m=+152.842495157 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-jjzct" (UID: "4655a83e-4980-4a5c-9f61-e305ae418822") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 14:12:02 crc kubenswrapper[4805]: I1203 14:12:02.724480 4805 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-kfh2f" podStartSLOduration=132.724460737 podStartE2EDuration="2m12.724460737s" podCreationTimestamp="2025-12-03 14:09:50 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 14:12:02.723952125 +0000 UTC m=+152.386869048" watchObservedRunningTime="2025-12-03 14:12:02.724460737 +0000 UTC m=+152.387377660" Dec 03 14:12:02 crc kubenswrapper[4805]: I1203 14:12:02.780273 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 14:12:02 crc kubenswrapper[4805]: E1203 14:12:02.781106 4805 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-03 14:12:03.281086847 +0000 UTC m=+152.944003770 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 14:12:02 crc kubenswrapper[4805]: I1203 14:12:02.821268 4805 patch_prober.go:28] interesting pod/router-default-5444994796-kxnqm container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 03 14:12:02 crc kubenswrapper[4805]: [-]has-synced failed: reason withheld Dec 03 14:12:02 crc kubenswrapper[4805]: [+]process-running ok Dec 03 14:12:02 crc kubenswrapper[4805]: healthz check failed Dec 03 14:12:02 crc kubenswrapper[4805]: I1203 14:12:02.821348 4805 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-kxnqm" podUID="b22cbc60-5975-42a4-85d3-6add6917fd88" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 03 14:12:02 crc kubenswrapper[4805]: I1203 14:12:02.882219 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-jjzct\" (UID: \"4655a83e-4980-4a5c-9f61-e305ae418822\") " pod="openshift-image-registry/image-registry-697d97f7c8-jjzct" Dec 03 14:12:02 crc kubenswrapper[4805]: E1203 14:12:02.882649 4805 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-03 14:12:03.38262163 +0000 UTC m=+153.045538553 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-jjzct" (UID: "4655a83e-4980-4a5c-9f61-e305ae418822") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 14:12:02 crc kubenswrapper[4805]: I1203 14:12:02.983319 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 14:12:02 crc kubenswrapper[4805]: E1203 14:12:02.983676 4805 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-03 14:12:03.483661232 +0000 UTC m=+153.146578155 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 14:12:03 crc kubenswrapper[4805]: I1203 14:12:03.085070 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-jjzct\" (UID: \"4655a83e-4980-4a5c-9f61-e305ae418822\") " pod="openshift-image-registry/image-registry-697d97f7c8-jjzct" Dec 03 14:12:03 crc kubenswrapper[4805]: E1203 14:12:03.085384 4805 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-03 14:12:03.585368689 +0000 UTC m=+153.248285612 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-jjzct" (UID: "4655a83e-4980-4a5c-9f61-e305ae418822") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 14:12:03 crc kubenswrapper[4805]: I1203 14:12:03.150126 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-krqsw" event={"ID":"8801ff25-d273-4354-ab39-16e979d0b73d","Type":"ContainerStarted","Data":"33bee8096721b261521f3d664f6daf6d06570934489bbbbffedacac1e245f13d"} Dec 03 14:12:03 crc kubenswrapper[4805]: I1203 14:12:03.150197 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-krqsw" event={"ID":"8801ff25-d273-4354-ab39-16e979d0b73d","Type":"ContainerStarted","Data":"4ac1a2a818238fdfe03415fea5f731c35b401cc81ba7e372d94f13387c4fb0fe"} Dec 03 14:12:03 crc kubenswrapper[4805]: I1203 14:12:03.152685 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-7nhvl" event={"ID":"ca425595-aae1-44f2-9470-84ba52167748","Type":"ContainerStarted","Data":"c9126d1256a3972cf1270c33873ff65f2ade0dc6422e1ea5471dbbefa5489670"} Dec 03 14:12:03 crc kubenswrapper[4805]: I1203 14:12:03.154549 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-b7wrl" event={"ID":"b44fb2fa-6155-40fc-b5f7-eb5d1fa391bb","Type":"ContainerStarted","Data":"d22df581d63cd4b8c5df5bf4dd9af72133f3b38046fc7f146a2e697860b96ba2"} Dec 03 14:12:03 crc kubenswrapper[4805]: I1203 14:12:03.154695 4805 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-b7wrl" Dec 03 14:12:03 crc kubenswrapper[4805]: I1203 14:12:03.157437 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-kfh2f" event={"ID":"0a7d0b69-f9a6-4a40-9981-3dd999c400d5","Type":"ContainerStarted","Data":"c5718cae6440fdb57c20e3c655f5dfec3a3dbfebb52d1708954bef697b969557"} Dec 03 14:12:03 crc kubenswrapper[4805]: I1203 14:12:03.160740 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-6n5xz" event={"ID":"470bfa35-69f4-4865-ac37-0aa43d229b05","Type":"ContainerStarted","Data":"51e14bcce8f8b0519d444a24ae6bf9d594501c912986cb183f0716e263981e7b"} Dec 03 14:12:03 crc kubenswrapper[4805]: I1203 14:12:03.172340 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-t4kl8" event={"ID":"e600aefb-0839-423f-8ac8-1916400cfa92","Type":"ContainerStarted","Data":"4262254ac67787b09ed37edc43c7691e5c080321fbe9733a63c53fd65da9e978"} Dec 03 14:12:03 crc kubenswrapper[4805]: I1203 14:12:03.176248 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-j8pwh" event={"ID":"220a8b3d-e5fa-486f-ad11-4fb9a43ab092","Type":"ContainerStarted","Data":"8e9ab2308964b034bb4de6b42a77e15123ba2188b31492a275d41183e1538f6b"} Dec 03 14:12:03 crc kubenswrapper[4805]: I1203 14:12:03.183094 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-g46rj" event={"ID":"f71c6d49-bfbf-4c57-a4f8-5cd22d1038cb","Type":"ContainerStarted","Data":"a3ecb1899201d80d672867320d9b5eb24494cac2b9a234fcfc6a4c9e68861804"} Dec 03 14:12:03 crc kubenswrapper[4805]: I1203 14:12:03.185797 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 14:12:03 crc kubenswrapper[4805]: E1203 14:12:03.185930 4805 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-03 14:12:03.685914489 +0000 UTC m=+153.348831412 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 14:12:03 crc kubenswrapper[4805]: I1203 14:12:03.186015 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-jjzct\" (UID: \"4655a83e-4980-4a5c-9f61-e305ae418822\") " pod="openshift-image-registry/image-registry-697d97f7c8-jjzct" Dec 03 14:12:03 crc kubenswrapper[4805]: E1203 14:12:03.186260 4805 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-03 14:12:03.686253437 +0000 UTC m=+153.349170360 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-jjzct" (UID: "4655a83e-4980-4a5c-9f61-e305ae418822") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 14:12:03 crc kubenswrapper[4805]: I1203 14:12:03.189040 4805 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-apiserver/apiserver-76f77b778f-7nhvl" podStartSLOduration=134.189013353 podStartE2EDuration="2m14.189013353s" podCreationTimestamp="2025-12-03 14:09:49 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 14:12:03.186462382 +0000 UTC m=+152.849379305" watchObservedRunningTime="2025-12-03 14:12:03.189013353 +0000 UTC m=+152.851930276" Dec 03 14:12:03 crc kubenswrapper[4805]: I1203 14:12:03.194389 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns-operator/dns-operator-744455d44c-rjf9m" event={"ID":"8e437287-3b92-4c6a-abe9-84c786ba3220","Type":"ContainerStarted","Data":"2da73f2bce9c3efa2b61d80deb86499ee8eeb9604b0374e6321832c1f90e3ee1"} Dec 03 14:12:03 crc kubenswrapper[4805]: I1203 14:12:03.206702 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-h5tv7" event={"ID":"4aa1a2c6-8700-4f70-b804-f57371df7025","Type":"ContainerStarted","Data":"8e1201c16ac23c70f848b0a409a63670936a00a15a8f8add7fd929f7e3eba036"} Dec 03 14:12:03 crc kubenswrapper[4805]: I1203 14:12:03.214667 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/dns-default-7wnwf" event={"ID":"b1313b26-f0b6-462b-8210-c19b0d4065a4","Type":"ContainerStarted","Data":"21369c5a54f67323428a01f8467a7297de8d0333fcaae0b28bd2e174c3fad8cd"} Dec 03 14:12:03 crc kubenswrapper[4805]: I1203 14:12:03.217902 4805 patch_prober.go:28] interesting pod/marketplace-operator-79b997595-spjlp container/marketplace-operator namespace/openshift-marketplace: Readiness probe status=failure output="Get \"http://10.217.0.18:8080/healthz\": dial tcp 10.217.0.18:8080: connect: connection refused" start-of-body= Dec 03 14:12:03 crc kubenswrapper[4805]: I1203 14:12:03.217955 4805 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-marketplace/marketplace-operator-79b997595-spjlp" podUID="66812e7e-aa85-4045-8e7a-8967b84f8849" containerName="marketplace-operator" probeResult="failure" output="Get \"http://10.217.0.18:8080/healthz\": dial tcp 10.217.0.18:8080: connect: connection refused" Dec 03 14:12:03 crc kubenswrapper[4805]: I1203 14:12:03.218441 4805 patch_prober.go:28] interesting pod/downloads-7954f5f757-s25zl container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.11:8080/\": dial tcp 10.217.0.11:8080: connect: connection refused" start-of-body= Dec 03 14:12:03 crc kubenswrapper[4805]: I1203 14:12:03.218494 4805 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-s25zl" podUID="6c2d6fbc-771d-4aff-b1e8-3eda17204931" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.11:8080/\": dial tcp 10.217.0.11:8080: connect: connection refused" Dec 03 14:12:03 crc kubenswrapper[4805]: I1203 14:12:03.227387 4805 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-ss8kj" Dec 03 14:12:03 crc kubenswrapper[4805]: I1203 14:12:03.231102 4805 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-b7wrl" podStartSLOduration=133.231077998 podStartE2EDuration="2m13.231077998s" podCreationTimestamp="2025-12-03 14:09:50 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 14:12:03.230483714 +0000 UTC m=+152.893400637" watchObservedRunningTime="2025-12-03 14:12:03.231077998 +0000 UTC m=+152.893994921" Dec 03 14:12:03 crc kubenswrapper[4805]: I1203 14:12:03.270975 4805 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-j8pwh" podStartSLOduration=133.270953792 podStartE2EDuration="2m13.270953792s" podCreationTimestamp="2025-12-03 14:09:50 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 14:12:03.264417528 +0000 UTC m=+152.927334451" watchObservedRunningTime="2025-12-03 14:12:03.270953792 +0000 UTC m=+152.933870715" Dec 03 14:12:03 crc kubenswrapper[4805]: I1203 14:12:03.287668 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 14:12:03 crc kubenswrapper[4805]: E1203 14:12:03.288044 4805 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-03 14:12:03.788007996 +0000 UTC m=+153.450924929 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 14:12:03 crc kubenswrapper[4805]: I1203 14:12:03.288330 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-jjzct\" (UID: \"4655a83e-4980-4a5c-9f61-e305ae418822\") " pod="openshift-image-registry/image-registry-697d97f7c8-jjzct" Dec 03 14:12:03 crc kubenswrapper[4805]: E1203 14:12:03.290488 4805 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-03 14:12:03.790468074 +0000 UTC m=+153.453385017 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-jjzct" (UID: "4655a83e-4980-4a5c-9f61-e305ae418822") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 14:12:03 crc kubenswrapper[4805]: I1203 14:12:03.303000 4805 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-t4kl8" podStartSLOduration=133.30297575 podStartE2EDuration="2m13.30297575s" podCreationTimestamp="2025-12-03 14:09:50 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 14:12:03.299679462 +0000 UTC m=+152.962596395" watchObservedRunningTime="2025-12-03 14:12:03.30297575 +0000 UTC m=+152.965892673" Dec 03 14:12:03 crc kubenswrapper[4805]: I1203 14:12:03.323916 4805 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-w6nm5"] Dec 03 14:12:03 crc kubenswrapper[4805]: I1203 14:12:03.324403 4805 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-dns-operator/dns-operator-744455d44c-rjf9m" podStartSLOduration=133.324385637 podStartE2EDuration="2m13.324385637s" podCreationTimestamp="2025-12-03 14:09:50 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 14:12:03.322345829 +0000 UTC m=+152.985262742" watchObservedRunningTime="2025-12-03 14:12:03.324385637 +0000 UTC m=+152.987302550" Dec 03 14:12:03 crc kubenswrapper[4805]: I1203 14:12:03.324998 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-w6nm5" Dec 03 14:12:03 crc kubenswrapper[4805]: I1203 14:12:03.329160 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"community-operators-dockercfg-dmngl" Dec 03 14:12:03 crc kubenswrapper[4805]: I1203 14:12:03.350356 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-w6nm5"] Dec 03 14:12:03 crc kubenswrapper[4805]: I1203 14:12:03.391029 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 14:12:03 crc kubenswrapper[4805]: E1203 14:12:03.391511 4805 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-03 14:12:03.891483325 +0000 UTC m=+153.554400248 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 14:12:03 crc kubenswrapper[4805]: I1203 14:12:03.392375 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-jjzct\" (UID: \"4655a83e-4980-4a5c-9f61-e305ae418822\") " pod="openshift-image-registry/image-registry-697d97f7c8-jjzct" Dec 03 14:12:03 crc kubenswrapper[4805]: E1203 14:12:03.412071 4805 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-03 14:12:03.912053532 +0000 UTC m=+153.574970445 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-jjzct" (UID: "4655a83e-4980-4a5c-9f61-e305ae418822") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 14:12:03 crc kubenswrapper[4805]: I1203 14:12:03.479867 4805 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-8gt4x"] Dec 03 14:12:03 crc kubenswrapper[4805]: I1203 14:12:03.480742 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-8gt4x" Dec 03 14:12:03 crc kubenswrapper[4805]: I1203 14:12:03.491259 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"certified-operators-dockercfg-4rs5g" Dec 03 14:12:03 crc kubenswrapper[4805]: I1203 14:12:03.495937 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 14:12:03 crc kubenswrapper[4805]: I1203 14:12:03.496105 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-b5wvv\" (UniqueName: \"kubernetes.io/projected/344cb196-3baa-48c0-abcb-7e46cbce614d-kube-api-access-b5wvv\") pod \"community-operators-w6nm5\" (UID: \"344cb196-3baa-48c0-abcb-7e46cbce614d\") " pod="openshift-marketplace/community-operators-w6nm5" Dec 03 14:12:03 crc kubenswrapper[4805]: I1203 14:12:03.496166 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/344cb196-3baa-48c0-abcb-7e46cbce614d-catalog-content\") pod \"community-operators-w6nm5\" (UID: \"344cb196-3baa-48c0-abcb-7e46cbce614d\") " pod="openshift-marketplace/community-operators-w6nm5" Dec 03 14:12:03 crc kubenswrapper[4805]: I1203 14:12:03.496226 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/344cb196-3baa-48c0-abcb-7e46cbce614d-utilities\") pod \"community-operators-w6nm5\" (UID: \"344cb196-3baa-48c0-abcb-7e46cbce614d\") " pod="openshift-marketplace/community-operators-w6nm5" Dec 03 14:12:03 crc kubenswrapper[4805]: E1203 14:12:03.496518 4805 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-03 14:12:03.996498241 +0000 UTC m=+153.659415164 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 14:12:03 crc kubenswrapper[4805]: I1203 14:12:03.510408 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-8gt4x"] Dec 03 14:12:03 crc kubenswrapper[4805]: I1203 14:12:03.598916 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-n626l\" (UniqueName: \"kubernetes.io/projected/b2e4ac3a-e225-46ae-874d-4cf931c13464-kube-api-access-n626l\") pod \"certified-operators-8gt4x\" (UID: \"b2e4ac3a-e225-46ae-874d-4cf931c13464\") " pod="openshift-marketplace/certified-operators-8gt4x" Dec 03 14:12:03 crc kubenswrapper[4805]: I1203 14:12:03.598971 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b2e4ac3a-e225-46ae-874d-4cf931c13464-utilities\") pod \"certified-operators-8gt4x\" (UID: \"b2e4ac3a-e225-46ae-874d-4cf931c13464\") " pod="openshift-marketplace/certified-operators-8gt4x" Dec 03 14:12:03 crc kubenswrapper[4805]: I1203 14:12:03.599019 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/344cb196-3baa-48c0-abcb-7e46cbce614d-utilities\") pod \"community-operators-w6nm5\" (UID: \"344cb196-3baa-48c0-abcb-7e46cbce614d\") " pod="openshift-marketplace/community-operators-w6nm5" Dec 03 14:12:03 crc kubenswrapper[4805]: I1203 14:12:03.599066 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b2e4ac3a-e225-46ae-874d-4cf931c13464-catalog-content\") pod \"certified-operators-8gt4x\" (UID: \"b2e4ac3a-e225-46ae-874d-4cf931c13464\") " pod="openshift-marketplace/certified-operators-8gt4x" Dec 03 14:12:03 crc kubenswrapper[4805]: I1203 14:12:03.599095 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-b5wvv\" (UniqueName: \"kubernetes.io/projected/344cb196-3baa-48c0-abcb-7e46cbce614d-kube-api-access-b5wvv\") pod \"community-operators-w6nm5\" (UID: \"344cb196-3baa-48c0-abcb-7e46cbce614d\") " pod="openshift-marketplace/community-operators-w6nm5" Dec 03 14:12:03 crc kubenswrapper[4805]: I1203 14:12:03.599124 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-jjzct\" (UID: \"4655a83e-4980-4a5c-9f61-e305ae418822\") " pod="openshift-image-registry/image-registry-697d97f7c8-jjzct" Dec 03 14:12:03 crc kubenswrapper[4805]: I1203 14:12:03.599154 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/344cb196-3baa-48c0-abcb-7e46cbce614d-catalog-content\") pod \"community-operators-w6nm5\" (UID: \"344cb196-3baa-48c0-abcb-7e46cbce614d\") " pod="openshift-marketplace/community-operators-w6nm5" Dec 03 14:12:03 crc kubenswrapper[4805]: I1203 14:12:03.599720 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/344cb196-3baa-48c0-abcb-7e46cbce614d-catalog-content\") pod \"community-operators-w6nm5\" (UID: \"344cb196-3baa-48c0-abcb-7e46cbce614d\") " pod="openshift-marketplace/community-operators-w6nm5" Dec 03 14:12:03 crc kubenswrapper[4805]: I1203 14:12:03.600045 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/344cb196-3baa-48c0-abcb-7e46cbce614d-utilities\") pod \"community-operators-w6nm5\" (UID: \"344cb196-3baa-48c0-abcb-7e46cbce614d\") " pod="openshift-marketplace/community-operators-w6nm5" Dec 03 14:12:03 crc kubenswrapper[4805]: E1203 14:12:03.600608 4805 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-03 14:12:04.100593945 +0000 UTC m=+153.763510868 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-jjzct" (UID: "4655a83e-4980-4a5c-9f61-e305ae418822") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 14:12:03 crc kubenswrapper[4805]: I1203 14:12:03.625584 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-b5wvv\" (UniqueName: \"kubernetes.io/projected/344cb196-3baa-48c0-abcb-7e46cbce614d-kube-api-access-b5wvv\") pod \"community-operators-w6nm5\" (UID: \"344cb196-3baa-48c0-abcb-7e46cbce614d\") " pod="openshift-marketplace/community-operators-w6nm5" Dec 03 14:12:03 crc kubenswrapper[4805]: I1203 14:12:03.649039 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-w6nm5" Dec 03 14:12:03 crc kubenswrapper[4805]: I1203 14:12:03.684575 4805 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-fbrtf"] Dec 03 14:12:03 crc kubenswrapper[4805]: I1203 14:12:03.685570 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-fbrtf" Dec 03 14:12:03 crc kubenswrapper[4805]: I1203 14:12:03.699816 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 14:12:03 crc kubenswrapper[4805]: I1203 14:12:03.700103 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-n626l\" (UniqueName: \"kubernetes.io/projected/b2e4ac3a-e225-46ae-874d-4cf931c13464-kube-api-access-n626l\") pod \"certified-operators-8gt4x\" (UID: \"b2e4ac3a-e225-46ae-874d-4cf931c13464\") " pod="openshift-marketplace/certified-operators-8gt4x" Dec 03 14:12:03 crc kubenswrapper[4805]: I1203 14:12:03.700154 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b2e4ac3a-e225-46ae-874d-4cf931c13464-utilities\") pod \"certified-operators-8gt4x\" (UID: \"b2e4ac3a-e225-46ae-874d-4cf931c13464\") " pod="openshift-marketplace/certified-operators-8gt4x" Dec 03 14:12:03 crc kubenswrapper[4805]: I1203 14:12:03.700222 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b2e4ac3a-e225-46ae-874d-4cf931c13464-catalog-content\") pod \"certified-operators-8gt4x\" (UID: \"b2e4ac3a-e225-46ae-874d-4cf931c13464\") " pod="openshift-marketplace/certified-operators-8gt4x" Dec 03 14:12:03 crc kubenswrapper[4805]: I1203 14:12:03.700735 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b2e4ac3a-e225-46ae-874d-4cf931c13464-catalog-content\") pod \"certified-operators-8gt4x\" (UID: \"b2e4ac3a-e225-46ae-874d-4cf931c13464\") " pod="openshift-marketplace/certified-operators-8gt4x" Dec 03 14:12:03 crc kubenswrapper[4805]: E1203 14:12:03.700821 4805 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-03 14:12:04.200802807 +0000 UTC m=+153.863719720 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 14:12:03 crc kubenswrapper[4805]: I1203 14:12:03.701454 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b2e4ac3a-e225-46ae-874d-4cf931c13464-utilities\") pod \"certified-operators-8gt4x\" (UID: \"b2e4ac3a-e225-46ae-874d-4cf931c13464\") " pod="openshift-marketplace/certified-operators-8gt4x" Dec 03 14:12:03 crc kubenswrapper[4805]: I1203 14:12:03.702438 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-fbrtf"] Dec 03 14:12:03 crc kubenswrapper[4805]: I1203 14:12:03.732806 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-n626l\" (UniqueName: \"kubernetes.io/projected/b2e4ac3a-e225-46ae-874d-4cf931c13464-kube-api-access-n626l\") pod \"certified-operators-8gt4x\" (UID: \"b2e4ac3a-e225-46ae-874d-4cf931c13464\") " pod="openshift-marketplace/certified-operators-8gt4x" Dec 03 14:12:03 crc kubenswrapper[4805]: I1203 14:12:03.804176 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-8gt4x" Dec 03 14:12:03 crc kubenswrapper[4805]: I1203 14:12:03.804460 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/731aa7bc-cde1-4ff7-aa29-78093078514d-catalog-content\") pod \"community-operators-fbrtf\" (UID: \"731aa7bc-cde1-4ff7-aa29-78093078514d\") " pod="openshift-marketplace/community-operators-fbrtf" Dec 03 14:12:03 crc kubenswrapper[4805]: I1203 14:12:03.804507 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/731aa7bc-cde1-4ff7-aa29-78093078514d-utilities\") pod \"community-operators-fbrtf\" (UID: \"731aa7bc-cde1-4ff7-aa29-78093078514d\") " pod="openshift-marketplace/community-operators-fbrtf" Dec 03 14:12:03 crc kubenswrapper[4805]: I1203 14:12:03.804543 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ph6rn\" (UniqueName: \"kubernetes.io/projected/731aa7bc-cde1-4ff7-aa29-78093078514d-kube-api-access-ph6rn\") pod \"community-operators-fbrtf\" (UID: \"731aa7bc-cde1-4ff7-aa29-78093078514d\") " pod="openshift-marketplace/community-operators-fbrtf" Dec 03 14:12:03 crc kubenswrapper[4805]: I1203 14:12:03.804585 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-jjzct\" (UID: \"4655a83e-4980-4a5c-9f61-e305ae418822\") " pod="openshift-image-registry/image-registry-697d97f7c8-jjzct" Dec 03 14:12:03 crc kubenswrapper[4805]: E1203 14:12:03.804867 4805 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-03 14:12:04.304835529 +0000 UTC m=+153.967752442 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-jjzct" (UID: "4655a83e-4980-4a5c-9f61-e305ae418822") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 14:12:03 crc kubenswrapper[4805]: I1203 14:12:03.818390 4805 patch_prober.go:28] interesting pod/router-default-5444994796-kxnqm container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 03 14:12:03 crc kubenswrapper[4805]: [-]has-synced failed: reason withheld Dec 03 14:12:03 crc kubenswrapper[4805]: [+]process-running ok Dec 03 14:12:03 crc kubenswrapper[4805]: healthz check failed Dec 03 14:12:03 crc kubenswrapper[4805]: I1203 14:12:03.818441 4805 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-kxnqm" podUID="b22cbc60-5975-42a4-85d3-6add6917fd88" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 03 14:12:03 crc kubenswrapper[4805]: I1203 14:12:03.886914 4805 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-9cm4t"] Dec 03 14:12:03 crc kubenswrapper[4805]: I1203 14:12:03.887904 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-9cm4t" Dec 03 14:12:03 crc kubenswrapper[4805]: I1203 14:12:03.899015 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-9cm4t"] Dec 03 14:12:03 crc kubenswrapper[4805]: I1203 14:12:03.905559 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 14:12:03 crc kubenswrapper[4805]: E1203 14:12:03.905737 4805 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-03 14:12:04.405712627 +0000 UTC m=+154.068629550 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 14:12:03 crc kubenswrapper[4805]: I1203 14:12:03.905859 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-jjzct\" (UID: \"4655a83e-4980-4a5c-9f61-e305ae418822\") " pod="openshift-image-registry/image-registry-697d97f7c8-jjzct" Dec 03 14:12:03 crc kubenswrapper[4805]: I1203 14:12:03.905922 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/731aa7bc-cde1-4ff7-aa29-78093078514d-catalog-content\") pod \"community-operators-fbrtf\" (UID: \"731aa7bc-cde1-4ff7-aa29-78093078514d\") " pod="openshift-marketplace/community-operators-fbrtf" Dec 03 14:12:03 crc kubenswrapper[4805]: I1203 14:12:03.905953 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/731aa7bc-cde1-4ff7-aa29-78093078514d-utilities\") pod \"community-operators-fbrtf\" (UID: \"731aa7bc-cde1-4ff7-aa29-78093078514d\") " pod="openshift-marketplace/community-operators-fbrtf" Dec 03 14:12:03 crc kubenswrapper[4805]: I1203 14:12:03.906010 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ph6rn\" (UniqueName: \"kubernetes.io/projected/731aa7bc-cde1-4ff7-aa29-78093078514d-kube-api-access-ph6rn\") pod \"community-operators-fbrtf\" (UID: \"731aa7bc-cde1-4ff7-aa29-78093078514d\") " pod="openshift-marketplace/community-operators-fbrtf" Dec 03 14:12:03 crc kubenswrapper[4805]: I1203 14:12:03.906733 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/731aa7bc-cde1-4ff7-aa29-78093078514d-catalog-content\") pod \"community-operators-fbrtf\" (UID: \"731aa7bc-cde1-4ff7-aa29-78093078514d\") " pod="openshift-marketplace/community-operators-fbrtf" Dec 03 14:12:03 crc kubenswrapper[4805]: I1203 14:12:03.906887 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/731aa7bc-cde1-4ff7-aa29-78093078514d-utilities\") pod \"community-operators-fbrtf\" (UID: \"731aa7bc-cde1-4ff7-aa29-78093078514d\") " pod="openshift-marketplace/community-operators-fbrtf" Dec 03 14:12:03 crc kubenswrapper[4805]: E1203 14:12:03.907092 4805 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-03 14:12:04.407075499 +0000 UTC m=+154.069992492 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-jjzct" (UID: "4655a83e-4980-4a5c-9f61-e305ae418822") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 14:12:03 crc kubenswrapper[4805]: I1203 14:12:03.940196 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ph6rn\" (UniqueName: \"kubernetes.io/projected/731aa7bc-cde1-4ff7-aa29-78093078514d-kube-api-access-ph6rn\") pod \"community-operators-fbrtf\" (UID: \"731aa7bc-cde1-4ff7-aa29-78093078514d\") " pod="openshift-marketplace/community-operators-fbrtf" Dec 03 14:12:04 crc kubenswrapper[4805]: I1203 14:12:04.013367 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 14:12:04 crc kubenswrapper[4805]: E1203 14:12:04.013642 4805 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-03 14:12:04.513610791 +0000 UTC m=+154.176527714 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 14:12:04 crc kubenswrapper[4805]: I1203 14:12:04.013799 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/64f6384f-41e0-469e-8920-27e9e97b62db-utilities\") pod \"certified-operators-9cm4t\" (UID: \"64f6384f-41e0-469e-8920-27e9e97b62db\") " pod="openshift-marketplace/certified-operators-9cm4t" Dec 03 14:12:04 crc kubenswrapper[4805]: I1203 14:12:04.013820 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/64f6384f-41e0-469e-8920-27e9e97b62db-catalog-content\") pod \"certified-operators-9cm4t\" (UID: \"64f6384f-41e0-469e-8920-27e9e97b62db\") " pod="openshift-marketplace/certified-operators-9cm4t" Dec 03 14:12:04 crc kubenswrapper[4805]: I1203 14:12:04.013869 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-jjzct\" (UID: \"4655a83e-4980-4a5c-9f61-e305ae418822\") " pod="openshift-image-registry/image-registry-697d97f7c8-jjzct" Dec 03 14:12:04 crc kubenswrapper[4805]: I1203 14:12:04.013895 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rv4vh\" (UniqueName: \"kubernetes.io/projected/64f6384f-41e0-469e-8920-27e9e97b62db-kube-api-access-rv4vh\") pod \"certified-operators-9cm4t\" (UID: \"64f6384f-41e0-469e-8920-27e9e97b62db\") " pod="openshift-marketplace/certified-operators-9cm4t" Dec 03 14:12:04 crc kubenswrapper[4805]: E1203 14:12:04.014208 4805 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-03 14:12:04.514194095 +0000 UTC m=+154.177111018 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-jjzct" (UID: "4655a83e-4980-4a5c-9f61-e305ae418822") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 14:12:04 crc kubenswrapper[4805]: I1203 14:12:04.017129 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-fbrtf" Dec 03 14:12:04 crc kubenswrapper[4805]: I1203 14:12:04.123674 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 14:12:04 crc kubenswrapper[4805]: I1203 14:12:04.123853 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/64f6384f-41e0-469e-8920-27e9e97b62db-utilities\") pod \"certified-operators-9cm4t\" (UID: \"64f6384f-41e0-469e-8920-27e9e97b62db\") " pod="openshift-marketplace/certified-operators-9cm4t" Dec 03 14:12:04 crc kubenswrapper[4805]: I1203 14:12:04.123874 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/64f6384f-41e0-469e-8920-27e9e97b62db-catalog-content\") pod \"certified-operators-9cm4t\" (UID: \"64f6384f-41e0-469e-8920-27e9e97b62db\") " pod="openshift-marketplace/certified-operators-9cm4t" Dec 03 14:12:04 crc kubenswrapper[4805]: I1203 14:12:04.123914 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rv4vh\" (UniqueName: \"kubernetes.io/projected/64f6384f-41e0-469e-8920-27e9e97b62db-kube-api-access-rv4vh\") pod \"certified-operators-9cm4t\" (UID: \"64f6384f-41e0-469e-8920-27e9e97b62db\") " pod="openshift-marketplace/certified-operators-9cm4t" Dec 03 14:12:04 crc kubenswrapper[4805]: E1203 14:12:04.124262 4805 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-03 14:12:04.62424752 +0000 UTC m=+154.287164443 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 14:12:04 crc kubenswrapper[4805]: I1203 14:12:04.124610 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/64f6384f-41e0-469e-8920-27e9e97b62db-utilities\") pod \"certified-operators-9cm4t\" (UID: \"64f6384f-41e0-469e-8920-27e9e97b62db\") " pod="openshift-marketplace/certified-operators-9cm4t" Dec 03 14:12:04 crc kubenswrapper[4805]: I1203 14:12:04.124691 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/64f6384f-41e0-469e-8920-27e9e97b62db-catalog-content\") pod \"certified-operators-9cm4t\" (UID: \"64f6384f-41e0-469e-8920-27e9e97b62db\") " pod="openshift-marketplace/certified-operators-9cm4t" Dec 03 14:12:04 crc kubenswrapper[4805]: I1203 14:12:04.190257 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rv4vh\" (UniqueName: \"kubernetes.io/projected/64f6384f-41e0-469e-8920-27e9e97b62db-kube-api-access-rv4vh\") pod \"certified-operators-9cm4t\" (UID: \"64f6384f-41e0-469e-8920-27e9e97b62db\") " pod="openshift-marketplace/certified-operators-9cm4t" Dec 03 14:12:04 crc kubenswrapper[4805]: I1203 14:12:04.213165 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-9cm4t" Dec 03 14:12:04 crc kubenswrapper[4805]: I1203 14:12:04.225495 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-jjzct\" (UID: \"4655a83e-4980-4a5c-9f61-e305ae418822\") " pod="openshift-image-registry/image-registry-697d97f7c8-jjzct" Dec 03 14:12:04 crc kubenswrapper[4805]: E1203 14:12:04.225751 4805 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-03 14:12:04.725740912 +0000 UTC m=+154.388657835 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-jjzct" (UID: "4655a83e-4980-4a5c-9f61-e305ae418822") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 14:12:04 crc kubenswrapper[4805]: I1203 14:12:04.226149 4805 plugin_watcher.go:194] "Adding socket path or updating timestamp to desired state cache" path="/var/lib/kubelet/plugins_registry/kubevirt.io.hostpath-provisioner-reg.sock" Dec 03 14:12:04 crc kubenswrapper[4805]: I1203 14:12:04.329588 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 14:12:04 crc kubenswrapper[4805]: E1203 14:12:04.330424 4805 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-03 14:12:04.83040511 +0000 UTC m=+154.493322033 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 14:12:04 crc kubenswrapper[4805]: I1203 14:12:04.360219 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-krqsw" event={"ID":"8801ff25-d273-4354-ab39-16e979d0b73d","Type":"ContainerStarted","Data":"5fa69034bebc29ba27674b8b9e45ecb1adbf3322dca2436703ea988ddcf2f4b7"} Dec 03 14:12:04 crc kubenswrapper[4805]: I1203 14:12:04.365520 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-w6nm5"] Dec 03 14:12:04 crc kubenswrapper[4805]: I1203 14:12:04.376294 4805 generic.go:334] "Generic (PLEG): container finished" podID="8b954d96-a62c-434e-a340-8446d778c545" containerID="8f4f233b53a4e215b29e4cb9baef3ccf262fa81a4f6bddc1a28d3a846ab83332" exitCode=0 Dec 03 14:12:04 crc kubenswrapper[4805]: I1203 14:12:04.376947 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29412840-gj7pr" event={"ID":"8b954d96-a62c-434e-a340-8446d778c545","Type":"ContainerDied","Data":"8f4f233b53a4e215b29e4cb9baef3ccf262fa81a4f6bddc1a28d3a846ab83332"} Dec 03 14:12:04 crc kubenswrapper[4805]: W1203 14:12:04.378667 4805 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod344cb196_3baa_48c0_abcb_7e46cbce614d.slice/crio-9c86a4483d9c040cfd4bb2d0e7facf620e95b3e01e98f8356b5085dad93668fe WatchSource:0}: Error finding container 9c86a4483d9c040cfd4bb2d0e7facf620e95b3e01e98f8356b5085dad93668fe: Status 404 returned error can't find the container with id 9c86a4483d9c040cfd4bb2d0e7facf620e95b3e01e98f8356b5085dad93668fe Dec 03 14:12:04 crc kubenswrapper[4805]: I1203 14:12:04.400162 4805 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/marketplace-operator-79b997595-spjlp" Dec 03 14:12:04 crc kubenswrapper[4805]: I1203 14:12:04.415878 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-8gt4x"] Dec 03 14:12:04 crc kubenswrapper[4805]: I1203 14:12:04.432600 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-jjzct\" (UID: \"4655a83e-4980-4a5c-9f61-e305ae418822\") " pod="openshift-image-registry/image-registry-697d97f7c8-jjzct" Dec 03 14:12:04 crc kubenswrapper[4805]: E1203 14:12:04.438331 4805 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-03 14:12:04.938313704 +0000 UTC m=+154.601230707 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-jjzct" (UID: "4655a83e-4980-4a5c-9f61-e305ae418822") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 14:12:04 crc kubenswrapper[4805]: W1203 14:12:04.442021 4805 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podb2e4ac3a_e225_46ae_874d_4cf931c13464.slice/crio-fad96d1f140c18627af015f90094d1e33cfe8bc2801445395bcefd0696ea7cc3 WatchSource:0}: Error finding container fad96d1f140c18627af015f90094d1e33cfe8bc2801445395bcefd0696ea7cc3: Status 404 returned error can't find the container with id fad96d1f140c18627af015f90094d1e33cfe8bc2801445395bcefd0696ea7cc3 Dec 03 14:12:04 crc kubenswrapper[4805]: I1203 14:12:04.475831 4805 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="hostpath-provisioner/csi-hostpathplugin-krqsw" podStartSLOduration=9.475814582 podStartE2EDuration="9.475814582s" podCreationTimestamp="2025-12-03 14:11:55 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 14:12:04.445617377 +0000 UTC m=+154.108534300" watchObservedRunningTime="2025-12-03 14:12:04.475814582 +0000 UTC m=+154.138731505" Dec 03 14:12:04 crc kubenswrapper[4805]: I1203 14:12:04.546501 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 14:12:04 crc kubenswrapper[4805]: E1203 14:12:04.547155 4805 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-03 14:12:05.04712945 +0000 UTC m=+154.710046373 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 14:12:04 crc kubenswrapper[4805]: I1203 14:12:04.606019 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-fbrtf"] Dec 03 14:12:04 crc kubenswrapper[4805]: I1203 14:12:04.648238 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-jjzct\" (UID: \"4655a83e-4980-4a5c-9f61-e305ae418822\") " pod="openshift-image-registry/image-registry-697d97f7c8-jjzct" Dec 03 14:12:04 crc kubenswrapper[4805]: E1203 14:12:04.648603 4805 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-03 14:12:05.148588431 +0000 UTC m=+154.811505354 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-jjzct" (UID: "4655a83e-4980-4a5c-9f61-e305ae418822") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 14:12:04 crc kubenswrapper[4805]: I1203 14:12:04.701897 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-9cm4t"] Dec 03 14:12:04 crc kubenswrapper[4805]: W1203 14:12:04.709573 4805 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod64f6384f_41e0_469e_8920_27e9e97b62db.slice/crio-2044a0871b4ae6ac514cc53eb77fa1c70c3a799fe279d2d157c39f4aad5f407f WatchSource:0}: Error finding container 2044a0871b4ae6ac514cc53eb77fa1c70c3a799fe279d2d157c39f4aad5f407f: Status 404 returned error can't find the container with id 2044a0871b4ae6ac514cc53eb77fa1c70c3a799fe279d2d157c39f4aad5f407f Dec 03 14:12:04 crc kubenswrapper[4805]: I1203 14:12:04.749061 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 14:12:04 crc kubenswrapper[4805]: E1203 14:12:04.749256 4805 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-03 14:12:05.249228763 +0000 UTC m=+154.912145686 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 14:12:04 crc kubenswrapper[4805]: I1203 14:12:04.749325 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-jjzct\" (UID: \"4655a83e-4980-4a5c-9f61-e305ae418822\") " pod="openshift-image-registry/image-registry-697d97f7c8-jjzct" Dec 03 14:12:04 crc kubenswrapper[4805]: E1203 14:12:04.749652 4805 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-03 14:12:05.249645643 +0000 UTC m=+154.912562566 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-jjzct" (UID: "4655a83e-4980-4a5c-9f61-e305ae418822") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 14:12:04 crc kubenswrapper[4805]: I1203 14:12:04.819581 4805 patch_prober.go:28] interesting pod/router-default-5444994796-kxnqm container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 03 14:12:04 crc kubenswrapper[4805]: [-]has-synced failed: reason withheld Dec 03 14:12:04 crc kubenswrapper[4805]: [+]process-running ok Dec 03 14:12:04 crc kubenswrapper[4805]: healthz check failed Dec 03 14:12:04 crc kubenswrapper[4805]: I1203 14:12:04.819677 4805 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-kxnqm" podUID="b22cbc60-5975-42a4-85d3-6add6917fd88" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 03 14:12:04 crc kubenswrapper[4805]: I1203 14:12:04.850283 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 14:12:04 crc kubenswrapper[4805]: E1203 14:12:04.850414 4805 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-03 14:12:05.350396128 +0000 UTC m=+155.013313051 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 14:12:04 crc kubenswrapper[4805]: I1203 14:12:04.850504 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-jjzct\" (UID: \"4655a83e-4980-4a5c-9f61-e305ae418822\") " pod="openshift-image-registry/image-registry-697d97f7c8-jjzct" Dec 03 14:12:04 crc kubenswrapper[4805]: E1203 14:12:04.850807 4805 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-03 14:12:05.350799478 +0000 UTC m=+155.013716401 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-jjzct" (UID: "4655a83e-4980-4a5c-9f61-e305ae418822") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 14:12:04 crc kubenswrapper[4805]: I1203 14:12:04.863090 4805 reconciler.go:161] "OperationExecutor.RegisterPlugin started" plugin={"SocketPath":"/var/lib/kubelet/plugins_registry/kubevirt.io.hostpath-provisioner-reg.sock","Timestamp":"2025-12-03T14:12:04.226158442Z","Handler":null,"Name":""} Dec 03 14:12:04 crc kubenswrapper[4805]: I1203 14:12:04.952059 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 14:12:04 crc kubenswrapper[4805]: E1203 14:12:04.952214 4805 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-03 14:12:05.452190858 +0000 UTC m=+155.115107781 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 14:12:04 crc kubenswrapper[4805]: I1203 14:12:04.952343 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-jjzct\" (UID: \"4655a83e-4980-4a5c-9f61-e305ae418822\") " pod="openshift-image-registry/image-registry-697d97f7c8-jjzct" Dec 03 14:12:04 crc kubenswrapper[4805]: E1203 14:12:04.952596 4805 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-03 14:12:05.452588807 +0000 UTC m=+155.115505730 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-jjzct" (UID: "4655a83e-4980-4a5c-9f61-e305ae418822") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 14:12:05 crc kubenswrapper[4805]: I1203 14:12:05.052998 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 14:12:05 crc kubenswrapper[4805]: E1203 14:12:05.053140 4805 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-03 14:12:05.553121097 +0000 UTC m=+155.216038020 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 14:12:05 crc kubenswrapper[4805]: I1203 14:12:05.053258 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-jjzct\" (UID: \"4655a83e-4980-4a5c-9f61-e305ae418822\") " pod="openshift-image-registry/image-registry-697d97f7c8-jjzct" Dec 03 14:12:05 crc kubenswrapper[4805]: E1203 14:12:05.053585 4805 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-03 14:12:05.553575837 +0000 UTC m=+155.216492760 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-jjzct" (UID: "4655a83e-4980-4a5c-9f61-e305ae418822") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 14:12:05 crc kubenswrapper[4805]: I1203 14:12:05.154504 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 14:12:05 crc kubenswrapper[4805]: E1203 14:12:05.154825 4805 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-03 14:12:05.654792383 +0000 UTC m=+155.317709336 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 14:12:05 crc kubenswrapper[4805]: I1203 14:12:05.250220 4805 csi_plugin.go:100] kubernetes.io/csi: Trying to validate a new CSI Driver with name: kubevirt.io.hostpath-provisioner endpoint: /var/lib/kubelet/plugins/csi-hostpath/csi.sock versions: 1.0.0 Dec 03 14:12:05 crc kubenswrapper[4805]: I1203 14:12:05.250260 4805 csi_plugin.go:113] kubernetes.io/csi: Register new plugin with name: kubevirt.io.hostpath-provisioner at endpoint: /var/lib/kubelet/plugins/csi-hostpath/csi.sock Dec 03 14:12:05 crc kubenswrapper[4805]: I1203 14:12:05.255786 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-jjzct\" (UID: \"4655a83e-4980-4a5c-9f61-e305ae418822\") " pod="openshift-image-registry/image-registry-697d97f7c8-jjzct" Dec 03 14:12:05 crc kubenswrapper[4805]: I1203 14:12:05.364182 4805 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Dec 03 14:12:05 crc kubenswrapper[4805]: I1203 14:12:05.364242 4805 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-jjzct\" (UID: \"4655a83e-4980-4a5c-9f61-e305ae418822\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/globalmount\"" pod="openshift-image-registry/image-registry-697d97f7c8-jjzct" Dec 03 14:12:05 crc kubenswrapper[4805]: I1203 14:12:05.382224 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-w6nm5" event={"ID":"344cb196-3baa-48c0-abcb-7e46cbce614d","Type":"ContainerStarted","Data":"9c86a4483d9c040cfd4bb2d0e7facf620e95b3e01e98f8356b5085dad93668fe"} Dec 03 14:12:05 crc kubenswrapper[4805]: I1203 14:12:05.390100 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-9cm4t" event={"ID":"64f6384f-41e0-469e-8920-27e9e97b62db","Type":"ContainerStarted","Data":"2044a0871b4ae6ac514cc53eb77fa1c70c3a799fe279d2d157c39f4aad5f407f"} Dec 03 14:12:05 crc kubenswrapper[4805]: I1203 14:12:05.391828 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-fbrtf" event={"ID":"731aa7bc-cde1-4ff7-aa29-78093078514d","Type":"ContainerStarted","Data":"f04dc853bd11007db4faffdd39b6693fca131898415cc3f50ebddbd1166b1476"} Dec 03 14:12:05 crc kubenswrapper[4805]: I1203 14:12:05.397242 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-8gt4x" event={"ID":"b2e4ac3a-e225-46ae-874d-4cf931c13464","Type":"ContainerStarted","Data":"fad96d1f140c18627af015f90094d1e33cfe8bc2801445395bcefd0696ea7cc3"} Dec 03 14:12:05 crc kubenswrapper[4805]: I1203 14:12:05.408160 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-jjzct\" (UID: \"4655a83e-4980-4a5c-9f61-e305ae418822\") " pod="openshift-image-registry/image-registry-697d97f7c8-jjzct" Dec 03 14:12:05 crc kubenswrapper[4805]: I1203 14:12:05.458655 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 14:12:05 crc kubenswrapper[4805]: I1203 14:12:05.465646 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (OuterVolumeSpecName: "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8". PluginName "kubernetes.io/csi", VolumeGidValue "" Dec 03 14:12:05 crc kubenswrapper[4805]: I1203 14:12:05.473916 4805 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-m7r99"] Dec 03 14:12:05 crc kubenswrapper[4805]: I1203 14:12:05.475306 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-m7r99" Dec 03 14:12:05 crc kubenswrapper[4805]: I1203 14:12:05.477946 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-marketplace-dockercfg-x2ctb" Dec 03 14:12:05 crc kubenswrapper[4805]: I1203 14:12:05.487705 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-m7r99"] Dec 03 14:12:05 crc kubenswrapper[4805]: I1203 14:12:05.505401 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-jjzct" Dec 03 14:12:05 crc kubenswrapper[4805]: I1203 14:12:05.568557 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1e412276-9d78-4771-b447-b4e2b388e604-utilities\") pod \"redhat-marketplace-m7r99\" (UID: \"1e412276-9d78-4771-b447-b4e2b388e604\") " pod="openshift-marketplace/redhat-marketplace-m7r99" Dec 03 14:12:05 crc kubenswrapper[4805]: I1203 14:12:05.568631 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-h6c8m\" (UniqueName: \"kubernetes.io/projected/1e412276-9d78-4771-b447-b4e2b388e604-kube-api-access-h6c8m\") pod \"redhat-marketplace-m7r99\" (UID: \"1e412276-9d78-4771-b447-b4e2b388e604\") " pod="openshift-marketplace/redhat-marketplace-m7r99" Dec 03 14:12:05 crc kubenswrapper[4805]: I1203 14:12:05.568671 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1e412276-9d78-4771-b447-b4e2b388e604-catalog-content\") pod \"redhat-marketplace-m7r99\" (UID: \"1e412276-9d78-4771-b447-b4e2b388e604\") " pod="openshift-marketplace/redhat-marketplace-m7r99" Dec 03 14:12:05 crc kubenswrapper[4805]: I1203 14:12:05.669323 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1e412276-9d78-4771-b447-b4e2b388e604-utilities\") pod \"redhat-marketplace-m7r99\" (UID: \"1e412276-9d78-4771-b447-b4e2b388e604\") " pod="openshift-marketplace/redhat-marketplace-m7r99" Dec 03 14:12:05 crc kubenswrapper[4805]: I1203 14:12:05.669372 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-h6c8m\" (UniqueName: \"kubernetes.io/projected/1e412276-9d78-4771-b447-b4e2b388e604-kube-api-access-h6c8m\") pod \"redhat-marketplace-m7r99\" (UID: \"1e412276-9d78-4771-b447-b4e2b388e604\") " pod="openshift-marketplace/redhat-marketplace-m7r99" Dec 03 14:12:05 crc kubenswrapper[4805]: I1203 14:12:05.669394 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1e412276-9d78-4771-b447-b4e2b388e604-catalog-content\") pod \"redhat-marketplace-m7r99\" (UID: \"1e412276-9d78-4771-b447-b4e2b388e604\") " pod="openshift-marketplace/redhat-marketplace-m7r99" Dec 03 14:12:05 crc kubenswrapper[4805]: I1203 14:12:05.670225 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1e412276-9d78-4771-b447-b4e2b388e604-utilities\") pod \"redhat-marketplace-m7r99\" (UID: \"1e412276-9d78-4771-b447-b4e2b388e604\") " pod="openshift-marketplace/redhat-marketplace-m7r99" Dec 03 14:12:05 crc kubenswrapper[4805]: I1203 14:12:05.670608 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1e412276-9d78-4771-b447-b4e2b388e604-catalog-content\") pod \"redhat-marketplace-m7r99\" (UID: \"1e412276-9d78-4771-b447-b4e2b388e604\") " pod="openshift-marketplace/redhat-marketplace-m7r99" Dec 03 14:12:05 crc kubenswrapper[4805]: I1203 14:12:05.678418 4805 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29412840-gj7pr" Dec 03 14:12:05 crc kubenswrapper[4805]: I1203 14:12:05.692975 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-h6c8m\" (UniqueName: \"kubernetes.io/projected/1e412276-9d78-4771-b447-b4e2b388e604-kube-api-access-h6c8m\") pod \"redhat-marketplace-m7r99\" (UID: \"1e412276-9d78-4771-b447-b4e2b388e604\") " pod="openshift-marketplace/redhat-marketplace-m7r99" Dec 03 14:12:05 crc kubenswrapper[4805]: I1203 14:12:05.717079 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-jjzct"] Dec 03 14:12:05 crc kubenswrapper[4805]: W1203 14:12:05.724135 4805 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod4655a83e_4980_4a5c_9f61_e305ae418822.slice/crio-45e95adcdd0b0349f91467461f335ee50bd61445aaf7d6948daf31c5482d4d35 WatchSource:0}: Error finding container 45e95adcdd0b0349f91467461f335ee50bd61445aaf7d6948daf31c5482d4d35: Status 404 returned error can't find the container with id 45e95adcdd0b0349f91467461f335ee50bd61445aaf7d6948daf31c5482d4d35 Dec 03 14:12:05 crc kubenswrapper[4805]: I1203 14:12:05.770780 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/8b954d96-a62c-434e-a340-8446d778c545-secret-volume\") pod \"8b954d96-a62c-434e-a340-8446d778c545\" (UID: \"8b954d96-a62c-434e-a340-8446d778c545\") " Dec 03 14:12:05 crc kubenswrapper[4805]: I1203 14:12:05.770830 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7frlc\" (UniqueName: \"kubernetes.io/projected/8b954d96-a62c-434e-a340-8446d778c545-kube-api-access-7frlc\") pod \"8b954d96-a62c-434e-a340-8446d778c545\" (UID: \"8b954d96-a62c-434e-a340-8446d778c545\") " Dec 03 14:12:05 crc kubenswrapper[4805]: I1203 14:12:05.771001 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/8b954d96-a62c-434e-a340-8446d778c545-config-volume\") pod \"8b954d96-a62c-434e-a340-8446d778c545\" (UID: \"8b954d96-a62c-434e-a340-8446d778c545\") " Dec 03 14:12:05 crc kubenswrapper[4805]: I1203 14:12:05.772313 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8b954d96-a62c-434e-a340-8446d778c545-config-volume" (OuterVolumeSpecName: "config-volume") pod "8b954d96-a62c-434e-a340-8446d778c545" (UID: "8b954d96-a62c-434e-a340-8446d778c545"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 14:12:05 crc kubenswrapper[4805]: I1203 14:12:05.775614 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8b954d96-a62c-434e-a340-8446d778c545-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "8b954d96-a62c-434e-a340-8446d778c545" (UID: "8b954d96-a62c-434e-a340-8446d778c545"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 14:12:05 crc kubenswrapper[4805]: I1203 14:12:05.775661 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8b954d96-a62c-434e-a340-8446d778c545-kube-api-access-7frlc" (OuterVolumeSpecName: "kube-api-access-7frlc") pod "8b954d96-a62c-434e-a340-8446d778c545" (UID: "8b954d96-a62c-434e-a340-8446d778c545"). InnerVolumeSpecName "kube-api-access-7frlc". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 14:12:05 crc kubenswrapper[4805]: I1203 14:12:05.818690 4805 patch_prober.go:28] interesting pod/router-default-5444994796-kxnqm container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 03 14:12:05 crc kubenswrapper[4805]: [-]has-synced failed: reason withheld Dec 03 14:12:05 crc kubenswrapper[4805]: [+]process-running ok Dec 03 14:12:05 crc kubenswrapper[4805]: healthz check failed Dec 03 14:12:05 crc kubenswrapper[4805]: I1203 14:12:05.818762 4805 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-kxnqm" podUID="b22cbc60-5975-42a4-85d3-6add6917fd88" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 03 14:12:05 crc kubenswrapper[4805]: I1203 14:12:05.872640 4805 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-dlmr9"] Dec 03 14:12:05 crc kubenswrapper[4805]: E1203 14:12:05.872820 4805 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8b954d96-a62c-434e-a340-8446d778c545" containerName="collect-profiles" Dec 03 14:12:05 crc kubenswrapper[4805]: I1203 14:12:05.872831 4805 state_mem.go:107] "Deleted CPUSet assignment" podUID="8b954d96-a62c-434e-a340-8446d778c545" containerName="collect-profiles" Dec 03 14:12:05 crc kubenswrapper[4805]: I1203 14:12:05.872962 4805 memory_manager.go:354] "RemoveStaleState removing state" podUID="8b954d96-a62c-434e-a340-8446d778c545" containerName="collect-profiles" Dec 03 14:12:05 crc kubenswrapper[4805]: I1203 14:12:05.873071 4805 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/8b954d96-a62c-434e-a340-8446d778c545-config-volume\") on node \"crc\" DevicePath \"\"" Dec 03 14:12:05 crc kubenswrapper[4805]: I1203 14:12:05.873155 4805 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7frlc\" (UniqueName: \"kubernetes.io/projected/8b954d96-a62c-434e-a340-8446d778c545-kube-api-access-7frlc\") on node \"crc\" DevicePath \"\"" Dec 03 14:12:05 crc kubenswrapper[4805]: I1203 14:12:05.873187 4805 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/8b954d96-a62c-434e-a340-8446d778c545-secret-volume\") on node \"crc\" DevicePath \"\"" Dec 03 14:12:05 crc kubenswrapper[4805]: I1203 14:12:05.873813 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-dlmr9" Dec 03 14:12:05 crc kubenswrapper[4805]: I1203 14:12:05.882017 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-dlmr9"] Dec 03 14:12:05 crc kubenswrapper[4805]: I1203 14:12:05.974678 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/77e93612-7c30-4f0e-9347-6e1373dfaaf8-utilities\") pod \"redhat-marketplace-dlmr9\" (UID: \"77e93612-7c30-4f0e-9347-6e1373dfaaf8\") " pod="openshift-marketplace/redhat-marketplace-dlmr9" Dec 03 14:12:05 crc kubenswrapper[4805]: I1203 14:12:05.974735 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/77e93612-7c30-4f0e-9347-6e1373dfaaf8-catalog-content\") pod \"redhat-marketplace-dlmr9\" (UID: \"77e93612-7c30-4f0e-9347-6e1373dfaaf8\") " pod="openshift-marketplace/redhat-marketplace-dlmr9" Dec 03 14:12:05 crc kubenswrapper[4805]: I1203 14:12:05.974761 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-s6blv\" (UniqueName: \"kubernetes.io/projected/77e93612-7c30-4f0e-9347-6e1373dfaaf8-kube-api-access-s6blv\") pod \"redhat-marketplace-dlmr9\" (UID: \"77e93612-7c30-4f0e-9347-6e1373dfaaf8\") " pod="openshift-marketplace/redhat-marketplace-dlmr9" Dec 03 14:12:05 crc kubenswrapper[4805]: I1203 14:12:05.976813 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-m7r99" Dec 03 14:12:06 crc kubenswrapper[4805]: I1203 14:12:06.077137 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/77e93612-7c30-4f0e-9347-6e1373dfaaf8-utilities\") pod \"redhat-marketplace-dlmr9\" (UID: \"77e93612-7c30-4f0e-9347-6e1373dfaaf8\") " pod="openshift-marketplace/redhat-marketplace-dlmr9" Dec 03 14:12:06 crc kubenswrapper[4805]: I1203 14:12:06.077189 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/77e93612-7c30-4f0e-9347-6e1373dfaaf8-catalog-content\") pod \"redhat-marketplace-dlmr9\" (UID: \"77e93612-7c30-4f0e-9347-6e1373dfaaf8\") " pod="openshift-marketplace/redhat-marketplace-dlmr9" Dec 03 14:12:06 crc kubenswrapper[4805]: I1203 14:12:06.077222 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s6blv\" (UniqueName: \"kubernetes.io/projected/77e93612-7c30-4f0e-9347-6e1373dfaaf8-kube-api-access-s6blv\") pod \"redhat-marketplace-dlmr9\" (UID: \"77e93612-7c30-4f0e-9347-6e1373dfaaf8\") " pod="openshift-marketplace/redhat-marketplace-dlmr9" Dec 03 14:12:06 crc kubenswrapper[4805]: I1203 14:12:06.080520 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/77e93612-7c30-4f0e-9347-6e1373dfaaf8-utilities\") pod \"redhat-marketplace-dlmr9\" (UID: \"77e93612-7c30-4f0e-9347-6e1373dfaaf8\") " pod="openshift-marketplace/redhat-marketplace-dlmr9" Dec 03 14:12:06 crc kubenswrapper[4805]: I1203 14:12:06.080788 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/77e93612-7c30-4f0e-9347-6e1373dfaaf8-catalog-content\") pod \"redhat-marketplace-dlmr9\" (UID: \"77e93612-7c30-4f0e-9347-6e1373dfaaf8\") " pod="openshift-marketplace/redhat-marketplace-dlmr9" Dec 03 14:12:06 crc kubenswrapper[4805]: I1203 14:12:06.100013 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s6blv\" (UniqueName: \"kubernetes.io/projected/77e93612-7c30-4f0e-9347-6e1373dfaaf8-kube-api-access-s6blv\") pod \"redhat-marketplace-dlmr9\" (UID: \"77e93612-7c30-4f0e-9347-6e1373dfaaf8\") " pod="openshift-marketplace/redhat-marketplace-dlmr9" Dec 03 14:12:06 crc kubenswrapper[4805]: I1203 14:12:06.166442 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-m7r99"] Dec 03 14:12:06 crc kubenswrapper[4805]: W1203 14:12:06.173624 4805 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod1e412276_9d78_4771_b447_b4e2b388e604.slice/crio-4a3c5188215db263a0646b5984c0ff692c2829ba3f3a403761ac309db9f6b223 WatchSource:0}: Error finding container 4a3c5188215db263a0646b5984c0ff692c2829ba3f3a403761ac309db9f6b223: Status 404 returned error can't find the container with id 4a3c5188215db263a0646b5984c0ff692c2829ba3f3a403761ac309db9f6b223 Dec 03 14:12:06 crc kubenswrapper[4805]: I1203 14:12:06.193405 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-dlmr9" Dec 03 14:12:06 crc kubenswrapper[4805]: I1203 14:12:06.249411 4805 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-config-operator/openshift-config-operator-7777fb866f-7wzgf" Dec 03 14:12:06 crc kubenswrapper[4805]: I1203 14:12:06.407946 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-jjzct" event={"ID":"4655a83e-4980-4a5c-9f61-e305ae418822","Type":"ContainerStarted","Data":"2eb46bb7d9082e503a63d58a515289f423d646a019ee0ec11dd33ae69c6562a5"} Dec 03 14:12:06 crc kubenswrapper[4805]: I1203 14:12:06.407996 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-jjzct" event={"ID":"4655a83e-4980-4a5c-9f61-e305ae418822","Type":"ContainerStarted","Data":"45e95adcdd0b0349f91467461f335ee50bd61445aaf7d6948daf31c5482d4d35"} Dec 03 14:12:06 crc kubenswrapper[4805]: I1203 14:12:06.408057 4805 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-image-registry/image-registry-697d97f7c8-jjzct" Dec 03 14:12:06 crc kubenswrapper[4805]: I1203 14:12:06.412677 4805 generic.go:334] "Generic (PLEG): container finished" podID="b2e4ac3a-e225-46ae-874d-4cf931c13464" containerID="397f4b6319d8994e0a243b3377f230736eaef70f2f17cdcde055b94bc9582296" exitCode=0 Dec 03 14:12:06 crc kubenswrapper[4805]: I1203 14:12:06.412753 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-8gt4x" event={"ID":"b2e4ac3a-e225-46ae-874d-4cf931c13464","Type":"ContainerDied","Data":"397f4b6319d8994e0a243b3377f230736eaef70f2f17cdcde055b94bc9582296"} Dec 03 14:12:06 crc kubenswrapper[4805]: I1203 14:12:06.414658 4805 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 03 14:12:06 crc kubenswrapper[4805]: I1203 14:12:06.422214 4805 generic.go:334] "Generic (PLEG): container finished" podID="344cb196-3baa-48c0-abcb-7e46cbce614d" containerID="2fd386dcb356443b4d70345cf5a236e1680e0f2dce2c6a2c7a8ab14803d8d28e" exitCode=0 Dec 03 14:12:06 crc kubenswrapper[4805]: I1203 14:12:06.422384 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-w6nm5" event={"ID":"344cb196-3baa-48c0-abcb-7e46cbce614d","Type":"ContainerDied","Data":"2fd386dcb356443b4d70345cf5a236e1680e0f2dce2c6a2c7a8ab14803d8d28e"} Dec 03 14:12:06 crc kubenswrapper[4805]: I1203 14:12:06.427826 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-dlmr9"] Dec 03 14:12:06 crc kubenswrapper[4805]: I1203 14:12:06.428461 4805 generic.go:334] "Generic (PLEG): container finished" podID="1e412276-9d78-4771-b447-b4e2b388e604" containerID="01c7edcfff6ab9f5d32726a538449aba4ba91c0b84f5c45ed5b86a66f255feb5" exitCode=0 Dec 03 14:12:06 crc kubenswrapper[4805]: I1203 14:12:06.428546 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-m7r99" event={"ID":"1e412276-9d78-4771-b447-b4e2b388e604","Type":"ContainerDied","Data":"01c7edcfff6ab9f5d32726a538449aba4ba91c0b84f5c45ed5b86a66f255feb5"} Dec 03 14:12:06 crc kubenswrapper[4805]: I1203 14:12:06.428586 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-m7r99" event={"ID":"1e412276-9d78-4771-b447-b4e2b388e604","Type":"ContainerStarted","Data":"4a3c5188215db263a0646b5984c0ff692c2829ba3f3a403761ac309db9f6b223"} Dec 03 14:12:06 crc kubenswrapper[4805]: I1203 14:12:06.436369 4805 generic.go:334] "Generic (PLEG): container finished" podID="64f6384f-41e0-469e-8920-27e9e97b62db" containerID="0c30658b6b2df48a53b073c612c15a2418c09abafbcee03a2cb9f43d89c0c547" exitCode=0 Dec 03 14:12:06 crc kubenswrapper[4805]: I1203 14:12:06.436497 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-9cm4t" event={"ID":"64f6384f-41e0-469e-8920-27e9e97b62db","Type":"ContainerDied","Data":"0c30658b6b2df48a53b073c612c15a2418c09abafbcee03a2cb9f43d89c0c547"} Dec 03 14:12:06 crc kubenswrapper[4805]: I1203 14:12:06.438747 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29412840-gj7pr" event={"ID":"8b954d96-a62c-434e-a340-8446d778c545","Type":"ContainerDied","Data":"3d3293741e2741cb9b99650b37e0faa09813046145b469055b1661244883e262"} Dec 03 14:12:06 crc kubenswrapper[4805]: I1203 14:12:06.438787 4805 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="3d3293741e2741cb9b99650b37e0faa09813046145b469055b1661244883e262" Dec 03 14:12:06 crc kubenswrapper[4805]: I1203 14:12:06.438815 4805 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29412840-gj7pr" Dec 03 14:12:06 crc kubenswrapper[4805]: I1203 14:12:06.446703 4805 generic.go:334] "Generic (PLEG): container finished" podID="731aa7bc-cde1-4ff7-aa29-78093078514d" containerID="1c047e20d07ca13ec2cd6254198e5bb079268d59b37a71ae8421c01cc5167e62" exitCode=0 Dec 03 14:12:06 crc kubenswrapper[4805]: I1203 14:12:06.446756 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-fbrtf" event={"ID":"731aa7bc-cde1-4ff7-aa29-78093078514d","Type":"ContainerDied","Data":"1c047e20d07ca13ec2cd6254198e5bb079268d59b37a71ae8421c01cc5167e62"} Dec 03 14:12:06 crc kubenswrapper[4805]: I1203 14:12:06.469302 4805 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/image-registry-697d97f7c8-jjzct" podStartSLOduration=137.469273426 podStartE2EDuration="2m17.469273426s" podCreationTimestamp="2025-12-03 14:09:49 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 14:12:06.436317666 +0000 UTC m=+156.099234589" watchObservedRunningTime="2025-12-03 14:12:06.469273426 +0000 UTC m=+156.132190349" Dec 03 14:12:06 crc kubenswrapper[4805]: I1203 14:12:06.485670 4805 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-bd7j9"] Dec 03 14:12:06 crc kubenswrapper[4805]: I1203 14:12:06.488563 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-bd7j9" Dec 03 14:12:06 crc kubenswrapper[4805]: I1203 14:12:06.491469 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-bd7j9"] Dec 03 14:12:06 crc kubenswrapper[4805]: I1203 14:12:06.496252 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-operators-dockercfg-ct8rh" Dec 03 14:12:06 crc kubenswrapper[4805]: I1203 14:12:06.593562 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d0a05061-dab0-4a85-be73-b1d2acbc7b8f-catalog-content\") pod \"redhat-operators-bd7j9\" (UID: \"d0a05061-dab0-4a85-be73-b1d2acbc7b8f\") " pod="openshift-marketplace/redhat-operators-bd7j9" Dec 03 14:12:06 crc kubenswrapper[4805]: I1203 14:12:06.593892 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d0a05061-dab0-4a85-be73-b1d2acbc7b8f-utilities\") pod \"redhat-operators-bd7j9\" (UID: \"d0a05061-dab0-4a85-be73-b1d2acbc7b8f\") " pod="openshift-marketplace/redhat-operators-bd7j9" Dec 03 14:12:06 crc kubenswrapper[4805]: I1203 14:12:06.593973 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-74jhd\" (UniqueName: \"kubernetes.io/projected/d0a05061-dab0-4a85-be73-b1d2acbc7b8f-kube-api-access-74jhd\") pod \"redhat-operators-bd7j9\" (UID: \"d0a05061-dab0-4a85-be73-b1d2acbc7b8f\") " pod="openshift-marketplace/redhat-operators-bd7j9" Dec 03 14:12:06 crc kubenswrapper[4805]: I1203 14:12:06.673818 4805 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-dcw96"] Dec 03 14:12:06 crc kubenswrapper[4805]: I1203 14:12:06.674683 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-dcw96" Dec 03 14:12:06 crc kubenswrapper[4805]: I1203 14:12:06.685372 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-dcw96"] Dec 03 14:12:06 crc kubenswrapper[4805]: I1203 14:12:06.701642 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d0a05061-dab0-4a85-be73-b1d2acbc7b8f-catalog-content\") pod \"redhat-operators-bd7j9\" (UID: \"d0a05061-dab0-4a85-be73-b1d2acbc7b8f\") " pod="openshift-marketplace/redhat-operators-bd7j9" Dec 03 14:12:06 crc kubenswrapper[4805]: I1203 14:12:06.701776 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d0a05061-dab0-4a85-be73-b1d2acbc7b8f-utilities\") pod \"redhat-operators-bd7j9\" (UID: \"d0a05061-dab0-4a85-be73-b1d2acbc7b8f\") " pod="openshift-marketplace/redhat-operators-bd7j9" Dec 03 14:12:06 crc kubenswrapper[4805]: I1203 14:12:06.701802 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-74jhd\" (UniqueName: \"kubernetes.io/projected/d0a05061-dab0-4a85-be73-b1d2acbc7b8f-kube-api-access-74jhd\") pod \"redhat-operators-bd7j9\" (UID: \"d0a05061-dab0-4a85-be73-b1d2acbc7b8f\") " pod="openshift-marketplace/redhat-operators-bd7j9" Dec 03 14:12:06 crc kubenswrapper[4805]: I1203 14:12:06.702197 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d0a05061-dab0-4a85-be73-b1d2acbc7b8f-catalog-content\") pod \"redhat-operators-bd7j9\" (UID: \"d0a05061-dab0-4a85-be73-b1d2acbc7b8f\") " pod="openshift-marketplace/redhat-operators-bd7j9" Dec 03 14:12:06 crc kubenswrapper[4805]: I1203 14:12:06.702247 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d0a05061-dab0-4a85-be73-b1d2acbc7b8f-utilities\") pod \"redhat-operators-bd7j9\" (UID: \"d0a05061-dab0-4a85-be73-b1d2acbc7b8f\") " pod="openshift-marketplace/redhat-operators-bd7j9" Dec 03 14:12:06 crc kubenswrapper[4805]: I1203 14:12:06.711106 4805 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8f668bae-612b-4b75-9490-919e737c6a3b" path="/var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes" Dec 03 14:12:06 crc kubenswrapper[4805]: I1203 14:12:06.725102 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-74jhd\" (UniqueName: \"kubernetes.io/projected/d0a05061-dab0-4a85-be73-b1d2acbc7b8f-kube-api-access-74jhd\") pod \"redhat-operators-bd7j9\" (UID: \"d0a05061-dab0-4a85-be73-b1d2acbc7b8f\") " pod="openshift-marketplace/redhat-operators-bd7j9" Dec 03 14:12:06 crc kubenswrapper[4805]: I1203 14:12:06.803655 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8fd4f74e-3b48-4a4a-8898-e76c5926abab-catalog-content\") pod \"redhat-operators-dcw96\" (UID: \"8fd4f74e-3b48-4a4a-8898-e76c5926abab\") " pod="openshift-marketplace/redhat-operators-dcw96" Dec 03 14:12:06 crc kubenswrapper[4805]: I1203 14:12:06.803721 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8fd4f74e-3b48-4a4a-8898-e76c5926abab-utilities\") pod \"redhat-operators-dcw96\" (UID: \"8fd4f74e-3b48-4a4a-8898-e76c5926abab\") " pod="openshift-marketplace/redhat-operators-dcw96" Dec 03 14:12:06 crc kubenswrapper[4805]: I1203 14:12:06.803779 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-btdbn\" (UniqueName: \"kubernetes.io/projected/8fd4f74e-3b48-4a4a-8898-e76c5926abab-kube-api-access-btdbn\") pod \"redhat-operators-dcw96\" (UID: \"8fd4f74e-3b48-4a4a-8898-e76c5926abab\") " pod="openshift-marketplace/redhat-operators-dcw96" Dec 03 14:12:06 crc kubenswrapper[4805]: I1203 14:12:06.820160 4805 patch_prober.go:28] interesting pod/router-default-5444994796-kxnqm container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 03 14:12:06 crc kubenswrapper[4805]: [-]has-synced failed: reason withheld Dec 03 14:12:06 crc kubenswrapper[4805]: [+]process-running ok Dec 03 14:12:06 crc kubenswrapper[4805]: healthz check failed Dec 03 14:12:06 crc kubenswrapper[4805]: I1203 14:12:06.820234 4805 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-kxnqm" podUID="b22cbc60-5975-42a4-85d3-6add6917fd88" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 03 14:12:06 crc kubenswrapper[4805]: I1203 14:12:06.862656 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-bd7j9" Dec 03 14:12:06 crc kubenswrapper[4805]: I1203 14:12:06.904731 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8fd4f74e-3b48-4a4a-8898-e76c5926abab-catalog-content\") pod \"redhat-operators-dcw96\" (UID: \"8fd4f74e-3b48-4a4a-8898-e76c5926abab\") " pod="openshift-marketplace/redhat-operators-dcw96" Dec 03 14:12:06 crc kubenswrapper[4805]: I1203 14:12:06.904795 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8fd4f74e-3b48-4a4a-8898-e76c5926abab-utilities\") pod \"redhat-operators-dcw96\" (UID: \"8fd4f74e-3b48-4a4a-8898-e76c5926abab\") " pod="openshift-marketplace/redhat-operators-dcw96" Dec 03 14:12:06 crc kubenswrapper[4805]: I1203 14:12:06.904821 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-btdbn\" (UniqueName: \"kubernetes.io/projected/8fd4f74e-3b48-4a4a-8898-e76c5926abab-kube-api-access-btdbn\") pod \"redhat-operators-dcw96\" (UID: \"8fd4f74e-3b48-4a4a-8898-e76c5926abab\") " pod="openshift-marketplace/redhat-operators-dcw96" Dec 03 14:12:06 crc kubenswrapper[4805]: I1203 14:12:06.905665 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8fd4f74e-3b48-4a4a-8898-e76c5926abab-catalog-content\") pod \"redhat-operators-dcw96\" (UID: \"8fd4f74e-3b48-4a4a-8898-e76c5926abab\") " pod="openshift-marketplace/redhat-operators-dcw96" Dec 03 14:12:06 crc kubenswrapper[4805]: I1203 14:12:06.905681 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8fd4f74e-3b48-4a4a-8898-e76c5926abab-utilities\") pod \"redhat-operators-dcw96\" (UID: \"8fd4f74e-3b48-4a4a-8898-e76c5926abab\") " pod="openshift-marketplace/redhat-operators-dcw96" Dec 03 14:12:06 crc kubenswrapper[4805]: I1203 14:12:06.922555 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-btdbn\" (UniqueName: \"kubernetes.io/projected/8fd4f74e-3b48-4a4a-8898-e76c5926abab-kube-api-access-btdbn\") pod \"redhat-operators-dcw96\" (UID: \"8fd4f74e-3b48-4a4a-8898-e76c5926abab\") " pod="openshift-marketplace/redhat-operators-dcw96" Dec 03 14:12:07 crc kubenswrapper[4805]: I1203 14:12:06.995239 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-dcw96" Dec 03 14:12:07 crc kubenswrapper[4805]: I1203 14:12:07.017044 4805 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-apiserver/apiserver-76f77b778f-7nhvl" Dec 03 14:12:07 crc kubenswrapper[4805]: I1203 14:12:07.017113 4805 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-apiserver/apiserver-76f77b778f-7nhvl" Dec 03 14:12:07 crc kubenswrapper[4805]: I1203 14:12:07.025699 4805 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-apiserver/apiserver-76f77b778f-7nhvl" Dec 03 14:12:07 crc kubenswrapper[4805]: I1203 14:12:07.112080 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-bd7j9"] Dec 03 14:12:07 crc kubenswrapper[4805]: I1203 14:12:07.233219 4805 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console/console-f9d7485db-pp9tv" Dec 03 14:12:07 crc kubenswrapper[4805]: I1203 14:12:07.233464 4805 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-console/console-f9d7485db-pp9tv" Dec 03 14:12:07 crc kubenswrapper[4805]: I1203 14:12:07.235682 4805 patch_prober.go:28] interesting pod/console-f9d7485db-pp9tv container/console namespace/openshift-console: Startup probe status=failure output="Get \"https://10.217.0.27:8443/health\": dial tcp 10.217.0.27:8443: connect: connection refused" start-of-body= Dec 03 14:12:07 crc kubenswrapper[4805]: I1203 14:12:07.235720 4805 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-console/console-f9d7485db-pp9tv" podUID="e2034338-8a35-4fd2-9bbf-be733939fc25" containerName="console" probeResult="failure" output="Get \"https://10.217.0.27:8443/health\": dial tcp 10.217.0.27:8443: connect: connection refused" Dec 03 14:12:07 crc kubenswrapper[4805]: I1203 14:12:07.244097 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-dcw96"] Dec 03 14:12:07 crc kubenswrapper[4805]: W1203 14:12:07.257698 4805 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod8fd4f74e_3b48_4a4a_8898_e76c5926abab.slice/crio-f5c09446dc047ae6e72d08ffb758c00dbecc9e5462264e08716ecc22716616db WatchSource:0}: Error finding container f5c09446dc047ae6e72d08ffb758c00dbecc9e5462264e08716ecc22716616db: Status 404 returned error can't find the container with id f5c09446dc047ae6e72d08ffb758c00dbecc9e5462264e08716ecc22716616db Dec 03 14:12:07 crc kubenswrapper[4805]: I1203 14:12:07.370390 4805 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-kfh2f" Dec 03 14:12:07 crc kubenswrapper[4805]: I1203 14:12:07.383073 4805 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-kfh2f" Dec 03 14:12:07 crc kubenswrapper[4805]: I1203 14:12:07.470775 4805 generic.go:334] "Generic (PLEG): container finished" podID="d0a05061-dab0-4a85-be73-b1d2acbc7b8f" containerID="38be2c7eb01fb25fa571a38edfabdf0beede1d902fd2c2f2d3bf004010e331e8" exitCode=0 Dec 03 14:12:07 crc kubenswrapper[4805]: I1203 14:12:07.470914 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-bd7j9" event={"ID":"d0a05061-dab0-4a85-be73-b1d2acbc7b8f","Type":"ContainerDied","Data":"38be2c7eb01fb25fa571a38edfabdf0beede1d902fd2c2f2d3bf004010e331e8"} Dec 03 14:12:07 crc kubenswrapper[4805]: I1203 14:12:07.475753 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-bd7j9" event={"ID":"d0a05061-dab0-4a85-be73-b1d2acbc7b8f","Type":"ContainerStarted","Data":"77d1a04380488ca0fe6609e31fd04dd77e33e0987bd8b09f722d503c4dcaa265"} Dec 03 14:12:07 crc kubenswrapper[4805]: I1203 14:12:07.498258 4805 generic.go:334] "Generic (PLEG): container finished" podID="77e93612-7c30-4f0e-9347-6e1373dfaaf8" containerID="d9d2f8339d6ff1d43a7fe3e0a95de4dfa69ac0deced0520fc10abdffb92b64dc" exitCode=0 Dec 03 14:12:07 crc kubenswrapper[4805]: I1203 14:12:07.498440 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-dlmr9" event={"ID":"77e93612-7c30-4f0e-9347-6e1373dfaaf8","Type":"ContainerDied","Data":"d9d2f8339d6ff1d43a7fe3e0a95de4dfa69ac0deced0520fc10abdffb92b64dc"} Dec 03 14:12:07 crc kubenswrapper[4805]: I1203 14:12:07.498470 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-dlmr9" event={"ID":"77e93612-7c30-4f0e-9347-6e1373dfaaf8","Type":"ContainerStarted","Data":"60fe01cf43710b6e306d919ebd344e08920a9544ca27ddf906ddec2535a10f9a"} Dec 03 14:12:07 crc kubenswrapper[4805]: I1203 14:12:07.501887 4805 patch_prober.go:28] interesting pod/downloads-7954f5f757-s25zl container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.11:8080/\": dial tcp 10.217.0.11:8080: connect: connection refused" start-of-body= Dec 03 14:12:07 crc kubenswrapper[4805]: I1203 14:12:07.501886 4805 patch_prober.go:28] interesting pod/downloads-7954f5f757-s25zl container/download-server namespace/openshift-console: Liveness probe status=failure output="Get \"http://10.217.0.11:8080/\": dial tcp 10.217.0.11:8080: connect: connection refused" start-of-body= Dec 03 14:12:07 crc kubenswrapper[4805]: I1203 14:12:07.501924 4805 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-s25zl" podUID="6c2d6fbc-771d-4aff-b1e8-3eda17204931" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.11:8080/\": dial tcp 10.217.0.11:8080: connect: connection refused" Dec 03 14:12:07 crc kubenswrapper[4805]: I1203 14:12:07.501935 4805 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-console/downloads-7954f5f757-s25zl" podUID="6c2d6fbc-771d-4aff-b1e8-3eda17204931" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.11:8080/\": dial tcp 10.217.0.11:8080: connect: connection refused" Dec 03 14:12:07 crc kubenswrapper[4805]: I1203 14:12:07.502785 4805 generic.go:334] "Generic (PLEG): container finished" podID="8fd4f74e-3b48-4a4a-8898-e76c5926abab" containerID="55d20a2330bc6375f2d04585cfeb78ca414231ded021e913ef6546a6ce39ab03" exitCode=0 Dec 03 14:12:07 crc kubenswrapper[4805]: I1203 14:12:07.503791 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-dcw96" event={"ID":"8fd4f74e-3b48-4a4a-8898-e76c5926abab","Type":"ContainerDied","Data":"55d20a2330bc6375f2d04585cfeb78ca414231ded021e913ef6546a6ce39ab03"} Dec 03 14:12:07 crc kubenswrapper[4805]: I1203 14:12:07.503814 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-dcw96" event={"ID":"8fd4f74e-3b48-4a4a-8898-e76c5926abab","Type":"ContainerStarted","Data":"f5c09446dc047ae6e72d08ffb758c00dbecc9e5462264e08716ecc22716616db"} Dec 03 14:12:07 crc kubenswrapper[4805]: I1203 14:12:07.516718 4805 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-apiserver/apiserver-76f77b778f-7nhvl" Dec 03 14:12:07 crc kubenswrapper[4805]: I1203 14:12:07.828860 4805 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ingress/router-default-5444994796-kxnqm" Dec 03 14:12:07 crc kubenswrapper[4805]: I1203 14:12:07.836867 4805 patch_prober.go:28] interesting pod/router-default-5444994796-kxnqm container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 03 14:12:07 crc kubenswrapper[4805]: [-]has-synced failed: reason withheld Dec 03 14:12:07 crc kubenswrapper[4805]: [+]process-running ok Dec 03 14:12:07 crc kubenswrapper[4805]: healthz check failed Dec 03 14:12:07 crc kubenswrapper[4805]: I1203 14:12:07.836932 4805 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-kxnqm" podUID="b22cbc60-5975-42a4-85d3-6add6917fd88" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 03 14:12:08 crc kubenswrapper[4805]: I1203 14:12:08.066889 4805 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-controller-manager/revision-pruner-9-crc"] Dec 03 14:12:08 crc kubenswrapper[4805]: I1203 14:12:08.067848 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Dec 03 14:12:08 crc kubenswrapper[4805]: I1203 14:12:08.071062 4805 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager"/"kube-root-ca.crt" Dec 03 14:12:08 crc kubenswrapper[4805]: I1203 14:12:08.075450 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager"/"installer-sa-dockercfg-kjl2n" Dec 03 14:12:08 crc kubenswrapper[4805]: I1203 14:12:08.080756 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager/revision-pruner-9-crc"] Dec 03 14:12:08 crc kubenswrapper[4805]: I1203 14:12:08.134468 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/a6e8e8f2-46cf-4322-8fdb-057a28032e66-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"a6e8e8f2-46cf-4322-8fdb-057a28032e66\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Dec 03 14:12:08 crc kubenswrapper[4805]: I1203 14:12:08.134523 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/a6e8e8f2-46cf-4322-8fdb-057a28032e66-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"a6e8e8f2-46cf-4322-8fdb-057a28032e66\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Dec 03 14:12:08 crc kubenswrapper[4805]: I1203 14:12:08.236244 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/a6e8e8f2-46cf-4322-8fdb-057a28032e66-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"a6e8e8f2-46cf-4322-8fdb-057a28032e66\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Dec 03 14:12:08 crc kubenswrapper[4805]: I1203 14:12:08.236330 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/a6e8e8f2-46cf-4322-8fdb-057a28032e66-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"a6e8e8f2-46cf-4322-8fdb-057a28032e66\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Dec 03 14:12:08 crc kubenswrapper[4805]: I1203 14:12:08.236423 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/a6e8e8f2-46cf-4322-8fdb-057a28032e66-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"a6e8e8f2-46cf-4322-8fdb-057a28032e66\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Dec 03 14:12:08 crc kubenswrapper[4805]: I1203 14:12:08.274604 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/a6e8e8f2-46cf-4322-8fdb-057a28032e66-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"a6e8e8f2-46cf-4322-8fdb-057a28032e66\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Dec 03 14:12:08 crc kubenswrapper[4805]: I1203 14:12:08.408608 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Dec 03 14:12:08 crc kubenswrapper[4805]: I1203 14:12:08.820744 4805 patch_prober.go:28] interesting pod/router-default-5444994796-kxnqm container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 03 14:12:08 crc kubenswrapper[4805]: [-]has-synced failed: reason withheld Dec 03 14:12:08 crc kubenswrapper[4805]: [+]process-running ok Dec 03 14:12:08 crc kubenswrapper[4805]: healthz check failed Dec 03 14:12:08 crc kubenswrapper[4805]: I1203 14:12:08.821269 4805 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-kxnqm" podUID="b22cbc60-5975-42a4-85d3-6add6917fd88" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 03 14:12:08 crc kubenswrapper[4805]: I1203 14:12:08.879630 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager/revision-pruner-9-crc"] Dec 03 14:12:08 crc kubenswrapper[4805]: I1203 14:12:08.950033 4805 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/revision-pruner-8-crc"] Dec 03 14:12:08 crc kubenswrapper[4805]: I1203 14:12:08.950984 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Dec 03 14:12:08 crc kubenswrapper[4805]: I1203 14:12:08.954334 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/revision-pruner-8-crc"] Dec 03 14:12:08 crc kubenswrapper[4805]: I1203 14:12:08.955400 4805 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver"/"kube-root-ca.crt" Dec 03 14:12:08 crc kubenswrapper[4805]: I1203 14:12:08.955739 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver"/"installer-sa-dockercfg-5pr6n" Dec 03 14:12:09 crc kubenswrapper[4805]: I1203 14:12:09.051626 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/c9b110a7-f82f-4a3c-a43a-6650b2d8f2da-kube-api-access\") pod \"revision-pruner-8-crc\" (UID: \"c9b110a7-f82f-4a3c-a43a-6650b2d8f2da\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Dec 03 14:12:09 crc kubenswrapper[4805]: I1203 14:12:09.051703 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/c9b110a7-f82f-4a3c-a43a-6650b2d8f2da-kubelet-dir\") pod \"revision-pruner-8-crc\" (UID: \"c9b110a7-f82f-4a3c-a43a-6650b2d8f2da\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Dec 03 14:12:09 crc kubenswrapper[4805]: I1203 14:12:09.154427 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/c9b110a7-f82f-4a3c-a43a-6650b2d8f2da-kube-api-access\") pod \"revision-pruner-8-crc\" (UID: \"c9b110a7-f82f-4a3c-a43a-6650b2d8f2da\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Dec 03 14:12:09 crc kubenswrapper[4805]: I1203 14:12:09.154629 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/c9b110a7-f82f-4a3c-a43a-6650b2d8f2da-kubelet-dir\") pod \"revision-pruner-8-crc\" (UID: \"c9b110a7-f82f-4a3c-a43a-6650b2d8f2da\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Dec 03 14:12:09 crc kubenswrapper[4805]: I1203 14:12:09.154804 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/c9b110a7-f82f-4a3c-a43a-6650b2d8f2da-kubelet-dir\") pod \"revision-pruner-8-crc\" (UID: \"c9b110a7-f82f-4a3c-a43a-6650b2d8f2da\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Dec 03 14:12:09 crc kubenswrapper[4805]: I1203 14:12:09.176749 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/c9b110a7-f82f-4a3c-a43a-6650b2d8f2da-kube-api-access\") pod \"revision-pruner-8-crc\" (UID: \"c9b110a7-f82f-4a3c-a43a-6650b2d8f2da\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Dec 03 14:12:09 crc kubenswrapper[4805]: I1203 14:12:09.297144 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Dec 03 14:12:09 crc kubenswrapper[4805]: I1203 14:12:09.545887 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/revision-pruner-9-crc" event={"ID":"a6e8e8f2-46cf-4322-8fdb-057a28032e66","Type":"ContainerStarted","Data":"f03fc6f62aa8f755a998694a7ab47e4ce5b5289cd11b0ab2c2d220329b0d05b2"} Dec 03 14:12:09 crc kubenswrapper[4805]: I1203 14:12:09.785235 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/revision-pruner-8-crc"] Dec 03 14:12:09 crc kubenswrapper[4805]: I1203 14:12:09.823650 4805 patch_prober.go:28] interesting pod/router-default-5444994796-kxnqm container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 03 14:12:09 crc kubenswrapper[4805]: [-]has-synced failed: reason withheld Dec 03 14:12:09 crc kubenswrapper[4805]: [+]process-running ok Dec 03 14:12:09 crc kubenswrapper[4805]: healthz check failed Dec 03 14:12:09 crc kubenswrapper[4805]: I1203 14:12:09.823727 4805 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-kxnqm" podUID="b22cbc60-5975-42a4-85d3-6add6917fd88" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 03 14:12:10 crc kubenswrapper[4805]: I1203 14:12:10.167224 4805 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-dns/dns-default-7wnwf" Dec 03 14:12:10 crc kubenswrapper[4805]: I1203 14:12:10.560590 4805 generic.go:334] "Generic (PLEG): container finished" podID="a6e8e8f2-46cf-4322-8fdb-057a28032e66" containerID="3f23521fc52c130e7e0e2e74f21d817d35df0dab8fb209da46b8bfe96e23dd36" exitCode=0 Dec 03 14:12:10 crc kubenswrapper[4805]: I1203 14:12:10.560696 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/revision-pruner-9-crc" event={"ID":"a6e8e8f2-46cf-4322-8fdb-057a28032e66","Type":"ContainerDied","Data":"3f23521fc52c130e7e0e2e74f21d817d35df0dab8fb209da46b8bfe96e23dd36"} Dec 03 14:12:10 crc kubenswrapper[4805]: I1203 14:12:10.565336 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-8-crc" event={"ID":"c9b110a7-f82f-4a3c-a43a-6650b2d8f2da","Type":"ContainerStarted","Data":"13d5cb42a3f4fb8b95bd1117916529d554eb863a7d014ef1a1dc82ff7481e5c7"} Dec 03 14:12:10 crc kubenswrapper[4805]: I1203 14:12:10.565388 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-8-crc" event={"ID":"c9b110a7-f82f-4a3c-a43a-6650b2d8f2da","Type":"ContainerStarted","Data":"58e812ac15e6ffcc2cce5d3c0537eab4d370d2f585562d3488f98be3bf1b5237"} Dec 03 14:12:10 crc kubenswrapper[4805]: I1203 14:12:10.819329 4805 patch_prober.go:28] interesting pod/router-default-5444994796-kxnqm container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 03 14:12:10 crc kubenswrapper[4805]: [-]has-synced failed: reason withheld Dec 03 14:12:10 crc kubenswrapper[4805]: [+]process-running ok Dec 03 14:12:10 crc kubenswrapper[4805]: healthz check failed Dec 03 14:12:10 crc kubenswrapper[4805]: I1203 14:12:10.819431 4805 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-kxnqm" podUID="b22cbc60-5975-42a4-85d3-6add6917fd88" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 03 14:12:11 crc kubenswrapper[4805]: I1203 14:12:11.578296 4805 generic.go:334] "Generic (PLEG): container finished" podID="c9b110a7-f82f-4a3c-a43a-6650b2d8f2da" containerID="13d5cb42a3f4fb8b95bd1117916529d554eb863a7d014ef1a1dc82ff7481e5c7" exitCode=0 Dec 03 14:12:11 crc kubenswrapper[4805]: I1203 14:12:11.578372 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-8-crc" event={"ID":"c9b110a7-f82f-4a3c-a43a-6650b2d8f2da","Type":"ContainerDied","Data":"13d5cb42a3f4fb8b95bd1117916529d554eb863a7d014ef1a1dc82ff7481e5c7"} Dec 03 14:12:11 crc kubenswrapper[4805]: I1203 14:12:11.821343 4805 patch_prober.go:28] interesting pod/router-default-5444994796-kxnqm container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 03 14:12:11 crc kubenswrapper[4805]: [-]has-synced failed: reason withheld Dec 03 14:12:11 crc kubenswrapper[4805]: [+]process-running ok Dec 03 14:12:11 crc kubenswrapper[4805]: healthz check failed Dec 03 14:12:11 crc kubenswrapper[4805]: I1203 14:12:11.821736 4805 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-kxnqm" podUID="b22cbc60-5975-42a4-85d3-6add6917fd88" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 03 14:12:11 crc kubenswrapper[4805]: I1203 14:12:11.875820 4805 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Dec 03 14:12:12 crc kubenswrapper[4805]: I1203 14:12:12.023309 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/a6e8e8f2-46cf-4322-8fdb-057a28032e66-kubelet-dir\") pod \"a6e8e8f2-46cf-4322-8fdb-057a28032e66\" (UID: \"a6e8e8f2-46cf-4322-8fdb-057a28032e66\") " Dec 03 14:12:12 crc kubenswrapper[4805]: I1203 14:12:12.023413 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/a6e8e8f2-46cf-4322-8fdb-057a28032e66-kubelet-dir" (OuterVolumeSpecName: "kubelet-dir") pod "a6e8e8f2-46cf-4322-8fdb-057a28032e66" (UID: "a6e8e8f2-46cf-4322-8fdb-057a28032e66"). InnerVolumeSpecName "kubelet-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 03 14:12:12 crc kubenswrapper[4805]: I1203 14:12:12.023436 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/a6e8e8f2-46cf-4322-8fdb-057a28032e66-kube-api-access\") pod \"a6e8e8f2-46cf-4322-8fdb-057a28032e66\" (UID: \"a6e8e8f2-46cf-4322-8fdb-057a28032e66\") " Dec 03 14:12:12 crc kubenswrapper[4805]: I1203 14:12:12.023692 4805 reconciler_common.go:293] "Volume detached for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/a6e8e8f2-46cf-4322-8fdb-057a28032e66-kubelet-dir\") on node \"crc\" DevicePath \"\"" Dec 03 14:12:12 crc kubenswrapper[4805]: I1203 14:12:12.044319 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a6e8e8f2-46cf-4322-8fdb-057a28032e66-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "a6e8e8f2-46cf-4322-8fdb-057a28032e66" (UID: "a6e8e8f2-46cf-4322-8fdb-057a28032e66"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 14:12:12 crc kubenswrapper[4805]: I1203 14:12:12.125008 4805 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/a6e8e8f2-46cf-4322-8fdb-057a28032e66-kube-api-access\") on node \"crc\" DevicePath \"\"" Dec 03 14:12:12 crc kubenswrapper[4805]: I1203 14:12:12.589003 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/revision-pruner-9-crc" event={"ID":"a6e8e8f2-46cf-4322-8fdb-057a28032e66","Type":"ContainerDied","Data":"f03fc6f62aa8f755a998694a7ab47e4ce5b5289cd11b0ab2c2d220329b0d05b2"} Dec 03 14:12:12 crc kubenswrapper[4805]: I1203 14:12:12.589060 4805 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="f03fc6f62aa8f755a998694a7ab47e4ce5b5289cd11b0ab2c2d220329b0d05b2" Dec 03 14:12:12 crc kubenswrapper[4805]: I1203 14:12:12.589023 4805 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Dec 03 14:12:12 crc kubenswrapper[4805]: I1203 14:12:12.631852 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/6b36a3db-1047-487a-9d6c-4f51ffdd8cc1-metrics-certs\") pod \"network-metrics-daemon-4gd9r\" (UID: \"6b36a3db-1047-487a-9d6c-4f51ffdd8cc1\") " pod="openshift-multus/network-metrics-daemon-4gd9r" Dec 03 14:12:12 crc kubenswrapper[4805]: I1203 14:12:12.635908 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/6b36a3db-1047-487a-9d6c-4f51ffdd8cc1-metrics-certs\") pod \"network-metrics-daemon-4gd9r\" (UID: \"6b36a3db-1047-487a-9d6c-4f51ffdd8cc1\") " pod="openshift-multus/network-metrics-daemon-4gd9r" Dec 03 14:12:12 crc kubenswrapper[4805]: I1203 14:12:12.821178 4805 patch_prober.go:28] interesting pod/router-default-5444994796-kxnqm container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 03 14:12:12 crc kubenswrapper[4805]: [-]has-synced failed: reason withheld Dec 03 14:12:12 crc kubenswrapper[4805]: [+]process-running ok Dec 03 14:12:12 crc kubenswrapper[4805]: healthz check failed Dec 03 14:12:12 crc kubenswrapper[4805]: I1203 14:12:12.822356 4805 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-kxnqm" podUID="b22cbc60-5975-42a4-85d3-6add6917fd88" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 03 14:12:12 crc kubenswrapper[4805]: I1203 14:12:12.875627 4805 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Dec 03 14:12:12 crc kubenswrapper[4805]: I1203 14:12:12.914641 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-4gd9r" Dec 03 14:12:13 crc kubenswrapper[4805]: I1203 14:12:13.050021 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/c9b110a7-f82f-4a3c-a43a-6650b2d8f2da-kube-api-access\") pod \"c9b110a7-f82f-4a3c-a43a-6650b2d8f2da\" (UID: \"c9b110a7-f82f-4a3c-a43a-6650b2d8f2da\") " Dec 03 14:12:13 crc kubenswrapper[4805]: I1203 14:12:13.050166 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/c9b110a7-f82f-4a3c-a43a-6650b2d8f2da-kubelet-dir\") pod \"c9b110a7-f82f-4a3c-a43a-6650b2d8f2da\" (UID: \"c9b110a7-f82f-4a3c-a43a-6650b2d8f2da\") " Dec 03 14:12:13 crc kubenswrapper[4805]: I1203 14:12:13.050226 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/c9b110a7-f82f-4a3c-a43a-6650b2d8f2da-kubelet-dir" (OuterVolumeSpecName: "kubelet-dir") pod "c9b110a7-f82f-4a3c-a43a-6650b2d8f2da" (UID: "c9b110a7-f82f-4a3c-a43a-6650b2d8f2da"). InnerVolumeSpecName "kubelet-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 03 14:12:13 crc kubenswrapper[4805]: I1203 14:12:13.050527 4805 reconciler_common.go:293] "Volume detached for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/c9b110a7-f82f-4a3c-a43a-6650b2d8f2da-kubelet-dir\") on node \"crc\" DevicePath \"\"" Dec 03 14:12:13 crc kubenswrapper[4805]: I1203 14:12:13.055508 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c9b110a7-f82f-4a3c-a43a-6650b2d8f2da-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "c9b110a7-f82f-4a3c-a43a-6650b2d8f2da" (UID: "c9b110a7-f82f-4a3c-a43a-6650b2d8f2da"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 14:12:13 crc kubenswrapper[4805]: I1203 14:12:13.151644 4805 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/c9b110a7-f82f-4a3c-a43a-6650b2d8f2da-kube-api-access\") on node \"crc\" DevicePath \"\"" Dec 03 14:12:13 crc kubenswrapper[4805]: I1203 14:12:13.339030 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/network-metrics-daemon-4gd9r"] Dec 03 14:12:13 crc kubenswrapper[4805]: W1203 14:12:13.351912 4805 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod6b36a3db_1047_487a_9d6c_4f51ffdd8cc1.slice/crio-6ceafe6f705cf0010f0a955e540954184cbc4c17c99b5c683f045f24ad597265 WatchSource:0}: Error finding container 6ceafe6f705cf0010f0a955e540954184cbc4c17c99b5c683f045f24ad597265: Status 404 returned error can't find the container with id 6ceafe6f705cf0010f0a955e540954184cbc4c17c99b5c683f045f24ad597265 Dec 03 14:12:13 crc kubenswrapper[4805]: I1203 14:12:13.596701 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-8-crc" event={"ID":"c9b110a7-f82f-4a3c-a43a-6650b2d8f2da","Type":"ContainerDied","Data":"58e812ac15e6ffcc2cce5d3c0537eab4d370d2f585562d3488f98be3bf1b5237"} Dec 03 14:12:13 crc kubenswrapper[4805]: I1203 14:12:13.596761 4805 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="58e812ac15e6ffcc2cce5d3c0537eab4d370d2f585562d3488f98be3bf1b5237" Dec 03 14:12:13 crc kubenswrapper[4805]: I1203 14:12:13.596761 4805 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Dec 03 14:12:13 crc kubenswrapper[4805]: I1203 14:12:13.598412 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/network-metrics-daemon-4gd9r" event={"ID":"6b36a3db-1047-487a-9d6c-4f51ffdd8cc1","Type":"ContainerStarted","Data":"6ceafe6f705cf0010f0a955e540954184cbc4c17c99b5c683f045f24ad597265"} Dec 03 14:12:13 crc kubenswrapper[4805]: I1203 14:12:13.819013 4805 patch_prober.go:28] interesting pod/router-default-5444994796-kxnqm container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 03 14:12:13 crc kubenswrapper[4805]: [-]has-synced failed: reason withheld Dec 03 14:12:13 crc kubenswrapper[4805]: [+]process-running ok Dec 03 14:12:13 crc kubenswrapper[4805]: healthz check failed Dec 03 14:12:13 crc kubenswrapper[4805]: I1203 14:12:13.819114 4805 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-kxnqm" podUID="b22cbc60-5975-42a4-85d3-6add6917fd88" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 03 14:12:13 crc kubenswrapper[4805]: I1203 14:12:13.917871 4805 patch_prober.go:28] interesting pod/machine-config-daemon-gskh4 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 03 14:12:13 crc kubenswrapper[4805]: I1203 14:12:13.917927 4805 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-gskh4" podUID="6098937f-e3f6-45e8-a647-4994a79cd711" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 03 14:12:14 crc kubenswrapper[4805]: I1203 14:12:14.612682 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/network-metrics-daemon-4gd9r" event={"ID":"6b36a3db-1047-487a-9d6c-4f51ffdd8cc1","Type":"ContainerStarted","Data":"bacc487d3abb6781ba15fce34b4fa427a6f7ff1c5ab9b5633b898be1ce2fe582"} Dec 03 14:12:14 crc kubenswrapper[4805]: I1203 14:12:14.819281 4805 patch_prober.go:28] interesting pod/router-default-5444994796-kxnqm container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 03 14:12:14 crc kubenswrapper[4805]: [-]has-synced failed: reason withheld Dec 03 14:12:14 crc kubenswrapper[4805]: [+]process-running ok Dec 03 14:12:14 crc kubenswrapper[4805]: healthz check failed Dec 03 14:12:14 crc kubenswrapper[4805]: I1203 14:12:14.819345 4805 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-kxnqm" podUID="b22cbc60-5975-42a4-85d3-6add6917fd88" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 03 14:12:15 crc kubenswrapper[4805]: I1203 14:12:15.819080 4805 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-ingress/router-default-5444994796-kxnqm" Dec 03 14:12:15 crc kubenswrapper[4805]: I1203 14:12:15.822129 4805 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ingress/router-default-5444994796-kxnqm" Dec 03 14:12:17 crc kubenswrapper[4805]: I1203 14:12:17.234447 4805 patch_prober.go:28] interesting pod/console-f9d7485db-pp9tv container/console namespace/openshift-console: Startup probe status=failure output="Get \"https://10.217.0.27:8443/health\": dial tcp 10.217.0.27:8443: connect: connection refused" start-of-body= Dec 03 14:12:17 crc kubenswrapper[4805]: I1203 14:12:17.234517 4805 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-console/console-f9d7485db-pp9tv" podUID="e2034338-8a35-4fd2-9bbf-be733939fc25" containerName="console" probeResult="failure" output="Get \"https://10.217.0.27:8443/health\": dial tcp 10.217.0.27:8443: connect: connection refused" Dec 03 14:12:17 crc kubenswrapper[4805]: I1203 14:12:17.519244 4805 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console/downloads-7954f5f757-s25zl" Dec 03 14:12:25 crc kubenswrapper[4805]: I1203 14:12:25.513586 4805 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-image-registry/image-registry-697d97f7c8-jjzct" Dec 03 14:12:27 crc kubenswrapper[4805]: I1203 14:12:27.238357 4805 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-console/console-f9d7485db-pp9tv" Dec 03 14:12:27 crc kubenswrapper[4805]: I1203 14:12:27.241392 4805 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console/console-f9d7485db-pp9tv" Dec 03 14:12:37 crc kubenswrapper[4805]: I1203 14:12:37.500225 4805 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 14:12:37 crc kubenswrapper[4805]: E1203 14:12:37.736972 4805 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/community-operator-index:v4.18" Dec 03 14:12:37 crc kubenswrapper[4805]: E1203 14:12:37.737182 4805 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/community-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-b5wvv,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod community-operators-w6nm5_openshift-marketplace(344cb196-3baa-48c0-abcb-7e46cbce614d): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Dec 03 14:12:37 crc kubenswrapper[4805]: E1203 14:12:37.738461 4805 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/community-operators-w6nm5" podUID="344cb196-3baa-48c0-abcb-7e46cbce614d" Dec 03 14:12:38 crc kubenswrapper[4805]: I1203 14:12:38.130971 4805 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-b7wrl" Dec 03 14:12:38 crc kubenswrapper[4805]: E1203 14:12:38.646603 4805 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"\"" pod="openshift-marketplace/community-operators-w6nm5" podUID="344cb196-3baa-48c0-abcb-7e46cbce614d" Dec 03 14:12:41 crc kubenswrapper[4805]: E1203 14:12:41.457212 4805 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/redhat-marketplace-index:v4.18" Dec 03 14:12:41 crc kubenswrapper[4805]: E1203 14:12:41.457435 4805 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/redhat-marketplace-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-h6c8m,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod redhat-marketplace-m7r99_openshift-marketplace(1e412276-9d78-4771-b447-b4e2b388e604): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Dec 03 14:12:41 crc kubenswrapper[4805]: E1203 14:12:41.458643 4805 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/redhat-marketplace-m7r99" podUID="1e412276-9d78-4771-b447-b4e2b388e604" Dec 03 14:12:43 crc kubenswrapper[4805]: E1203 14:12:43.010173 4805 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/redhat-operator-index:v4.18" Dec 03 14:12:43 crc kubenswrapper[4805]: E1203 14:12:43.010582 4805 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/redhat-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-74jhd,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod redhat-operators-bd7j9_openshift-marketplace(d0a05061-dab0-4a85-be73-b1d2acbc7b8f): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Dec 03 14:12:43 crc kubenswrapper[4805]: E1203 14:12:43.011770 4805 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/redhat-operators-bd7j9" podUID="d0a05061-dab0-4a85-be73-b1d2acbc7b8f" Dec 03 14:12:43 crc kubenswrapper[4805]: E1203 14:12:43.466892 4805 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-marketplace-m7r99" podUID="1e412276-9d78-4771-b447-b4e2b388e604" Dec 03 14:12:43 crc kubenswrapper[4805]: I1203 14:12:43.484792 4805 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/revision-pruner-9-crc"] Dec 03 14:12:43 crc kubenswrapper[4805]: E1203 14:12:43.485073 4805 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a6e8e8f2-46cf-4322-8fdb-057a28032e66" containerName="pruner" Dec 03 14:12:43 crc kubenswrapper[4805]: I1203 14:12:43.485087 4805 state_mem.go:107] "Deleted CPUSet assignment" podUID="a6e8e8f2-46cf-4322-8fdb-057a28032e66" containerName="pruner" Dec 03 14:12:43 crc kubenswrapper[4805]: E1203 14:12:43.485102 4805 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c9b110a7-f82f-4a3c-a43a-6650b2d8f2da" containerName="pruner" Dec 03 14:12:43 crc kubenswrapper[4805]: I1203 14:12:43.485110 4805 state_mem.go:107] "Deleted CPUSet assignment" podUID="c9b110a7-f82f-4a3c-a43a-6650b2d8f2da" containerName="pruner" Dec 03 14:12:43 crc kubenswrapper[4805]: I1203 14:12:43.486097 4805 memory_manager.go:354] "RemoveStaleState removing state" podUID="a6e8e8f2-46cf-4322-8fdb-057a28032e66" containerName="pruner" Dec 03 14:12:43 crc kubenswrapper[4805]: I1203 14:12:43.486149 4805 memory_manager.go:354] "RemoveStaleState removing state" podUID="c9b110a7-f82f-4a3c-a43a-6650b2d8f2da" containerName="pruner" Dec 03 14:12:43 crc kubenswrapper[4805]: I1203 14:12:43.487430 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-9-crc" Dec 03 14:12:43 crc kubenswrapper[4805]: I1203 14:12:43.489802 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/revision-pruner-9-crc"] Dec 03 14:12:43 crc kubenswrapper[4805]: I1203 14:12:43.491041 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver"/"installer-sa-dockercfg-5pr6n" Dec 03 14:12:43 crc kubenswrapper[4805]: I1203 14:12:43.491300 4805 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver"/"kube-root-ca.crt" Dec 03 14:12:43 crc kubenswrapper[4805]: I1203 14:12:43.534999 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/81b74538-a789-43ab-9a72-fc8fd397371e-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"81b74538-a789-43ab-9a72-fc8fd397371e\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Dec 03 14:12:43 crc kubenswrapper[4805]: I1203 14:12:43.535064 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/81b74538-a789-43ab-9a72-fc8fd397371e-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"81b74538-a789-43ab-9a72-fc8fd397371e\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Dec 03 14:12:43 crc kubenswrapper[4805]: I1203 14:12:43.635627 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/81b74538-a789-43ab-9a72-fc8fd397371e-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"81b74538-a789-43ab-9a72-fc8fd397371e\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Dec 03 14:12:43 crc kubenswrapper[4805]: I1203 14:12:43.636263 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/81b74538-a789-43ab-9a72-fc8fd397371e-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"81b74538-a789-43ab-9a72-fc8fd397371e\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Dec 03 14:12:43 crc kubenswrapper[4805]: I1203 14:12:43.635786 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/81b74538-a789-43ab-9a72-fc8fd397371e-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"81b74538-a789-43ab-9a72-fc8fd397371e\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Dec 03 14:12:43 crc kubenswrapper[4805]: I1203 14:12:43.659337 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/81b74538-a789-43ab-9a72-fc8fd397371e-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"81b74538-a789-43ab-9a72-fc8fd397371e\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Dec 03 14:12:43 crc kubenswrapper[4805]: E1203 14:12:43.781461 4805 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-operators-bd7j9" podUID="d0a05061-dab0-4a85-be73-b1d2acbc7b8f" Dec 03 14:12:43 crc kubenswrapper[4805]: I1203 14:12:43.839265 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-9-crc" Dec 03 14:12:43 crc kubenswrapper[4805]: I1203 14:12:43.918734 4805 patch_prober.go:28] interesting pod/machine-config-daemon-gskh4 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 03 14:12:43 crc kubenswrapper[4805]: I1203 14:12:43.918775 4805 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-gskh4" podUID="6098937f-e3f6-45e8-a647-4994a79cd711" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 03 14:12:44 crc kubenswrapper[4805]: I1203 14:12:44.027438 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/revision-pruner-9-crc"] Dec 03 14:12:44 crc kubenswrapper[4805]: W1203 14:12:44.036148 4805 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-pod81b74538_a789_43ab_9a72_fc8fd397371e.slice/crio-0c56297eff21c4137446818a57bb4c409792207d4fb005201e4ef4307372fc8f WatchSource:0}: Error finding container 0c56297eff21c4137446818a57bb4c409792207d4fb005201e4ef4307372fc8f: Status 404 returned error can't find the container with id 0c56297eff21c4137446818a57bb4c409792207d4fb005201e4ef4307372fc8f Dec 03 14:12:44 crc kubenswrapper[4805]: E1203 14:12:44.239315 4805 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/redhat-operator-index:v4.18" Dec 03 14:12:44 crc kubenswrapper[4805]: E1203 14:12:44.239495 4805 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/redhat-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-btdbn,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod redhat-operators-dcw96_openshift-marketplace(8fd4f74e-3b48-4a4a-8898-e76c5926abab): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Dec 03 14:12:44 crc kubenswrapper[4805]: E1203 14:12:44.240752 4805 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/redhat-operators-dcw96" podUID="8fd4f74e-3b48-4a4a-8898-e76c5926abab" Dec 03 14:12:44 crc kubenswrapper[4805]: E1203 14:12:44.396116 4805 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/certified-operator-index:v4.18" Dec 03 14:12:44 crc kubenswrapper[4805]: E1203 14:12:44.396321 4805 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/certified-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-rv4vh,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod certified-operators-9cm4t_openshift-marketplace(64f6384f-41e0-469e-8920-27e9e97b62db): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Dec 03 14:12:44 crc kubenswrapper[4805]: E1203 14:12:44.397523 4805 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/certified-operators-9cm4t" podUID="64f6384f-41e0-469e-8920-27e9e97b62db" Dec 03 14:12:44 crc kubenswrapper[4805]: E1203 14:12:44.581199 4805 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/redhat-marketplace-index:v4.18" Dec 03 14:12:44 crc kubenswrapper[4805]: E1203 14:12:44.581460 4805 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/redhat-marketplace-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-s6blv,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod redhat-marketplace-dlmr9_openshift-marketplace(77e93612-7c30-4f0e-9347-6e1373dfaaf8): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Dec 03 14:12:44 crc kubenswrapper[4805]: E1203 14:12:44.582904 4805 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/redhat-marketplace-dlmr9" podUID="77e93612-7c30-4f0e-9347-6e1373dfaaf8" Dec 03 14:12:44 crc kubenswrapper[4805]: I1203 14:12:44.785042 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-9-crc" event={"ID":"81b74538-a789-43ab-9a72-fc8fd397371e","Type":"ContainerStarted","Data":"6105095ec33e4c58e75a44853d7d0621f84125df06667a856f278a15c1c4785b"} Dec 03 14:12:44 crc kubenswrapper[4805]: I1203 14:12:44.785099 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-9-crc" event={"ID":"81b74538-a789-43ab-9a72-fc8fd397371e","Type":"ContainerStarted","Data":"0c56297eff21c4137446818a57bb4c409792207d4fb005201e4ef4307372fc8f"} Dec 03 14:12:44 crc kubenswrapper[4805]: I1203 14:12:44.789575 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/network-metrics-daemon-4gd9r" event={"ID":"6b36a3db-1047-487a-9d6c-4f51ffdd8cc1","Type":"ContainerStarted","Data":"54e94228d33c004d969dd7b6e9c55829e471143b4603ad00de70e59bfada0c6a"} Dec 03 14:12:44 crc kubenswrapper[4805]: E1203 14:12:44.790589 4805 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-marketplace-dlmr9" podUID="77e93612-7c30-4f0e-9347-6e1373dfaaf8" Dec 03 14:12:44 crc kubenswrapper[4805]: E1203 14:12:44.791786 4805 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-operators-dcw96" podUID="8fd4f74e-3b48-4a4a-8898-e76c5926abab" Dec 03 14:12:44 crc kubenswrapper[4805]: E1203 14:12:44.794427 4805 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"\"" pod="openshift-marketplace/certified-operators-9cm4t" podUID="64f6384f-41e0-469e-8920-27e9e97b62db" Dec 03 14:12:44 crc kubenswrapper[4805]: I1203 14:12:44.802160 4805 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver/revision-pruner-9-crc" podStartSLOduration=1.8021392729999999 podStartE2EDuration="1.802139273s" podCreationTimestamp="2025-12-03 14:12:43 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 14:12:44.80164732 +0000 UTC m=+194.464564243" watchObservedRunningTime="2025-12-03 14:12:44.802139273 +0000 UTC m=+194.465056196" Dec 03 14:12:44 crc kubenswrapper[4805]: I1203 14:12:44.818820 4805 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/network-metrics-daemon-4gd9r" podStartSLOduration=175.818804367 podStartE2EDuration="2m55.818804367s" podCreationTimestamp="2025-12-03 14:09:49 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 14:12:44.81515525 +0000 UTC m=+194.478072173" watchObservedRunningTime="2025-12-03 14:12:44.818804367 +0000 UTC m=+194.481721290" Dec 03 14:12:45 crc kubenswrapper[4805]: I1203 14:12:45.795405 4805 generic.go:334] "Generic (PLEG): container finished" podID="81b74538-a789-43ab-9a72-fc8fd397371e" containerID="6105095ec33e4c58e75a44853d7d0621f84125df06667a856f278a15c1c4785b" exitCode=0 Dec 03 14:12:45 crc kubenswrapper[4805]: I1203 14:12:45.795559 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-9-crc" event={"ID":"81b74538-a789-43ab-9a72-fc8fd397371e","Type":"ContainerDied","Data":"6105095ec33e4c58e75a44853d7d0621f84125df06667a856f278a15c1c4785b"} Dec 03 14:12:45 crc kubenswrapper[4805]: I1203 14:12:45.797448 4805 generic.go:334] "Generic (PLEG): container finished" podID="731aa7bc-cde1-4ff7-aa29-78093078514d" containerID="6dddb2c849fd648636450ea53af9b98562b33acfe94c38f5f21888d7ee4f41ab" exitCode=0 Dec 03 14:12:45 crc kubenswrapper[4805]: I1203 14:12:45.797506 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-fbrtf" event={"ID":"731aa7bc-cde1-4ff7-aa29-78093078514d","Type":"ContainerDied","Data":"6dddb2c849fd648636450ea53af9b98562b33acfe94c38f5f21888d7ee4f41ab"} Dec 03 14:12:45 crc kubenswrapper[4805]: I1203 14:12:45.799888 4805 generic.go:334] "Generic (PLEG): container finished" podID="b2e4ac3a-e225-46ae-874d-4cf931c13464" containerID="8ab28724d5fa789a469456ce08a8c43bf28fa705458279faad7ea7f330f14a6b" exitCode=0 Dec 03 14:12:45 crc kubenswrapper[4805]: I1203 14:12:45.799989 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-8gt4x" event={"ID":"b2e4ac3a-e225-46ae-874d-4cf931c13464","Type":"ContainerDied","Data":"8ab28724d5fa789a469456ce08a8c43bf28fa705458279faad7ea7f330f14a6b"} Dec 03 14:12:46 crc kubenswrapper[4805]: I1203 14:12:46.806723 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-8gt4x" event={"ID":"b2e4ac3a-e225-46ae-874d-4cf931c13464","Type":"ContainerStarted","Data":"ae4620a4d37207bd0317f1311c70528b552e6f40a2fac0b242d743282dadf663"} Dec 03 14:12:46 crc kubenswrapper[4805]: I1203 14:12:46.810336 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-fbrtf" event={"ID":"731aa7bc-cde1-4ff7-aa29-78093078514d","Type":"ContainerStarted","Data":"8e81b2f0ad0e09adc68b67e411190b3013f19fdb8456ff563baad34de23a42a7"} Dec 03 14:12:46 crc kubenswrapper[4805]: I1203 14:12:46.828612 4805 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-8gt4x" podStartSLOduration=3.828388424 podStartE2EDuration="43.828590228s" podCreationTimestamp="2025-12-03 14:12:03 +0000 UTC" firstStartedPulling="2025-12-03 14:12:06.414433548 +0000 UTC m=+156.077350471" lastFinishedPulling="2025-12-03 14:12:46.414635352 +0000 UTC m=+196.077552275" observedRunningTime="2025-12-03 14:12:46.826754099 +0000 UTC m=+196.489671022" watchObservedRunningTime="2025-12-03 14:12:46.828590228 +0000 UTC m=+196.491507151" Dec 03 14:12:46 crc kubenswrapper[4805]: I1203 14:12:46.847408 4805 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-fbrtf" podStartSLOduration=4.01208516 podStartE2EDuration="43.847391369s" podCreationTimestamp="2025-12-03 14:12:03 +0000 UTC" firstStartedPulling="2025-12-03 14:12:06.448235828 +0000 UTC m=+156.111152751" lastFinishedPulling="2025-12-03 14:12:46.283542037 +0000 UTC m=+195.946458960" observedRunningTime="2025-12-03 14:12:46.845653333 +0000 UTC m=+196.508570256" watchObservedRunningTime="2025-12-03 14:12:46.847391369 +0000 UTC m=+196.510308292" Dec 03 14:12:47 crc kubenswrapper[4805]: I1203 14:12:47.063127 4805 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-9-crc" Dec 03 14:12:47 crc kubenswrapper[4805]: I1203 14:12:47.087069 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/81b74538-a789-43ab-9a72-fc8fd397371e-kubelet-dir\") pod \"81b74538-a789-43ab-9a72-fc8fd397371e\" (UID: \"81b74538-a789-43ab-9a72-fc8fd397371e\") " Dec 03 14:12:47 crc kubenswrapper[4805]: I1203 14:12:47.087142 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/81b74538-a789-43ab-9a72-fc8fd397371e-kube-api-access\") pod \"81b74538-a789-43ab-9a72-fc8fd397371e\" (UID: \"81b74538-a789-43ab-9a72-fc8fd397371e\") " Dec 03 14:12:47 crc kubenswrapper[4805]: I1203 14:12:47.087887 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/81b74538-a789-43ab-9a72-fc8fd397371e-kubelet-dir" (OuterVolumeSpecName: "kubelet-dir") pod "81b74538-a789-43ab-9a72-fc8fd397371e" (UID: "81b74538-a789-43ab-9a72-fc8fd397371e"). InnerVolumeSpecName "kubelet-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 03 14:12:47 crc kubenswrapper[4805]: I1203 14:12:47.098014 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/81b74538-a789-43ab-9a72-fc8fd397371e-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "81b74538-a789-43ab-9a72-fc8fd397371e" (UID: "81b74538-a789-43ab-9a72-fc8fd397371e"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 14:12:47 crc kubenswrapper[4805]: I1203 14:12:47.187960 4805 reconciler_common.go:293] "Volume detached for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/81b74538-a789-43ab-9a72-fc8fd397371e-kubelet-dir\") on node \"crc\" DevicePath \"\"" Dec 03 14:12:47 crc kubenswrapper[4805]: I1203 14:12:47.187988 4805 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/81b74538-a789-43ab-9a72-fc8fd397371e-kube-api-access\") on node \"crc\" DevicePath \"\"" Dec 03 14:12:47 crc kubenswrapper[4805]: I1203 14:12:47.820392 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-9-crc" event={"ID":"81b74538-a789-43ab-9a72-fc8fd397371e","Type":"ContainerDied","Data":"0c56297eff21c4137446818a57bb4c409792207d4fb005201e4ef4307372fc8f"} Dec 03 14:12:47 crc kubenswrapper[4805]: I1203 14:12:47.820662 4805 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="0c56297eff21c4137446818a57bb4c409792207d4fb005201e4ef4307372fc8f" Dec 03 14:12:47 crc kubenswrapper[4805]: I1203 14:12:47.820461 4805 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-9-crc" Dec 03 14:12:49 crc kubenswrapper[4805]: I1203 14:12:49.656557 4805 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/installer-9-crc"] Dec 03 14:12:49 crc kubenswrapper[4805]: E1203 14:12:49.659050 4805 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="81b74538-a789-43ab-9a72-fc8fd397371e" containerName="pruner" Dec 03 14:12:49 crc kubenswrapper[4805]: I1203 14:12:49.659178 4805 state_mem.go:107] "Deleted CPUSet assignment" podUID="81b74538-a789-43ab-9a72-fc8fd397371e" containerName="pruner" Dec 03 14:12:49 crc kubenswrapper[4805]: I1203 14:12:49.659518 4805 memory_manager.go:354] "RemoveStaleState removing state" podUID="81b74538-a789-43ab-9a72-fc8fd397371e" containerName="pruner" Dec 03 14:12:49 crc kubenswrapper[4805]: I1203 14:12:49.661873 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/installer-9-crc"] Dec 03 14:12:49 crc kubenswrapper[4805]: I1203 14:12:49.661934 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/installer-9-crc" Dec 03 14:12:49 crc kubenswrapper[4805]: I1203 14:12:49.664922 4805 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver"/"kube-root-ca.crt" Dec 03 14:12:49 crc kubenswrapper[4805]: I1203 14:12:49.666608 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver"/"installer-sa-dockercfg-5pr6n" Dec 03 14:12:49 crc kubenswrapper[4805]: I1203 14:12:49.820031 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/62ba53fc-fcf4-47be-a415-ea016529bac1-kube-api-access\") pod \"installer-9-crc\" (UID: \"62ba53fc-fcf4-47be-a415-ea016529bac1\") " pod="openshift-kube-apiserver/installer-9-crc" Dec 03 14:12:49 crc kubenswrapper[4805]: I1203 14:12:49.820236 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/62ba53fc-fcf4-47be-a415-ea016529bac1-kubelet-dir\") pod \"installer-9-crc\" (UID: \"62ba53fc-fcf4-47be-a415-ea016529bac1\") " pod="openshift-kube-apiserver/installer-9-crc" Dec 03 14:12:49 crc kubenswrapper[4805]: I1203 14:12:49.820286 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/62ba53fc-fcf4-47be-a415-ea016529bac1-var-lock\") pod \"installer-9-crc\" (UID: \"62ba53fc-fcf4-47be-a415-ea016529bac1\") " pod="openshift-kube-apiserver/installer-9-crc" Dec 03 14:12:49 crc kubenswrapper[4805]: I1203 14:12:49.921464 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/62ba53fc-fcf4-47be-a415-ea016529bac1-kubelet-dir\") pod \"installer-9-crc\" (UID: \"62ba53fc-fcf4-47be-a415-ea016529bac1\") " pod="openshift-kube-apiserver/installer-9-crc" Dec 03 14:12:49 crc kubenswrapper[4805]: I1203 14:12:49.921507 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/62ba53fc-fcf4-47be-a415-ea016529bac1-var-lock\") pod \"installer-9-crc\" (UID: \"62ba53fc-fcf4-47be-a415-ea016529bac1\") " pod="openshift-kube-apiserver/installer-9-crc" Dec 03 14:12:49 crc kubenswrapper[4805]: I1203 14:12:49.921575 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/62ba53fc-fcf4-47be-a415-ea016529bac1-kube-api-access\") pod \"installer-9-crc\" (UID: \"62ba53fc-fcf4-47be-a415-ea016529bac1\") " pod="openshift-kube-apiserver/installer-9-crc" Dec 03 14:12:49 crc kubenswrapper[4805]: I1203 14:12:49.921577 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/62ba53fc-fcf4-47be-a415-ea016529bac1-kubelet-dir\") pod \"installer-9-crc\" (UID: \"62ba53fc-fcf4-47be-a415-ea016529bac1\") " pod="openshift-kube-apiserver/installer-9-crc" Dec 03 14:12:49 crc kubenswrapper[4805]: I1203 14:12:49.921675 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/62ba53fc-fcf4-47be-a415-ea016529bac1-var-lock\") pod \"installer-9-crc\" (UID: \"62ba53fc-fcf4-47be-a415-ea016529bac1\") " pod="openshift-kube-apiserver/installer-9-crc" Dec 03 14:12:49 crc kubenswrapper[4805]: I1203 14:12:49.943513 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/62ba53fc-fcf4-47be-a415-ea016529bac1-kube-api-access\") pod \"installer-9-crc\" (UID: \"62ba53fc-fcf4-47be-a415-ea016529bac1\") " pod="openshift-kube-apiserver/installer-9-crc" Dec 03 14:12:49 crc kubenswrapper[4805]: I1203 14:12:49.982075 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/installer-9-crc" Dec 03 14:12:50 crc kubenswrapper[4805]: I1203 14:12:50.175303 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/installer-9-crc"] Dec 03 14:12:50 crc kubenswrapper[4805]: I1203 14:12:50.843199 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/installer-9-crc" event={"ID":"62ba53fc-fcf4-47be-a415-ea016529bac1","Type":"ContainerStarted","Data":"bda7e3bf266699bbd48c9f409b7901d9f99f3f853398099c4cd493857174e4ca"} Dec 03 14:12:51 crc kubenswrapper[4805]: I1203 14:12:51.850641 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/installer-9-crc" event={"ID":"62ba53fc-fcf4-47be-a415-ea016529bac1","Type":"ContainerStarted","Data":"8a65ec5b7de3262599fc740a2a5b215b18fa383d37c67b71f2c75048b5b4deda"} Dec 03 14:12:52 crc kubenswrapper[4805]: I1203 14:12:52.859223 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-w6nm5" event={"ID":"344cb196-3baa-48c0-abcb-7e46cbce614d","Type":"ContainerStarted","Data":"aae5371377082372f9fd92376d8cc23f7bb2d5195e09b55895849e7a7ef4ca6b"} Dec 03 14:12:52 crc kubenswrapper[4805]: I1203 14:12:52.877786 4805 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver/installer-9-crc" podStartSLOduration=3.87776423 podStartE2EDuration="3.87776423s" podCreationTimestamp="2025-12-03 14:12:49 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 14:12:51.866278155 +0000 UTC m=+201.529195098" watchObservedRunningTime="2025-12-03 14:12:52.87776423 +0000 UTC m=+202.540681153" Dec 03 14:12:53 crc kubenswrapper[4805]: I1203 14:12:53.805717 4805 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-8gt4x" Dec 03 14:12:53 crc kubenswrapper[4805]: I1203 14:12:53.807127 4805 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-8gt4x" Dec 03 14:12:53 crc kubenswrapper[4805]: I1203 14:12:53.865577 4805 generic.go:334] "Generic (PLEG): container finished" podID="344cb196-3baa-48c0-abcb-7e46cbce614d" containerID="aae5371377082372f9fd92376d8cc23f7bb2d5195e09b55895849e7a7ef4ca6b" exitCode=0 Dec 03 14:12:53 crc kubenswrapper[4805]: I1203 14:12:53.865695 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-w6nm5" event={"ID":"344cb196-3baa-48c0-abcb-7e46cbce614d","Type":"ContainerDied","Data":"aae5371377082372f9fd92376d8cc23f7bb2d5195e09b55895849e7a7ef4ca6b"} Dec 03 14:12:53 crc kubenswrapper[4805]: I1203 14:12:53.866490 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-w6nm5" event={"ID":"344cb196-3baa-48c0-abcb-7e46cbce614d","Type":"ContainerStarted","Data":"d7a629c994143e2afd38ad34222b48e2bbde341dc173df63ac2114c8591c19f7"} Dec 03 14:12:53 crc kubenswrapper[4805]: I1203 14:12:53.872450 4805 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-8gt4x" Dec 03 14:12:53 crc kubenswrapper[4805]: I1203 14:12:53.891286 4805 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-w6nm5" podStartSLOduration=3.669435358 podStartE2EDuration="50.891268611s" podCreationTimestamp="2025-12-03 14:12:03 +0000 UTC" firstStartedPulling="2025-12-03 14:12:06.42507853 +0000 UTC m=+156.087995453" lastFinishedPulling="2025-12-03 14:12:53.646911783 +0000 UTC m=+203.309828706" observedRunningTime="2025-12-03 14:12:53.888146318 +0000 UTC m=+203.551063261" watchObservedRunningTime="2025-12-03 14:12:53.891268611 +0000 UTC m=+203.554185534" Dec 03 14:12:53 crc kubenswrapper[4805]: I1203 14:12:53.924640 4805 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-8gt4x" Dec 03 14:12:54 crc kubenswrapper[4805]: I1203 14:12:54.018075 4805 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-fbrtf" Dec 03 14:12:54 crc kubenswrapper[4805]: I1203 14:12:54.018320 4805 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-fbrtf" Dec 03 14:12:54 crc kubenswrapper[4805]: I1203 14:12:54.054141 4805 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-fbrtf" Dec 03 14:12:54 crc kubenswrapper[4805]: I1203 14:12:54.910971 4805 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-fbrtf" Dec 03 14:12:56 crc kubenswrapper[4805]: I1203 14:12:56.103497 4805 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-fbrtf"] Dec 03 14:12:57 crc kubenswrapper[4805]: I1203 14:12:57.883378 4805 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-fbrtf" podUID="731aa7bc-cde1-4ff7-aa29-78093078514d" containerName="registry-server" containerID="cri-o://8e81b2f0ad0e09adc68b67e411190b3013f19fdb8456ff563baad34de23a42a7" gracePeriod=2 Dec 03 14:12:58 crc kubenswrapper[4805]: I1203 14:12:58.891133 4805 generic.go:334] "Generic (PLEG): container finished" podID="1e412276-9d78-4771-b447-b4e2b388e604" containerID="d3a6280f659aef72bcb760b2b10ced59c5868f476392736b6f3c84c58c46fa06" exitCode=0 Dec 03 14:12:58 crc kubenswrapper[4805]: I1203 14:12:58.891211 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-m7r99" event={"ID":"1e412276-9d78-4771-b447-b4e2b388e604","Type":"ContainerDied","Data":"d3a6280f659aef72bcb760b2b10ced59c5868f476392736b6f3c84c58c46fa06"} Dec 03 14:12:58 crc kubenswrapper[4805]: I1203 14:12:58.894012 4805 generic.go:334] "Generic (PLEG): container finished" podID="77e93612-7c30-4f0e-9347-6e1373dfaaf8" containerID="381c8419c7601aad161bac4e1319af388398957804435159738857cb6591b293" exitCode=0 Dec 03 14:12:58 crc kubenswrapper[4805]: I1203 14:12:58.894055 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-dlmr9" event={"ID":"77e93612-7c30-4f0e-9347-6e1373dfaaf8","Type":"ContainerDied","Data":"381c8419c7601aad161bac4e1319af388398957804435159738857cb6591b293"} Dec 03 14:12:58 crc kubenswrapper[4805]: I1203 14:12:58.899684 4805 generic.go:334] "Generic (PLEG): container finished" podID="731aa7bc-cde1-4ff7-aa29-78093078514d" containerID="8e81b2f0ad0e09adc68b67e411190b3013f19fdb8456ff563baad34de23a42a7" exitCode=0 Dec 03 14:12:58 crc kubenswrapper[4805]: I1203 14:12:58.899727 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-fbrtf" event={"ID":"731aa7bc-cde1-4ff7-aa29-78093078514d","Type":"ContainerDied","Data":"8e81b2f0ad0e09adc68b67e411190b3013f19fdb8456ff563baad34de23a42a7"} Dec 03 14:12:59 crc kubenswrapper[4805]: I1203 14:12:59.139538 4805 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-fbrtf" Dec 03 14:12:59 crc kubenswrapper[4805]: I1203 14:12:59.251993 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/731aa7bc-cde1-4ff7-aa29-78093078514d-utilities\") pod \"731aa7bc-cde1-4ff7-aa29-78093078514d\" (UID: \"731aa7bc-cde1-4ff7-aa29-78093078514d\") " Dec 03 14:12:59 crc kubenswrapper[4805]: I1203 14:12:59.252066 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/731aa7bc-cde1-4ff7-aa29-78093078514d-catalog-content\") pod \"731aa7bc-cde1-4ff7-aa29-78093078514d\" (UID: \"731aa7bc-cde1-4ff7-aa29-78093078514d\") " Dec 03 14:12:59 crc kubenswrapper[4805]: I1203 14:12:59.253136 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/731aa7bc-cde1-4ff7-aa29-78093078514d-utilities" (OuterVolumeSpecName: "utilities") pod "731aa7bc-cde1-4ff7-aa29-78093078514d" (UID: "731aa7bc-cde1-4ff7-aa29-78093078514d"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 14:12:59 crc kubenswrapper[4805]: I1203 14:12:59.259964 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ph6rn\" (UniqueName: \"kubernetes.io/projected/731aa7bc-cde1-4ff7-aa29-78093078514d-kube-api-access-ph6rn\") pod \"731aa7bc-cde1-4ff7-aa29-78093078514d\" (UID: \"731aa7bc-cde1-4ff7-aa29-78093078514d\") " Dec 03 14:12:59 crc kubenswrapper[4805]: I1203 14:12:59.260263 4805 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/731aa7bc-cde1-4ff7-aa29-78093078514d-utilities\") on node \"crc\" DevicePath \"\"" Dec 03 14:12:59 crc kubenswrapper[4805]: I1203 14:12:59.266203 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/731aa7bc-cde1-4ff7-aa29-78093078514d-kube-api-access-ph6rn" (OuterVolumeSpecName: "kube-api-access-ph6rn") pod "731aa7bc-cde1-4ff7-aa29-78093078514d" (UID: "731aa7bc-cde1-4ff7-aa29-78093078514d"). InnerVolumeSpecName "kube-api-access-ph6rn". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 14:12:59 crc kubenswrapper[4805]: I1203 14:12:59.312563 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/731aa7bc-cde1-4ff7-aa29-78093078514d-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "731aa7bc-cde1-4ff7-aa29-78093078514d" (UID: "731aa7bc-cde1-4ff7-aa29-78093078514d"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 14:12:59 crc kubenswrapper[4805]: I1203 14:12:59.361923 4805 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/731aa7bc-cde1-4ff7-aa29-78093078514d-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 03 14:12:59 crc kubenswrapper[4805]: I1203 14:12:59.362090 4805 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ph6rn\" (UniqueName: \"kubernetes.io/projected/731aa7bc-cde1-4ff7-aa29-78093078514d-kube-api-access-ph6rn\") on node \"crc\" DevicePath \"\"" Dec 03 14:12:59 crc kubenswrapper[4805]: I1203 14:12:59.907776 4805 generic.go:334] "Generic (PLEG): container finished" podID="d0a05061-dab0-4a85-be73-b1d2acbc7b8f" containerID="545b83ba10b4fa150c471ebb69ce19ffc8b1ba8aa337a88331459f13f345068a" exitCode=0 Dec 03 14:12:59 crc kubenswrapper[4805]: I1203 14:12:59.907867 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-bd7j9" event={"ID":"d0a05061-dab0-4a85-be73-b1d2acbc7b8f","Type":"ContainerDied","Data":"545b83ba10b4fa150c471ebb69ce19ffc8b1ba8aa337a88331459f13f345068a"} Dec 03 14:12:59 crc kubenswrapper[4805]: I1203 14:12:59.911328 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-m7r99" event={"ID":"1e412276-9d78-4771-b447-b4e2b388e604","Type":"ContainerStarted","Data":"1613fa3f6b59024d57deed3610696ad2566be4c7d5110cec4a2f3f22a3a00311"} Dec 03 14:12:59 crc kubenswrapper[4805]: I1203 14:12:59.915464 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-dlmr9" event={"ID":"77e93612-7c30-4f0e-9347-6e1373dfaaf8","Type":"ContainerStarted","Data":"3b67b5a47b1aba08949400ad9b5eedb7ca2cd9e0411c6efc3b9165d04befe24c"} Dec 03 14:12:59 crc kubenswrapper[4805]: I1203 14:12:59.917889 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-fbrtf" event={"ID":"731aa7bc-cde1-4ff7-aa29-78093078514d","Type":"ContainerDied","Data":"f04dc853bd11007db4faffdd39b6693fca131898415cc3f50ebddbd1166b1476"} Dec 03 14:12:59 crc kubenswrapper[4805]: I1203 14:12:59.917939 4805 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-fbrtf" Dec 03 14:12:59 crc kubenswrapper[4805]: I1203 14:12:59.917944 4805 scope.go:117] "RemoveContainer" containerID="8e81b2f0ad0e09adc68b67e411190b3013f19fdb8456ff563baad34de23a42a7" Dec 03 14:12:59 crc kubenswrapper[4805]: I1203 14:12:59.934225 4805 scope.go:117] "RemoveContainer" containerID="6dddb2c849fd648636450ea53af9b98562b33acfe94c38f5f21888d7ee4f41ab" Dec 03 14:12:59 crc kubenswrapper[4805]: I1203 14:12:59.954137 4805 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-m7r99" podStartSLOduration=2.034132658 podStartE2EDuration="54.95412128s" podCreationTimestamp="2025-12-03 14:12:05 +0000 UTC" firstStartedPulling="2025-12-03 14:12:06.431178515 +0000 UTC m=+156.094095438" lastFinishedPulling="2025-12-03 14:12:59.351167137 +0000 UTC m=+209.014084060" observedRunningTime="2025-12-03 14:12:59.951448088 +0000 UTC m=+209.614365011" watchObservedRunningTime="2025-12-03 14:12:59.95412128 +0000 UTC m=+209.617038203" Dec 03 14:12:59 crc kubenswrapper[4805]: I1203 14:12:59.966065 4805 scope.go:117] "RemoveContainer" containerID="1c047e20d07ca13ec2cd6254198e5bb079268d59b37a71ae8421c01cc5167e62" Dec 03 14:12:59 crc kubenswrapper[4805]: I1203 14:12:59.972587 4805 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-dlmr9" podStartSLOduration=3.135020625 podStartE2EDuration="54.9725672s" podCreationTimestamp="2025-12-03 14:12:05 +0000 UTC" firstStartedPulling="2025-12-03 14:12:07.50112122 +0000 UTC m=+157.164038143" lastFinishedPulling="2025-12-03 14:12:59.338667795 +0000 UTC m=+209.001584718" observedRunningTime="2025-12-03 14:12:59.969964071 +0000 UTC m=+209.632881014" watchObservedRunningTime="2025-12-03 14:12:59.9725672 +0000 UTC m=+209.635484133" Dec 03 14:12:59 crc kubenswrapper[4805]: I1203 14:12:59.987410 4805 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-fbrtf"] Dec 03 14:12:59 crc kubenswrapper[4805]: I1203 14:12:59.993096 4805 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-fbrtf"] Dec 03 14:13:00 crc kubenswrapper[4805]: I1203 14:13:00.702885 4805 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="731aa7bc-cde1-4ff7-aa29-78093078514d" path="/var/lib/kubelet/pods/731aa7bc-cde1-4ff7-aa29-78093078514d/volumes" Dec 03 14:13:00 crc kubenswrapper[4805]: I1203 14:13:00.923347 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-bd7j9" event={"ID":"d0a05061-dab0-4a85-be73-b1d2acbc7b8f","Type":"ContainerStarted","Data":"48f11437168216c2fe0dbb6d16be8cd4176c1f7d852e1837bb92258c4b0f67e3"} Dec 03 14:13:00 crc kubenswrapper[4805]: I1203 14:13:00.945593 4805 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-bd7j9" podStartSLOduration=1.799509901 podStartE2EDuration="54.945572294s" podCreationTimestamp="2025-12-03 14:12:06 +0000 UTC" firstStartedPulling="2025-12-03 14:12:07.48545986 +0000 UTC m=+157.148376783" lastFinishedPulling="2025-12-03 14:13:00.631522253 +0000 UTC m=+210.294439176" observedRunningTime="2025-12-03 14:13:00.94281158 +0000 UTC m=+210.605728503" watchObservedRunningTime="2025-12-03 14:13:00.945572294 +0000 UTC m=+210.608489227" Dec 03 14:13:01 crc kubenswrapper[4805]: I1203 14:13:01.931907 4805 generic.go:334] "Generic (PLEG): container finished" podID="64f6384f-41e0-469e-8920-27e9e97b62db" containerID="d5453cf390b7015789f699303295f36cf786a15f78885a09b9c40bd9adf0c125" exitCode=0 Dec 03 14:13:01 crc kubenswrapper[4805]: I1203 14:13:01.931978 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-9cm4t" event={"ID":"64f6384f-41e0-469e-8920-27e9e97b62db","Type":"ContainerDied","Data":"d5453cf390b7015789f699303295f36cf786a15f78885a09b9c40bd9adf0c125"} Dec 03 14:13:01 crc kubenswrapper[4805]: I1203 14:13:01.934820 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-dcw96" event={"ID":"8fd4f74e-3b48-4a4a-8898-e76c5926abab","Type":"ContainerStarted","Data":"fe565be9850fa082dece91006e313b97f73b64633d1e70e948b4e30bf6193c74"} Dec 03 14:13:02 crc kubenswrapper[4805]: I1203 14:13:02.949519 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-9cm4t" event={"ID":"64f6384f-41e0-469e-8920-27e9e97b62db","Type":"ContainerStarted","Data":"65a3e60d7a131cd0d239b955d8911d4d763c2d8da4f4420acf1c8963654a31e6"} Dec 03 14:13:02 crc kubenswrapper[4805]: I1203 14:13:02.952986 4805 generic.go:334] "Generic (PLEG): container finished" podID="8fd4f74e-3b48-4a4a-8898-e76c5926abab" containerID="fe565be9850fa082dece91006e313b97f73b64633d1e70e948b4e30bf6193c74" exitCode=0 Dec 03 14:13:02 crc kubenswrapper[4805]: I1203 14:13:02.953035 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-dcw96" event={"ID":"8fd4f74e-3b48-4a4a-8898-e76c5926abab","Type":"ContainerDied","Data":"fe565be9850fa082dece91006e313b97f73b64633d1e70e948b4e30bf6193c74"} Dec 03 14:13:02 crc kubenswrapper[4805]: I1203 14:13:02.968617 4805 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-9cm4t" podStartSLOduration=4.024201363 podStartE2EDuration="59.968598739s" podCreationTimestamp="2025-12-03 14:12:03 +0000 UTC" firstStartedPulling="2025-12-03 14:12:06.438319644 +0000 UTC m=+156.101236567" lastFinishedPulling="2025-12-03 14:13:02.38271702 +0000 UTC m=+212.045633943" observedRunningTime="2025-12-03 14:13:02.965822295 +0000 UTC m=+212.628739218" watchObservedRunningTime="2025-12-03 14:13:02.968598739 +0000 UTC m=+212.631515662" Dec 03 14:13:03 crc kubenswrapper[4805]: I1203 14:13:03.650328 4805 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-w6nm5" Dec 03 14:13:03 crc kubenswrapper[4805]: I1203 14:13:03.650882 4805 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-w6nm5" Dec 03 14:13:03 crc kubenswrapper[4805]: I1203 14:13:03.694442 4805 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-w6nm5" Dec 03 14:13:03 crc kubenswrapper[4805]: I1203 14:13:03.959785 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-dcw96" event={"ID":"8fd4f74e-3b48-4a4a-8898-e76c5926abab","Type":"ContainerStarted","Data":"5042f6082cb2397962c0570413a8bd4e5d2d8eee4d625cd906acc4f25bad7ce3"} Dec 03 14:13:03 crc kubenswrapper[4805]: I1203 14:13:03.982121 4805 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-dcw96" podStartSLOduration=2.051624322 podStartE2EDuration="57.982103429s" podCreationTimestamp="2025-12-03 14:12:06 +0000 UTC" firstStartedPulling="2025-12-03 14:12:07.507093222 +0000 UTC m=+157.170010145" lastFinishedPulling="2025-12-03 14:13:03.437572329 +0000 UTC m=+213.100489252" observedRunningTime="2025-12-03 14:13:03.980685571 +0000 UTC m=+213.643602494" watchObservedRunningTime="2025-12-03 14:13:03.982103429 +0000 UTC m=+213.645020352" Dec 03 14:13:04 crc kubenswrapper[4805]: I1203 14:13:04.005106 4805 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-w6nm5" Dec 03 14:13:04 crc kubenswrapper[4805]: I1203 14:13:04.213887 4805 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-9cm4t" Dec 03 14:13:04 crc kubenswrapper[4805]: I1203 14:13:04.213953 4805 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-9cm4t" Dec 03 14:13:04 crc kubenswrapper[4805]: I1203 14:13:04.257455 4805 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-9cm4t" Dec 03 14:13:05 crc kubenswrapper[4805]: I1203 14:13:05.977801 4805 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-m7r99" Dec 03 14:13:05 crc kubenswrapper[4805]: I1203 14:13:05.978093 4805 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-m7r99" Dec 03 14:13:06 crc kubenswrapper[4805]: I1203 14:13:06.025306 4805 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-m7r99" Dec 03 14:13:06 crc kubenswrapper[4805]: I1203 14:13:06.194090 4805 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-dlmr9" Dec 03 14:13:06 crc kubenswrapper[4805]: I1203 14:13:06.194148 4805 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-dlmr9" Dec 03 14:13:06 crc kubenswrapper[4805]: I1203 14:13:06.252637 4805 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-dlmr9" Dec 03 14:13:06 crc kubenswrapper[4805]: I1203 14:13:06.863345 4805 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-bd7j9" Dec 03 14:13:06 crc kubenswrapper[4805]: I1203 14:13:06.863716 4805 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-bd7j9" Dec 03 14:13:06 crc kubenswrapper[4805]: I1203 14:13:06.901987 4805 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-bd7j9" Dec 03 14:13:06 crc kubenswrapper[4805]: I1203 14:13:06.995959 4805 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-dcw96" Dec 03 14:13:06 crc kubenswrapper[4805]: I1203 14:13:06.996446 4805 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-dcw96" Dec 03 14:13:07 crc kubenswrapper[4805]: I1203 14:13:07.010906 4805 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-dlmr9" Dec 03 14:13:07 crc kubenswrapper[4805]: I1203 14:13:07.012858 4805 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-m7r99" Dec 03 14:13:07 crc kubenswrapper[4805]: I1203 14:13:07.020886 4805 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-bd7j9" Dec 03 14:13:08 crc kubenswrapper[4805]: I1203 14:13:08.033758 4805 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-dcw96" podUID="8fd4f74e-3b48-4a4a-8898-e76c5926abab" containerName="registry-server" probeResult="failure" output=< Dec 03 14:13:08 crc kubenswrapper[4805]: timeout: failed to connect service ":50051" within 1s Dec 03 14:13:08 crc kubenswrapper[4805]: > Dec 03 14:13:10 crc kubenswrapper[4805]: I1203 14:13:10.106187 4805 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-dlmr9"] Dec 03 14:13:10 crc kubenswrapper[4805]: I1203 14:13:10.106750 4805 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-dlmr9" podUID="77e93612-7c30-4f0e-9347-6e1373dfaaf8" containerName="registry-server" containerID="cri-o://3b67b5a47b1aba08949400ad9b5eedb7ca2cd9e0411c6efc3b9165d04befe24c" gracePeriod=2 Dec 03 14:13:11 crc kubenswrapper[4805]: I1203 14:13:10.999968 4805 generic.go:334] "Generic (PLEG): container finished" podID="77e93612-7c30-4f0e-9347-6e1373dfaaf8" containerID="3b67b5a47b1aba08949400ad9b5eedb7ca2cd9e0411c6efc3b9165d04befe24c" exitCode=0 Dec 03 14:13:11 crc kubenswrapper[4805]: I1203 14:13:11.000017 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-dlmr9" event={"ID":"77e93612-7c30-4f0e-9347-6e1373dfaaf8","Type":"ContainerDied","Data":"3b67b5a47b1aba08949400ad9b5eedb7ca2cd9e0411c6efc3b9165d04befe24c"} Dec 03 14:13:11 crc kubenswrapper[4805]: I1203 14:13:11.918796 4805 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-dlmr9" Dec 03 14:13:12 crc kubenswrapper[4805]: I1203 14:13:12.013708 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-dlmr9" event={"ID":"77e93612-7c30-4f0e-9347-6e1373dfaaf8","Type":"ContainerDied","Data":"60fe01cf43710b6e306d919ebd344e08920a9544ca27ddf906ddec2535a10f9a"} Dec 03 14:13:12 crc kubenswrapper[4805]: I1203 14:13:12.013759 4805 scope.go:117] "RemoveContainer" containerID="3b67b5a47b1aba08949400ad9b5eedb7ca2cd9e0411c6efc3b9165d04befe24c" Dec 03 14:13:12 crc kubenswrapper[4805]: I1203 14:13:12.013922 4805 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-dlmr9" Dec 03 14:13:12 crc kubenswrapper[4805]: I1203 14:13:12.015943 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-s6blv\" (UniqueName: \"kubernetes.io/projected/77e93612-7c30-4f0e-9347-6e1373dfaaf8-kube-api-access-s6blv\") pod \"77e93612-7c30-4f0e-9347-6e1373dfaaf8\" (UID: \"77e93612-7c30-4f0e-9347-6e1373dfaaf8\") " Dec 03 14:13:12 crc kubenswrapper[4805]: I1203 14:13:12.016052 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/77e93612-7c30-4f0e-9347-6e1373dfaaf8-catalog-content\") pod \"77e93612-7c30-4f0e-9347-6e1373dfaaf8\" (UID: \"77e93612-7c30-4f0e-9347-6e1373dfaaf8\") " Dec 03 14:13:12 crc kubenswrapper[4805]: I1203 14:13:12.016101 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/77e93612-7c30-4f0e-9347-6e1373dfaaf8-utilities\") pod \"77e93612-7c30-4f0e-9347-6e1373dfaaf8\" (UID: \"77e93612-7c30-4f0e-9347-6e1373dfaaf8\") " Dec 03 14:13:12 crc kubenswrapper[4805]: I1203 14:13:12.016936 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/77e93612-7c30-4f0e-9347-6e1373dfaaf8-utilities" (OuterVolumeSpecName: "utilities") pod "77e93612-7c30-4f0e-9347-6e1373dfaaf8" (UID: "77e93612-7c30-4f0e-9347-6e1373dfaaf8"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 14:13:12 crc kubenswrapper[4805]: I1203 14:13:12.021641 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/77e93612-7c30-4f0e-9347-6e1373dfaaf8-kube-api-access-s6blv" (OuterVolumeSpecName: "kube-api-access-s6blv") pod "77e93612-7c30-4f0e-9347-6e1373dfaaf8" (UID: "77e93612-7c30-4f0e-9347-6e1373dfaaf8"). InnerVolumeSpecName "kube-api-access-s6blv". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 14:13:12 crc kubenswrapper[4805]: I1203 14:13:12.034440 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/77e93612-7c30-4f0e-9347-6e1373dfaaf8-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "77e93612-7c30-4f0e-9347-6e1373dfaaf8" (UID: "77e93612-7c30-4f0e-9347-6e1373dfaaf8"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 14:13:12 crc kubenswrapper[4805]: I1203 14:13:12.046551 4805 scope.go:117] "RemoveContainer" containerID="381c8419c7601aad161bac4e1319af388398957804435159738857cb6591b293" Dec 03 14:13:12 crc kubenswrapper[4805]: I1203 14:13:12.065280 4805 scope.go:117] "RemoveContainer" containerID="d9d2f8339d6ff1d43a7fe3e0a95de4dfa69ac0deced0520fc10abdffb92b64dc" Dec 03 14:13:12 crc kubenswrapper[4805]: I1203 14:13:12.117345 4805 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-s6blv\" (UniqueName: \"kubernetes.io/projected/77e93612-7c30-4f0e-9347-6e1373dfaaf8-kube-api-access-s6blv\") on node \"crc\" DevicePath \"\"" Dec 03 14:13:12 crc kubenswrapper[4805]: I1203 14:13:12.117383 4805 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/77e93612-7c30-4f0e-9347-6e1373dfaaf8-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 03 14:13:12 crc kubenswrapper[4805]: I1203 14:13:12.117425 4805 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/77e93612-7c30-4f0e-9347-6e1373dfaaf8-utilities\") on node \"crc\" DevicePath \"\"" Dec 03 14:13:12 crc kubenswrapper[4805]: I1203 14:13:12.343768 4805 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-dlmr9"] Dec 03 14:13:12 crc kubenswrapper[4805]: I1203 14:13:12.346123 4805 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-dlmr9"] Dec 03 14:13:12 crc kubenswrapper[4805]: I1203 14:13:12.703408 4805 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="77e93612-7c30-4f0e-9347-6e1373dfaaf8" path="/var/lib/kubelet/pods/77e93612-7c30-4f0e-9347-6e1373dfaaf8/volumes" Dec 03 14:13:13 crc kubenswrapper[4805]: I1203 14:13:13.917215 4805 patch_prober.go:28] interesting pod/machine-config-daemon-gskh4 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 03 14:13:13 crc kubenswrapper[4805]: I1203 14:13:13.917538 4805 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-gskh4" podUID="6098937f-e3f6-45e8-a647-4994a79cd711" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 03 14:13:13 crc kubenswrapper[4805]: I1203 14:13:13.917584 4805 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-gskh4" Dec 03 14:13:13 crc kubenswrapper[4805]: I1203 14:13:13.918195 4805 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"9ad6b68a42fd33ee167e67873c6eed3f11cdf6dc71d22e952c021436eb967dd9"} pod="openshift-machine-config-operator/machine-config-daemon-gskh4" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 03 14:13:13 crc kubenswrapper[4805]: I1203 14:13:13.918291 4805 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-gskh4" podUID="6098937f-e3f6-45e8-a647-4994a79cd711" containerName="machine-config-daemon" containerID="cri-o://9ad6b68a42fd33ee167e67873c6eed3f11cdf6dc71d22e952c021436eb967dd9" gracePeriod=600 Dec 03 14:13:14 crc kubenswrapper[4805]: I1203 14:13:14.255389 4805 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-9cm4t" Dec 03 14:13:14 crc kubenswrapper[4805]: I1203 14:13:14.903788 4805 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-9cm4t"] Dec 03 14:13:15 crc kubenswrapper[4805]: I1203 14:13:15.030074 4805 generic.go:334] "Generic (PLEG): container finished" podID="6098937f-e3f6-45e8-a647-4994a79cd711" containerID="9ad6b68a42fd33ee167e67873c6eed3f11cdf6dc71d22e952c021436eb967dd9" exitCode=0 Dec 03 14:13:15 crc kubenswrapper[4805]: I1203 14:13:15.030159 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-gskh4" event={"ID":"6098937f-e3f6-45e8-a647-4994a79cd711","Type":"ContainerDied","Data":"9ad6b68a42fd33ee167e67873c6eed3f11cdf6dc71d22e952c021436eb967dd9"} Dec 03 14:13:15 crc kubenswrapper[4805]: I1203 14:13:15.030201 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-gskh4" event={"ID":"6098937f-e3f6-45e8-a647-4994a79cd711","Type":"ContainerStarted","Data":"550f8bfd5120af37ce2c9f018ac5ef6ec16519c28f68395f823626d542873ea8"} Dec 03 14:13:15 crc kubenswrapper[4805]: I1203 14:13:15.030274 4805 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-9cm4t" podUID="64f6384f-41e0-469e-8920-27e9e97b62db" containerName="registry-server" containerID="cri-o://65a3e60d7a131cd0d239b955d8911d4d763c2d8da4f4420acf1c8963654a31e6" gracePeriod=2 Dec 03 14:13:16 crc kubenswrapper[4805]: I1203 14:13:16.038492 4805 generic.go:334] "Generic (PLEG): container finished" podID="64f6384f-41e0-469e-8920-27e9e97b62db" containerID="65a3e60d7a131cd0d239b955d8911d4d763c2d8da4f4420acf1c8963654a31e6" exitCode=0 Dec 03 14:13:16 crc kubenswrapper[4805]: I1203 14:13:16.038584 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-9cm4t" event={"ID":"64f6384f-41e0-469e-8920-27e9e97b62db","Type":"ContainerDied","Data":"65a3e60d7a131cd0d239b955d8911d4d763c2d8da4f4420acf1c8963654a31e6"} Dec 03 14:13:16 crc kubenswrapper[4805]: I1203 14:13:16.476157 4805 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-9cm4t" Dec 03 14:13:16 crc kubenswrapper[4805]: I1203 14:13:16.631149 4805 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-ntv6j"] Dec 03 14:13:16 crc kubenswrapper[4805]: I1203 14:13:16.666030 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/64f6384f-41e0-469e-8920-27e9e97b62db-catalog-content\") pod \"64f6384f-41e0-469e-8920-27e9e97b62db\" (UID: \"64f6384f-41e0-469e-8920-27e9e97b62db\") " Dec 03 14:13:16 crc kubenswrapper[4805]: I1203 14:13:16.666109 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/64f6384f-41e0-469e-8920-27e9e97b62db-utilities\") pod \"64f6384f-41e0-469e-8920-27e9e97b62db\" (UID: \"64f6384f-41e0-469e-8920-27e9e97b62db\") " Dec 03 14:13:16 crc kubenswrapper[4805]: I1203 14:13:16.666145 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rv4vh\" (UniqueName: \"kubernetes.io/projected/64f6384f-41e0-469e-8920-27e9e97b62db-kube-api-access-rv4vh\") pod \"64f6384f-41e0-469e-8920-27e9e97b62db\" (UID: \"64f6384f-41e0-469e-8920-27e9e97b62db\") " Dec 03 14:13:16 crc kubenswrapper[4805]: I1203 14:13:16.667337 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/64f6384f-41e0-469e-8920-27e9e97b62db-utilities" (OuterVolumeSpecName: "utilities") pod "64f6384f-41e0-469e-8920-27e9e97b62db" (UID: "64f6384f-41e0-469e-8920-27e9e97b62db"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 14:13:16 crc kubenswrapper[4805]: I1203 14:13:16.680141 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/64f6384f-41e0-469e-8920-27e9e97b62db-kube-api-access-rv4vh" (OuterVolumeSpecName: "kube-api-access-rv4vh") pod "64f6384f-41e0-469e-8920-27e9e97b62db" (UID: "64f6384f-41e0-469e-8920-27e9e97b62db"). InnerVolumeSpecName "kube-api-access-rv4vh". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 14:13:16 crc kubenswrapper[4805]: I1203 14:13:16.721274 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/64f6384f-41e0-469e-8920-27e9e97b62db-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "64f6384f-41e0-469e-8920-27e9e97b62db" (UID: "64f6384f-41e0-469e-8920-27e9e97b62db"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 14:13:16 crc kubenswrapper[4805]: I1203 14:13:16.770177 4805 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/64f6384f-41e0-469e-8920-27e9e97b62db-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 03 14:13:16 crc kubenswrapper[4805]: I1203 14:13:16.770216 4805 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/64f6384f-41e0-469e-8920-27e9e97b62db-utilities\") on node \"crc\" DevicePath \"\"" Dec 03 14:13:16 crc kubenswrapper[4805]: I1203 14:13:16.770231 4805 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rv4vh\" (UniqueName: \"kubernetes.io/projected/64f6384f-41e0-469e-8920-27e9e97b62db-kube-api-access-rv4vh\") on node \"crc\" DevicePath \"\"" Dec 03 14:13:17 crc kubenswrapper[4805]: I1203 14:13:17.036131 4805 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-dcw96" Dec 03 14:13:17 crc kubenswrapper[4805]: I1203 14:13:17.045488 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-9cm4t" event={"ID":"64f6384f-41e0-469e-8920-27e9e97b62db","Type":"ContainerDied","Data":"2044a0871b4ae6ac514cc53eb77fa1c70c3a799fe279d2d157c39f4aad5f407f"} Dec 03 14:13:17 crc kubenswrapper[4805]: I1203 14:13:17.045542 4805 scope.go:117] "RemoveContainer" containerID="65a3e60d7a131cd0d239b955d8911d4d763c2d8da4f4420acf1c8963654a31e6" Dec 03 14:13:17 crc kubenswrapper[4805]: I1203 14:13:17.045666 4805 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-9cm4t" Dec 03 14:13:17 crc kubenswrapper[4805]: I1203 14:13:17.065430 4805 scope.go:117] "RemoveContainer" containerID="d5453cf390b7015789f699303295f36cf786a15f78885a09b9c40bd9adf0c125" Dec 03 14:13:17 crc kubenswrapper[4805]: I1203 14:13:17.070339 4805 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-9cm4t"] Dec 03 14:13:17 crc kubenswrapper[4805]: I1203 14:13:17.074689 4805 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-9cm4t"] Dec 03 14:13:17 crc kubenswrapper[4805]: I1203 14:13:17.075870 4805 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-dcw96" Dec 03 14:13:17 crc kubenswrapper[4805]: I1203 14:13:17.091273 4805 scope.go:117] "RemoveContainer" containerID="0c30658b6b2df48a53b073c612c15a2418c09abafbcee03a2cb9f43d89c0c547" Dec 03 14:13:18 crc kubenswrapper[4805]: I1203 14:13:18.703020 4805 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="64f6384f-41e0-469e-8920-27e9e97b62db" path="/var/lib/kubelet/pods/64f6384f-41e0-469e-8920-27e9e97b62db/volumes" Dec 03 14:13:19 crc kubenswrapper[4805]: I1203 14:13:19.304437 4805 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-dcw96"] Dec 03 14:13:19 crc kubenswrapper[4805]: I1203 14:13:19.305043 4805 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-dcw96" podUID="8fd4f74e-3b48-4a4a-8898-e76c5926abab" containerName="registry-server" containerID="cri-o://5042f6082cb2397962c0570413a8bd4e5d2d8eee4d625cd906acc4f25bad7ce3" gracePeriod=2 Dec 03 14:13:19 crc kubenswrapper[4805]: I1203 14:13:19.661859 4805 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-dcw96" Dec 03 14:13:19 crc kubenswrapper[4805]: I1203 14:13:19.806624 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8fd4f74e-3b48-4a4a-8898-e76c5926abab-utilities\") pod \"8fd4f74e-3b48-4a4a-8898-e76c5926abab\" (UID: \"8fd4f74e-3b48-4a4a-8898-e76c5926abab\") " Dec 03 14:13:19 crc kubenswrapper[4805]: I1203 14:13:19.806689 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8fd4f74e-3b48-4a4a-8898-e76c5926abab-catalog-content\") pod \"8fd4f74e-3b48-4a4a-8898-e76c5926abab\" (UID: \"8fd4f74e-3b48-4a4a-8898-e76c5926abab\") " Dec 03 14:13:19 crc kubenswrapper[4805]: I1203 14:13:19.806733 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-btdbn\" (UniqueName: \"kubernetes.io/projected/8fd4f74e-3b48-4a4a-8898-e76c5926abab-kube-api-access-btdbn\") pod \"8fd4f74e-3b48-4a4a-8898-e76c5926abab\" (UID: \"8fd4f74e-3b48-4a4a-8898-e76c5926abab\") " Dec 03 14:13:19 crc kubenswrapper[4805]: I1203 14:13:19.807497 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8fd4f74e-3b48-4a4a-8898-e76c5926abab-utilities" (OuterVolumeSpecName: "utilities") pod "8fd4f74e-3b48-4a4a-8898-e76c5926abab" (UID: "8fd4f74e-3b48-4a4a-8898-e76c5926abab"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 14:13:19 crc kubenswrapper[4805]: I1203 14:13:19.814620 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8fd4f74e-3b48-4a4a-8898-e76c5926abab-kube-api-access-btdbn" (OuterVolumeSpecName: "kube-api-access-btdbn") pod "8fd4f74e-3b48-4a4a-8898-e76c5926abab" (UID: "8fd4f74e-3b48-4a4a-8898-e76c5926abab"). InnerVolumeSpecName "kube-api-access-btdbn". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 14:13:19 crc kubenswrapper[4805]: I1203 14:13:19.908880 4805 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8fd4f74e-3b48-4a4a-8898-e76c5926abab-utilities\") on node \"crc\" DevicePath \"\"" Dec 03 14:13:19 crc kubenswrapper[4805]: I1203 14:13:19.908928 4805 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-btdbn\" (UniqueName: \"kubernetes.io/projected/8fd4f74e-3b48-4a4a-8898-e76c5926abab-kube-api-access-btdbn\") on node \"crc\" DevicePath \"\"" Dec 03 14:13:19 crc kubenswrapper[4805]: I1203 14:13:19.912546 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8fd4f74e-3b48-4a4a-8898-e76c5926abab-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "8fd4f74e-3b48-4a4a-8898-e76c5926abab" (UID: "8fd4f74e-3b48-4a4a-8898-e76c5926abab"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 14:13:20 crc kubenswrapper[4805]: I1203 14:13:20.009946 4805 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8fd4f74e-3b48-4a4a-8898-e76c5926abab-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 03 14:13:20 crc kubenswrapper[4805]: I1203 14:13:20.069947 4805 generic.go:334] "Generic (PLEG): container finished" podID="8fd4f74e-3b48-4a4a-8898-e76c5926abab" containerID="5042f6082cb2397962c0570413a8bd4e5d2d8eee4d625cd906acc4f25bad7ce3" exitCode=0 Dec 03 14:13:20 crc kubenswrapper[4805]: I1203 14:13:20.069992 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-dcw96" event={"ID":"8fd4f74e-3b48-4a4a-8898-e76c5926abab","Type":"ContainerDied","Data":"5042f6082cb2397962c0570413a8bd4e5d2d8eee4d625cd906acc4f25bad7ce3"} Dec 03 14:13:20 crc kubenswrapper[4805]: I1203 14:13:20.070020 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-dcw96" event={"ID":"8fd4f74e-3b48-4a4a-8898-e76c5926abab","Type":"ContainerDied","Data":"f5c09446dc047ae6e72d08ffb758c00dbecc9e5462264e08716ecc22716616db"} Dec 03 14:13:20 crc kubenswrapper[4805]: I1203 14:13:20.070036 4805 scope.go:117] "RemoveContainer" containerID="5042f6082cb2397962c0570413a8bd4e5d2d8eee4d625cd906acc4f25bad7ce3" Dec 03 14:13:20 crc kubenswrapper[4805]: I1203 14:13:20.070091 4805 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-dcw96" Dec 03 14:13:20 crc kubenswrapper[4805]: I1203 14:13:20.091612 4805 scope.go:117] "RemoveContainer" containerID="fe565be9850fa082dece91006e313b97f73b64633d1e70e948b4e30bf6193c74" Dec 03 14:13:20 crc kubenswrapper[4805]: I1203 14:13:20.115905 4805 scope.go:117] "RemoveContainer" containerID="55d20a2330bc6375f2d04585cfeb78ca414231ded021e913ef6546a6ce39ab03" Dec 03 14:13:20 crc kubenswrapper[4805]: I1203 14:13:20.117776 4805 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-dcw96"] Dec 03 14:13:20 crc kubenswrapper[4805]: I1203 14:13:20.121428 4805 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-dcw96"] Dec 03 14:13:20 crc kubenswrapper[4805]: I1203 14:13:20.485137 4805 scope.go:117] "RemoveContainer" containerID="5042f6082cb2397962c0570413a8bd4e5d2d8eee4d625cd906acc4f25bad7ce3" Dec 03 14:13:20 crc kubenswrapper[4805]: E1203 14:13:20.485894 4805 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5042f6082cb2397962c0570413a8bd4e5d2d8eee4d625cd906acc4f25bad7ce3\": container with ID starting with 5042f6082cb2397962c0570413a8bd4e5d2d8eee4d625cd906acc4f25bad7ce3 not found: ID does not exist" containerID="5042f6082cb2397962c0570413a8bd4e5d2d8eee4d625cd906acc4f25bad7ce3" Dec 03 14:13:20 crc kubenswrapper[4805]: I1203 14:13:20.486153 4805 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5042f6082cb2397962c0570413a8bd4e5d2d8eee4d625cd906acc4f25bad7ce3"} err="failed to get container status \"5042f6082cb2397962c0570413a8bd4e5d2d8eee4d625cd906acc4f25bad7ce3\": rpc error: code = NotFound desc = could not find container \"5042f6082cb2397962c0570413a8bd4e5d2d8eee4d625cd906acc4f25bad7ce3\": container with ID starting with 5042f6082cb2397962c0570413a8bd4e5d2d8eee4d625cd906acc4f25bad7ce3 not found: ID does not exist" Dec 03 14:13:20 crc kubenswrapper[4805]: I1203 14:13:20.486199 4805 scope.go:117] "RemoveContainer" containerID="fe565be9850fa082dece91006e313b97f73b64633d1e70e948b4e30bf6193c74" Dec 03 14:13:20 crc kubenswrapper[4805]: E1203 14:13:20.487066 4805 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"fe565be9850fa082dece91006e313b97f73b64633d1e70e948b4e30bf6193c74\": container with ID starting with fe565be9850fa082dece91006e313b97f73b64633d1e70e948b4e30bf6193c74 not found: ID does not exist" containerID="fe565be9850fa082dece91006e313b97f73b64633d1e70e948b4e30bf6193c74" Dec 03 14:13:20 crc kubenswrapper[4805]: I1203 14:13:20.487106 4805 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"fe565be9850fa082dece91006e313b97f73b64633d1e70e948b4e30bf6193c74"} err="failed to get container status \"fe565be9850fa082dece91006e313b97f73b64633d1e70e948b4e30bf6193c74\": rpc error: code = NotFound desc = could not find container \"fe565be9850fa082dece91006e313b97f73b64633d1e70e948b4e30bf6193c74\": container with ID starting with fe565be9850fa082dece91006e313b97f73b64633d1e70e948b4e30bf6193c74 not found: ID does not exist" Dec 03 14:13:20 crc kubenswrapper[4805]: I1203 14:13:20.487138 4805 scope.go:117] "RemoveContainer" containerID="55d20a2330bc6375f2d04585cfeb78ca414231ded021e913ef6546a6ce39ab03" Dec 03 14:13:20 crc kubenswrapper[4805]: E1203 14:13:20.487611 4805 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"55d20a2330bc6375f2d04585cfeb78ca414231ded021e913ef6546a6ce39ab03\": container with ID starting with 55d20a2330bc6375f2d04585cfeb78ca414231ded021e913ef6546a6ce39ab03 not found: ID does not exist" containerID="55d20a2330bc6375f2d04585cfeb78ca414231ded021e913ef6546a6ce39ab03" Dec 03 14:13:20 crc kubenswrapper[4805]: I1203 14:13:20.487644 4805 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"55d20a2330bc6375f2d04585cfeb78ca414231ded021e913ef6546a6ce39ab03"} err="failed to get container status \"55d20a2330bc6375f2d04585cfeb78ca414231ded021e913ef6546a6ce39ab03\": rpc error: code = NotFound desc = could not find container \"55d20a2330bc6375f2d04585cfeb78ca414231ded021e913ef6546a6ce39ab03\": container with ID starting with 55d20a2330bc6375f2d04585cfeb78ca414231ded021e913ef6546a6ce39ab03 not found: ID does not exist" Dec 03 14:13:20 crc kubenswrapper[4805]: I1203 14:13:20.711617 4805 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8fd4f74e-3b48-4a4a-8898-e76c5926abab" path="/var/lib/kubelet/pods/8fd4f74e-3b48-4a4a-8898-e76c5926abab/volumes" Dec 03 14:13:28 crc kubenswrapper[4805]: I1203 14:13:28.509531 4805 kubelet.go:2421] "SyncLoop ADD" source="file" pods=["openshift-kube-apiserver/kube-apiserver-startup-monitor-crc"] Dec 03 14:13:28 crc kubenswrapper[4805]: E1203 14:13:28.510421 4805 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8fd4f74e-3b48-4a4a-8898-e76c5926abab" containerName="extract-utilities" Dec 03 14:13:28 crc kubenswrapper[4805]: I1203 14:13:28.510439 4805 state_mem.go:107] "Deleted CPUSet assignment" podUID="8fd4f74e-3b48-4a4a-8898-e76c5926abab" containerName="extract-utilities" Dec 03 14:13:28 crc kubenswrapper[4805]: E1203 14:13:28.510454 4805 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="731aa7bc-cde1-4ff7-aa29-78093078514d" containerName="registry-server" Dec 03 14:13:28 crc kubenswrapper[4805]: I1203 14:13:28.510461 4805 state_mem.go:107] "Deleted CPUSet assignment" podUID="731aa7bc-cde1-4ff7-aa29-78093078514d" containerName="registry-server" Dec 03 14:13:28 crc kubenswrapper[4805]: E1203 14:13:28.510477 4805 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="77e93612-7c30-4f0e-9347-6e1373dfaaf8" containerName="extract-content" Dec 03 14:13:28 crc kubenswrapper[4805]: I1203 14:13:28.510484 4805 state_mem.go:107] "Deleted CPUSet assignment" podUID="77e93612-7c30-4f0e-9347-6e1373dfaaf8" containerName="extract-content" Dec 03 14:13:28 crc kubenswrapper[4805]: E1203 14:13:28.510494 4805 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8fd4f74e-3b48-4a4a-8898-e76c5926abab" containerName="registry-server" Dec 03 14:13:28 crc kubenswrapper[4805]: I1203 14:13:28.510502 4805 state_mem.go:107] "Deleted CPUSet assignment" podUID="8fd4f74e-3b48-4a4a-8898-e76c5926abab" containerName="registry-server" Dec 03 14:13:28 crc kubenswrapper[4805]: E1203 14:13:28.510511 4805 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="64f6384f-41e0-469e-8920-27e9e97b62db" containerName="extract-utilities" Dec 03 14:13:28 crc kubenswrapper[4805]: I1203 14:13:28.510518 4805 state_mem.go:107] "Deleted CPUSet assignment" podUID="64f6384f-41e0-469e-8920-27e9e97b62db" containerName="extract-utilities" Dec 03 14:13:28 crc kubenswrapper[4805]: E1203 14:13:28.510527 4805 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="731aa7bc-cde1-4ff7-aa29-78093078514d" containerName="extract-content" Dec 03 14:13:28 crc kubenswrapper[4805]: I1203 14:13:28.510534 4805 state_mem.go:107] "Deleted CPUSet assignment" podUID="731aa7bc-cde1-4ff7-aa29-78093078514d" containerName="extract-content" Dec 03 14:13:28 crc kubenswrapper[4805]: E1203 14:13:28.510544 4805 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="77e93612-7c30-4f0e-9347-6e1373dfaaf8" containerName="extract-utilities" Dec 03 14:13:28 crc kubenswrapper[4805]: I1203 14:13:28.510550 4805 state_mem.go:107] "Deleted CPUSet assignment" podUID="77e93612-7c30-4f0e-9347-6e1373dfaaf8" containerName="extract-utilities" Dec 03 14:13:28 crc kubenswrapper[4805]: E1203 14:13:28.510562 4805 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8fd4f74e-3b48-4a4a-8898-e76c5926abab" containerName="extract-content" Dec 03 14:13:28 crc kubenswrapper[4805]: I1203 14:13:28.510568 4805 state_mem.go:107] "Deleted CPUSet assignment" podUID="8fd4f74e-3b48-4a4a-8898-e76c5926abab" containerName="extract-content" Dec 03 14:13:28 crc kubenswrapper[4805]: E1203 14:13:28.510579 4805 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="64f6384f-41e0-469e-8920-27e9e97b62db" containerName="registry-server" Dec 03 14:13:28 crc kubenswrapper[4805]: I1203 14:13:28.510587 4805 state_mem.go:107] "Deleted CPUSet assignment" podUID="64f6384f-41e0-469e-8920-27e9e97b62db" containerName="registry-server" Dec 03 14:13:28 crc kubenswrapper[4805]: E1203 14:13:28.510598 4805 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="64f6384f-41e0-469e-8920-27e9e97b62db" containerName="extract-content" Dec 03 14:13:28 crc kubenswrapper[4805]: I1203 14:13:28.510606 4805 state_mem.go:107] "Deleted CPUSet assignment" podUID="64f6384f-41e0-469e-8920-27e9e97b62db" containerName="extract-content" Dec 03 14:13:28 crc kubenswrapper[4805]: E1203 14:13:28.510618 4805 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="731aa7bc-cde1-4ff7-aa29-78093078514d" containerName="extract-utilities" Dec 03 14:13:28 crc kubenswrapper[4805]: I1203 14:13:28.510626 4805 state_mem.go:107] "Deleted CPUSet assignment" podUID="731aa7bc-cde1-4ff7-aa29-78093078514d" containerName="extract-utilities" Dec 03 14:13:28 crc kubenswrapper[4805]: E1203 14:13:28.510636 4805 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="77e93612-7c30-4f0e-9347-6e1373dfaaf8" containerName="registry-server" Dec 03 14:13:28 crc kubenswrapper[4805]: I1203 14:13:28.510642 4805 state_mem.go:107] "Deleted CPUSet assignment" podUID="77e93612-7c30-4f0e-9347-6e1373dfaaf8" containerName="registry-server" Dec 03 14:13:28 crc kubenswrapper[4805]: I1203 14:13:28.510753 4805 memory_manager.go:354] "RemoveStaleState removing state" podUID="64f6384f-41e0-469e-8920-27e9e97b62db" containerName="registry-server" Dec 03 14:13:28 crc kubenswrapper[4805]: I1203 14:13:28.510765 4805 memory_manager.go:354] "RemoveStaleState removing state" podUID="731aa7bc-cde1-4ff7-aa29-78093078514d" containerName="registry-server" Dec 03 14:13:28 crc kubenswrapper[4805]: I1203 14:13:28.510775 4805 memory_manager.go:354] "RemoveStaleState removing state" podUID="8fd4f74e-3b48-4a4a-8898-e76c5926abab" containerName="registry-server" Dec 03 14:13:28 crc kubenswrapper[4805]: I1203 14:13:28.510784 4805 memory_manager.go:354] "RemoveStaleState removing state" podUID="77e93612-7c30-4f0e-9347-6e1373dfaaf8" containerName="registry-server" Dec 03 14:13:28 crc kubenswrapper[4805]: I1203 14:13:28.511174 4805 kubelet.go:2431] "SyncLoop REMOVE" source="file" pods=["openshift-kube-apiserver/kube-apiserver-crc"] Dec 03 14:13:28 crc kubenswrapper[4805]: I1203 14:13:28.511480 4805 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" containerID="cri-o://3e5a34636f0a8c4d9179074596d28bc2eb7a504573398c596a637defd829907a" gracePeriod=15 Dec 03 14:13:28 crc kubenswrapper[4805]: I1203 14:13:28.511558 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 03 14:13:28 crc kubenswrapper[4805]: I1203 14:13:28.511559 4805 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-insecure-readyz" containerID="cri-o://bd92a160eb530c5687ebd7cd5b83b41820a2b808a0fd9514e5ace9d237e1bd26" gracePeriod=15 Dec 03 14:13:28 crc kubenswrapper[4805]: I1203 14:13:28.511614 4805 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-regeneration-controller" containerID="cri-o://29cd95be3f964415943c873e7914ffcea4abe50dfdf17f8a9cd03fc38df35981" gracePeriod=15 Dec 03 14:13:28 crc kubenswrapper[4805]: I1203 14:13:28.511703 4805 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-syncer" containerID="cri-o://951173b3d3d581275a6f6e43209b5e0611a701ca7a10f81504d723e24a2d21a3" gracePeriod=15 Dec 03 14:13:28 crc kubenswrapper[4805]: I1203 14:13:28.512043 4805 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" containerID="cri-o://e64730957bbbb2feb4d506e0ed9731a0594bf66fa1e8fcdb8ec944ded6308eed" gracePeriod=15 Dec 03 14:13:28 crc kubenswrapper[4805]: I1203 14:13:28.512898 4805 kubelet.go:2421] "SyncLoop ADD" source="file" pods=["openshift-kube-apiserver/kube-apiserver-crc"] Dec 03 14:13:28 crc kubenswrapper[4805]: E1203 14:13:28.513598 4805 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="setup" Dec 03 14:13:28 crc kubenswrapper[4805]: I1203 14:13:28.513625 4805 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="setup" Dec 03 14:13:28 crc kubenswrapper[4805]: E1203 14:13:28.513650 4805 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-regeneration-controller" Dec 03 14:13:28 crc kubenswrapper[4805]: I1203 14:13:28.513667 4805 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-regeneration-controller" Dec 03 14:13:28 crc kubenswrapper[4805]: E1203 14:13:28.513692 4805 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-insecure-readyz" Dec 03 14:13:28 crc kubenswrapper[4805]: I1203 14:13:28.513708 4805 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-insecure-readyz" Dec 03 14:13:28 crc kubenswrapper[4805]: E1203 14:13:28.513734 4805 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-syncer" Dec 03 14:13:28 crc kubenswrapper[4805]: I1203 14:13:28.513748 4805 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-syncer" Dec 03 14:13:28 crc kubenswrapper[4805]: E1203 14:13:28.513768 4805 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" Dec 03 14:13:28 crc kubenswrapper[4805]: I1203 14:13:28.513783 4805 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" Dec 03 14:13:28 crc kubenswrapper[4805]: E1203 14:13:28.513814 4805 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Dec 03 14:13:28 crc kubenswrapper[4805]: I1203 14:13:28.513830 4805 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Dec 03 14:13:28 crc kubenswrapper[4805]: E1203 14:13:28.513885 4805 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Dec 03 14:13:28 crc kubenswrapper[4805]: I1203 14:13:28.513902 4805 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Dec 03 14:13:28 crc kubenswrapper[4805]: I1203 14:13:28.514158 4805 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Dec 03 14:13:28 crc kubenswrapper[4805]: I1203 14:13:28.514183 4805 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-regeneration-controller" Dec 03 14:13:28 crc kubenswrapper[4805]: I1203 14:13:28.514207 4805 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-syncer" Dec 03 14:13:28 crc kubenswrapper[4805]: I1203 14:13:28.514235 4805 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Dec 03 14:13:28 crc kubenswrapper[4805]: I1203 14:13:28.514289 4805 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-insecure-readyz" Dec 03 14:13:28 crc kubenswrapper[4805]: I1203 14:13:28.514307 4805 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" Dec 03 14:13:28 crc kubenswrapper[4805]: I1203 14:13:28.523430 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"manifests\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 03 14:13:28 crc kubenswrapper[4805]: I1203 14:13:28.523566 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 03 14:13:28 crc kubenswrapper[4805]: I1203 14:13:28.523617 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 03 14:13:28 crc kubenswrapper[4805]: I1203 14:13:28.523665 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 03 14:13:28 crc kubenswrapper[4805]: I1203 14:13:28.523718 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 03 14:13:28 crc kubenswrapper[4805]: I1203 14:13:28.523906 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 03 14:13:28 crc kubenswrapper[4805]: I1203 14:13:28.523983 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 03 14:13:28 crc kubenswrapper[4805]: I1203 14:13:28.524023 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 03 14:13:28 crc kubenswrapper[4805]: I1203 14:13:28.628828 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 03 14:13:28 crc kubenswrapper[4805]: I1203 14:13:28.628908 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 03 14:13:28 crc kubenswrapper[4805]: I1203 14:13:28.628939 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 03 14:13:28 crc kubenswrapper[4805]: I1203 14:13:28.628995 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"manifests\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 03 14:13:28 crc kubenswrapper[4805]: I1203 14:13:28.629023 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 03 14:13:28 crc kubenswrapper[4805]: I1203 14:13:28.629051 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 03 14:13:28 crc kubenswrapper[4805]: I1203 14:13:28.629072 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 03 14:13:28 crc kubenswrapper[4805]: I1203 14:13:28.629094 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 03 14:13:28 crc kubenswrapper[4805]: I1203 14:13:28.629164 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 03 14:13:28 crc kubenswrapper[4805]: I1203 14:13:28.629203 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 03 14:13:28 crc kubenswrapper[4805]: I1203 14:13:28.629230 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 03 14:13:28 crc kubenswrapper[4805]: I1203 14:13:28.629254 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 03 14:13:28 crc kubenswrapper[4805]: I1203 14:13:28.629277 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"manifests\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 03 14:13:28 crc kubenswrapper[4805]: I1203 14:13:28.629300 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 03 14:13:28 crc kubenswrapper[4805]: I1203 14:13:28.629325 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 03 14:13:28 crc kubenswrapper[4805]: I1203 14:13:28.629348 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 03 14:13:29 crc kubenswrapper[4805]: I1203 14:13:29.130129 4805 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/0.log" Dec 03 14:13:29 crc kubenswrapper[4805]: I1203 14:13:29.132095 4805 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-cert-syncer/0.log" Dec 03 14:13:29 crc kubenswrapper[4805]: I1203 14:13:29.133127 4805 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="e64730957bbbb2feb4d506e0ed9731a0594bf66fa1e8fcdb8ec944ded6308eed" exitCode=0 Dec 03 14:13:29 crc kubenswrapper[4805]: I1203 14:13:29.133155 4805 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="bd92a160eb530c5687ebd7cd5b83b41820a2b808a0fd9514e5ace9d237e1bd26" exitCode=0 Dec 03 14:13:29 crc kubenswrapper[4805]: I1203 14:13:29.133163 4805 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="29cd95be3f964415943c873e7914ffcea4abe50dfdf17f8a9cd03fc38df35981" exitCode=0 Dec 03 14:13:29 crc kubenswrapper[4805]: I1203 14:13:29.133172 4805 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="951173b3d3d581275a6f6e43209b5e0611a701ca7a10f81504d723e24a2d21a3" exitCode=2 Dec 03 14:13:29 crc kubenswrapper[4805]: I1203 14:13:29.133243 4805 scope.go:117] "RemoveContainer" containerID="5da55f769e23ad52bac9c77da62c78e70fa34db8ed1a435e4eaf4dad3fd2b68a" Dec 03 14:13:29 crc kubenswrapper[4805]: I1203 14:13:29.136411 4805 generic.go:334] "Generic (PLEG): container finished" podID="62ba53fc-fcf4-47be-a415-ea016529bac1" containerID="8a65ec5b7de3262599fc740a2a5b215b18fa383d37c67b71f2c75048b5b4deda" exitCode=0 Dec 03 14:13:29 crc kubenswrapper[4805]: I1203 14:13:29.136446 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/installer-9-crc" event={"ID":"62ba53fc-fcf4-47be-a415-ea016529bac1","Type":"ContainerDied","Data":"8a65ec5b7de3262599fc740a2a5b215b18fa383d37c67b71f2c75048b5b4deda"} Dec 03 14:13:29 crc kubenswrapper[4805]: I1203 14:13:29.137628 4805 status_manager.go:851] "Failed to get status for pod" podUID="62ba53fc-fcf4-47be-a415-ea016529bac1" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.2:6443: connect: connection refused" Dec 03 14:13:30 crc kubenswrapper[4805]: I1203 14:13:30.149765 4805 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-cert-syncer/0.log" Dec 03 14:13:30 crc kubenswrapper[4805]: I1203 14:13:30.420261 4805 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/installer-9-crc" Dec 03 14:13:30 crc kubenswrapper[4805]: I1203 14:13:30.421140 4805 status_manager.go:851] "Failed to get status for pod" podUID="62ba53fc-fcf4-47be-a415-ea016529bac1" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.2:6443: connect: connection refused" Dec 03 14:13:30 crc kubenswrapper[4805]: I1203 14:13:30.551044 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/62ba53fc-fcf4-47be-a415-ea016529bac1-kubelet-dir\") pod \"62ba53fc-fcf4-47be-a415-ea016529bac1\" (UID: \"62ba53fc-fcf4-47be-a415-ea016529bac1\") " Dec 03 14:13:30 crc kubenswrapper[4805]: I1203 14:13:30.551174 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/62ba53fc-fcf4-47be-a415-ea016529bac1-kubelet-dir" (OuterVolumeSpecName: "kubelet-dir") pod "62ba53fc-fcf4-47be-a415-ea016529bac1" (UID: "62ba53fc-fcf4-47be-a415-ea016529bac1"). InnerVolumeSpecName "kubelet-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 03 14:13:30 crc kubenswrapper[4805]: I1203 14:13:30.551606 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/62ba53fc-fcf4-47be-a415-ea016529bac1-var-lock" (OuterVolumeSpecName: "var-lock") pod "62ba53fc-fcf4-47be-a415-ea016529bac1" (UID: "62ba53fc-fcf4-47be-a415-ea016529bac1"). InnerVolumeSpecName "var-lock". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 03 14:13:30 crc kubenswrapper[4805]: I1203 14:13:30.551658 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/62ba53fc-fcf4-47be-a415-ea016529bac1-var-lock\") pod \"62ba53fc-fcf4-47be-a415-ea016529bac1\" (UID: \"62ba53fc-fcf4-47be-a415-ea016529bac1\") " Dec 03 14:13:30 crc kubenswrapper[4805]: I1203 14:13:30.553274 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/62ba53fc-fcf4-47be-a415-ea016529bac1-kube-api-access\") pod \"62ba53fc-fcf4-47be-a415-ea016529bac1\" (UID: \"62ba53fc-fcf4-47be-a415-ea016529bac1\") " Dec 03 14:13:30 crc kubenswrapper[4805]: I1203 14:13:30.553696 4805 reconciler_common.go:293] "Volume detached for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/62ba53fc-fcf4-47be-a415-ea016529bac1-var-lock\") on node \"crc\" DevicePath \"\"" Dec 03 14:13:30 crc kubenswrapper[4805]: I1203 14:13:30.553737 4805 reconciler_common.go:293] "Volume detached for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/62ba53fc-fcf4-47be-a415-ea016529bac1-kubelet-dir\") on node \"crc\" DevicePath \"\"" Dec 03 14:13:30 crc kubenswrapper[4805]: I1203 14:13:30.557001 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/62ba53fc-fcf4-47be-a415-ea016529bac1-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "62ba53fc-fcf4-47be-a415-ea016529bac1" (UID: "62ba53fc-fcf4-47be-a415-ea016529bac1"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 14:13:30 crc kubenswrapper[4805]: I1203 14:13:30.655480 4805 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/62ba53fc-fcf4-47be-a415-ea016529bac1-kube-api-access\") on node \"crc\" DevicePath \"\"" Dec 03 14:13:30 crc kubenswrapper[4805]: I1203 14:13:30.706984 4805 status_manager.go:851] "Failed to get status for pod" podUID="62ba53fc-fcf4-47be-a415-ea016529bac1" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.2:6443: connect: connection refused" Dec 03 14:13:30 crc kubenswrapper[4805]: E1203 14:13:30.769685 4805 desired_state_of_world_populator.go:312] "Error processing volume" err="error processing PVC openshift-image-registry/crc-image-registry-storage: failed to fetch PVC from API server: Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-image-registry/persistentvolumeclaims/crc-image-registry-storage\": dial tcp 38.102.83.2:6443: connect: connection refused" pod="openshift-image-registry/image-registry-697d97f7c8-jjzct" volumeName="registry-storage" Dec 03 14:13:30 crc kubenswrapper[4805]: I1203 14:13:30.897987 4805 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-cert-syncer/0.log" Dec 03 14:13:30 crc kubenswrapper[4805]: I1203 14:13:30.899046 4805 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 03 14:13:30 crc kubenswrapper[4805]: I1203 14:13:30.899608 4805 status_manager.go:851] "Failed to get status for pod" podUID="f4b27818a5e8e43d0dc095d08835c792" pod="openshift-kube-apiserver/kube-apiserver-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.2:6443: connect: connection refused" Dec 03 14:13:30 crc kubenswrapper[4805]: I1203 14:13:30.900054 4805 status_manager.go:851] "Failed to get status for pod" podUID="62ba53fc-fcf4-47be-a415-ea016529bac1" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.2:6443: connect: connection refused" Dec 03 14:13:31 crc kubenswrapper[4805]: I1203 14:13:31.060117 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") pod \"f4b27818a5e8e43d0dc095d08835c792\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " Dec 03 14:13:31 crc kubenswrapper[4805]: I1203 14:13:31.060178 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") pod \"f4b27818a5e8e43d0dc095d08835c792\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " Dec 03 14:13:31 crc kubenswrapper[4805]: I1203 14:13:31.060223 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") pod \"f4b27818a5e8e43d0dc095d08835c792\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " Dec 03 14:13:31 crc kubenswrapper[4805]: I1203 14:13:31.060505 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir" (OuterVolumeSpecName: "cert-dir") pod "f4b27818a5e8e43d0dc095d08835c792" (UID: "f4b27818a5e8e43d0dc095d08835c792"). InnerVolumeSpecName "cert-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 03 14:13:31 crc kubenswrapper[4805]: I1203 14:13:31.060543 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir" (OuterVolumeSpecName: "resource-dir") pod "f4b27818a5e8e43d0dc095d08835c792" (UID: "f4b27818a5e8e43d0dc095d08835c792"). InnerVolumeSpecName "resource-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 03 14:13:31 crc kubenswrapper[4805]: I1203 14:13:31.060563 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir" (OuterVolumeSpecName: "audit-dir") pod "f4b27818a5e8e43d0dc095d08835c792" (UID: "f4b27818a5e8e43d0dc095d08835c792"). InnerVolumeSpecName "audit-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 03 14:13:31 crc kubenswrapper[4805]: I1203 14:13:31.159240 4805 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-cert-syncer/0.log" Dec 03 14:13:31 crc kubenswrapper[4805]: I1203 14:13:31.159915 4805 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="3e5a34636f0a8c4d9179074596d28bc2eb7a504573398c596a637defd829907a" exitCode=0 Dec 03 14:13:31 crc kubenswrapper[4805]: I1203 14:13:31.160025 4805 scope.go:117] "RemoveContainer" containerID="e64730957bbbb2feb4d506e0ed9731a0594bf66fa1e8fcdb8ec944ded6308eed" Dec 03 14:13:31 crc kubenswrapper[4805]: I1203 14:13:31.160222 4805 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 03 14:13:31 crc kubenswrapper[4805]: I1203 14:13:31.163400 4805 reconciler_common.go:293] "Volume detached for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") on node \"crc\" DevicePath \"\"" Dec 03 14:13:31 crc kubenswrapper[4805]: I1203 14:13:31.163462 4805 reconciler_common.go:293] "Volume detached for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") on node \"crc\" DevicePath \"\"" Dec 03 14:13:31 crc kubenswrapper[4805]: I1203 14:13:31.163486 4805 reconciler_common.go:293] "Volume detached for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") on node \"crc\" DevicePath \"\"" Dec 03 14:13:31 crc kubenswrapper[4805]: I1203 14:13:31.168088 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/installer-9-crc" event={"ID":"62ba53fc-fcf4-47be-a415-ea016529bac1","Type":"ContainerDied","Data":"bda7e3bf266699bbd48c9f409b7901d9f99f3f853398099c4cd493857174e4ca"} Dec 03 14:13:31 crc kubenswrapper[4805]: I1203 14:13:31.168171 4805 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="bda7e3bf266699bbd48c9f409b7901d9f99f3f853398099c4cd493857174e4ca" Dec 03 14:13:31 crc kubenswrapper[4805]: I1203 14:13:31.168285 4805 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/installer-9-crc" Dec 03 14:13:31 crc kubenswrapper[4805]: I1203 14:13:31.174653 4805 status_manager.go:851] "Failed to get status for pod" podUID="62ba53fc-fcf4-47be-a415-ea016529bac1" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.2:6443: connect: connection refused" Dec 03 14:13:31 crc kubenswrapper[4805]: I1203 14:13:31.175202 4805 status_manager.go:851] "Failed to get status for pod" podUID="f4b27818a5e8e43d0dc095d08835c792" pod="openshift-kube-apiserver/kube-apiserver-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.2:6443: connect: connection refused" Dec 03 14:13:31 crc kubenswrapper[4805]: I1203 14:13:31.189815 4805 status_manager.go:851] "Failed to get status for pod" podUID="f4b27818a5e8e43d0dc095d08835c792" pod="openshift-kube-apiserver/kube-apiserver-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.2:6443: connect: connection refused" Dec 03 14:13:31 crc kubenswrapper[4805]: I1203 14:13:31.190278 4805 status_manager.go:851] "Failed to get status for pod" podUID="62ba53fc-fcf4-47be-a415-ea016529bac1" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.2:6443: connect: connection refused" Dec 03 14:13:31 crc kubenswrapper[4805]: I1203 14:13:31.194996 4805 scope.go:117] "RemoveContainer" containerID="bd92a160eb530c5687ebd7cd5b83b41820a2b808a0fd9514e5ace9d237e1bd26" Dec 03 14:13:31 crc kubenswrapper[4805]: I1203 14:13:31.214375 4805 scope.go:117] "RemoveContainer" containerID="29cd95be3f964415943c873e7914ffcea4abe50dfdf17f8a9cd03fc38df35981" Dec 03 14:13:31 crc kubenswrapper[4805]: I1203 14:13:31.235020 4805 scope.go:117] "RemoveContainer" containerID="951173b3d3d581275a6f6e43209b5e0611a701ca7a10f81504d723e24a2d21a3" Dec 03 14:13:31 crc kubenswrapper[4805]: I1203 14:13:31.253926 4805 scope.go:117] "RemoveContainer" containerID="3e5a34636f0a8c4d9179074596d28bc2eb7a504573398c596a637defd829907a" Dec 03 14:13:31 crc kubenswrapper[4805]: I1203 14:13:31.275589 4805 scope.go:117] "RemoveContainer" containerID="af9c940135012c629447528c61743ed1e58708ab57ce76604c6da0479aa459cf" Dec 03 14:13:31 crc kubenswrapper[4805]: I1203 14:13:31.293912 4805 scope.go:117] "RemoveContainer" containerID="e64730957bbbb2feb4d506e0ed9731a0594bf66fa1e8fcdb8ec944ded6308eed" Dec 03 14:13:31 crc kubenswrapper[4805]: E1203 14:13:31.294440 4805 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e64730957bbbb2feb4d506e0ed9731a0594bf66fa1e8fcdb8ec944ded6308eed\": container with ID starting with e64730957bbbb2feb4d506e0ed9731a0594bf66fa1e8fcdb8ec944ded6308eed not found: ID does not exist" containerID="e64730957bbbb2feb4d506e0ed9731a0594bf66fa1e8fcdb8ec944ded6308eed" Dec 03 14:13:31 crc kubenswrapper[4805]: I1203 14:13:31.294518 4805 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e64730957bbbb2feb4d506e0ed9731a0594bf66fa1e8fcdb8ec944ded6308eed"} err="failed to get container status \"e64730957bbbb2feb4d506e0ed9731a0594bf66fa1e8fcdb8ec944ded6308eed\": rpc error: code = NotFound desc = could not find container \"e64730957bbbb2feb4d506e0ed9731a0594bf66fa1e8fcdb8ec944ded6308eed\": container with ID starting with e64730957bbbb2feb4d506e0ed9731a0594bf66fa1e8fcdb8ec944ded6308eed not found: ID does not exist" Dec 03 14:13:31 crc kubenswrapper[4805]: I1203 14:13:31.294561 4805 scope.go:117] "RemoveContainer" containerID="bd92a160eb530c5687ebd7cd5b83b41820a2b808a0fd9514e5ace9d237e1bd26" Dec 03 14:13:31 crc kubenswrapper[4805]: E1203 14:13:31.295247 4805 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"bd92a160eb530c5687ebd7cd5b83b41820a2b808a0fd9514e5ace9d237e1bd26\": container with ID starting with bd92a160eb530c5687ebd7cd5b83b41820a2b808a0fd9514e5ace9d237e1bd26 not found: ID does not exist" containerID="bd92a160eb530c5687ebd7cd5b83b41820a2b808a0fd9514e5ace9d237e1bd26" Dec 03 14:13:31 crc kubenswrapper[4805]: I1203 14:13:31.295304 4805 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"bd92a160eb530c5687ebd7cd5b83b41820a2b808a0fd9514e5ace9d237e1bd26"} err="failed to get container status \"bd92a160eb530c5687ebd7cd5b83b41820a2b808a0fd9514e5ace9d237e1bd26\": rpc error: code = NotFound desc = could not find container \"bd92a160eb530c5687ebd7cd5b83b41820a2b808a0fd9514e5ace9d237e1bd26\": container with ID starting with bd92a160eb530c5687ebd7cd5b83b41820a2b808a0fd9514e5ace9d237e1bd26 not found: ID does not exist" Dec 03 14:13:31 crc kubenswrapper[4805]: I1203 14:13:31.295340 4805 scope.go:117] "RemoveContainer" containerID="29cd95be3f964415943c873e7914ffcea4abe50dfdf17f8a9cd03fc38df35981" Dec 03 14:13:31 crc kubenswrapper[4805]: E1203 14:13:31.296306 4805 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"29cd95be3f964415943c873e7914ffcea4abe50dfdf17f8a9cd03fc38df35981\": container with ID starting with 29cd95be3f964415943c873e7914ffcea4abe50dfdf17f8a9cd03fc38df35981 not found: ID does not exist" containerID="29cd95be3f964415943c873e7914ffcea4abe50dfdf17f8a9cd03fc38df35981" Dec 03 14:13:31 crc kubenswrapper[4805]: I1203 14:13:31.296347 4805 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"29cd95be3f964415943c873e7914ffcea4abe50dfdf17f8a9cd03fc38df35981"} err="failed to get container status \"29cd95be3f964415943c873e7914ffcea4abe50dfdf17f8a9cd03fc38df35981\": rpc error: code = NotFound desc = could not find container \"29cd95be3f964415943c873e7914ffcea4abe50dfdf17f8a9cd03fc38df35981\": container with ID starting with 29cd95be3f964415943c873e7914ffcea4abe50dfdf17f8a9cd03fc38df35981 not found: ID does not exist" Dec 03 14:13:31 crc kubenswrapper[4805]: I1203 14:13:31.296374 4805 scope.go:117] "RemoveContainer" containerID="951173b3d3d581275a6f6e43209b5e0611a701ca7a10f81504d723e24a2d21a3" Dec 03 14:13:31 crc kubenswrapper[4805]: E1203 14:13:31.297566 4805 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"951173b3d3d581275a6f6e43209b5e0611a701ca7a10f81504d723e24a2d21a3\": container with ID starting with 951173b3d3d581275a6f6e43209b5e0611a701ca7a10f81504d723e24a2d21a3 not found: ID does not exist" containerID="951173b3d3d581275a6f6e43209b5e0611a701ca7a10f81504d723e24a2d21a3" Dec 03 14:13:31 crc kubenswrapper[4805]: I1203 14:13:31.297630 4805 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"951173b3d3d581275a6f6e43209b5e0611a701ca7a10f81504d723e24a2d21a3"} err="failed to get container status \"951173b3d3d581275a6f6e43209b5e0611a701ca7a10f81504d723e24a2d21a3\": rpc error: code = NotFound desc = could not find container \"951173b3d3d581275a6f6e43209b5e0611a701ca7a10f81504d723e24a2d21a3\": container with ID starting with 951173b3d3d581275a6f6e43209b5e0611a701ca7a10f81504d723e24a2d21a3 not found: ID does not exist" Dec 03 14:13:31 crc kubenswrapper[4805]: I1203 14:13:31.297669 4805 scope.go:117] "RemoveContainer" containerID="3e5a34636f0a8c4d9179074596d28bc2eb7a504573398c596a637defd829907a" Dec 03 14:13:31 crc kubenswrapper[4805]: E1203 14:13:31.298254 4805 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3e5a34636f0a8c4d9179074596d28bc2eb7a504573398c596a637defd829907a\": container with ID starting with 3e5a34636f0a8c4d9179074596d28bc2eb7a504573398c596a637defd829907a not found: ID does not exist" containerID="3e5a34636f0a8c4d9179074596d28bc2eb7a504573398c596a637defd829907a" Dec 03 14:13:31 crc kubenswrapper[4805]: I1203 14:13:31.298315 4805 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3e5a34636f0a8c4d9179074596d28bc2eb7a504573398c596a637defd829907a"} err="failed to get container status \"3e5a34636f0a8c4d9179074596d28bc2eb7a504573398c596a637defd829907a\": rpc error: code = NotFound desc = could not find container \"3e5a34636f0a8c4d9179074596d28bc2eb7a504573398c596a637defd829907a\": container with ID starting with 3e5a34636f0a8c4d9179074596d28bc2eb7a504573398c596a637defd829907a not found: ID does not exist" Dec 03 14:13:31 crc kubenswrapper[4805]: I1203 14:13:31.298345 4805 scope.go:117] "RemoveContainer" containerID="af9c940135012c629447528c61743ed1e58708ab57ce76604c6da0479aa459cf" Dec 03 14:13:31 crc kubenswrapper[4805]: E1203 14:13:31.298892 4805 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"af9c940135012c629447528c61743ed1e58708ab57ce76604c6da0479aa459cf\": container with ID starting with af9c940135012c629447528c61743ed1e58708ab57ce76604c6da0479aa459cf not found: ID does not exist" containerID="af9c940135012c629447528c61743ed1e58708ab57ce76604c6da0479aa459cf" Dec 03 14:13:31 crc kubenswrapper[4805]: I1203 14:13:31.298951 4805 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"af9c940135012c629447528c61743ed1e58708ab57ce76604c6da0479aa459cf"} err="failed to get container status \"af9c940135012c629447528c61743ed1e58708ab57ce76604c6da0479aa459cf\": rpc error: code = NotFound desc = could not find container \"af9c940135012c629447528c61743ed1e58708ab57ce76604c6da0479aa459cf\": container with ID starting with af9c940135012c629447528c61743ed1e58708ab57ce76604c6da0479aa459cf not found: ID does not exist" Dec 03 14:13:32 crc kubenswrapper[4805]: I1203 14:13:32.703244 4805 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f4b27818a5e8e43d0dc095d08835c792" path="/var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/volumes" Dec 03 14:13:33 crc kubenswrapper[4805]: E1203 14:13:33.555484 4805 kubelet.go:1929] "Failed creating a mirror pod for" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods\": dial tcp 38.102.83.2:6443: connect: connection refused" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 03 14:13:33 crc kubenswrapper[4805]: I1203 14:13:33.556191 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 03 14:13:33 crc kubenswrapper[4805]: E1203 14:13:33.589625 4805 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/events\": dial tcp 38.102.83.2:6443: connect: connection refused" event="&Event{ObjectMeta:{kube-apiserver-startup-monitor-crc.187dba118fc72fad openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-startup-monitor-crc,UID:f85e55b1a89d02b0cb034b1ea31ed45a,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{startup-monitor},},Reason:Pulled,Message:Container image \"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2025-12-03 14:13:33.588508589 +0000 UTC m=+243.251425542,LastTimestamp:2025-12-03 14:13:33.588508589 +0000 UTC m=+243.251425542,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Dec 03 14:13:34 crc kubenswrapper[4805]: I1203 14:13:34.192956 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" event={"ID":"f85e55b1a89d02b0cb034b1ea31ed45a","Type":"ContainerStarted","Data":"def67344e1a1d03baa3748a24cca7799fa8707eac5d82957414ba33076fe1d29"} Dec 03 14:13:34 crc kubenswrapper[4805]: I1203 14:13:34.193312 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" event={"ID":"f85e55b1a89d02b0cb034b1ea31ed45a","Type":"ContainerStarted","Data":"2c3c2e470490bc1fa72ca9dfb39b90db47b2af82fed5b0dfe477962062a76632"} Dec 03 14:13:34 crc kubenswrapper[4805]: E1203 14:13:34.194087 4805 kubelet.go:1929] "Failed creating a mirror pod for" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods\": dial tcp 38.102.83.2:6443: connect: connection refused" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 03 14:13:34 crc kubenswrapper[4805]: I1203 14:13:34.194234 4805 status_manager.go:851] "Failed to get status for pod" podUID="62ba53fc-fcf4-47be-a415-ea016529bac1" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.2:6443: connect: connection refused" Dec 03 14:13:36 crc kubenswrapper[4805]: E1203 14:13:36.378668 4805 controller.go:195] "Failed to update lease" err="Put \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.2:6443: connect: connection refused" Dec 03 14:13:36 crc kubenswrapper[4805]: E1203 14:13:36.379539 4805 controller.go:195] "Failed to update lease" err="Put \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.2:6443: connect: connection refused" Dec 03 14:13:36 crc kubenswrapper[4805]: E1203 14:13:36.380198 4805 controller.go:195] "Failed to update lease" err="Put \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.2:6443: connect: connection refused" Dec 03 14:13:36 crc kubenswrapper[4805]: E1203 14:13:36.380893 4805 controller.go:195] "Failed to update lease" err="Put \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.2:6443: connect: connection refused" Dec 03 14:13:36 crc kubenswrapper[4805]: E1203 14:13:36.381394 4805 controller.go:195] "Failed to update lease" err="Put \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.2:6443: connect: connection refused" Dec 03 14:13:36 crc kubenswrapper[4805]: I1203 14:13:36.381462 4805 controller.go:115] "failed to update lease using latest lease, fallback to ensure lease" err="failed 5 attempts to update lease" Dec 03 14:13:36 crc kubenswrapper[4805]: E1203 14:13:36.382073 4805 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.2:6443: connect: connection refused" interval="200ms" Dec 03 14:13:36 crc kubenswrapper[4805]: E1203 14:13:36.583650 4805 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.2:6443: connect: connection refused" interval="400ms" Dec 03 14:13:36 crc kubenswrapper[4805]: E1203 14:13:36.984993 4805 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.2:6443: connect: connection refused" interval="800ms" Dec 03 14:13:37 crc kubenswrapper[4805]: E1203 14:13:37.748458 4805 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/events\": dial tcp 38.102.83.2:6443: connect: connection refused" event="&Event{ObjectMeta:{kube-apiserver-startup-monitor-crc.187dba118fc72fad openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-startup-monitor-crc,UID:f85e55b1a89d02b0cb034b1ea31ed45a,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{startup-monitor},},Reason:Pulled,Message:Container image \"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2025-12-03 14:13:33.588508589 +0000 UTC m=+243.251425542,LastTimestamp:2025-12-03 14:13:33.588508589 +0000 UTC m=+243.251425542,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Dec 03 14:13:37 crc kubenswrapper[4805]: E1203 14:13:37.786079 4805 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.2:6443: connect: connection refused" interval="1.6s" Dec 03 14:13:39 crc kubenswrapper[4805]: E1203 14:13:39.387380 4805 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.2:6443: connect: connection refused" interval="3.2s" Dec 03 14:13:40 crc kubenswrapper[4805]: I1203 14:13:40.694152 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 03 14:13:40 crc kubenswrapper[4805]: I1203 14:13:40.697908 4805 status_manager.go:851] "Failed to get status for pod" podUID="62ba53fc-fcf4-47be-a415-ea016529bac1" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.2:6443: connect: connection refused" Dec 03 14:13:40 crc kubenswrapper[4805]: I1203 14:13:40.698415 4805 status_manager.go:851] "Failed to get status for pod" podUID="62ba53fc-fcf4-47be-a415-ea016529bac1" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.2:6443: connect: connection refused" Dec 03 14:13:40 crc kubenswrapper[4805]: I1203 14:13:40.716347 4805 kubelet.go:1909] "Trying to delete pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="ced8291d-0ce5-4e74-9eaf-2bf033774039" Dec 03 14:13:40 crc kubenswrapper[4805]: I1203 14:13:40.716402 4805 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="ced8291d-0ce5-4e74-9eaf-2bf033774039" Dec 03 14:13:40 crc kubenswrapper[4805]: E1203 14:13:40.717212 4805 mirror_client.go:138] "Failed deleting a mirror pod" err="Delete \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.2:6443: connect: connection refused" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 03 14:13:40 crc kubenswrapper[4805]: I1203 14:13:40.717914 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 03 14:13:40 crc kubenswrapper[4805]: W1203 14:13:40.747640 4805 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod71bb4a3aecc4ba5b26c4b7318770ce13.slice/crio-06e9e6a601cec6c053b69a1999594412a0d1eb91f61fb6704b51e5c69c4bc8b2 WatchSource:0}: Error finding container 06e9e6a601cec6c053b69a1999594412a0d1eb91f61fb6704b51e5c69c4bc8b2: Status 404 returned error can't find the container with id 06e9e6a601cec6c053b69a1999594412a0d1eb91f61fb6704b51e5c69c4bc8b2 Dec 03 14:13:41 crc kubenswrapper[4805]: I1203 14:13:41.238673 4805 generic.go:334] "Generic (PLEG): container finished" podID="71bb4a3aecc4ba5b26c4b7318770ce13" containerID="621d927753654f08d773f3a19050c10ec1fe580e9b9b69fb9c1514f27e78ef43" exitCode=0 Dec 03 14:13:41 crc kubenswrapper[4805]: I1203 14:13:41.238729 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerDied","Data":"621d927753654f08d773f3a19050c10ec1fe580e9b9b69fb9c1514f27e78ef43"} Dec 03 14:13:41 crc kubenswrapper[4805]: I1203 14:13:41.238804 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"06e9e6a601cec6c053b69a1999594412a0d1eb91f61fb6704b51e5c69c4bc8b2"} Dec 03 14:13:41 crc kubenswrapper[4805]: I1203 14:13:41.239346 4805 kubelet.go:1909] "Trying to delete pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="ced8291d-0ce5-4e74-9eaf-2bf033774039" Dec 03 14:13:41 crc kubenswrapper[4805]: I1203 14:13:41.239387 4805 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="ced8291d-0ce5-4e74-9eaf-2bf033774039" Dec 03 14:13:41 crc kubenswrapper[4805]: I1203 14:13:41.239705 4805 status_manager.go:851] "Failed to get status for pod" podUID="62ba53fc-fcf4-47be-a415-ea016529bac1" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.2:6443: connect: connection refused" Dec 03 14:13:41 crc kubenswrapper[4805]: E1203 14:13:41.239880 4805 mirror_client.go:138] "Failed deleting a mirror pod" err="Delete \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.2:6443: connect: connection refused" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 03 14:13:41 crc kubenswrapper[4805]: I1203 14:13:41.243665 4805 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-controller-manager_kube-controller-manager-crc_f614b9022728cf315e60c057852e563e/kube-controller-manager/0.log" Dec 03 14:13:41 crc kubenswrapper[4805]: I1203 14:13:41.243739 4805 generic.go:334] "Generic (PLEG): container finished" podID="f614b9022728cf315e60c057852e563e" containerID="8673a29c0e16f4e707d79f0610722ccabb023ff2b4a18d7cf9d305183817d952" exitCode=1 Dec 03 14:13:41 crc kubenswrapper[4805]: I1203 14:13:41.243773 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerDied","Data":"8673a29c0e16f4e707d79f0610722ccabb023ff2b4a18d7cf9d305183817d952"} Dec 03 14:13:41 crc kubenswrapper[4805]: I1203 14:13:41.244399 4805 scope.go:117] "RemoveContainer" containerID="8673a29c0e16f4e707d79f0610722ccabb023ff2b4a18d7cf9d305183817d952" Dec 03 14:13:41 crc kubenswrapper[4805]: I1203 14:13:41.244535 4805 status_manager.go:851] "Failed to get status for pod" podUID="f614b9022728cf315e60c057852e563e" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-controller-manager/pods/kube-controller-manager-crc\": dial tcp 38.102.83.2:6443: connect: connection refused" Dec 03 14:13:41 crc kubenswrapper[4805]: I1203 14:13:41.245345 4805 status_manager.go:851] "Failed to get status for pod" podUID="62ba53fc-fcf4-47be-a415-ea016529bac1" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.2:6443: connect: connection refused" Dec 03 14:13:41 crc kubenswrapper[4805]: I1203 14:13:41.654493 4805 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-authentication/oauth-openshift-558db77b4-ntv6j" podUID="1658d8e1-ac63-40fe-98ce-900a9f9aeae5" containerName="oauth-openshift" containerID="cri-o://bf5f45a82b82b4c46cfce651ba0e5f2273e655f4722c7485eed466b0c536c77f" gracePeriod=15 Dec 03 14:13:42 crc kubenswrapper[4805]: I1203 14:13:42.119861 4805 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-ntv6j" Dec 03 14:13:42 crc kubenswrapper[4805]: I1203 14:13:42.205435 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/1658d8e1-ac63-40fe-98ce-900a9f9aeae5-v4-0-config-user-template-provider-selection\") pod \"1658d8e1-ac63-40fe-98ce-900a9f9aeae5\" (UID: \"1658d8e1-ac63-40fe-98ce-900a9f9aeae5\") " Dec 03 14:13:42 crc kubenswrapper[4805]: I1203 14:13:42.205548 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/1658d8e1-ac63-40fe-98ce-900a9f9aeae5-v4-0-config-system-session\") pod \"1658d8e1-ac63-40fe-98ce-900a9f9aeae5\" (UID: \"1658d8e1-ac63-40fe-98ce-900a9f9aeae5\") " Dec 03 14:13:42 crc kubenswrapper[4805]: I1203 14:13:42.205590 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/1658d8e1-ac63-40fe-98ce-900a9f9aeae5-v4-0-config-system-cliconfig\") pod \"1658d8e1-ac63-40fe-98ce-900a9f9aeae5\" (UID: \"1658d8e1-ac63-40fe-98ce-900a9f9aeae5\") " Dec 03 14:13:42 crc kubenswrapper[4805]: I1203 14:13:42.205626 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/1658d8e1-ac63-40fe-98ce-900a9f9aeae5-v4-0-config-system-ocp-branding-template\") pod \"1658d8e1-ac63-40fe-98ce-900a9f9aeae5\" (UID: \"1658d8e1-ac63-40fe-98ce-900a9f9aeae5\") " Dec 03 14:13:42 crc kubenswrapper[4805]: I1203 14:13:42.205660 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/1658d8e1-ac63-40fe-98ce-900a9f9aeae5-v4-0-config-system-router-certs\") pod \"1658d8e1-ac63-40fe-98ce-900a9f9aeae5\" (UID: \"1658d8e1-ac63-40fe-98ce-900a9f9aeae5\") " Dec 03 14:13:42 crc kubenswrapper[4805]: I1203 14:13:42.205684 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/1658d8e1-ac63-40fe-98ce-900a9f9aeae5-v4-0-config-system-trusted-ca-bundle\") pod \"1658d8e1-ac63-40fe-98ce-900a9f9aeae5\" (UID: \"1658d8e1-ac63-40fe-98ce-900a9f9aeae5\") " Dec 03 14:13:42 crc kubenswrapper[4805]: I1203 14:13:42.205722 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dxgkw\" (UniqueName: \"kubernetes.io/projected/1658d8e1-ac63-40fe-98ce-900a9f9aeae5-kube-api-access-dxgkw\") pod \"1658d8e1-ac63-40fe-98ce-900a9f9aeae5\" (UID: \"1658d8e1-ac63-40fe-98ce-900a9f9aeae5\") " Dec 03 14:13:42 crc kubenswrapper[4805]: I1203 14:13:42.205747 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/1658d8e1-ac63-40fe-98ce-900a9f9aeae5-audit-dir\") pod \"1658d8e1-ac63-40fe-98ce-900a9f9aeae5\" (UID: \"1658d8e1-ac63-40fe-98ce-900a9f9aeae5\") " Dec 03 14:13:42 crc kubenswrapper[4805]: I1203 14:13:42.205777 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/1658d8e1-ac63-40fe-98ce-900a9f9aeae5-v4-0-config-system-service-ca\") pod \"1658d8e1-ac63-40fe-98ce-900a9f9aeae5\" (UID: \"1658d8e1-ac63-40fe-98ce-900a9f9aeae5\") " Dec 03 14:13:42 crc kubenswrapper[4805]: I1203 14:13:42.205805 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/1658d8e1-ac63-40fe-98ce-900a9f9aeae5-audit-policies\") pod \"1658d8e1-ac63-40fe-98ce-900a9f9aeae5\" (UID: \"1658d8e1-ac63-40fe-98ce-900a9f9aeae5\") " Dec 03 14:13:42 crc kubenswrapper[4805]: I1203 14:13:42.205827 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/1658d8e1-ac63-40fe-98ce-900a9f9aeae5-v4-0-config-user-idp-0-file-data\") pod \"1658d8e1-ac63-40fe-98ce-900a9f9aeae5\" (UID: \"1658d8e1-ac63-40fe-98ce-900a9f9aeae5\") " Dec 03 14:13:42 crc kubenswrapper[4805]: I1203 14:13:42.205914 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/1658d8e1-ac63-40fe-98ce-900a9f9aeae5-v4-0-config-system-serving-cert\") pod \"1658d8e1-ac63-40fe-98ce-900a9f9aeae5\" (UID: \"1658d8e1-ac63-40fe-98ce-900a9f9aeae5\") " Dec 03 14:13:42 crc kubenswrapper[4805]: I1203 14:13:42.205958 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/1658d8e1-ac63-40fe-98ce-900a9f9aeae5-v4-0-config-user-template-login\") pod \"1658d8e1-ac63-40fe-98ce-900a9f9aeae5\" (UID: \"1658d8e1-ac63-40fe-98ce-900a9f9aeae5\") " Dec 03 14:13:42 crc kubenswrapper[4805]: I1203 14:13:42.205996 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/1658d8e1-ac63-40fe-98ce-900a9f9aeae5-v4-0-config-user-template-error\") pod \"1658d8e1-ac63-40fe-98ce-900a9f9aeae5\" (UID: \"1658d8e1-ac63-40fe-98ce-900a9f9aeae5\") " Dec 03 14:13:42 crc kubenswrapper[4805]: I1203 14:13:42.206420 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1658d8e1-ac63-40fe-98ce-900a9f9aeae5-v4-0-config-system-cliconfig" (OuterVolumeSpecName: "v4-0-config-system-cliconfig") pod "1658d8e1-ac63-40fe-98ce-900a9f9aeae5" (UID: "1658d8e1-ac63-40fe-98ce-900a9f9aeae5"). InnerVolumeSpecName "v4-0-config-system-cliconfig". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 14:13:42 crc kubenswrapper[4805]: I1203 14:13:42.219195 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1658d8e1-ac63-40fe-98ce-900a9f9aeae5-v4-0-config-system-trusted-ca-bundle" (OuterVolumeSpecName: "v4-0-config-system-trusted-ca-bundle") pod "1658d8e1-ac63-40fe-98ce-900a9f9aeae5" (UID: "1658d8e1-ac63-40fe-98ce-900a9f9aeae5"). InnerVolumeSpecName "v4-0-config-system-trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 14:13:42 crc kubenswrapper[4805]: I1203 14:13:42.219469 4805 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/1658d8e1-ac63-40fe-98ce-900a9f9aeae5-v4-0-config-system-cliconfig\") on node \"crc\" DevicePath \"\"" Dec 03 14:13:42 crc kubenswrapper[4805]: I1203 14:13:42.219492 4805 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/1658d8e1-ac63-40fe-98ce-900a9f9aeae5-v4-0-config-system-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 14:13:42 crc kubenswrapper[4805]: I1203 14:13:42.223157 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1658d8e1-ac63-40fe-98ce-900a9f9aeae5-v4-0-config-user-template-provider-selection" (OuterVolumeSpecName: "v4-0-config-user-template-provider-selection") pod "1658d8e1-ac63-40fe-98ce-900a9f9aeae5" (UID: "1658d8e1-ac63-40fe-98ce-900a9f9aeae5"). InnerVolumeSpecName "v4-0-config-user-template-provider-selection". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 14:13:42 crc kubenswrapper[4805]: I1203 14:13:42.223298 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1658d8e1-ac63-40fe-98ce-900a9f9aeae5-audit-policies" (OuterVolumeSpecName: "audit-policies") pod "1658d8e1-ac63-40fe-98ce-900a9f9aeae5" (UID: "1658d8e1-ac63-40fe-98ce-900a9f9aeae5"). InnerVolumeSpecName "audit-policies". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 14:13:42 crc kubenswrapper[4805]: I1203 14:13:42.226096 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1658d8e1-ac63-40fe-98ce-900a9f9aeae5-v4-0-config-system-session" (OuterVolumeSpecName: "v4-0-config-system-session") pod "1658d8e1-ac63-40fe-98ce-900a9f9aeae5" (UID: "1658d8e1-ac63-40fe-98ce-900a9f9aeae5"). InnerVolumeSpecName "v4-0-config-system-session". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 14:13:42 crc kubenswrapper[4805]: I1203 14:13:42.225724 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/1658d8e1-ac63-40fe-98ce-900a9f9aeae5-audit-dir" (OuterVolumeSpecName: "audit-dir") pod "1658d8e1-ac63-40fe-98ce-900a9f9aeae5" (UID: "1658d8e1-ac63-40fe-98ce-900a9f9aeae5"). InnerVolumeSpecName "audit-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 03 14:13:42 crc kubenswrapper[4805]: I1203 14:13:42.226484 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1658d8e1-ac63-40fe-98ce-900a9f9aeae5-v4-0-config-user-template-error" (OuterVolumeSpecName: "v4-0-config-user-template-error") pod "1658d8e1-ac63-40fe-98ce-900a9f9aeae5" (UID: "1658d8e1-ac63-40fe-98ce-900a9f9aeae5"). InnerVolumeSpecName "v4-0-config-user-template-error". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 14:13:42 crc kubenswrapper[4805]: I1203 14:13:42.226658 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1658d8e1-ac63-40fe-98ce-900a9f9aeae5-v4-0-config-system-service-ca" (OuterVolumeSpecName: "v4-0-config-system-service-ca") pod "1658d8e1-ac63-40fe-98ce-900a9f9aeae5" (UID: "1658d8e1-ac63-40fe-98ce-900a9f9aeae5"). InnerVolumeSpecName "v4-0-config-system-service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 14:13:42 crc kubenswrapper[4805]: I1203 14:13:42.228606 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1658d8e1-ac63-40fe-98ce-900a9f9aeae5-v4-0-config-system-ocp-branding-template" (OuterVolumeSpecName: "v4-0-config-system-ocp-branding-template") pod "1658d8e1-ac63-40fe-98ce-900a9f9aeae5" (UID: "1658d8e1-ac63-40fe-98ce-900a9f9aeae5"). InnerVolumeSpecName "v4-0-config-system-ocp-branding-template". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 14:13:42 crc kubenswrapper[4805]: I1203 14:13:42.232005 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1658d8e1-ac63-40fe-98ce-900a9f9aeae5-v4-0-config-system-router-certs" (OuterVolumeSpecName: "v4-0-config-system-router-certs") pod "1658d8e1-ac63-40fe-98ce-900a9f9aeae5" (UID: "1658d8e1-ac63-40fe-98ce-900a9f9aeae5"). InnerVolumeSpecName "v4-0-config-system-router-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 14:13:42 crc kubenswrapper[4805]: I1203 14:13:42.232899 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1658d8e1-ac63-40fe-98ce-900a9f9aeae5-v4-0-config-user-idp-0-file-data" (OuterVolumeSpecName: "v4-0-config-user-idp-0-file-data") pod "1658d8e1-ac63-40fe-98ce-900a9f9aeae5" (UID: "1658d8e1-ac63-40fe-98ce-900a9f9aeae5"). InnerVolumeSpecName "v4-0-config-user-idp-0-file-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 14:13:42 crc kubenswrapper[4805]: I1203 14:13:42.233463 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1658d8e1-ac63-40fe-98ce-900a9f9aeae5-v4-0-config-user-template-login" (OuterVolumeSpecName: "v4-0-config-user-template-login") pod "1658d8e1-ac63-40fe-98ce-900a9f9aeae5" (UID: "1658d8e1-ac63-40fe-98ce-900a9f9aeae5"). InnerVolumeSpecName "v4-0-config-user-template-login". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 14:13:42 crc kubenswrapper[4805]: I1203 14:13:42.233727 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1658d8e1-ac63-40fe-98ce-900a9f9aeae5-v4-0-config-system-serving-cert" (OuterVolumeSpecName: "v4-0-config-system-serving-cert") pod "1658d8e1-ac63-40fe-98ce-900a9f9aeae5" (UID: "1658d8e1-ac63-40fe-98ce-900a9f9aeae5"). InnerVolumeSpecName "v4-0-config-system-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 14:13:42 crc kubenswrapper[4805]: I1203 14:13:42.236668 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1658d8e1-ac63-40fe-98ce-900a9f9aeae5-kube-api-access-dxgkw" (OuterVolumeSpecName: "kube-api-access-dxgkw") pod "1658d8e1-ac63-40fe-98ce-900a9f9aeae5" (UID: "1658d8e1-ac63-40fe-98ce-900a9f9aeae5"). InnerVolumeSpecName "kube-api-access-dxgkw". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 14:13:42 crc kubenswrapper[4805]: I1203 14:13:42.251134 4805 generic.go:334] "Generic (PLEG): container finished" podID="1658d8e1-ac63-40fe-98ce-900a9f9aeae5" containerID="bf5f45a82b82b4c46cfce651ba0e5f2273e655f4722c7485eed466b0c536c77f" exitCode=0 Dec 03 14:13:42 crc kubenswrapper[4805]: I1203 14:13:42.251203 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-ntv6j" event={"ID":"1658d8e1-ac63-40fe-98ce-900a9f9aeae5","Type":"ContainerDied","Data":"bf5f45a82b82b4c46cfce651ba0e5f2273e655f4722c7485eed466b0c536c77f"} Dec 03 14:13:42 crc kubenswrapper[4805]: I1203 14:13:42.251235 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-ntv6j" event={"ID":"1658d8e1-ac63-40fe-98ce-900a9f9aeae5","Type":"ContainerDied","Data":"b58229b467e31512c2984b233d11b240a14366ce5c3263aeab9af7a21498a9a6"} Dec 03 14:13:42 crc kubenswrapper[4805]: I1203 14:13:42.251256 4805 scope.go:117] "RemoveContainer" containerID="bf5f45a82b82b4c46cfce651ba0e5f2273e655f4722c7485eed466b0c536c77f" Dec 03 14:13:42 crc kubenswrapper[4805]: I1203 14:13:42.251374 4805 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-ntv6j" Dec 03 14:13:42 crc kubenswrapper[4805]: I1203 14:13:42.261517 4805 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-controller-manager_kube-controller-manager-crc_f614b9022728cf315e60c057852e563e/kube-controller-manager/0.log" Dec 03 14:13:42 crc kubenswrapper[4805]: I1203 14:13:42.261779 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"604de07374b8d85ab0785f21f409ddd439bfadd9af01803476055e82f903166a"} Dec 03 14:13:42 crc kubenswrapper[4805]: I1203 14:13:42.265668 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"9eda093f9202f9c335cdab9a716acf7c8cf8ab42f1477b31f6a155c25008cb20"} Dec 03 14:13:42 crc kubenswrapper[4805]: I1203 14:13:42.265952 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"9c0df6a00c2ef5877d4f17b86ae6c9c45b2e94cde271f6e0feea9546446b9f1e"} Dec 03 14:13:42 crc kubenswrapper[4805]: I1203 14:13:42.266055 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"f5a7a58320bad5c0038e02bcb9971f3da0da7ceaacb0a62d4e254c322b046552"} Dec 03 14:13:42 crc kubenswrapper[4805]: I1203 14:13:42.274555 4805 scope.go:117] "RemoveContainer" containerID="bf5f45a82b82b4c46cfce651ba0e5f2273e655f4722c7485eed466b0c536c77f" Dec 03 14:13:42 crc kubenswrapper[4805]: E1203 14:13:42.278085 4805 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"bf5f45a82b82b4c46cfce651ba0e5f2273e655f4722c7485eed466b0c536c77f\": container with ID starting with bf5f45a82b82b4c46cfce651ba0e5f2273e655f4722c7485eed466b0c536c77f not found: ID does not exist" containerID="bf5f45a82b82b4c46cfce651ba0e5f2273e655f4722c7485eed466b0c536c77f" Dec 03 14:13:42 crc kubenswrapper[4805]: I1203 14:13:42.278200 4805 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"bf5f45a82b82b4c46cfce651ba0e5f2273e655f4722c7485eed466b0c536c77f"} err="failed to get container status \"bf5f45a82b82b4c46cfce651ba0e5f2273e655f4722c7485eed466b0c536c77f\": rpc error: code = NotFound desc = could not find container \"bf5f45a82b82b4c46cfce651ba0e5f2273e655f4722c7485eed466b0c536c77f\": container with ID starting with bf5f45a82b82b4c46cfce651ba0e5f2273e655f4722c7485eed466b0c536c77f not found: ID does not exist" Dec 03 14:13:42 crc kubenswrapper[4805]: I1203 14:13:42.320900 4805 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/1658d8e1-ac63-40fe-98ce-900a9f9aeae5-v4-0-config-system-service-ca\") on node \"crc\" DevicePath \"\"" Dec 03 14:13:42 crc kubenswrapper[4805]: I1203 14:13:42.320932 4805 reconciler_common.go:293] "Volume detached for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/1658d8e1-ac63-40fe-98ce-900a9f9aeae5-audit-policies\") on node \"crc\" DevicePath \"\"" Dec 03 14:13:42 crc kubenswrapper[4805]: I1203 14:13:42.320947 4805 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/1658d8e1-ac63-40fe-98ce-900a9f9aeae5-v4-0-config-user-idp-0-file-data\") on node \"crc\" DevicePath \"\"" Dec 03 14:13:42 crc kubenswrapper[4805]: I1203 14:13:42.320956 4805 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/1658d8e1-ac63-40fe-98ce-900a9f9aeae5-v4-0-config-system-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 03 14:13:42 crc kubenswrapper[4805]: I1203 14:13:42.320966 4805 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/1658d8e1-ac63-40fe-98ce-900a9f9aeae5-v4-0-config-user-template-login\") on node \"crc\" DevicePath \"\"" Dec 03 14:13:42 crc kubenswrapper[4805]: I1203 14:13:42.320976 4805 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/1658d8e1-ac63-40fe-98ce-900a9f9aeae5-v4-0-config-user-template-error\") on node \"crc\" DevicePath \"\"" Dec 03 14:13:42 crc kubenswrapper[4805]: I1203 14:13:42.320984 4805 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/1658d8e1-ac63-40fe-98ce-900a9f9aeae5-v4-0-config-user-template-provider-selection\") on node \"crc\" DevicePath \"\"" Dec 03 14:13:42 crc kubenswrapper[4805]: I1203 14:13:42.320994 4805 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/1658d8e1-ac63-40fe-98ce-900a9f9aeae5-v4-0-config-system-session\") on node \"crc\" DevicePath \"\"" Dec 03 14:13:42 crc kubenswrapper[4805]: I1203 14:13:42.321002 4805 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/1658d8e1-ac63-40fe-98ce-900a9f9aeae5-v4-0-config-system-ocp-branding-template\") on node \"crc\" DevicePath \"\"" Dec 03 14:13:42 crc kubenswrapper[4805]: I1203 14:13:42.321012 4805 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/1658d8e1-ac63-40fe-98ce-900a9f9aeae5-v4-0-config-system-router-certs\") on node \"crc\" DevicePath \"\"" Dec 03 14:13:42 crc kubenswrapper[4805]: I1203 14:13:42.321021 4805 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dxgkw\" (UniqueName: \"kubernetes.io/projected/1658d8e1-ac63-40fe-98ce-900a9f9aeae5-kube-api-access-dxgkw\") on node \"crc\" DevicePath \"\"" Dec 03 14:13:42 crc kubenswrapper[4805]: I1203 14:13:42.321029 4805 reconciler_common.go:293] "Volume detached for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/1658d8e1-ac63-40fe-98ce-900a9f9aeae5-audit-dir\") on node \"crc\" DevicePath \"\"" Dec 03 14:13:42 crc kubenswrapper[4805]: I1203 14:13:42.393910 4805 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 03 14:13:42 crc kubenswrapper[4805]: I1203 14:13:42.394344 4805 patch_prober.go:28] interesting pod/kube-controller-manager-crc container/kube-controller-manager namespace/openshift-kube-controller-manager: Startup probe status=failure output="Get \"https://192.168.126.11:10257/healthz\": dial tcp 192.168.126.11:10257: connect: connection refused" start-of-body= Dec 03 14:13:42 crc kubenswrapper[4805]: I1203 14:13:42.394400 4805 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podUID="f614b9022728cf315e60c057852e563e" containerName="kube-controller-manager" probeResult="failure" output="Get \"https://192.168.126.11:10257/healthz\": dial tcp 192.168.126.11:10257: connect: connection refused" Dec 03 14:13:43 crc kubenswrapper[4805]: I1203 14:13:43.274632 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"28f0de72749a64fcba1a111ba7101cd0d537bad1ee649cad96f3bbc46150aa62"} Dec 03 14:13:43 crc kubenswrapper[4805]: I1203 14:13:43.274910 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"c3bc47f6c73548f4f4eb0ccb2cfdf3e6aa201be57e7aea82e45be6e03f9e0652"} Dec 03 14:13:43 crc kubenswrapper[4805]: I1203 14:13:43.275077 4805 kubelet.go:1909] "Trying to delete pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="ced8291d-0ce5-4e74-9eaf-2bf033774039" Dec 03 14:13:43 crc kubenswrapper[4805]: I1203 14:13:43.275106 4805 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="ced8291d-0ce5-4e74-9eaf-2bf033774039" Dec 03 14:13:45 crc kubenswrapper[4805]: I1203 14:13:45.718752 4805 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 03 14:13:45 crc kubenswrapper[4805]: I1203 14:13:45.718855 4805 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 03 14:13:45 crc kubenswrapper[4805]: I1203 14:13:45.728053 4805 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 03 14:13:48 crc kubenswrapper[4805]: I1203 14:13:48.286392 4805 kubelet.go:1914] "Deleted mirror pod because it is outdated" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 03 14:13:48 crc kubenswrapper[4805]: I1203 14:13:48.305209 4805 kubelet.go:1909] "Trying to delete pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="ced8291d-0ce5-4e74-9eaf-2bf033774039" Dec 03 14:13:48 crc kubenswrapper[4805]: I1203 14:13:48.305237 4805 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="ced8291d-0ce5-4e74-9eaf-2bf033774039" Dec 03 14:13:48 crc kubenswrapper[4805]: I1203 14:13:48.305228 4805 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 03 14:13:48 crc kubenswrapper[4805]: I1203 14:13:48.310523 4805 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 03 14:13:49 crc kubenswrapper[4805]: I1203 14:13:49.308801 4805 kubelet.go:1909] "Trying to delete pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="ced8291d-0ce5-4e74-9eaf-2bf033774039" Dec 03 14:13:49 crc kubenswrapper[4805]: I1203 14:13:49.308828 4805 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="ced8291d-0ce5-4e74-9eaf-2bf033774039" Dec 03 14:13:50 crc kubenswrapper[4805]: I1203 14:13:50.052971 4805 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 03 14:13:50 crc kubenswrapper[4805]: I1203 14:13:50.314449 4805 kubelet.go:1909] "Trying to delete pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="ced8291d-0ce5-4e74-9eaf-2bf033774039" Dec 03 14:13:50 crc kubenswrapper[4805]: I1203 14:13:50.314484 4805 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="ced8291d-0ce5-4e74-9eaf-2bf033774039" Dec 03 14:13:50 crc kubenswrapper[4805]: I1203 14:13:50.709950 4805 status_manager.go:861] "Pod was deleted and then recreated, skipping status update" pod="openshift-kube-apiserver/kube-apiserver-crc" oldPodUID="71bb4a3aecc4ba5b26c4b7318770ce13" podUID="4f23ee18-1376-4681-9f28-cd3a94f2ccfc" Dec 03 14:13:52 crc kubenswrapper[4805]: I1203 14:13:52.392197 4805 patch_prober.go:28] interesting pod/kube-controller-manager-crc container/kube-controller-manager namespace/openshift-kube-controller-manager: Startup probe status=failure output="Get \"https://192.168.126.11:10257/healthz\": dial tcp 192.168.126.11:10257: connect: connection refused" start-of-body= Dec 03 14:13:52 crc kubenswrapper[4805]: I1203 14:13:52.392301 4805 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podUID="f614b9022728cf315e60c057852e563e" containerName="kube-controller-manager" probeResult="failure" output="Get \"https://192.168.126.11:10257/healthz\": dial tcp 192.168.126.11:10257: connect: connection refused" Dec 03 14:13:57 crc kubenswrapper[4805]: I1203 14:13:57.423979 4805 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-service-ca.crt" Dec 03 14:13:58 crc kubenswrapper[4805]: I1203 14:13:58.032083 4805 reflector.go:368] Caches populated for *v1.Secret from object-"hostpath-provisioner"/"csi-hostpath-provisioner-sa-dockercfg-qd74k" Dec 03 14:13:58 crc kubenswrapper[4805]: I1203 14:13:58.743383 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"marketplace-operator-dockercfg-5nsgg" Dec 03 14:13:58 crc kubenswrapper[4805]: I1203 14:13:58.791491 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"machine-api-operator-tls" Dec 03 14:13:58 crc kubenswrapper[4805]: I1203 14:13:58.833403 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"marketplace-operator-metrics" Dec 03 14:13:58 crc kubenswrapper[4805]: I1203 14:13:58.882598 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"cluster-image-registry-operator-dockercfg-m4qtx" Dec 03 14:13:58 crc kubenswrapper[4805]: I1203 14:13:58.976933 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"openshift-controller-manager-sa-dockercfg-msq4c" Dec 03 14:13:59 crc kubenswrapper[4805]: I1203 14:13:59.510821 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-config-operator"/"openshift-config-operator-dockercfg-7pc5z" Dec 03 14:13:59 crc kubenswrapper[4805]: I1203 14:13:59.539501 4805 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"kube-root-ca.crt" Dec 03 14:13:59 crc kubenswrapper[4805]: I1203 14:13:59.587322 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-stats-default" Dec 03 14:13:59 crc kubenswrapper[4805]: I1203 14:13:59.621268 4805 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-canary"/"openshift-service-ca.crt" Dec 03 14:13:59 crc kubenswrapper[4805]: I1203 14:13:59.631910 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"image-registry-operator-tls" Dec 03 14:13:59 crc kubenswrapper[4805]: I1203 14:13:59.771285 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator"/"kube-storage-version-migrator-sa-dockercfg-5xfcg" Dec 03 14:14:00 crc kubenswrapper[4805]: I1203 14:14:00.979481 4805 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"openshift-service-ca.crt" Dec 03 14:14:01 crc kubenswrapper[4805]: I1203 14:14:01.005066 4805 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"default-cni-sysctl-allowlist" Dec 03 14:14:01 crc kubenswrapper[4805]: I1203 14:14:01.109263 4805 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"openshift-service-ca.crt" Dec 03 14:14:01 crc kubenswrapper[4805]: I1203 14:14:01.311789 4805 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"openshift-service-ca.crt" Dec 03 14:14:01 crc kubenswrapper[4805]: I1203 14:14:01.481543 4805 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns-operator"/"kube-root-ca.crt" Dec 03 14:14:01 crc kubenswrapper[4805]: I1203 14:14:01.661891 4805 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"openshift-service-ca.crt" Dec 03 14:14:01 crc kubenswrapper[4805]: I1203 14:14:01.770706 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-serving-cert" Dec 03 14:14:01 crc kubenswrapper[4805]: I1203 14:14:01.987588 4805 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"audit-1" Dec 03 14:14:02 crc kubenswrapper[4805]: I1203 14:14:02.018772 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-operator"/"ingress-operator-dockercfg-7lnqk" Dec 03 14:14:02 crc kubenswrapper[4805]: I1203 14:14:02.019689 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-serving-cert" Dec 03 14:14:02 crc kubenswrapper[4805]: I1203 14:14:02.080327 4805 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"kube-root-ca.crt" Dec 03 14:14:02 crc kubenswrapper[4805]: I1203 14:14:02.173530 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-control-plane-metrics-cert" Dec 03 14:14:02 crc kubenswrapper[4805]: I1203 14:14:02.190760 4805 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-config" Dec 03 14:14:02 crc kubenswrapper[4805]: I1203 14:14:02.292494 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca"/"service-ca-dockercfg-pn86c" Dec 03 14:14:02 crc kubenswrapper[4805]: I1203 14:14:02.391356 4805 patch_prober.go:28] interesting pod/kube-controller-manager-crc container/kube-controller-manager namespace/openshift-kube-controller-manager: Startup probe status=failure output="Get \"https://192.168.126.11:10257/healthz\": dial tcp 192.168.126.11:10257: connect: connection refused" start-of-body= Dec 03 14:14:02 crc kubenswrapper[4805]: I1203 14:14:02.391698 4805 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podUID="f614b9022728cf315e60c057852e563e" containerName="kube-controller-manager" probeResult="failure" output="Get \"https://192.168.126.11:10257/healthz\": dial tcp 192.168.126.11:10257: connect: connection refused" Dec 03 14:14:02 crc kubenswrapper[4805]: I1203 14:14:02.391991 4805 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 03 14:14:02 crc kubenswrapper[4805]: I1203 14:14:02.392981 4805 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="kube-controller-manager" containerStatusID={"Type":"cri-o","ID":"604de07374b8d85ab0785f21f409ddd439bfadd9af01803476055e82f903166a"} pod="openshift-kube-controller-manager/kube-controller-manager-crc" containerMessage="Container kube-controller-manager failed startup probe, will be restarted" Dec 03 14:14:02 crc kubenswrapper[4805]: I1203 14:14:02.393407 4805 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podUID="f614b9022728cf315e60c057852e563e" containerName="kube-controller-manager" containerID="cri-o://604de07374b8d85ab0785f21f409ddd439bfadd9af01803476055e82f903166a" gracePeriod=30 Dec 03 14:14:02 crc kubenswrapper[4805]: I1203 14:14:02.417688 4805 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"kube-root-ca.crt" Dec 03 14:14:02 crc kubenswrapper[4805]: I1203 14:14:02.423459 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"olm-operator-serviceaccount-dockercfg-rq7zk" Dec 03 14:14:02 crc kubenswrapper[4805]: I1203 14:14:02.596536 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-dockercfg-f62pw" Dec 03 14:14:02 crc kubenswrapper[4805]: I1203 14:14:02.683430 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"olm-operator-serving-cert" Dec 03 14:14:02 crc kubenswrapper[4805]: I1203 14:14:02.685044 4805 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"kube-rbac-proxy" Dec 03 14:14:02 crc kubenswrapper[4805]: I1203 14:14:02.738372 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-client" Dec 03 14:14:02 crc kubenswrapper[4805]: I1203 14:14:02.741035 4805 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"image-registry-certificates" Dec 03 14:14:02 crc kubenswrapper[4805]: I1203 14:14:02.761248 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"route-controller-manager-sa-dockercfg-h2zr2" Dec 03 14:14:02 crc kubenswrapper[4805]: I1203 14:14:02.790985 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-marketplace-dockercfg-x2ctb" Dec 03 14:14:02 crc kubenswrapper[4805]: I1203 14:14:02.907356 4805 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"kube-root-ca.crt" Dec 03 14:14:02 crc kubenswrapper[4805]: I1203 14:14:02.936529 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-daemon-dockercfg-r5tcq" Dec 03 14:14:03 crc kubenswrapper[4805]: I1203 14:14:03.348724 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"serving-cert" Dec 03 14:14:03 crc kubenswrapper[4805]: I1203 14:14:03.360789 4805 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"kube-root-ca.crt" Dec 03 14:14:03 crc kubenswrapper[4805]: I1203 14:14:03.385688 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-serving-cert" Dec 03 14:14:03 crc kubenswrapper[4805]: I1203 14:14:03.498773 4805 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver-operator"/"kube-root-ca.crt" Dec 03 14:14:03 crc kubenswrapper[4805]: I1203 14:14:03.596537 4805 reflector.go:368] Caches populated for *v1.ConfigMap from object-"hostpath-provisioner"/"openshift-service-ca.crt" Dec 03 14:14:03 crc kubenswrapper[4805]: I1203 14:14:03.637146 4805 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"machine-approver-config" Dec 03 14:14:03 crc kubenswrapper[4805]: I1203 14:14:03.714296 4805 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-version"/"kube-root-ca.crt" Dec 03 14:14:03 crc kubenswrapper[4805]: I1203 14:14:03.800309 4805 reflector.go:368] Caches populated for *v1.Service from k8s.io/client-go/informers/factory.go:160 Dec 03 14:14:03 crc kubenswrapper[4805]: I1203 14:14:03.837443 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"control-plane-machine-set-operator-dockercfg-k9rxt" Dec 03 14:14:03 crc kubenswrapper[4805]: I1203 14:14:03.890716 4805 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"openshift-service-ca.crt" Dec 03 14:14:03 crc kubenswrapper[4805]: I1203 14:14:03.954500 4805 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"config" Dec 03 14:14:04 crc kubenswrapper[4805]: I1203 14:14:04.068517 4805 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"audit-1" Dec 03 14:14:04 crc kubenswrapper[4805]: I1203 14:14:04.114237 4805 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"kube-root-ca.crt" Dec 03 14:14:04 crc kubenswrapper[4805]: I1203 14:14:04.210538 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"dns-dockercfg-jwfmh" Dec 03 14:14:04 crc kubenswrapper[4805]: I1203 14:14:04.231291 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"node-bootstrapper-token" Dec 03 14:14:04 crc kubenswrapper[4805]: I1203 14:14:04.263773 4805 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"authentication-operator-config" Dec 03 14:14:04 crc kubenswrapper[4805]: I1203 14:14:04.297155 4805 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"kube-root-ca.crt" Dec 03 14:14:04 crc kubenswrapper[4805]: I1203 14:14:04.324106 4805 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"openshift-service-ca.crt" Dec 03 14:14:04 crc kubenswrapper[4805]: I1203 14:14:04.327426 4805 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"image-import-ca" Dec 03 14:14:04 crc kubenswrapper[4805]: I1203 14:14:04.328903 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"etcd-client" Dec 03 14:14:04 crc kubenswrapper[4805]: I1203 14:14:04.363556 4805 reflector.go:368] Caches populated for *v1.CSIDriver from k8s.io/client-go/informers/factory.go:160 Dec 03 14:14:04 crc kubenswrapper[4805]: I1203 14:14:04.399173 4805 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"service-ca" Dec 03 14:14:04 crc kubenswrapper[4805]: I1203 14:14:04.462248 4805 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"ovnkube-config" Dec 03 14:14:04 crc kubenswrapper[4805]: I1203 14:14:04.464769 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator-operator"/"kube-storage-version-migrator-operator-dockercfg-2bh8d" Dec 03 14:14:04 crc kubenswrapper[4805]: I1203 14:14:04.507037 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication-operator"/"authentication-operator-dockercfg-mz9bj" Dec 03 14:14:04 crc kubenswrapper[4805]: I1203 14:14:04.697140 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-dockercfg-gkqpw" Dec 03 14:14:04 crc kubenswrapper[4805]: I1203 14:14:04.729448 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"node-ca-dockercfg-4777p" Dec 03 14:14:04 crc kubenswrapper[4805]: I1203 14:14:04.961616 4805 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"openshift-service-ca.crt" Dec 03 14:14:05 crc kubenswrapper[4805]: I1203 14:14:05.033295 4805 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"openshift-service-ca.crt" Dec 03 14:14:05 crc kubenswrapper[4805]: I1203 14:14:05.064558 4805 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"kube-root-ca.crt" Dec 03 14:14:05 crc kubenswrapper[4805]: I1203 14:14:05.075349 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"mco-proxy-tls" Dec 03 14:14:05 crc kubenswrapper[4805]: I1203 14:14:05.086909 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator-operator"/"serving-cert" Dec 03 14:14:05 crc kubenswrapper[4805]: I1203 14:14:05.116272 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-dockercfg-x57mr" Dec 03 14:14:05 crc kubenswrapper[4805]: I1203 14:14:05.123341 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-scheduler-operator"/"kube-scheduler-operator-serving-cert" Dec 03 14:14:05 crc kubenswrapper[4805]: I1203 14:14:05.145209 4805 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"console-operator-config" Dec 03 14:14:05 crc kubenswrapper[4805]: I1203 14:14:05.153679 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-ancillary-tools-dockercfg-vnmsz" Dec 03 14:14:05 crc kubenswrapper[4805]: I1203 14:14:05.177324 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication-operator"/"serving-cert" Dec 03 14:14:05 crc kubenswrapper[4805]: I1203 14:14:05.200202 4805 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"kube-root-ca.crt" Dec 03 14:14:05 crc kubenswrapper[4805]: I1203 14:14:05.325163 4805 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"iptables-alerter-script" Dec 03 14:14:05 crc kubenswrapper[4805]: I1203 14:14:05.339829 4805 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"openshift-service-ca.crt" Dec 03 14:14:05 crc kubenswrapper[4805]: I1203 14:14:05.355502 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-node-metrics-cert" Dec 03 14:14:05 crc kubenswrapper[4805]: I1203 14:14:05.374771 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console-operator"/"serving-cert" Dec 03 14:14:05 crc kubenswrapper[4805]: I1203 14:14:05.477734 4805 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"openshift-service-ca.crt" Dec 03 14:14:05 crc kubenswrapper[4805]: I1203 14:14:05.519558 4805 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"service-ca-bundle" Dec 03 14:14:05 crc kubenswrapper[4805]: I1203 14:14:05.640443 4805 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-scheduler-operator"/"kube-root-ca.crt" Dec 03 14:14:05 crc kubenswrapper[4805]: I1203 14:14:05.664546 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-operator-dockercfg-98p87" Dec 03 14:14:05 crc kubenswrapper[4805]: I1203 14:14:05.683277 4805 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"kube-root-ca.crt" Dec 03 14:14:05 crc kubenswrapper[4805]: I1203 14:14:05.766563 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-canary"/"default-dockercfg-2llfx" Dec 03 14:14:05 crc kubenswrapper[4805]: I1203 14:14:05.773428 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-machine-approver"/"machine-approver-sa-dockercfg-nl2j4" Dec 03 14:14:05 crc kubenswrapper[4805]: I1203 14:14:05.829897 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns-operator"/"metrics-tls" Dec 03 14:14:05 crc kubenswrapper[4805]: I1203 14:14:05.832623 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-server-dockercfg-qx5rd" Dec 03 14:14:05 crc kubenswrapper[4805]: I1203 14:14:05.888118 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"oauth-apiserver-sa-dockercfg-6r2bq" Dec 03 14:14:05 crc kubenswrapper[4805]: I1203 14:14:05.972913 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-kubernetes-control-plane-dockercfg-gs7dd" Dec 03 14:14:06 crc kubenswrapper[4805]: I1203 14:14:06.063297 4805 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-scheduler-operator"/"openshift-kube-scheduler-operator-config" Dec 03 14:14:06 crc kubenswrapper[4805]: I1203 14:14:06.115130 4805 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"config" Dec 03 14:14:06 crc kubenswrapper[4805]: I1203 14:14:06.124657 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-certs-default" Dec 03 14:14:06 crc kubenswrapper[4805]: I1203 14:14:06.149757 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-canary"/"canary-serving-cert" Dec 03 14:14:06 crc kubenswrapper[4805]: I1203 14:14:06.201131 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"packageserver-service-cert" Dec 03 14:14:06 crc kubenswrapper[4805]: I1203 14:14:06.269802 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"node-resolver-dockercfg-kz9s7" Dec 03 14:14:06 crc kubenswrapper[4805]: I1203 14:14:06.327546 4805 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"openshift-service-ca.crt" Dec 03 14:14:06 crc kubenswrapper[4805]: I1203 14:14:06.328671 4805 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-global-ca" Dec 03 14:14:06 crc kubenswrapper[4805]: I1203 14:14:06.347660 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"community-operators-dockercfg-dmngl" Dec 03 14:14:06 crc kubenswrapper[4805]: I1203 14:14:06.372677 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"serving-cert" Dec 03 14:14:06 crc kubenswrapper[4805]: I1203 14:14:06.525300 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"serving-cert" Dec 03 14:14:06 crc kubenswrapper[4805]: I1203 14:14:06.564137 4805 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-config" Dec 03 14:14:06 crc kubenswrapper[4805]: I1203 14:14:06.616739 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"installation-pull-secrets" Dec 03 14:14:06 crc kubenswrapper[4805]: I1203 14:14:06.653225 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-console"/"networking-console-plugin-cert" Dec 03 14:14:06 crc kubenswrapper[4805]: I1203 14:14:06.721725 4805 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"openshift-service-ca.crt" Dec 03 14:14:06 crc kubenswrapper[4805]: I1203 14:14:06.768283 4805 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"kube-root-ca.crt" Dec 03 14:14:07 crc kubenswrapper[4805]: I1203 14:14:07.034719 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-node-identity"/"network-node-identity-cert" Dec 03 14:14:07 crc kubenswrapper[4805]: I1203 14:14:07.092253 4805 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-samples-operator"/"kube-root-ca.crt" Dec 03 14:14:07 crc kubenswrapper[4805]: I1203 14:14:07.226029 4805 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"openshift-service-ca.crt" Dec 03 14:14:07 crc kubenswrapper[4805]: I1203 14:14:07.235110 4805 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"openshift-service-ca.crt" Dec 03 14:14:07 crc kubenswrapper[4805]: I1203 14:14:07.257304 4805 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"config" Dec 03 14:14:07 crc kubenswrapper[4805]: I1203 14:14:07.259391 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-controller-dockercfg-c2lfx" Dec 03 14:14:07 crc kubenswrapper[4805]: I1203 14:14:07.286650 4805 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator"/"kube-root-ca.crt" Dec 03 14:14:07 crc kubenswrapper[4805]: I1203 14:14:07.330858 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"default-dockercfg-2q5b6" Dec 03 14:14:07 crc kubenswrapper[4805]: I1203 14:14:07.416543 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"machine-api-operator-dockercfg-mfbb7" Dec 03 14:14:07 crc kubenswrapper[4805]: I1203 14:14:07.476756 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-dockercfg-zdk86" Dec 03 14:14:07 crc kubenswrapper[4805]: I1203 14:14:07.531592 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"pprof-cert" Dec 03 14:14:07 crc kubenswrapper[4805]: I1203 14:14:07.665873 4805 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"service-ca-bundle" Dec 03 14:14:07 crc kubenswrapper[4805]: I1203 14:14:07.677442 4805 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"kube-rbac-proxy" Dec 03 14:14:07 crc kubenswrapper[4805]: I1203 14:14:07.681643 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-serving-cert" Dec 03 14:14:07 crc kubenswrapper[4805]: I1203 14:14:07.682876 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"dns-default-metrics-tls" Dec 03 14:14:07 crc kubenswrapper[4805]: I1203 14:14:07.769102 4805 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"kube-root-ca.crt" Dec 03 14:14:07 crc kubenswrapper[4805]: I1203 14:14:07.823204 4805 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"kube-root-ca.crt" Dec 03 14:14:07 crc kubenswrapper[4805]: I1203 14:14:07.949996 4805 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"openshift-service-ca.crt" Dec 03 14:14:07 crc kubenswrapper[4805]: I1203 14:14:07.973657 4805 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"env-overrides" Dec 03 14:14:07 crc kubenswrapper[4805]: I1203 14:14:07.980686 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-operator-serving-cert" Dec 03 14:14:07 crc kubenswrapper[4805]: I1203 14:14:07.982158 4805 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"openshift-service-ca.crt" Dec 03 14:14:08 crc kubenswrapper[4805]: I1203 14:14:08.100265 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"etcd-client" Dec 03 14:14:08 crc kubenswrapper[4805]: I1203 14:14:08.127146 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-config-operator"/"config-operator-serving-cert" Dec 03 14:14:08 crc kubenswrapper[4805]: I1203 14:14:08.127226 4805 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"ovnkube-identity-cm" Dec 03 14:14:08 crc kubenswrapper[4805]: I1203 14:14:08.275597 4805 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"cni-copy-resources" Dec 03 14:14:08 crc kubenswrapper[4805]: I1203 14:14:08.289710 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"metrics-daemon-sa-dockercfg-d427c" Dec 03 14:14:08 crc kubenswrapper[4805]: I1203 14:14:08.292686 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-dockercfg-xtcjv" Dec 03 14:14:08 crc kubenswrapper[4805]: I1203 14:14:08.378247 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-metrics-certs-default" Dec 03 14:14:08 crc kubenswrapper[4805]: I1203 14:14:08.408594 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"image-registry-tls" Dec 03 14:14:08 crc kubenswrapper[4805]: I1203 14:14:08.429679 4805 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"trusted-ca-bundle" Dec 03 14:14:08 crc kubenswrapper[4805]: I1203 14:14:08.544004 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-operator"/"metrics-tls" Dec 03 14:14:08 crc kubenswrapper[4805]: I1203 14:14:08.566388 4805 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"client-ca" Dec 03 14:14:08 crc kubenswrapper[4805]: I1203 14:14:08.580093 4805 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-canary"/"kube-root-ca.crt" Dec 03 14:14:08 crc kubenswrapper[4805]: I1203 14:14:08.659371 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"default-dockercfg-chnjx" Dec 03 14:14:08 crc kubenswrapper[4805]: I1203 14:14:08.779646 4805 reflector.go:368] Caches populated for *v1.Node from k8s.io/client-go/informers/factory.go:160 Dec 03 14:14:08 crc kubenswrapper[4805]: I1203 14:14:08.853208 4805 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-operator-config" Dec 03 14:14:08 crc kubenswrapper[4805]: I1203 14:14:08.876163 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-samples-operator"/"cluster-samples-operator-dockercfg-xpp9w" Dec 03 14:14:08 crc kubenswrapper[4805]: I1203 14:14:08.885279 4805 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"openshift-service-ca.crt" Dec 03 14:14:08 crc kubenswrapper[4805]: I1203 14:14:08.928093 4805 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"etcd-serving-ca" Dec 03 14:14:08 crc kubenswrapper[4805]: I1203 14:14:08.980659 4805 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"kube-root-ca.crt" Dec 03 14:14:08 crc kubenswrapper[4805]: I1203 14:14:08.991204 4805 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"kube-root-ca.crt" Dec 03 14:14:09 crc kubenswrapper[4805]: I1203 14:14:09.096818 4805 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"service-ca-operator-config" Dec 03 14:14:09 crc kubenswrapper[4805]: I1203 14:14:09.122719 4805 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"signing-cabundle" Dec 03 14:14:09 crc kubenswrapper[4805]: I1203 14:14:09.155367 4805 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-config" Dec 03 14:14:09 crc kubenswrapper[4805]: I1203 14:14:09.198898 4805 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"multus-daemon-config" Dec 03 14:14:09 crc kubenswrapper[4805]: I1203 14:14:09.237158 4805 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-console"/"networking-console-plugin" Dec 03 14:14:09 crc kubenswrapper[4805]: I1203 14:14:09.310167 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"catalog-operator-serving-cert" Dec 03 14:14:09 crc kubenswrapper[4805]: I1203 14:14:09.383386 4805 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"kube-root-ca.crt" Dec 03 14:14:09 crc kubenswrapper[4805]: I1203 14:14:09.476791 4805 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"env-overrides" Dec 03 14:14:09 crc kubenswrapper[4805]: I1203 14:14:09.510959 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"registry-dockercfg-kzzsd" Dec 03 14:14:09 crc kubenswrapper[4805]: I1203 14:14:09.514751 4805 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-version"/"openshift-service-ca.crt" Dec 03 14:14:09 crc kubenswrapper[4805]: I1203 14:14:09.757661 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-kubernetes-node-dockercfg-pwtwl" Dec 03 14:14:09 crc kubenswrapper[4805]: I1203 14:14:09.773084 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-server-tls" Dec 03 14:14:09 crc kubenswrapper[4805]: I1203 14:14:09.802658 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-operator-dockercfg-r9srn" Dec 03 14:14:09 crc kubenswrapper[4805]: I1203 14:14:09.815972 4805 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"client-ca" Dec 03 14:14:09 crc kubenswrapper[4805]: I1203 14:14:09.854259 4805 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"ovnkube-script-lib" Dec 03 14:14:09 crc kubenswrapper[4805]: I1203 14:14:09.893452 4805 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-diagnostics"/"kube-root-ca.crt" Dec 03 14:14:09 crc kubenswrapper[4805]: I1203 14:14:09.974621 4805 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"openshift-service-ca.crt" Dec 03 14:14:10 crc kubenswrapper[4805]: I1203 14:14:10.034762 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-samples-operator"/"samples-operator-tls" Dec 03 14:14:10 crc kubenswrapper[4805]: I1203 14:14:10.047357 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-version"/"default-dockercfg-gxtc4" Dec 03 14:14:10 crc kubenswrapper[4805]: I1203 14:14:10.427545 4805 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"kube-root-ca.crt" Dec 03 14:14:10 crc kubenswrapper[4805]: I1203 14:14:10.431117 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-oauth-config" Dec 03 14:14:10 crc kubenswrapper[4805]: I1203 14:14:10.519775 4805 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"openshift-service-ca.crt" Dec 03 14:14:10 crc kubenswrapper[4805]: I1203 14:14:10.535626 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"encryption-config-1" Dec 03 14:14:10 crc kubenswrapper[4805]: I1203 14:14:10.577069 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-machine-approver"/"machine-approver-tls" Dec 03 14:14:10 crc kubenswrapper[4805]: I1203 14:14:10.588257 4805 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"openshift-service-ca.crt" Dec 03 14:14:10 crc kubenswrapper[4805]: I1203 14:14:10.594857 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"mcc-proxy-tls" Dec 03 14:14:10 crc kubenswrapper[4805]: I1203 14:14:10.635831 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-operator"/"metrics-tls" Dec 03 14:14:10 crc kubenswrapper[4805]: I1203 14:14:10.702298 4805 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"machine-api-operator-images" Dec 03 14:14:10 crc kubenswrapper[4805]: I1203 14:14:10.722711 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca-operator"/"service-ca-operator-dockercfg-rg9jl" Dec 03 14:14:10 crc kubenswrapper[4805]: I1203 14:14:10.831626 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns-operator"/"dns-operator-dockercfg-9mqw5" Dec 03 14:14:11 crc kubenswrapper[4805]: I1203 14:14:11.129258 4805 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-samples-operator"/"openshift-service-ca.crt" Dec 03 14:14:11 crc kubenswrapper[4805]: I1203 14:14:11.208446 4805 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns-operator"/"openshift-service-ca.crt" Dec 03 14:14:11 crc kubenswrapper[4805]: I1203 14:14:11.241650 4805 reflector.go:368] Caches populated for *v1.ConfigMap from object-"hostpath-provisioner"/"kube-root-ca.crt" Dec 03 14:14:11 crc kubenswrapper[4805]: I1203 14:14:11.382325 4805 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"kube-root-ca.crt" Dec 03 14:14:11 crc kubenswrapper[4805]: I1203 14:14:11.436636 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-ac-dockercfg-9lkdf" Dec 03 14:14:11 crc kubenswrapper[4805]: I1203 14:14:11.512694 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-admission-controller-secret" Dec 03 14:14:11 crc kubenswrapper[4805]: I1203 14:14:11.534021 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca-operator"/"serving-cert" Dec 03 14:14:11 crc kubenswrapper[4805]: I1203 14:14:11.549463 4805 reflector.go:368] Caches populated for *v1.Pod from pkg/kubelet/config/apiserver.go:66 Dec 03 14:14:11 crc kubenswrapper[4805]: I1203 14:14:11.555032 4805 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-ntv6j","openshift-kube-apiserver/kube-apiserver-crc"] Dec 03 14:14:11 crc kubenswrapper[4805]: I1203 14:14:11.555107 4805 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/kube-apiserver-crc"] Dec 03 14:14:11 crc kubenswrapper[4805]: I1203 14:14:11.561764 4805 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 03 14:14:11 crc kubenswrapper[4805]: I1203 14:14:11.572564 4805 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver/kube-apiserver-crc" podStartSLOduration=23.572549529 podStartE2EDuration="23.572549529s" podCreationTimestamp="2025-12-03 14:13:48 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 14:14:11.570327819 +0000 UTC m=+281.233244762" watchObservedRunningTime="2025-12-03 14:14:11.572549529 +0000 UTC m=+281.235466452" Dec 03 14:14:11 crc kubenswrapper[4805]: I1203 14:14:11.615388 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-serving-cert" Dec 03 14:14:11 crc kubenswrapper[4805]: I1203 14:14:11.713805 4805 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-diagnostics"/"openshift-service-ca.crt" Dec 03 14:14:11 crc kubenswrapper[4805]: I1203 14:14:11.828464 4805 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-config-operator"/"openshift-service-ca.crt" Dec 03 14:14:11 crc kubenswrapper[4805]: I1203 14:14:11.874335 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"metrics-daemon-secret" Dec 03 14:14:11 crc kubenswrapper[4805]: I1203 14:14:11.876553 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca"/"signing-key" Dec 03 14:14:11 crc kubenswrapper[4805]: I1203 14:14:11.927371 4805 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-config-operator"/"kube-root-ca.crt" Dec 03 14:14:11 crc kubenswrapper[4805]: I1203 14:14:11.934530 4805 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"kube-root-ca.crt" Dec 03 14:14:11 crc kubenswrapper[4805]: I1203 14:14:11.957888 4805 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"kube-root-ca.crt" Dec 03 14:14:11 crc kubenswrapper[4805]: I1203 14:14:11.993998 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"openshift-apiserver-sa-dockercfg-djjff" Dec 03 14:14:12 crc kubenswrapper[4805]: I1203 14:14:12.007068 4805 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"machine-config-operator-images" Dec 03 14:14:12 crc kubenswrapper[4805]: I1203 14:14:12.019804 4805 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"trusted-ca-bundle" Dec 03 14:14:12 crc kubenswrapper[4805]: I1203 14:14:12.145354 4805 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"openshift-service-ca.crt" Dec 03 14:14:12 crc kubenswrapper[4805]: I1203 14:14:12.160670 4805 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"kube-root-ca.crt" Dec 03 14:14:12 crc kubenswrapper[4805]: I1203 14:14:12.177249 4805 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"dns-default" Dec 03 14:14:12 crc kubenswrapper[4805]: I1203 14:14:12.184507 4805 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator"/"openshift-service-ca.crt" Dec 03 14:14:12 crc kubenswrapper[4805]: I1203 14:14:12.239734 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-dockercfg-vw8fw" Dec 03 14:14:12 crc kubenswrapper[4805]: I1203 14:14:12.259240 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"control-plane-machine-set-operator-tls" Dec 03 14:14:12 crc kubenswrapper[4805]: I1203 14:14:12.274995 4805 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"trusted-ca" Dec 03 14:14:12 crc kubenswrapper[4805]: I1203 14:14:12.411461 4805 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"openshift-service-ca.crt" Dec 03 14:14:12 crc kubenswrapper[4805]: I1203 14:14:12.439325 4805 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"kube-rbac-proxy" Dec 03 14:14:12 crc kubenswrapper[4805]: I1203 14:14:12.442490 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"certified-operators-dockercfg-4rs5g" Dec 03 14:14:12 crc kubenswrapper[4805]: I1203 14:14:12.457040 4805 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-config" Dec 03 14:14:12 crc kubenswrapper[4805]: I1203 14:14:12.563173 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"proxy-tls" Dec 03 14:14:12 crc kubenswrapper[4805]: I1203 14:14:12.655017 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"encryption-config-1" Dec 03 14:14:12 crc kubenswrapper[4805]: I1203 14:14:12.676396 4805 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"kube-root-ca.crt" Dec 03 14:14:12 crc kubenswrapper[4805]: I1203 14:14:12.707428 4805 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1658d8e1-ac63-40fe-98ce-900a9f9aeae5" path="/var/lib/kubelet/pods/1658d8e1-ac63-40fe-98ce-900a9f9aeae5/volumes" Dec 03 14:14:12 crc kubenswrapper[4805]: I1203 14:14:12.729075 4805 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"marketplace-trusted-ca" Dec 03 14:14:12 crc kubenswrapper[4805]: I1203 14:14:12.837867 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-version"/"cluster-version-operator-serving-cert" Dec 03 14:14:12 crc kubenswrapper[4805]: I1203 14:14:12.912401 4805 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-ca-bundle" Dec 03 14:14:12 crc kubenswrapper[4805]: I1203 14:14:12.973885 4805 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"openshift-service-ca.crt" Dec 03 14:14:13 crc kubenswrapper[4805]: I1203 14:14:13.130902 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"package-server-manager-serving-cert" Dec 03 14:14:13 crc kubenswrapper[4805]: I1203 14:14:13.146666 4805 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"kube-root-ca.crt" Dec 03 14:14:13 crc kubenswrapper[4805]: I1203 14:14:13.240193 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-operators-dockercfg-ct8rh" Dec 03 14:14:13 crc kubenswrapper[4805]: I1203 14:14:13.316760 4805 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"kube-root-ca.crt" Dec 03 14:14:13 crc kubenswrapper[4805]: I1203 14:14:13.618044 4805 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"openshift-service-ca.crt" Dec 03 14:14:13 crc kubenswrapper[4805]: I1203 14:14:13.849744 4805 reflector.go:368] Caches populated for *v1.RuntimeClass from k8s.io/client-go/informers/factory.go:160 Dec 03 14:14:13 crc kubenswrapper[4805]: I1203 14:14:13.950477 4805 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-authentication/oauth-openshift-9565f95f5-7jbmm"] Dec 03 14:14:13 crc kubenswrapper[4805]: E1203 14:14:13.950798 4805 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1658d8e1-ac63-40fe-98ce-900a9f9aeae5" containerName="oauth-openshift" Dec 03 14:14:13 crc kubenswrapper[4805]: I1203 14:14:13.950818 4805 state_mem.go:107] "Deleted CPUSet assignment" podUID="1658d8e1-ac63-40fe-98ce-900a9f9aeae5" containerName="oauth-openshift" Dec 03 14:14:13 crc kubenswrapper[4805]: E1203 14:14:13.950899 4805 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="62ba53fc-fcf4-47be-a415-ea016529bac1" containerName="installer" Dec 03 14:14:13 crc kubenswrapper[4805]: I1203 14:14:13.950912 4805 state_mem.go:107] "Deleted CPUSet assignment" podUID="62ba53fc-fcf4-47be-a415-ea016529bac1" containerName="installer" Dec 03 14:14:13 crc kubenswrapper[4805]: I1203 14:14:13.951103 4805 memory_manager.go:354] "RemoveStaleState removing state" podUID="62ba53fc-fcf4-47be-a415-ea016529bac1" containerName="installer" Dec 03 14:14:13 crc kubenswrapper[4805]: I1203 14:14:13.951141 4805 memory_manager.go:354] "RemoveStaleState removing state" podUID="1658d8e1-ac63-40fe-98ce-900a9f9aeae5" containerName="oauth-openshift" Dec 03 14:14:13 crc kubenswrapper[4805]: I1203 14:14:13.952027 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-9565f95f5-7jbmm" Dec 03 14:14:13 crc kubenswrapper[4805]: I1203 14:14:13.956959 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-serving-cert" Dec 03 14:14:13 crc kubenswrapper[4805]: I1203 14:14:13.957277 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-router-certs" Dec 03 14:14:13 crc kubenswrapper[4805]: I1203 14:14:13.958338 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-error" Dec 03 14:14:13 crc kubenswrapper[4805]: I1203 14:14:13.958349 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-provider-selection" Dec 03 14:14:13 crc kubenswrapper[4805]: I1203 14:14:13.958379 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-session" Dec 03 14:14:13 crc kubenswrapper[4805]: I1203 14:14:13.959378 4805 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"kube-root-ca.crt" Dec 03 14:14:13 crc kubenswrapper[4805]: I1203 14:14:13.959672 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"oauth-openshift-dockercfg-znhcc" Dec 03 14:14:13 crc kubenswrapper[4805]: I1203 14:14:13.959707 4805 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"audit" Dec 03 14:14:13 crc kubenswrapper[4805]: I1203 14:14:13.963059 4805 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-service-ca" Dec 03 14:14:13 crc kubenswrapper[4805]: I1203 14:14:13.963270 4805 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-cliconfig" Dec 03 14:14:13 crc kubenswrapper[4805]: I1203 14:14:13.967044 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-idp-0-file-data" Dec 03 14:14:13 crc kubenswrapper[4805]: I1203 14:14:13.968757 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-9565f95f5-7jbmm"] Dec 03 14:14:13 crc kubenswrapper[4805]: I1203 14:14:13.972710 4805 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"openshift-service-ca.crt" Dec 03 14:14:13 crc kubenswrapper[4805]: I1203 14:14:13.973215 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-login" Dec 03 14:14:13 crc kubenswrapper[4805]: I1203 14:14:13.978099 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-scheduler-operator"/"openshift-kube-scheduler-operator-dockercfg-qt55r" Dec 03 14:14:13 crc kubenswrapper[4805]: I1203 14:14:13.981674 4805 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-trusted-ca-bundle" Dec 03 14:14:13 crc kubenswrapper[4805]: I1203 14:14:13.990986 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-ocp-branding-template" Dec 03 14:14:13 crc kubenswrapper[4805]: I1203 14:14:13.999920 4805 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"kube-root-ca.crt" Dec 03 14:14:14 crc kubenswrapper[4805]: I1203 14:14:14.128297 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/58a56dbb-271d-49d7-bce9-21a6aef96265-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-9565f95f5-7jbmm\" (UID: \"58a56dbb-271d-49d7-bce9-21a6aef96265\") " pod="openshift-authentication/oauth-openshift-9565f95f5-7jbmm" Dec 03 14:14:14 crc kubenswrapper[4805]: I1203 14:14:14.128386 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/58a56dbb-271d-49d7-bce9-21a6aef96265-audit-policies\") pod \"oauth-openshift-9565f95f5-7jbmm\" (UID: \"58a56dbb-271d-49d7-bce9-21a6aef96265\") " pod="openshift-authentication/oauth-openshift-9565f95f5-7jbmm" Dec 03 14:14:14 crc kubenswrapper[4805]: I1203 14:14:14.128433 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/58a56dbb-271d-49d7-bce9-21a6aef96265-v4-0-config-system-serving-cert\") pod \"oauth-openshift-9565f95f5-7jbmm\" (UID: \"58a56dbb-271d-49d7-bce9-21a6aef96265\") " pod="openshift-authentication/oauth-openshift-9565f95f5-7jbmm" Dec 03 14:14:14 crc kubenswrapper[4805]: I1203 14:14:14.128480 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/58a56dbb-271d-49d7-bce9-21a6aef96265-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-9565f95f5-7jbmm\" (UID: \"58a56dbb-271d-49d7-bce9-21a6aef96265\") " pod="openshift-authentication/oauth-openshift-9565f95f5-7jbmm" Dec 03 14:14:14 crc kubenswrapper[4805]: I1203 14:14:14.128552 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/58a56dbb-271d-49d7-bce9-21a6aef96265-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-9565f95f5-7jbmm\" (UID: \"58a56dbb-271d-49d7-bce9-21a6aef96265\") " pod="openshift-authentication/oauth-openshift-9565f95f5-7jbmm" Dec 03 14:14:14 crc kubenswrapper[4805]: I1203 14:14:14.128718 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/58a56dbb-271d-49d7-bce9-21a6aef96265-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-9565f95f5-7jbmm\" (UID: \"58a56dbb-271d-49d7-bce9-21a6aef96265\") " pod="openshift-authentication/oauth-openshift-9565f95f5-7jbmm" Dec 03 14:14:14 crc kubenswrapper[4805]: I1203 14:14:14.128827 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/58a56dbb-271d-49d7-bce9-21a6aef96265-v4-0-config-system-session\") pod \"oauth-openshift-9565f95f5-7jbmm\" (UID: \"58a56dbb-271d-49d7-bce9-21a6aef96265\") " pod="openshift-authentication/oauth-openshift-9565f95f5-7jbmm" Dec 03 14:14:14 crc kubenswrapper[4805]: I1203 14:14:14.128902 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/58a56dbb-271d-49d7-bce9-21a6aef96265-v4-0-config-user-template-error\") pod \"oauth-openshift-9565f95f5-7jbmm\" (UID: \"58a56dbb-271d-49d7-bce9-21a6aef96265\") " pod="openshift-authentication/oauth-openshift-9565f95f5-7jbmm" Dec 03 14:14:14 crc kubenswrapper[4805]: I1203 14:14:14.128988 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/58a56dbb-271d-49d7-bce9-21a6aef96265-v4-0-config-system-service-ca\") pod \"oauth-openshift-9565f95f5-7jbmm\" (UID: \"58a56dbb-271d-49d7-bce9-21a6aef96265\") " pod="openshift-authentication/oauth-openshift-9565f95f5-7jbmm" Dec 03 14:14:14 crc kubenswrapper[4805]: I1203 14:14:14.129101 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/58a56dbb-271d-49d7-bce9-21a6aef96265-audit-dir\") pod \"oauth-openshift-9565f95f5-7jbmm\" (UID: \"58a56dbb-271d-49d7-bce9-21a6aef96265\") " pod="openshift-authentication/oauth-openshift-9565f95f5-7jbmm" Dec 03 14:14:14 crc kubenswrapper[4805]: I1203 14:14:14.129169 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/58a56dbb-271d-49d7-bce9-21a6aef96265-v4-0-config-system-router-certs\") pod \"oauth-openshift-9565f95f5-7jbmm\" (UID: \"58a56dbb-271d-49d7-bce9-21a6aef96265\") " pod="openshift-authentication/oauth-openshift-9565f95f5-7jbmm" Dec 03 14:14:14 crc kubenswrapper[4805]: I1203 14:14:14.129204 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qpt9w\" (UniqueName: \"kubernetes.io/projected/58a56dbb-271d-49d7-bce9-21a6aef96265-kube-api-access-qpt9w\") pod \"oauth-openshift-9565f95f5-7jbmm\" (UID: \"58a56dbb-271d-49d7-bce9-21a6aef96265\") " pod="openshift-authentication/oauth-openshift-9565f95f5-7jbmm" Dec 03 14:14:14 crc kubenswrapper[4805]: I1203 14:14:14.129253 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/58a56dbb-271d-49d7-bce9-21a6aef96265-v4-0-config-user-template-login\") pod \"oauth-openshift-9565f95f5-7jbmm\" (UID: \"58a56dbb-271d-49d7-bce9-21a6aef96265\") " pod="openshift-authentication/oauth-openshift-9565f95f5-7jbmm" Dec 03 14:14:14 crc kubenswrapper[4805]: I1203 14:14:14.129328 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/58a56dbb-271d-49d7-bce9-21a6aef96265-v4-0-config-system-cliconfig\") pod \"oauth-openshift-9565f95f5-7jbmm\" (UID: \"58a56dbb-271d-49d7-bce9-21a6aef96265\") " pod="openshift-authentication/oauth-openshift-9565f95f5-7jbmm" Dec 03 14:14:14 crc kubenswrapper[4805]: I1203 14:14:14.146333 4805 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-service-ca-bundle" Dec 03 14:14:14 crc kubenswrapper[4805]: I1203 14:14:14.231348 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/58a56dbb-271d-49d7-bce9-21a6aef96265-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-9565f95f5-7jbmm\" (UID: \"58a56dbb-271d-49d7-bce9-21a6aef96265\") " pod="openshift-authentication/oauth-openshift-9565f95f5-7jbmm" Dec 03 14:14:14 crc kubenswrapper[4805]: I1203 14:14:14.231435 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/58a56dbb-271d-49d7-bce9-21a6aef96265-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-9565f95f5-7jbmm\" (UID: \"58a56dbb-271d-49d7-bce9-21a6aef96265\") " pod="openshift-authentication/oauth-openshift-9565f95f5-7jbmm" Dec 03 14:14:14 crc kubenswrapper[4805]: I1203 14:14:14.231483 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/58a56dbb-271d-49d7-bce9-21a6aef96265-v4-0-config-system-session\") pod \"oauth-openshift-9565f95f5-7jbmm\" (UID: \"58a56dbb-271d-49d7-bce9-21a6aef96265\") " pod="openshift-authentication/oauth-openshift-9565f95f5-7jbmm" Dec 03 14:14:14 crc kubenswrapper[4805]: I1203 14:14:14.231519 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/58a56dbb-271d-49d7-bce9-21a6aef96265-v4-0-config-user-template-error\") pod \"oauth-openshift-9565f95f5-7jbmm\" (UID: \"58a56dbb-271d-49d7-bce9-21a6aef96265\") " pod="openshift-authentication/oauth-openshift-9565f95f5-7jbmm" Dec 03 14:14:14 crc kubenswrapper[4805]: I1203 14:14:14.231557 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/58a56dbb-271d-49d7-bce9-21a6aef96265-v4-0-config-system-service-ca\") pod \"oauth-openshift-9565f95f5-7jbmm\" (UID: \"58a56dbb-271d-49d7-bce9-21a6aef96265\") " pod="openshift-authentication/oauth-openshift-9565f95f5-7jbmm" Dec 03 14:14:14 crc kubenswrapper[4805]: I1203 14:14:14.231648 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/58a56dbb-271d-49d7-bce9-21a6aef96265-audit-dir\") pod \"oauth-openshift-9565f95f5-7jbmm\" (UID: \"58a56dbb-271d-49d7-bce9-21a6aef96265\") " pod="openshift-authentication/oauth-openshift-9565f95f5-7jbmm" Dec 03 14:14:14 crc kubenswrapper[4805]: I1203 14:14:14.231698 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/58a56dbb-271d-49d7-bce9-21a6aef96265-v4-0-config-system-router-certs\") pod \"oauth-openshift-9565f95f5-7jbmm\" (UID: \"58a56dbb-271d-49d7-bce9-21a6aef96265\") " pod="openshift-authentication/oauth-openshift-9565f95f5-7jbmm" Dec 03 14:14:14 crc kubenswrapper[4805]: I1203 14:14:14.231730 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qpt9w\" (UniqueName: \"kubernetes.io/projected/58a56dbb-271d-49d7-bce9-21a6aef96265-kube-api-access-qpt9w\") pod \"oauth-openshift-9565f95f5-7jbmm\" (UID: \"58a56dbb-271d-49d7-bce9-21a6aef96265\") " pod="openshift-authentication/oauth-openshift-9565f95f5-7jbmm" Dec 03 14:14:14 crc kubenswrapper[4805]: I1203 14:14:14.231770 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/58a56dbb-271d-49d7-bce9-21a6aef96265-v4-0-config-user-template-login\") pod \"oauth-openshift-9565f95f5-7jbmm\" (UID: \"58a56dbb-271d-49d7-bce9-21a6aef96265\") " pod="openshift-authentication/oauth-openshift-9565f95f5-7jbmm" Dec 03 14:14:14 crc kubenswrapper[4805]: I1203 14:14:14.231807 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/58a56dbb-271d-49d7-bce9-21a6aef96265-v4-0-config-system-cliconfig\") pod \"oauth-openshift-9565f95f5-7jbmm\" (UID: \"58a56dbb-271d-49d7-bce9-21a6aef96265\") " pod="openshift-authentication/oauth-openshift-9565f95f5-7jbmm" Dec 03 14:14:14 crc kubenswrapper[4805]: I1203 14:14:14.231909 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/58a56dbb-271d-49d7-bce9-21a6aef96265-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-9565f95f5-7jbmm\" (UID: \"58a56dbb-271d-49d7-bce9-21a6aef96265\") " pod="openshift-authentication/oauth-openshift-9565f95f5-7jbmm" Dec 03 14:14:14 crc kubenswrapper[4805]: I1203 14:14:14.231955 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/58a56dbb-271d-49d7-bce9-21a6aef96265-audit-policies\") pod \"oauth-openshift-9565f95f5-7jbmm\" (UID: \"58a56dbb-271d-49d7-bce9-21a6aef96265\") " pod="openshift-authentication/oauth-openshift-9565f95f5-7jbmm" Dec 03 14:14:14 crc kubenswrapper[4805]: I1203 14:14:14.231986 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/58a56dbb-271d-49d7-bce9-21a6aef96265-v4-0-config-system-serving-cert\") pod \"oauth-openshift-9565f95f5-7jbmm\" (UID: \"58a56dbb-271d-49d7-bce9-21a6aef96265\") " pod="openshift-authentication/oauth-openshift-9565f95f5-7jbmm" Dec 03 14:14:14 crc kubenswrapper[4805]: I1203 14:14:14.232030 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/58a56dbb-271d-49d7-bce9-21a6aef96265-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-9565f95f5-7jbmm\" (UID: \"58a56dbb-271d-49d7-bce9-21a6aef96265\") " pod="openshift-authentication/oauth-openshift-9565f95f5-7jbmm" Dec 03 14:14:14 crc kubenswrapper[4805]: I1203 14:14:14.233351 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/58a56dbb-271d-49d7-bce9-21a6aef96265-audit-dir\") pod \"oauth-openshift-9565f95f5-7jbmm\" (UID: \"58a56dbb-271d-49d7-bce9-21a6aef96265\") " pod="openshift-authentication/oauth-openshift-9565f95f5-7jbmm" Dec 03 14:14:14 crc kubenswrapper[4805]: I1203 14:14:14.234964 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/58a56dbb-271d-49d7-bce9-21a6aef96265-v4-0-config-system-service-ca\") pod \"oauth-openshift-9565f95f5-7jbmm\" (UID: \"58a56dbb-271d-49d7-bce9-21a6aef96265\") " pod="openshift-authentication/oauth-openshift-9565f95f5-7jbmm" Dec 03 14:14:14 crc kubenswrapper[4805]: I1203 14:14:14.236269 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/58a56dbb-271d-49d7-bce9-21a6aef96265-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-9565f95f5-7jbmm\" (UID: \"58a56dbb-271d-49d7-bce9-21a6aef96265\") " pod="openshift-authentication/oauth-openshift-9565f95f5-7jbmm" Dec 03 14:14:14 crc kubenswrapper[4805]: I1203 14:14:14.236701 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/58a56dbb-271d-49d7-bce9-21a6aef96265-audit-policies\") pod \"oauth-openshift-9565f95f5-7jbmm\" (UID: \"58a56dbb-271d-49d7-bce9-21a6aef96265\") " pod="openshift-authentication/oauth-openshift-9565f95f5-7jbmm" Dec 03 14:14:14 crc kubenswrapper[4805]: I1203 14:14:14.237667 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/58a56dbb-271d-49d7-bce9-21a6aef96265-v4-0-config-system-cliconfig\") pod \"oauth-openshift-9565f95f5-7jbmm\" (UID: \"58a56dbb-271d-49d7-bce9-21a6aef96265\") " pod="openshift-authentication/oauth-openshift-9565f95f5-7jbmm" Dec 03 14:14:14 crc kubenswrapper[4805]: I1203 14:14:14.241711 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/58a56dbb-271d-49d7-bce9-21a6aef96265-v4-0-config-user-template-error\") pod \"oauth-openshift-9565f95f5-7jbmm\" (UID: \"58a56dbb-271d-49d7-bce9-21a6aef96265\") " pod="openshift-authentication/oauth-openshift-9565f95f5-7jbmm" Dec 03 14:14:14 crc kubenswrapper[4805]: I1203 14:14:14.242193 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/58a56dbb-271d-49d7-bce9-21a6aef96265-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-9565f95f5-7jbmm\" (UID: \"58a56dbb-271d-49d7-bce9-21a6aef96265\") " pod="openshift-authentication/oauth-openshift-9565f95f5-7jbmm" Dec 03 14:14:14 crc kubenswrapper[4805]: I1203 14:14:14.242206 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/58a56dbb-271d-49d7-bce9-21a6aef96265-v4-0-config-system-router-certs\") pod \"oauth-openshift-9565f95f5-7jbmm\" (UID: \"58a56dbb-271d-49d7-bce9-21a6aef96265\") " pod="openshift-authentication/oauth-openshift-9565f95f5-7jbmm" Dec 03 14:14:14 crc kubenswrapper[4805]: I1203 14:14:14.242751 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/58a56dbb-271d-49d7-bce9-21a6aef96265-v4-0-config-system-serving-cert\") pod \"oauth-openshift-9565f95f5-7jbmm\" (UID: \"58a56dbb-271d-49d7-bce9-21a6aef96265\") " pod="openshift-authentication/oauth-openshift-9565f95f5-7jbmm" Dec 03 14:14:14 crc kubenswrapper[4805]: I1203 14:14:14.244273 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/58a56dbb-271d-49d7-bce9-21a6aef96265-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-9565f95f5-7jbmm\" (UID: \"58a56dbb-271d-49d7-bce9-21a6aef96265\") " pod="openshift-authentication/oauth-openshift-9565f95f5-7jbmm" Dec 03 14:14:14 crc kubenswrapper[4805]: I1203 14:14:14.245069 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/58a56dbb-271d-49d7-bce9-21a6aef96265-v4-0-config-user-template-login\") pod \"oauth-openshift-9565f95f5-7jbmm\" (UID: \"58a56dbb-271d-49d7-bce9-21a6aef96265\") " pod="openshift-authentication/oauth-openshift-9565f95f5-7jbmm" Dec 03 14:14:14 crc kubenswrapper[4805]: I1203 14:14:14.245258 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/58a56dbb-271d-49d7-bce9-21a6aef96265-v4-0-config-system-session\") pod \"oauth-openshift-9565f95f5-7jbmm\" (UID: \"58a56dbb-271d-49d7-bce9-21a6aef96265\") " pod="openshift-authentication/oauth-openshift-9565f95f5-7jbmm" Dec 03 14:14:14 crc kubenswrapper[4805]: I1203 14:14:14.251098 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/58a56dbb-271d-49d7-bce9-21a6aef96265-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-9565f95f5-7jbmm\" (UID: \"58a56dbb-271d-49d7-bce9-21a6aef96265\") " pod="openshift-authentication/oauth-openshift-9565f95f5-7jbmm" Dec 03 14:14:14 crc kubenswrapper[4805]: I1203 14:14:14.253354 4805 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"trusted-ca" Dec 03 14:14:14 crc kubenswrapper[4805]: I1203 14:14:14.267049 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qpt9w\" (UniqueName: \"kubernetes.io/projected/58a56dbb-271d-49d7-bce9-21a6aef96265-kube-api-access-qpt9w\") pod \"oauth-openshift-9565f95f5-7jbmm\" (UID: \"58a56dbb-271d-49d7-bce9-21a6aef96265\") " pod="openshift-authentication/oauth-openshift-9565f95f5-7jbmm" Dec 03 14:14:14 crc kubenswrapper[4805]: I1203 14:14:14.283587 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-9565f95f5-7jbmm" Dec 03 14:14:14 crc kubenswrapper[4805]: I1203 14:14:14.315079 4805 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"console-config" Dec 03 14:14:14 crc kubenswrapper[4805]: I1203 14:14:14.530061 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-9565f95f5-7jbmm"] Dec 03 14:14:14 crc kubenswrapper[4805]: I1203 14:14:14.530737 4805 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"trusted-ca" Dec 03 14:14:15 crc kubenswrapper[4805]: I1203 14:14:15.023384 4805 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"oauth-serving-cert" Dec 03 14:14:15 crc kubenswrapper[4805]: I1203 14:14:15.047400 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console-operator"/"console-operator-dockercfg-4xjcr" Dec 03 14:14:15 crc kubenswrapper[4805]: I1203 14:14:15.165665 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"serving-cert" Dec 03 14:14:15 crc kubenswrapper[4805]: I1203 14:14:15.203992 4805 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"trusted-ca-bundle" Dec 03 14:14:15 crc kubenswrapper[4805]: I1203 14:14:15.254893 4805 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager-operator"/"kube-root-ca.crt" Dec 03 14:14:15 crc kubenswrapper[4805]: I1203 14:14:15.469212 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-9565f95f5-7jbmm" event={"ID":"58a56dbb-271d-49d7-bce9-21a6aef96265","Type":"ContainerStarted","Data":"b546ec1516efcbc932d5a18fc1a934c2eefabe5f0a60016cd025184380924519"} Dec 03 14:14:15 crc kubenswrapper[4805]: I1203 14:14:15.469283 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-9565f95f5-7jbmm" event={"ID":"58a56dbb-271d-49d7-bce9-21a6aef96265","Type":"ContainerStarted","Data":"f4db3951ca5f8d1687d874ec68c9478a95ef57531bd33455501ff36256b3a847"} Dec 03 14:14:15 crc kubenswrapper[4805]: I1203 14:14:15.469630 4805 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-authentication/oauth-openshift-9565f95f5-7jbmm" Dec 03 14:14:15 crc kubenswrapper[4805]: I1203 14:14:15.476191 4805 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-authentication/oauth-openshift-9565f95f5-7jbmm" Dec 03 14:14:15 crc kubenswrapper[4805]: I1203 14:14:15.502562 4805 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-authentication/oauth-openshift-9565f95f5-7jbmm" podStartSLOduration=59.502545125 podStartE2EDuration="59.502545125s" podCreationTimestamp="2025-12-03 14:13:16 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 14:14:15.49719115 +0000 UTC m=+285.160108093" watchObservedRunningTime="2025-12-03 14:14:15.502545125 +0000 UTC m=+285.165462058" Dec 03 14:14:15 crc kubenswrapper[4805]: I1203 14:14:15.731720 4805 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"kube-root-ca.crt" Dec 03 14:14:15 crc kubenswrapper[4805]: I1203 14:14:15.796476 4805 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"openshift-service-ca.crt" Dec 03 14:14:16 crc kubenswrapper[4805]: I1203 14:14:16.243662 4805 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"etcd-serving-ca" Dec 03 14:14:16 crc kubenswrapper[4805]: I1203 14:14:16.633701 4805 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"trusted-ca-bundle" Dec 03 14:14:17 crc kubenswrapper[4805]: I1203 14:14:17.017028 4805 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"config" Dec 03 14:14:21 crc kubenswrapper[4805]: I1203 14:14:21.927962 4805 kubelet.go:2431] "SyncLoop REMOVE" source="file" pods=["openshift-kube-apiserver/kube-apiserver-startup-monitor-crc"] Dec 03 14:14:21 crc kubenswrapper[4805]: I1203 14:14:21.928168 4805 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" containerName="startup-monitor" containerID="cri-o://def67344e1a1d03baa3748a24cca7799fa8707eac5d82957414ba33076fe1d29" gracePeriod=5 Dec 03 14:14:27 crc kubenswrapper[4805]: I1203 14:14:27.523981 4805 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-startup-monitor-crc_f85e55b1a89d02b0cb034b1ea31ed45a/startup-monitor/0.log" Dec 03 14:14:27 crc kubenswrapper[4805]: I1203 14:14:27.524391 4805 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 03 14:14:27 crc kubenswrapper[4805]: I1203 14:14:27.549473 4805 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-startup-monitor-crc_f85e55b1a89d02b0cb034b1ea31ed45a/startup-monitor/0.log" Dec 03 14:14:27 crc kubenswrapper[4805]: I1203 14:14:27.549549 4805 generic.go:334] "Generic (PLEG): container finished" podID="f85e55b1a89d02b0cb034b1ea31ed45a" containerID="def67344e1a1d03baa3748a24cca7799fa8707eac5d82957414ba33076fe1d29" exitCode=137 Dec 03 14:14:27 crc kubenswrapper[4805]: I1203 14:14:27.549590 4805 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 03 14:14:27 crc kubenswrapper[4805]: I1203 14:14:27.549642 4805 scope.go:117] "RemoveContainer" containerID="def67344e1a1d03baa3748a24cca7799fa8707eac5d82957414ba33076fe1d29" Dec 03 14:14:27 crc kubenswrapper[4805]: I1203 14:14:27.567566 4805 scope.go:117] "RemoveContainer" containerID="def67344e1a1d03baa3748a24cca7799fa8707eac5d82957414ba33076fe1d29" Dec 03 14:14:27 crc kubenswrapper[4805]: E1203 14:14:27.567985 4805 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"def67344e1a1d03baa3748a24cca7799fa8707eac5d82957414ba33076fe1d29\": container with ID starting with def67344e1a1d03baa3748a24cca7799fa8707eac5d82957414ba33076fe1d29 not found: ID does not exist" containerID="def67344e1a1d03baa3748a24cca7799fa8707eac5d82957414ba33076fe1d29" Dec 03 14:14:27 crc kubenswrapper[4805]: I1203 14:14:27.568038 4805 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"def67344e1a1d03baa3748a24cca7799fa8707eac5d82957414ba33076fe1d29"} err="failed to get container status \"def67344e1a1d03baa3748a24cca7799fa8707eac5d82957414ba33076fe1d29\": rpc error: code = NotFound desc = could not find container \"def67344e1a1d03baa3748a24cca7799fa8707eac5d82957414ba33076fe1d29\": container with ID starting with def67344e1a1d03baa3748a24cca7799fa8707eac5d82957414ba33076fe1d29 not found: ID does not exist" Dec 03 14:14:27 crc kubenswrapper[4805]: I1203 14:14:27.613502 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"manifests\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests\") pod \"f85e55b1a89d02b0cb034b1ea31ed45a\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " Dec 03 14:14:27 crc kubenswrapper[4805]: I1203 14:14:27.613613 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pod-resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir\") pod \"f85e55b1a89d02b0cb034b1ea31ed45a\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " Dec 03 14:14:27 crc kubenswrapper[4805]: I1203 14:14:27.613625 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests" (OuterVolumeSpecName: "manifests") pod "f85e55b1a89d02b0cb034b1ea31ed45a" (UID: "f85e55b1a89d02b0cb034b1ea31ed45a"). InnerVolumeSpecName "manifests". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 03 14:14:27 crc kubenswrapper[4805]: I1203 14:14:27.613676 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log\") pod \"f85e55b1a89d02b0cb034b1ea31ed45a\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " Dec 03 14:14:27 crc kubenswrapper[4805]: I1203 14:14:27.613752 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir\") pod \"f85e55b1a89d02b0cb034b1ea31ed45a\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " Dec 03 14:14:27 crc kubenswrapper[4805]: I1203 14:14:27.613794 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock\") pod \"f85e55b1a89d02b0cb034b1ea31ed45a\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " Dec 03 14:14:27 crc kubenswrapper[4805]: I1203 14:14:27.613857 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log" (OuterVolumeSpecName: "var-log") pod "f85e55b1a89d02b0cb034b1ea31ed45a" (UID: "f85e55b1a89d02b0cb034b1ea31ed45a"). InnerVolumeSpecName "var-log". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 03 14:14:27 crc kubenswrapper[4805]: I1203 14:14:27.613922 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir" (OuterVolumeSpecName: "resource-dir") pod "f85e55b1a89d02b0cb034b1ea31ed45a" (UID: "f85e55b1a89d02b0cb034b1ea31ed45a"). InnerVolumeSpecName "resource-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 03 14:14:27 crc kubenswrapper[4805]: I1203 14:14:27.614010 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock" (OuterVolumeSpecName: "var-lock") pod "f85e55b1a89d02b0cb034b1ea31ed45a" (UID: "f85e55b1a89d02b0cb034b1ea31ed45a"). InnerVolumeSpecName "var-lock". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 03 14:14:27 crc kubenswrapper[4805]: I1203 14:14:27.614325 4805 reconciler_common.go:293] "Volume detached for volume \"manifests\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests\") on node \"crc\" DevicePath \"\"" Dec 03 14:14:27 crc kubenswrapper[4805]: I1203 14:14:27.614361 4805 reconciler_common.go:293] "Volume detached for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log\") on node \"crc\" DevicePath \"\"" Dec 03 14:14:27 crc kubenswrapper[4805]: I1203 14:14:27.614386 4805 reconciler_common.go:293] "Volume detached for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir\") on node \"crc\" DevicePath \"\"" Dec 03 14:14:27 crc kubenswrapper[4805]: I1203 14:14:27.614408 4805 reconciler_common.go:293] "Volume detached for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock\") on node \"crc\" DevicePath \"\"" Dec 03 14:14:27 crc kubenswrapper[4805]: I1203 14:14:27.621984 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir" (OuterVolumeSpecName: "pod-resource-dir") pod "f85e55b1a89d02b0cb034b1ea31ed45a" (UID: "f85e55b1a89d02b0cb034b1ea31ed45a"). InnerVolumeSpecName "pod-resource-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 03 14:14:27 crc kubenswrapper[4805]: I1203 14:14:27.715499 4805 reconciler_common.go:293] "Volume detached for volume \"pod-resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir\") on node \"crc\" DevicePath \"\"" Dec 03 14:14:28 crc kubenswrapper[4805]: I1203 14:14:28.705009 4805 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" path="/var/lib/kubelet/pods/f85e55b1a89d02b0cb034b1ea31ed45a/volumes" Dec 03 14:14:30 crc kubenswrapper[4805]: I1203 14:14:30.553111 4805 cert_rotation.go:91] certificate rotation detected, shutting down client connections to start using new credentials Dec 03 14:14:32 crc kubenswrapper[4805]: I1203 14:14:32.589086 4805 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-controller-manager_kube-controller-manager-crc_f614b9022728cf315e60c057852e563e/kube-controller-manager/1.log" Dec 03 14:14:32 crc kubenswrapper[4805]: I1203 14:14:32.591228 4805 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-controller-manager_kube-controller-manager-crc_f614b9022728cf315e60c057852e563e/kube-controller-manager/0.log" Dec 03 14:14:32 crc kubenswrapper[4805]: I1203 14:14:32.591411 4805 generic.go:334] "Generic (PLEG): container finished" podID="f614b9022728cf315e60c057852e563e" containerID="604de07374b8d85ab0785f21f409ddd439bfadd9af01803476055e82f903166a" exitCode=137 Dec 03 14:14:32 crc kubenswrapper[4805]: I1203 14:14:32.591487 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerDied","Data":"604de07374b8d85ab0785f21f409ddd439bfadd9af01803476055e82f903166a"} Dec 03 14:14:32 crc kubenswrapper[4805]: I1203 14:14:32.591699 4805 scope.go:117] "RemoveContainer" containerID="8673a29c0e16f4e707d79f0610722ccabb023ff2b4a18d7cf9d305183817d952" Dec 03 14:14:33 crc kubenswrapper[4805]: I1203 14:14:33.600244 4805 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-controller-manager_kube-controller-manager-crc_f614b9022728cf315e60c057852e563e/kube-controller-manager/1.log" Dec 03 14:14:33 crc kubenswrapper[4805]: I1203 14:14:33.602092 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"884a070c7a28e41241d8e52f2c5d97c880fcc658c364c6df9de5aaa4275db1a0"} Dec 03 14:14:40 crc kubenswrapper[4805]: I1203 14:14:40.052919 4805 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 03 14:14:42 crc kubenswrapper[4805]: I1203 14:14:42.391710 4805 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 03 14:14:42 crc kubenswrapper[4805]: I1203 14:14:42.397465 4805 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 03 14:14:49 crc kubenswrapper[4805]: I1203 14:14:49.521388 4805 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-k7rgc"] Dec 03 14:14:49 crc kubenswrapper[4805]: I1203 14:14:49.522419 4805 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-controller-manager/controller-manager-879f6c89f-k7rgc" podUID="b54908df-f04f-4515-ab36-842f0aef4a99" containerName="controller-manager" containerID="cri-o://4129f75f5d57f0e656ccb2de4e65efc62a7b3c8132a246f5fb03c630d83ee6ab" gracePeriod=30 Dec 03 14:14:49 crc kubenswrapper[4805]: I1203 14:14:49.525185 4805 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-nsz7s"] Dec 03 14:14:49 crc kubenswrapper[4805]: I1203 14:14:49.525389 4805 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-nsz7s" podUID="dfa6979a-2473-469b-ac56-9086e35a63f4" containerName="route-controller-manager" containerID="cri-o://6093b96c1fb59901b9b95e4d8fc91991e09c62d3131912bd96f23954fdca62c2" gracePeriod=30 Dec 03 14:14:49 crc kubenswrapper[4805]: I1203 14:14:49.715665 4805 generic.go:334] "Generic (PLEG): container finished" podID="b54908df-f04f-4515-ab36-842f0aef4a99" containerID="4129f75f5d57f0e656ccb2de4e65efc62a7b3c8132a246f5fb03c630d83ee6ab" exitCode=0 Dec 03 14:14:49 crc kubenswrapper[4805]: I1203 14:14:49.715809 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-k7rgc" event={"ID":"b54908df-f04f-4515-ab36-842f0aef4a99","Type":"ContainerDied","Data":"4129f75f5d57f0e656ccb2de4e65efc62a7b3c8132a246f5fb03c630d83ee6ab"} Dec 03 14:14:49 crc kubenswrapper[4805]: I1203 14:14:49.728640 4805 generic.go:334] "Generic (PLEG): container finished" podID="dfa6979a-2473-469b-ac56-9086e35a63f4" containerID="6093b96c1fb59901b9b95e4d8fc91991e09c62d3131912bd96f23954fdca62c2" exitCode=0 Dec 03 14:14:49 crc kubenswrapper[4805]: I1203 14:14:49.728694 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-nsz7s" event={"ID":"dfa6979a-2473-469b-ac56-9086e35a63f4","Type":"ContainerDied","Data":"6093b96c1fb59901b9b95e4d8fc91991e09c62d3131912bd96f23954fdca62c2"} Dec 03 14:14:49 crc kubenswrapper[4805]: I1203 14:14:49.890164 4805 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-k7rgc" Dec 03 14:14:49 crc kubenswrapper[4805]: I1203 14:14:49.962960 4805 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-nsz7s" Dec 03 14:14:50 crc kubenswrapper[4805]: I1203 14:14:50.012379 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/b54908df-f04f-4515-ab36-842f0aef4a99-proxy-ca-bundles\") pod \"b54908df-f04f-4515-ab36-842f0aef4a99\" (UID: \"b54908df-f04f-4515-ab36-842f0aef4a99\") " Dec 03 14:14:50 crc kubenswrapper[4805]: I1203 14:14:50.012427 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b54908df-f04f-4515-ab36-842f0aef4a99-config\") pod \"b54908df-f04f-4515-ab36-842f0aef4a99\" (UID: \"b54908df-f04f-4515-ab36-842f0aef4a99\") " Dec 03 14:14:50 crc kubenswrapper[4805]: I1203 14:14:50.012455 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/b54908df-f04f-4515-ab36-842f0aef4a99-client-ca\") pod \"b54908df-f04f-4515-ab36-842f0aef4a99\" (UID: \"b54908df-f04f-4515-ab36-842f0aef4a99\") " Dec 03 14:14:50 crc kubenswrapper[4805]: I1203 14:14:50.012522 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-89vsl\" (UniqueName: \"kubernetes.io/projected/b54908df-f04f-4515-ab36-842f0aef4a99-kube-api-access-89vsl\") pod \"b54908df-f04f-4515-ab36-842f0aef4a99\" (UID: \"b54908df-f04f-4515-ab36-842f0aef4a99\") " Dec 03 14:14:50 crc kubenswrapper[4805]: I1203 14:14:50.012563 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/b54908df-f04f-4515-ab36-842f0aef4a99-serving-cert\") pod \"b54908df-f04f-4515-ab36-842f0aef4a99\" (UID: \"b54908df-f04f-4515-ab36-842f0aef4a99\") " Dec 03 14:14:50 crc kubenswrapper[4805]: I1203 14:14:50.013163 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b54908df-f04f-4515-ab36-842f0aef4a99-proxy-ca-bundles" (OuterVolumeSpecName: "proxy-ca-bundles") pod "b54908df-f04f-4515-ab36-842f0aef4a99" (UID: "b54908df-f04f-4515-ab36-842f0aef4a99"). InnerVolumeSpecName "proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 14:14:50 crc kubenswrapper[4805]: I1203 14:14:50.013168 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b54908df-f04f-4515-ab36-842f0aef4a99-client-ca" (OuterVolumeSpecName: "client-ca") pod "b54908df-f04f-4515-ab36-842f0aef4a99" (UID: "b54908df-f04f-4515-ab36-842f0aef4a99"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 14:14:50 crc kubenswrapper[4805]: I1203 14:14:50.013456 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b54908df-f04f-4515-ab36-842f0aef4a99-config" (OuterVolumeSpecName: "config") pod "b54908df-f04f-4515-ab36-842f0aef4a99" (UID: "b54908df-f04f-4515-ab36-842f0aef4a99"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 14:14:50 crc kubenswrapper[4805]: I1203 14:14:50.018248 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b54908df-f04f-4515-ab36-842f0aef4a99-kube-api-access-89vsl" (OuterVolumeSpecName: "kube-api-access-89vsl") pod "b54908df-f04f-4515-ab36-842f0aef4a99" (UID: "b54908df-f04f-4515-ab36-842f0aef4a99"). InnerVolumeSpecName "kube-api-access-89vsl". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 14:14:50 crc kubenswrapper[4805]: I1203 14:14:50.018412 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b54908df-f04f-4515-ab36-842f0aef4a99-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "b54908df-f04f-4515-ab36-842f0aef4a99" (UID: "b54908df-f04f-4515-ab36-842f0aef4a99"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 14:14:50 crc kubenswrapper[4805]: I1203 14:14:50.057449 4805 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 03 14:14:50 crc kubenswrapper[4805]: I1203 14:14:50.114158 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/dfa6979a-2473-469b-ac56-9086e35a63f4-client-ca\") pod \"dfa6979a-2473-469b-ac56-9086e35a63f4\" (UID: \"dfa6979a-2473-469b-ac56-9086e35a63f4\") " Dec 03 14:14:50 crc kubenswrapper[4805]: I1203 14:14:50.114246 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-n7759\" (UniqueName: \"kubernetes.io/projected/dfa6979a-2473-469b-ac56-9086e35a63f4-kube-api-access-n7759\") pod \"dfa6979a-2473-469b-ac56-9086e35a63f4\" (UID: \"dfa6979a-2473-469b-ac56-9086e35a63f4\") " Dec 03 14:14:50 crc kubenswrapper[4805]: I1203 14:14:50.114310 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/dfa6979a-2473-469b-ac56-9086e35a63f4-serving-cert\") pod \"dfa6979a-2473-469b-ac56-9086e35a63f4\" (UID: \"dfa6979a-2473-469b-ac56-9086e35a63f4\") " Dec 03 14:14:50 crc kubenswrapper[4805]: I1203 14:14:50.114339 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/dfa6979a-2473-469b-ac56-9086e35a63f4-config\") pod \"dfa6979a-2473-469b-ac56-9086e35a63f4\" (UID: \"dfa6979a-2473-469b-ac56-9086e35a63f4\") " Dec 03 14:14:50 crc kubenswrapper[4805]: I1203 14:14:50.114565 4805 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-89vsl\" (UniqueName: \"kubernetes.io/projected/b54908df-f04f-4515-ab36-842f0aef4a99-kube-api-access-89vsl\") on node \"crc\" DevicePath \"\"" Dec 03 14:14:50 crc kubenswrapper[4805]: I1203 14:14:50.114583 4805 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/b54908df-f04f-4515-ab36-842f0aef4a99-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 03 14:14:50 crc kubenswrapper[4805]: I1203 14:14:50.114594 4805 reconciler_common.go:293] "Volume detached for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/b54908df-f04f-4515-ab36-842f0aef4a99-proxy-ca-bundles\") on node \"crc\" DevicePath \"\"" Dec 03 14:14:50 crc kubenswrapper[4805]: I1203 14:14:50.114606 4805 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b54908df-f04f-4515-ab36-842f0aef4a99-config\") on node \"crc\" DevicePath \"\"" Dec 03 14:14:50 crc kubenswrapper[4805]: I1203 14:14:50.114617 4805 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/b54908df-f04f-4515-ab36-842f0aef4a99-client-ca\") on node \"crc\" DevicePath \"\"" Dec 03 14:14:50 crc kubenswrapper[4805]: I1203 14:14:50.115102 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/dfa6979a-2473-469b-ac56-9086e35a63f4-client-ca" (OuterVolumeSpecName: "client-ca") pod "dfa6979a-2473-469b-ac56-9086e35a63f4" (UID: "dfa6979a-2473-469b-ac56-9086e35a63f4"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 14:14:50 crc kubenswrapper[4805]: I1203 14:14:50.115201 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/dfa6979a-2473-469b-ac56-9086e35a63f4-config" (OuterVolumeSpecName: "config") pod "dfa6979a-2473-469b-ac56-9086e35a63f4" (UID: "dfa6979a-2473-469b-ac56-9086e35a63f4"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 14:14:50 crc kubenswrapper[4805]: I1203 14:14:50.121314 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/dfa6979a-2473-469b-ac56-9086e35a63f4-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "dfa6979a-2473-469b-ac56-9086e35a63f4" (UID: "dfa6979a-2473-469b-ac56-9086e35a63f4"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 14:14:50 crc kubenswrapper[4805]: I1203 14:14:50.121418 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/dfa6979a-2473-469b-ac56-9086e35a63f4-kube-api-access-n7759" (OuterVolumeSpecName: "kube-api-access-n7759") pod "dfa6979a-2473-469b-ac56-9086e35a63f4" (UID: "dfa6979a-2473-469b-ac56-9086e35a63f4"). InnerVolumeSpecName "kube-api-access-n7759". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 14:14:50 crc kubenswrapper[4805]: I1203 14:14:50.215248 4805 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/dfa6979a-2473-469b-ac56-9086e35a63f4-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 03 14:14:50 crc kubenswrapper[4805]: I1203 14:14:50.215284 4805 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/dfa6979a-2473-469b-ac56-9086e35a63f4-config\") on node \"crc\" DevicePath \"\"" Dec 03 14:14:50 crc kubenswrapper[4805]: I1203 14:14:50.215293 4805 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/dfa6979a-2473-469b-ac56-9086e35a63f4-client-ca\") on node \"crc\" DevicePath \"\"" Dec 03 14:14:50 crc kubenswrapper[4805]: I1203 14:14:50.215303 4805 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-n7759\" (UniqueName: \"kubernetes.io/projected/dfa6979a-2473-469b-ac56-9086e35a63f4-kube-api-access-n7759\") on node \"crc\" DevicePath \"\"" Dec 03 14:14:50 crc kubenswrapper[4805]: I1203 14:14:50.741745 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-nsz7s" event={"ID":"dfa6979a-2473-469b-ac56-9086e35a63f4","Type":"ContainerDied","Data":"ff8b6532555301b7ddcd5390cb0fd97c161d7b3fd3aa87e3f6e8d0d6a140f457"} Dec 03 14:14:50 crc kubenswrapper[4805]: I1203 14:14:50.741807 4805 scope.go:117] "RemoveContainer" containerID="6093b96c1fb59901b9b95e4d8fc91991e09c62d3131912bd96f23954fdca62c2" Dec 03 14:14:50 crc kubenswrapper[4805]: I1203 14:14:50.741959 4805 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-nsz7s" Dec 03 14:14:50 crc kubenswrapper[4805]: I1203 14:14:50.744522 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-k7rgc" event={"ID":"b54908df-f04f-4515-ab36-842f0aef4a99","Type":"ContainerDied","Data":"1aebe087880b686ed928502ca004a539fe527c336faf9330357b68468b25841c"} Dec 03 14:14:50 crc kubenswrapper[4805]: I1203 14:14:50.744611 4805 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-k7rgc" Dec 03 14:14:50 crc kubenswrapper[4805]: I1203 14:14:50.771899 4805 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-k7rgc"] Dec 03 14:14:50 crc kubenswrapper[4805]: I1203 14:14:50.776787 4805 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-k7rgc"] Dec 03 14:14:50 crc kubenswrapper[4805]: I1203 14:14:50.781795 4805 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-nsz7s"] Dec 03 14:14:50 crc kubenswrapper[4805]: I1203 14:14:50.786095 4805 scope.go:117] "RemoveContainer" containerID="4129f75f5d57f0e656ccb2de4e65efc62a7b3c8132a246f5fb03c630d83ee6ab" Dec 03 14:14:50 crc kubenswrapper[4805]: I1203 14:14:50.789302 4805 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-nsz7s"] Dec 03 14:14:50 crc kubenswrapper[4805]: I1203 14:14:50.966612 4805 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager/controller-manager-794995848b-95m55"] Dec 03 14:14:50 crc kubenswrapper[4805]: E1203 14:14:50.966885 4805 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b54908df-f04f-4515-ab36-842f0aef4a99" containerName="controller-manager" Dec 03 14:14:50 crc kubenswrapper[4805]: I1203 14:14:50.966901 4805 state_mem.go:107] "Deleted CPUSet assignment" podUID="b54908df-f04f-4515-ab36-842f0aef4a99" containerName="controller-manager" Dec 03 14:14:50 crc kubenswrapper[4805]: E1203 14:14:50.966915 4805 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" containerName="startup-monitor" Dec 03 14:14:50 crc kubenswrapper[4805]: I1203 14:14:50.966924 4805 state_mem.go:107] "Deleted CPUSet assignment" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" containerName="startup-monitor" Dec 03 14:14:50 crc kubenswrapper[4805]: E1203 14:14:50.966942 4805 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="dfa6979a-2473-469b-ac56-9086e35a63f4" containerName="route-controller-manager" Dec 03 14:14:50 crc kubenswrapper[4805]: I1203 14:14:50.966951 4805 state_mem.go:107] "Deleted CPUSet assignment" podUID="dfa6979a-2473-469b-ac56-9086e35a63f4" containerName="route-controller-manager" Dec 03 14:14:50 crc kubenswrapper[4805]: I1203 14:14:50.967076 4805 memory_manager.go:354] "RemoveStaleState removing state" podUID="b54908df-f04f-4515-ab36-842f0aef4a99" containerName="controller-manager" Dec 03 14:14:50 crc kubenswrapper[4805]: I1203 14:14:50.967090 4805 memory_manager.go:354] "RemoveStaleState removing state" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" containerName="startup-monitor" Dec 03 14:14:50 crc kubenswrapper[4805]: I1203 14:14:50.967103 4805 memory_manager.go:354] "RemoveStaleState removing state" podUID="dfa6979a-2473-469b-ac56-9086e35a63f4" containerName="route-controller-manager" Dec 03 14:14:50 crc kubenswrapper[4805]: I1203 14:14:50.967511 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-794995848b-95m55" Dec 03 14:14:50 crc kubenswrapper[4805]: I1203 14:14:50.969580 4805 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-route-controller-manager/route-controller-manager-685b95576d-cnn5h"] Dec 03 14:14:50 crc kubenswrapper[4805]: I1203 14:14:50.970040 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-685b95576d-cnn5h" Dec 03 14:14:50 crc kubenswrapper[4805]: I1203 14:14:50.974576 4805 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"client-ca" Dec 03 14:14:50 crc kubenswrapper[4805]: I1203 14:14:50.974599 4805 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-service-ca.crt" Dec 03 14:14:50 crc kubenswrapper[4805]: I1203 14:14:50.974805 4805 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"kube-root-ca.crt" Dec 03 14:14:50 crc kubenswrapper[4805]: I1203 14:14:50.974834 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"openshift-controller-manager-sa-dockercfg-msq4c" Dec 03 14:14:50 crc kubenswrapper[4805]: I1203 14:14:50.974943 4805 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"openshift-service-ca.crt" Dec 03 14:14:50 crc kubenswrapper[4805]: I1203 14:14:50.975064 4805 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"config" Dec 03 14:14:50 crc kubenswrapper[4805]: I1203 14:14:50.975074 4805 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"kube-root-ca.crt" Dec 03 14:14:50 crc kubenswrapper[4805]: I1203 14:14:50.975978 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"route-controller-manager-sa-dockercfg-h2zr2" Dec 03 14:14:50 crc kubenswrapper[4805]: I1203 14:14:50.976729 4805 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"client-ca" Dec 03 14:14:50 crc kubenswrapper[4805]: I1203 14:14:50.977694 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"serving-cert" Dec 03 14:14:50 crc kubenswrapper[4805]: I1203 14:14:50.977881 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"serving-cert" Dec 03 14:14:50 crc kubenswrapper[4805]: I1203 14:14:50.980825 4805 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"config" Dec 03 14:14:50 crc kubenswrapper[4805]: I1203 14:14:50.983864 4805 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-global-ca" Dec 03 14:14:50 crc kubenswrapper[4805]: I1203 14:14:50.985229 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-685b95576d-cnn5h"] Dec 03 14:14:50 crc kubenswrapper[4805]: I1203 14:14:50.989017 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-794995848b-95m55"] Dec 03 14:14:51 crc kubenswrapper[4805]: I1203 14:14:51.129571 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/46633c53-ce91-4bbe-a650-b4138cfb8ef1-serving-cert\") pod \"controller-manager-794995848b-95m55\" (UID: \"46633c53-ce91-4bbe-a650-b4138cfb8ef1\") " pod="openshift-controller-manager/controller-manager-794995848b-95m55" Dec 03 14:14:51 crc kubenswrapper[4805]: I1203 14:14:51.129636 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/46633c53-ce91-4bbe-a650-b4138cfb8ef1-proxy-ca-bundles\") pod \"controller-manager-794995848b-95m55\" (UID: \"46633c53-ce91-4bbe-a650-b4138cfb8ef1\") " pod="openshift-controller-manager/controller-manager-794995848b-95m55" Dec 03 14:14:51 crc kubenswrapper[4805]: I1203 14:14:51.129696 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hr9lp\" (UniqueName: \"kubernetes.io/projected/46633c53-ce91-4bbe-a650-b4138cfb8ef1-kube-api-access-hr9lp\") pod \"controller-manager-794995848b-95m55\" (UID: \"46633c53-ce91-4bbe-a650-b4138cfb8ef1\") " pod="openshift-controller-manager/controller-manager-794995848b-95m55" Dec 03 14:14:51 crc kubenswrapper[4805]: I1203 14:14:51.129741 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/e3df4b05-07b7-4fe8-9a83-ecc033fa42b3-client-ca\") pod \"route-controller-manager-685b95576d-cnn5h\" (UID: \"e3df4b05-07b7-4fe8-9a83-ecc033fa42b3\") " pod="openshift-route-controller-manager/route-controller-manager-685b95576d-cnn5h" Dec 03 14:14:51 crc kubenswrapper[4805]: I1203 14:14:51.129764 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2tkhw\" (UniqueName: \"kubernetes.io/projected/e3df4b05-07b7-4fe8-9a83-ecc033fa42b3-kube-api-access-2tkhw\") pod \"route-controller-manager-685b95576d-cnn5h\" (UID: \"e3df4b05-07b7-4fe8-9a83-ecc033fa42b3\") " pod="openshift-route-controller-manager/route-controller-manager-685b95576d-cnn5h" Dec 03 14:14:51 crc kubenswrapper[4805]: I1203 14:14:51.129923 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/46633c53-ce91-4bbe-a650-b4138cfb8ef1-config\") pod \"controller-manager-794995848b-95m55\" (UID: \"46633c53-ce91-4bbe-a650-b4138cfb8ef1\") " pod="openshift-controller-manager/controller-manager-794995848b-95m55" Dec 03 14:14:51 crc kubenswrapper[4805]: I1203 14:14:51.129967 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e3df4b05-07b7-4fe8-9a83-ecc033fa42b3-config\") pod \"route-controller-manager-685b95576d-cnn5h\" (UID: \"e3df4b05-07b7-4fe8-9a83-ecc033fa42b3\") " pod="openshift-route-controller-manager/route-controller-manager-685b95576d-cnn5h" Dec 03 14:14:51 crc kubenswrapper[4805]: I1203 14:14:51.129992 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/46633c53-ce91-4bbe-a650-b4138cfb8ef1-client-ca\") pod \"controller-manager-794995848b-95m55\" (UID: \"46633c53-ce91-4bbe-a650-b4138cfb8ef1\") " pod="openshift-controller-manager/controller-manager-794995848b-95m55" Dec 03 14:14:51 crc kubenswrapper[4805]: I1203 14:14:51.130023 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e3df4b05-07b7-4fe8-9a83-ecc033fa42b3-serving-cert\") pod \"route-controller-manager-685b95576d-cnn5h\" (UID: \"e3df4b05-07b7-4fe8-9a83-ecc033fa42b3\") " pod="openshift-route-controller-manager/route-controller-manager-685b95576d-cnn5h" Dec 03 14:14:51 crc kubenswrapper[4805]: I1203 14:14:51.230810 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e3df4b05-07b7-4fe8-9a83-ecc033fa42b3-serving-cert\") pod \"route-controller-manager-685b95576d-cnn5h\" (UID: \"e3df4b05-07b7-4fe8-9a83-ecc033fa42b3\") " pod="openshift-route-controller-manager/route-controller-manager-685b95576d-cnn5h" Dec 03 14:14:51 crc kubenswrapper[4805]: I1203 14:14:51.231093 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/46633c53-ce91-4bbe-a650-b4138cfb8ef1-serving-cert\") pod \"controller-manager-794995848b-95m55\" (UID: \"46633c53-ce91-4bbe-a650-b4138cfb8ef1\") " pod="openshift-controller-manager/controller-manager-794995848b-95m55" Dec 03 14:14:51 crc kubenswrapper[4805]: I1203 14:14:51.231113 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/46633c53-ce91-4bbe-a650-b4138cfb8ef1-proxy-ca-bundles\") pod \"controller-manager-794995848b-95m55\" (UID: \"46633c53-ce91-4bbe-a650-b4138cfb8ef1\") " pod="openshift-controller-manager/controller-manager-794995848b-95m55" Dec 03 14:14:51 crc kubenswrapper[4805]: I1203 14:14:51.231146 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hr9lp\" (UniqueName: \"kubernetes.io/projected/46633c53-ce91-4bbe-a650-b4138cfb8ef1-kube-api-access-hr9lp\") pod \"controller-manager-794995848b-95m55\" (UID: \"46633c53-ce91-4bbe-a650-b4138cfb8ef1\") " pod="openshift-controller-manager/controller-manager-794995848b-95m55" Dec 03 14:14:51 crc kubenswrapper[4805]: I1203 14:14:51.231174 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/e3df4b05-07b7-4fe8-9a83-ecc033fa42b3-client-ca\") pod \"route-controller-manager-685b95576d-cnn5h\" (UID: \"e3df4b05-07b7-4fe8-9a83-ecc033fa42b3\") " pod="openshift-route-controller-manager/route-controller-manager-685b95576d-cnn5h" Dec 03 14:14:51 crc kubenswrapper[4805]: I1203 14:14:51.231192 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2tkhw\" (UniqueName: \"kubernetes.io/projected/e3df4b05-07b7-4fe8-9a83-ecc033fa42b3-kube-api-access-2tkhw\") pod \"route-controller-manager-685b95576d-cnn5h\" (UID: \"e3df4b05-07b7-4fe8-9a83-ecc033fa42b3\") " pod="openshift-route-controller-manager/route-controller-manager-685b95576d-cnn5h" Dec 03 14:14:51 crc kubenswrapper[4805]: I1203 14:14:51.231223 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/46633c53-ce91-4bbe-a650-b4138cfb8ef1-config\") pod \"controller-manager-794995848b-95m55\" (UID: \"46633c53-ce91-4bbe-a650-b4138cfb8ef1\") " pod="openshift-controller-manager/controller-manager-794995848b-95m55" Dec 03 14:14:51 crc kubenswrapper[4805]: I1203 14:14:51.231246 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e3df4b05-07b7-4fe8-9a83-ecc033fa42b3-config\") pod \"route-controller-manager-685b95576d-cnn5h\" (UID: \"e3df4b05-07b7-4fe8-9a83-ecc033fa42b3\") " pod="openshift-route-controller-manager/route-controller-manager-685b95576d-cnn5h" Dec 03 14:14:51 crc kubenswrapper[4805]: I1203 14:14:51.231262 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/46633c53-ce91-4bbe-a650-b4138cfb8ef1-client-ca\") pod \"controller-manager-794995848b-95m55\" (UID: \"46633c53-ce91-4bbe-a650-b4138cfb8ef1\") " pod="openshift-controller-manager/controller-manager-794995848b-95m55" Dec 03 14:14:51 crc kubenswrapper[4805]: I1203 14:14:51.232075 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/46633c53-ce91-4bbe-a650-b4138cfb8ef1-client-ca\") pod \"controller-manager-794995848b-95m55\" (UID: \"46633c53-ce91-4bbe-a650-b4138cfb8ef1\") " pod="openshift-controller-manager/controller-manager-794995848b-95m55" Dec 03 14:14:51 crc kubenswrapper[4805]: I1203 14:14:51.232284 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/e3df4b05-07b7-4fe8-9a83-ecc033fa42b3-client-ca\") pod \"route-controller-manager-685b95576d-cnn5h\" (UID: \"e3df4b05-07b7-4fe8-9a83-ecc033fa42b3\") " pod="openshift-route-controller-manager/route-controller-manager-685b95576d-cnn5h" Dec 03 14:14:51 crc kubenswrapper[4805]: I1203 14:14:51.232404 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/46633c53-ce91-4bbe-a650-b4138cfb8ef1-proxy-ca-bundles\") pod \"controller-manager-794995848b-95m55\" (UID: \"46633c53-ce91-4bbe-a650-b4138cfb8ef1\") " pod="openshift-controller-manager/controller-manager-794995848b-95m55" Dec 03 14:14:51 crc kubenswrapper[4805]: I1203 14:14:51.232714 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/46633c53-ce91-4bbe-a650-b4138cfb8ef1-config\") pod \"controller-manager-794995848b-95m55\" (UID: \"46633c53-ce91-4bbe-a650-b4138cfb8ef1\") " pod="openshift-controller-manager/controller-manager-794995848b-95m55" Dec 03 14:14:51 crc kubenswrapper[4805]: I1203 14:14:51.232761 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e3df4b05-07b7-4fe8-9a83-ecc033fa42b3-config\") pod \"route-controller-manager-685b95576d-cnn5h\" (UID: \"e3df4b05-07b7-4fe8-9a83-ecc033fa42b3\") " pod="openshift-route-controller-manager/route-controller-manager-685b95576d-cnn5h" Dec 03 14:14:51 crc kubenswrapper[4805]: I1203 14:14:51.235823 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e3df4b05-07b7-4fe8-9a83-ecc033fa42b3-serving-cert\") pod \"route-controller-manager-685b95576d-cnn5h\" (UID: \"e3df4b05-07b7-4fe8-9a83-ecc033fa42b3\") " pod="openshift-route-controller-manager/route-controller-manager-685b95576d-cnn5h" Dec 03 14:14:51 crc kubenswrapper[4805]: I1203 14:14:51.236499 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/46633c53-ce91-4bbe-a650-b4138cfb8ef1-serving-cert\") pod \"controller-manager-794995848b-95m55\" (UID: \"46633c53-ce91-4bbe-a650-b4138cfb8ef1\") " pod="openshift-controller-manager/controller-manager-794995848b-95m55" Dec 03 14:14:51 crc kubenswrapper[4805]: I1203 14:14:51.260704 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2tkhw\" (UniqueName: \"kubernetes.io/projected/e3df4b05-07b7-4fe8-9a83-ecc033fa42b3-kube-api-access-2tkhw\") pod \"route-controller-manager-685b95576d-cnn5h\" (UID: \"e3df4b05-07b7-4fe8-9a83-ecc033fa42b3\") " pod="openshift-route-controller-manager/route-controller-manager-685b95576d-cnn5h" Dec 03 14:14:51 crc kubenswrapper[4805]: I1203 14:14:51.263436 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hr9lp\" (UniqueName: \"kubernetes.io/projected/46633c53-ce91-4bbe-a650-b4138cfb8ef1-kube-api-access-hr9lp\") pod \"controller-manager-794995848b-95m55\" (UID: \"46633c53-ce91-4bbe-a650-b4138cfb8ef1\") " pod="openshift-controller-manager/controller-manager-794995848b-95m55" Dec 03 14:14:51 crc kubenswrapper[4805]: I1203 14:14:51.311754 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-794995848b-95m55" Dec 03 14:14:51 crc kubenswrapper[4805]: I1203 14:14:51.321248 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-685b95576d-cnn5h" Dec 03 14:14:51 crc kubenswrapper[4805]: I1203 14:14:51.570852 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-685b95576d-cnn5h"] Dec 03 14:14:51 crc kubenswrapper[4805]: W1203 14:14:51.580144 4805 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pode3df4b05_07b7_4fe8_9a83_ecc033fa42b3.slice/crio-c04492978f5d09dc3a6e4242ec0f21aaba593ed1cad0f5842addd41fc8d98f4e WatchSource:0}: Error finding container c04492978f5d09dc3a6e4242ec0f21aaba593ed1cad0f5842addd41fc8d98f4e: Status 404 returned error can't find the container with id c04492978f5d09dc3a6e4242ec0f21aaba593ed1cad0f5842addd41fc8d98f4e Dec 03 14:14:51 crc kubenswrapper[4805]: I1203 14:14:51.599177 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-794995848b-95m55"] Dec 03 14:14:51 crc kubenswrapper[4805]: I1203 14:14:51.753357 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-794995848b-95m55" event={"ID":"46633c53-ce91-4bbe-a650-b4138cfb8ef1","Type":"ContainerStarted","Data":"f7cac3bc95c2d86e6002ac07d2b1c4a47a19ee1e04d0d49d6214aa3dadf84082"} Dec 03 14:14:51 crc kubenswrapper[4805]: I1203 14:14:51.753396 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-794995848b-95m55" event={"ID":"46633c53-ce91-4bbe-a650-b4138cfb8ef1","Type":"ContainerStarted","Data":"2c699d8733dd380a42b8e7f1babee8c964d71d7e68a13e6d6e02f8cdfecf3fa8"} Dec 03 14:14:51 crc kubenswrapper[4805]: I1203 14:14:51.754555 4805 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-controller-manager/controller-manager-794995848b-95m55" Dec 03 14:14:51 crc kubenswrapper[4805]: I1203 14:14:51.755510 4805 patch_prober.go:28] interesting pod/controller-manager-794995848b-95m55 container/controller-manager namespace/openshift-controller-manager: Readiness probe status=failure output="Get \"https://10.217.0.57:8443/healthz\": dial tcp 10.217.0.57:8443: connect: connection refused" start-of-body= Dec 03 14:14:51 crc kubenswrapper[4805]: I1203 14:14:51.755543 4805 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-controller-manager/controller-manager-794995848b-95m55" podUID="46633c53-ce91-4bbe-a650-b4138cfb8ef1" containerName="controller-manager" probeResult="failure" output="Get \"https://10.217.0.57:8443/healthz\": dial tcp 10.217.0.57:8443: connect: connection refused" Dec 03 14:14:51 crc kubenswrapper[4805]: I1203 14:14:51.762300 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-685b95576d-cnn5h" event={"ID":"e3df4b05-07b7-4fe8-9a83-ecc033fa42b3","Type":"ContainerStarted","Data":"e9cbdec719e14a87485c13a4dba1d66d153cfcc6084372df1d5445fae49d0116"} Dec 03 14:14:51 crc kubenswrapper[4805]: I1203 14:14:51.762356 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-685b95576d-cnn5h" event={"ID":"e3df4b05-07b7-4fe8-9a83-ecc033fa42b3","Type":"ContainerStarted","Data":"c04492978f5d09dc3a6e4242ec0f21aaba593ed1cad0f5842addd41fc8d98f4e"} Dec 03 14:14:51 crc kubenswrapper[4805]: I1203 14:14:51.762984 4805 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-route-controller-manager/route-controller-manager-685b95576d-cnn5h" Dec 03 14:14:51 crc kubenswrapper[4805]: I1203 14:14:51.764084 4805 patch_prober.go:28] interesting pod/route-controller-manager-685b95576d-cnn5h container/route-controller-manager namespace/openshift-route-controller-manager: Readiness probe status=failure output="Get \"https://10.217.0.58:8443/healthz\": dial tcp 10.217.0.58:8443: connect: connection refused" start-of-body= Dec 03 14:14:51 crc kubenswrapper[4805]: I1203 14:14:51.764132 4805 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-route-controller-manager/route-controller-manager-685b95576d-cnn5h" podUID="e3df4b05-07b7-4fe8-9a83-ecc033fa42b3" containerName="route-controller-manager" probeResult="failure" output="Get \"https://10.217.0.58:8443/healthz\": dial tcp 10.217.0.58:8443: connect: connection refused" Dec 03 14:14:51 crc kubenswrapper[4805]: I1203 14:14:51.783923 4805 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager/controller-manager-794995848b-95m55" podStartSLOduration=2.7839024500000003 podStartE2EDuration="2.78390245s" podCreationTimestamp="2025-12-03 14:14:49 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 14:14:51.779774378 +0000 UTC m=+321.442691301" watchObservedRunningTime="2025-12-03 14:14:51.78390245 +0000 UTC m=+321.446819383" Dec 03 14:14:51 crc kubenswrapper[4805]: I1203 14:14:51.795290 4805 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-route-controller-manager/route-controller-manager-685b95576d-cnn5h" podStartSLOduration=2.7952728589999998 podStartE2EDuration="2.795272859s" podCreationTimestamp="2025-12-03 14:14:49 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 14:14:51.79493676 +0000 UTC m=+321.457853683" watchObservedRunningTime="2025-12-03 14:14:51.795272859 +0000 UTC m=+321.458189792" Dec 03 14:14:52 crc kubenswrapper[4805]: I1203 14:14:52.701461 4805 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b54908df-f04f-4515-ab36-842f0aef4a99" path="/var/lib/kubelet/pods/b54908df-f04f-4515-ab36-842f0aef4a99/volumes" Dec 03 14:14:52 crc kubenswrapper[4805]: I1203 14:14:52.702451 4805 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="dfa6979a-2473-469b-ac56-9086e35a63f4" path="/var/lib/kubelet/pods/dfa6979a-2473-469b-ac56-9086e35a63f4/volumes" Dec 03 14:14:52 crc kubenswrapper[4805]: I1203 14:14:52.771475 4805 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-route-controller-manager/route-controller-manager-685b95576d-cnn5h" Dec 03 14:14:52 crc kubenswrapper[4805]: I1203 14:14:52.779421 4805 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-controller-manager/controller-manager-794995848b-95m55" Dec 03 14:14:53 crc kubenswrapper[4805]: I1203 14:14:53.162642 4805 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-685b95576d-cnn5h"] Dec 03 14:14:54 crc kubenswrapper[4805]: I1203 14:14:54.778756 4805 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-route-controller-manager/route-controller-manager-685b95576d-cnn5h" podUID="e3df4b05-07b7-4fe8-9a83-ecc033fa42b3" containerName="route-controller-manager" containerID="cri-o://e9cbdec719e14a87485c13a4dba1d66d153cfcc6084372df1d5445fae49d0116" gracePeriod=30 Dec 03 14:14:55 crc kubenswrapper[4805]: I1203 14:14:55.171727 4805 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-685b95576d-cnn5h" Dec 03 14:14:55 crc kubenswrapper[4805]: I1203 14:14:55.203573 4805 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-route-controller-manager/route-controller-manager-7c545ff4c9-4nnf9"] Dec 03 14:14:55 crc kubenswrapper[4805]: E1203 14:14:55.203767 4805 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e3df4b05-07b7-4fe8-9a83-ecc033fa42b3" containerName="route-controller-manager" Dec 03 14:14:55 crc kubenswrapper[4805]: I1203 14:14:55.203777 4805 state_mem.go:107] "Deleted CPUSet assignment" podUID="e3df4b05-07b7-4fe8-9a83-ecc033fa42b3" containerName="route-controller-manager" Dec 03 14:14:55 crc kubenswrapper[4805]: I1203 14:14:55.203876 4805 memory_manager.go:354] "RemoveStaleState removing state" podUID="e3df4b05-07b7-4fe8-9a83-ecc033fa42b3" containerName="route-controller-manager" Dec 03 14:14:55 crc kubenswrapper[4805]: I1203 14:14:55.204199 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-7c545ff4c9-4nnf9" Dec 03 14:14:55 crc kubenswrapper[4805]: I1203 14:14:55.247985 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-7c545ff4c9-4nnf9"] Dec 03 14:14:55 crc kubenswrapper[4805]: I1203 14:14:55.286002 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e3df4b05-07b7-4fe8-9a83-ecc033fa42b3-config\") pod \"e3df4b05-07b7-4fe8-9a83-ecc033fa42b3\" (UID: \"e3df4b05-07b7-4fe8-9a83-ecc033fa42b3\") " Dec 03 14:14:55 crc kubenswrapper[4805]: I1203 14:14:55.286337 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/e3df4b05-07b7-4fe8-9a83-ecc033fa42b3-client-ca\") pod \"e3df4b05-07b7-4fe8-9a83-ecc033fa42b3\" (UID: \"e3df4b05-07b7-4fe8-9a83-ecc033fa42b3\") " Dec 03 14:14:55 crc kubenswrapper[4805]: I1203 14:14:55.286461 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e3df4b05-07b7-4fe8-9a83-ecc033fa42b3-serving-cert\") pod \"e3df4b05-07b7-4fe8-9a83-ecc033fa42b3\" (UID: \"e3df4b05-07b7-4fe8-9a83-ecc033fa42b3\") " Dec 03 14:14:55 crc kubenswrapper[4805]: I1203 14:14:55.286507 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2tkhw\" (UniqueName: \"kubernetes.io/projected/e3df4b05-07b7-4fe8-9a83-ecc033fa42b3-kube-api-access-2tkhw\") pod \"e3df4b05-07b7-4fe8-9a83-ecc033fa42b3\" (UID: \"e3df4b05-07b7-4fe8-9a83-ecc033fa42b3\") " Dec 03 14:14:55 crc kubenswrapper[4805]: I1203 14:14:55.286803 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c6b1cd48-f5b4-49c0-89cd-baf2b1904f6a-config\") pod \"route-controller-manager-7c545ff4c9-4nnf9\" (UID: \"c6b1cd48-f5b4-49c0-89cd-baf2b1904f6a\") " pod="openshift-route-controller-manager/route-controller-manager-7c545ff4c9-4nnf9" Dec 03 14:14:55 crc kubenswrapper[4805]: I1203 14:14:55.286910 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7tb8m\" (UniqueName: \"kubernetes.io/projected/c6b1cd48-f5b4-49c0-89cd-baf2b1904f6a-kube-api-access-7tb8m\") pod \"route-controller-manager-7c545ff4c9-4nnf9\" (UID: \"c6b1cd48-f5b4-49c0-89cd-baf2b1904f6a\") " pod="openshift-route-controller-manager/route-controller-manager-7c545ff4c9-4nnf9" Dec 03 14:14:55 crc kubenswrapper[4805]: I1203 14:14:55.286933 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e3df4b05-07b7-4fe8-9a83-ecc033fa42b3-config" (OuterVolumeSpecName: "config") pod "e3df4b05-07b7-4fe8-9a83-ecc033fa42b3" (UID: "e3df4b05-07b7-4fe8-9a83-ecc033fa42b3"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 14:14:55 crc kubenswrapper[4805]: I1203 14:14:55.286952 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e3df4b05-07b7-4fe8-9a83-ecc033fa42b3-client-ca" (OuterVolumeSpecName: "client-ca") pod "e3df4b05-07b7-4fe8-9a83-ecc033fa42b3" (UID: "e3df4b05-07b7-4fe8-9a83-ecc033fa42b3"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 14:14:55 crc kubenswrapper[4805]: I1203 14:14:55.286959 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/c6b1cd48-f5b4-49c0-89cd-baf2b1904f6a-serving-cert\") pod \"route-controller-manager-7c545ff4c9-4nnf9\" (UID: \"c6b1cd48-f5b4-49c0-89cd-baf2b1904f6a\") " pod="openshift-route-controller-manager/route-controller-manager-7c545ff4c9-4nnf9" Dec 03 14:14:55 crc kubenswrapper[4805]: I1203 14:14:55.287012 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/c6b1cd48-f5b4-49c0-89cd-baf2b1904f6a-client-ca\") pod \"route-controller-manager-7c545ff4c9-4nnf9\" (UID: \"c6b1cd48-f5b4-49c0-89cd-baf2b1904f6a\") " pod="openshift-route-controller-manager/route-controller-manager-7c545ff4c9-4nnf9" Dec 03 14:14:55 crc kubenswrapper[4805]: I1203 14:14:55.287061 4805 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e3df4b05-07b7-4fe8-9a83-ecc033fa42b3-config\") on node \"crc\" DevicePath \"\"" Dec 03 14:14:55 crc kubenswrapper[4805]: I1203 14:14:55.287073 4805 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/e3df4b05-07b7-4fe8-9a83-ecc033fa42b3-client-ca\") on node \"crc\" DevicePath \"\"" Dec 03 14:14:55 crc kubenswrapper[4805]: I1203 14:14:55.292011 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e3df4b05-07b7-4fe8-9a83-ecc033fa42b3-kube-api-access-2tkhw" (OuterVolumeSpecName: "kube-api-access-2tkhw") pod "e3df4b05-07b7-4fe8-9a83-ecc033fa42b3" (UID: "e3df4b05-07b7-4fe8-9a83-ecc033fa42b3"). InnerVolumeSpecName "kube-api-access-2tkhw". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 14:14:55 crc kubenswrapper[4805]: I1203 14:14:55.295950 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e3df4b05-07b7-4fe8-9a83-ecc033fa42b3-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "e3df4b05-07b7-4fe8-9a83-ecc033fa42b3" (UID: "e3df4b05-07b7-4fe8-9a83-ecc033fa42b3"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 14:14:55 crc kubenswrapper[4805]: I1203 14:14:55.387557 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/c6b1cd48-f5b4-49c0-89cd-baf2b1904f6a-serving-cert\") pod \"route-controller-manager-7c545ff4c9-4nnf9\" (UID: \"c6b1cd48-f5b4-49c0-89cd-baf2b1904f6a\") " pod="openshift-route-controller-manager/route-controller-manager-7c545ff4c9-4nnf9" Dec 03 14:14:55 crc kubenswrapper[4805]: I1203 14:14:55.387607 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/c6b1cd48-f5b4-49c0-89cd-baf2b1904f6a-client-ca\") pod \"route-controller-manager-7c545ff4c9-4nnf9\" (UID: \"c6b1cd48-f5b4-49c0-89cd-baf2b1904f6a\") " pod="openshift-route-controller-manager/route-controller-manager-7c545ff4c9-4nnf9" Dec 03 14:14:55 crc kubenswrapper[4805]: I1203 14:14:55.387688 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c6b1cd48-f5b4-49c0-89cd-baf2b1904f6a-config\") pod \"route-controller-manager-7c545ff4c9-4nnf9\" (UID: \"c6b1cd48-f5b4-49c0-89cd-baf2b1904f6a\") " pod="openshift-route-controller-manager/route-controller-manager-7c545ff4c9-4nnf9" Dec 03 14:14:55 crc kubenswrapper[4805]: I1203 14:14:55.387721 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7tb8m\" (UniqueName: \"kubernetes.io/projected/c6b1cd48-f5b4-49c0-89cd-baf2b1904f6a-kube-api-access-7tb8m\") pod \"route-controller-manager-7c545ff4c9-4nnf9\" (UID: \"c6b1cd48-f5b4-49c0-89cd-baf2b1904f6a\") " pod="openshift-route-controller-manager/route-controller-manager-7c545ff4c9-4nnf9" Dec 03 14:14:55 crc kubenswrapper[4805]: I1203 14:14:55.387776 4805 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e3df4b05-07b7-4fe8-9a83-ecc033fa42b3-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 03 14:14:55 crc kubenswrapper[4805]: I1203 14:14:55.387791 4805 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2tkhw\" (UniqueName: \"kubernetes.io/projected/e3df4b05-07b7-4fe8-9a83-ecc033fa42b3-kube-api-access-2tkhw\") on node \"crc\" DevicePath \"\"" Dec 03 14:14:55 crc kubenswrapper[4805]: I1203 14:14:55.388733 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/c6b1cd48-f5b4-49c0-89cd-baf2b1904f6a-client-ca\") pod \"route-controller-manager-7c545ff4c9-4nnf9\" (UID: \"c6b1cd48-f5b4-49c0-89cd-baf2b1904f6a\") " pod="openshift-route-controller-manager/route-controller-manager-7c545ff4c9-4nnf9" Dec 03 14:14:55 crc kubenswrapper[4805]: I1203 14:14:55.389653 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c6b1cd48-f5b4-49c0-89cd-baf2b1904f6a-config\") pod \"route-controller-manager-7c545ff4c9-4nnf9\" (UID: \"c6b1cd48-f5b4-49c0-89cd-baf2b1904f6a\") " pod="openshift-route-controller-manager/route-controller-manager-7c545ff4c9-4nnf9" Dec 03 14:14:55 crc kubenswrapper[4805]: I1203 14:14:55.392322 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/c6b1cd48-f5b4-49c0-89cd-baf2b1904f6a-serving-cert\") pod \"route-controller-manager-7c545ff4c9-4nnf9\" (UID: \"c6b1cd48-f5b4-49c0-89cd-baf2b1904f6a\") " pod="openshift-route-controller-manager/route-controller-manager-7c545ff4c9-4nnf9" Dec 03 14:14:55 crc kubenswrapper[4805]: I1203 14:14:55.409990 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7tb8m\" (UniqueName: \"kubernetes.io/projected/c6b1cd48-f5b4-49c0-89cd-baf2b1904f6a-kube-api-access-7tb8m\") pod \"route-controller-manager-7c545ff4c9-4nnf9\" (UID: \"c6b1cd48-f5b4-49c0-89cd-baf2b1904f6a\") " pod="openshift-route-controller-manager/route-controller-manager-7c545ff4c9-4nnf9" Dec 03 14:14:55 crc kubenswrapper[4805]: I1203 14:14:55.574789 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-7c545ff4c9-4nnf9" Dec 03 14:14:55 crc kubenswrapper[4805]: I1203 14:14:55.794073 4805 generic.go:334] "Generic (PLEG): container finished" podID="e3df4b05-07b7-4fe8-9a83-ecc033fa42b3" containerID="e9cbdec719e14a87485c13a4dba1d66d153cfcc6084372df1d5445fae49d0116" exitCode=0 Dec 03 14:14:55 crc kubenswrapper[4805]: I1203 14:14:55.794136 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-685b95576d-cnn5h" event={"ID":"e3df4b05-07b7-4fe8-9a83-ecc033fa42b3","Type":"ContainerDied","Data":"e9cbdec719e14a87485c13a4dba1d66d153cfcc6084372df1d5445fae49d0116"} Dec 03 14:14:55 crc kubenswrapper[4805]: I1203 14:14:55.794161 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-685b95576d-cnn5h" event={"ID":"e3df4b05-07b7-4fe8-9a83-ecc033fa42b3","Type":"ContainerDied","Data":"c04492978f5d09dc3a6e4242ec0f21aaba593ed1cad0f5842addd41fc8d98f4e"} Dec 03 14:14:55 crc kubenswrapper[4805]: I1203 14:14:55.794179 4805 scope.go:117] "RemoveContainer" containerID="e9cbdec719e14a87485c13a4dba1d66d153cfcc6084372df1d5445fae49d0116" Dec 03 14:14:55 crc kubenswrapper[4805]: I1203 14:14:55.794260 4805 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-685b95576d-cnn5h" Dec 03 14:14:55 crc kubenswrapper[4805]: I1203 14:14:55.810579 4805 scope.go:117] "RemoveContainer" containerID="e9cbdec719e14a87485c13a4dba1d66d153cfcc6084372df1d5445fae49d0116" Dec 03 14:14:55 crc kubenswrapper[4805]: E1203 14:14:55.811056 4805 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e9cbdec719e14a87485c13a4dba1d66d153cfcc6084372df1d5445fae49d0116\": container with ID starting with e9cbdec719e14a87485c13a4dba1d66d153cfcc6084372df1d5445fae49d0116 not found: ID does not exist" containerID="e9cbdec719e14a87485c13a4dba1d66d153cfcc6084372df1d5445fae49d0116" Dec 03 14:14:55 crc kubenswrapper[4805]: I1203 14:14:55.811118 4805 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e9cbdec719e14a87485c13a4dba1d66d153cfcc6084372df1d5445fae49d0116"} err="failed to get container status \"e9cbdec719e14a87485c13a4dba1d66d153cfcc6084372df1d5445fae49d0116\": rpc error: code = NotFound desc = could not find container \"e9cbdec719e14a87485c13a4dba1d66d153cfcc6084372df1d5445fae49d0116\": container with ID starting with e9cbdec719e14a87485c13a4dba1d66d153cfcc6084372df1d5445fae49d0116 not found: ID does not exist" Dec 03 14:14:55 crc kubenswrapper[4805]: I1203 14:14:55.828685 4805 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-685b95576d-cnn5h"] Dec 03 14:14:55 crc kubenswrapper[4805]: I1203 14:14:55.831406 4805 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-685b95576d-cnn5h"] Dec 03 14:14:56 crc kubenswrapper[4805]: I1203 14:14:56.004481 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-7c545ff4c9-4nnf9"] Dec 03 14:14:56 crc kubenswrapper[4805]: I1203 14:14:56.701548 4805 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e3df4b05-07b7-4fe8-9a83-ecc033fa42b3" path="/var/lib/kubelet/pods/e3df4b05-07b7-4fe8-9a83-ecc033fa42b3/volumes" Dec 03 14:14:56 crc kubenswrapper[4805]: I1203 14:14:56.800118 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-7c545ff4c9-4nnf9" event={"ID":"c6b1cd48-f5b4-49c0-89cd-baf2b1904f6a","Type":"ContainerStarted","Data":"1f275fec9627bd8eefbbcd117ccb005f49b1e6d1f6f0787163d8984f290f61dd"} Dec 03 14:14:56 crc kubenswrapper[4805]: I1203 14:14:56.800518 4805 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-route-controller-manager/route-controller-manager-7c545ff4c9-4nnf9" Dec 03 14:14:56 crc kubenswrapper[4805]: I1203 14:14:56.800583 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-7c545ff4c9-4nnf9" event={"ID":"c6b1cd48-f5b4-49c0-89cd-baf2b1904f6a","Type":"ContainerStarted","Data":"22fde97a7319b70eef9fd4d41ea5f9d741f01d0cd5f7a0e2f71d1f5e97a344ba"} Dec 03 14:14:56 crc kubenswrapper[4805]: I1203 14:14:56.804641 4805 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-route-controller-manager/route-controller-manager-7c545ff4c9-4nnf9" Dec 03 14:14:56 crc kubenswrapper[4805]: I1203 14:14:56.817592 4805 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-route-controller-manager/route-controller-manager-7c545ff4c9-4nnf9" podStartSLOduration=3.817574263 podStartE2EDuration="3.817574263s" podCreationTimestamp="2025-12-03 14:14:53 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 14:14:56.814793067 +0000 UTC m=+326.477709990" watchObservedRunningTime="2025-12-03 14:14:56.817574263 +0000 UTC m=+326.480491206" Dec 03 14:14:58 crc kubenswrapper[4805]: I1203 14:14:58.705814 4805 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/image-registry-66df7c8f76-vltnj"] Dec 03 14:14:58 crc kubenswrapper[4805]: I1203 14:14:58.706717 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-66df7c8f76-vltnj" Dec 03 14:14:58 crc kubenswrapper[4805]: I1203 14:14:58.723377 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-66df7c8f76-vltnj"] Dec 03 14:14:58 crc kubenswrapper[4805]: I1203 14:14:58.830504 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-66df7c8f76-vltnj\" (UID: \"f1a37491-1665-4349-81cb-fc7268837984\") " pod="openshift-image-registry/image-registry-66df7c8f76-vltnj" Dec 03 14:14:58 crc kubenswrapper[4805]: I1203 14:14:58.830582 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/f1a37491-1665-4349-81cb-fc7268837984-installation-pull-secrets\") pod \"image-registry-66df7c8f76-vltnj\" (UID: \"f1a37491-1665-4349-81cb-fc7268837984\") " pod="openshift-image-registry/image-registry-66df7c8f76-vltnj" Dec 03 14:14:58 crc kubenswrapper[4805]: I1203 14:14:58.830614 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bqjc2\" (UniqueName: \"kubernetes.io/projected/f1a37491-1665-4349-81cb-fc7268837984-kube-api-access-bqjc2\") pod \"image-registry-66df7c8f76-vltnj\" (UID: \"f1a37491-1665-4349-81cb-fc7268837984\") " pod="openshift-image-registry/image-registry-66df7c8f76-vltnj" Dec 03 14:14:58 crc kubenswrapper[4805]: I1203 14:14:58.830659 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/f1a37491-1665-4349-81cb-fc7268837984-registry-tls\") pod \"image-registry-66df7c8f76-vltnj\" (UID: \"f1a37491-1665-4349-81cb-fc7268837984\") " pod="openshift-image-registry/image-registry-66df7c8f76-vltnj" Dec 03 14:14:58 crc kubenswrapper[4805]: I1203 14:14:58.830683 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/f1a37491-1665-4349-81cb-fc7268837984-registry-certificates\") pod \"image-registry-66df7c8f76-vltnj\" (UID: \"f1a37491-1665-4349-81cb-fc7268837984\") " pod="openshift-image-registry/image-registry-66df7c8f76-vltnj" Dec 03 14:14:58 crc kubenswrapper[4805]: I1203 14:14:58.830924 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/f1a37491-1665-4349-81cb-fc7268837984-bound-sa-token\") pod \"image-registry-66df7c8f76-vltnj\" (UID: \"f1a37491-1665-4349-81cb-fc7268837984\") " pod="openshift-image-registry/image-registry-66df7c8f76-vltnj" Dec 03 14:14:58 crc kubenswrapper[4805]: I1203 14:14:58.831162 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/f1a37491-1665-4349-81cb-fc7268837984-ca-trust-extracted\") pod \"image-registry-66df7c8f76-vltnj\" (UID: \"f1a37491-1665-4349-81cb-fc7268837984\") " pod="openshift-image-registry/image-registry-66df7c8f76-vltnj" Dec 03 14:14:58 crc kubenswrapper[4805]: I1203 14:14:58.831302 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/f1a37491-1665-4349-81cb-fc7268837984-trusted-ca\") pod \"image-registry-66df7c8f76-vltnj\" (UID: \"f1a37491-1665-4349-81cb-fc7268837984\") " pod="openshift-image-registry/image-registry-66df7c8f76-vltnj" Dec 03 14:14:58 crc kubenswrapper[4805]: I1203 14:14:58.852407 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-66df7c8f76-vltnj\" (UID: \"f1a37491-1665-4349-81cb-fc7268837984\") " pod="openshift-image-registry/image-registry-66df7c8f76-vltnj" Dec 03 14:14:58 crc kubenswrapper[4805]: I1203 14:14:58.933328 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/f1a37491-1665-4349-81cb-fc7268837984-installation-pull-secrets\") pod \"image-registry-66df7c8f76-vltnj\" (UID: \"f1a37491-1665-4349-81cb-fc7268837984\") " pod="openshift-image-registry/image-registry-66df7c8f76-vltnj" Dec 03 14:14:58 crc kubenswrapper[4805]: I1203 14:14:58.933408 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bqjc2\" (UniqueName: \"kubernetes.io/projected/f1a37491-1665-4349-81cb-fc7268837984-kube-api-access-bqjc2\") pod \"image-registry-66df7c8f76-vltnj\" (UID: \"f1a37491-1665-4349-81cb-fc7268837984\") " pod="openshift-image-registry/image-registry-66df7c8f76-vltnj" Dec 03 14:14:58 crc kubenswrapper[4805]: I1203 14:14:58.933464 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/f1a37491-1665-4349-81cb-fc7268837984-registry-tls\") pod \"image-registry-66df7c8f76-vltnj\" (UID: \"f1a37491-1665-4349-81cb-fc7268837984\") " pod="openshift-image-registry/image-registry-66df7c8f76-vltnj" Dec 03 14:14:58 crc kubenswrapper[4805]: I1203 14:14:58.933488 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/f1a37491-1665-4349-81cb-fc7268837984-registry-certificates\") pod \"image-registry-66df7c8f76-vltnj\" (UID: \"f1a37491-1665-4349-81cb-fc7268837984\") " pod="openshift-image-registry/image-registry-66df7c8f76-vltnj" Dec 03 14:14:58 crc kubenswrapper[4805]: I1203 14:14:58.933544 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/f1a37491-1665-4349-81cb-fc7268837984-bound-sa-token\") pod \"image-registry-66df7c8f76-vltnj\" (UID: \"f1a37491-1665-4349-81cb-fc7268837984\") " pod="openshift-image-registry/image-registry-66df7c8f76-vltnj" Dec 03 14:14:58 crc kubenswrapper[4805]: I1203 14:14:58.933587 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/f1a37491-1665-4349-81cb-fc7268837984-ca-trust-extracted\") pod \"image-registry-66df7c8f76-vltnj\" (UID: \"f1a37491-1665-4349-81cb-fc7268837984\") " pod="openshift-image-registry/image-registry-66df7c8f76-vltnj" Dec 03 14:14:58 crc kubenswrapper[4805]: I1203 14:14:58.933618 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/f1a37491-1665-4349-81cb-fc7268837984-trusted-ca\") pod \"image-registry-66df7c8f76-vltnj\" (UID: \"f1a37491-1665-4349-81cb-fc7268837984\") " pod="openshift-image-registry/image-registry-66df7c8f76-vltnj" Dec 03 14:14:58 crc kubenswrapper[4805]: I1203 14:14:58.935551 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/f1a37491-1665-4349-81cb-fc7268837984-trusted-ca\") pod \"image-registry-66df7c8f76-vltnj\" (UID: \"f1a37491-1665-4349-81cb-fc7268837984\") " pod="openshift-image-registry/image-registry-66df7c8f76-vltnj" Dec 03 14:14:58 crc kubenswrapper[4805]: I1203 14:14:58.936027 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/f1a37491-1665-4349-81cb-fc7268837984-ca-trust-extracted\") pod \"image-registry-66df7c8f76-vltnj\" (UID: \"f1a37491-1665-4349-81cb-fc7268837984\") " pod="openshift-image-registry/image-registry-66df7c8f76-vltnj" Dec 03 14:14:58 crc kubenswrapper[4805]: I1203 14:14:58.936410 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/f1a37491-1665-4349-81cb-fc7268837984-registry-certificates\") pod \"image-registry-66df7c8f76-vltnj\" (UID: \"f1a37491-1665-4349-81cb-fc7268837984\") " pod="openshift-image-registry/image-registry-66df7c8f76-vltnj" Dec 03 14:14:58 crc kubenswrapper[4805]: I1203 14:14:58.946120 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/f1a37491-1665-4349-81cb-fc7268837984-registry-tls\") pod \"image-registry-66df7c8f76-vltnj\" (UID: \"f1a37491-1665-4349-81cb-fc7268837984\") " pod="openshift-image-registry/image-registry-66df7c8f76-vltnj" Dec 03 14:14:58 crc kubenswrapper[4805]: I1203 14:14:58.946142 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/f1a37491-1665-4349-81cb-fc7268837984-installation-pull-secrets\") pod \"image-registry-66df7c8f76-vltnj\" (UID: \"f1a37491-1665-4349-81cb-fc7268837984\") " pod="openshift-image-registry/image-registry-66df7c8f76-vltnj" Dec 03 14:14:58 crc kubenswrapper[4805]: I1203 14:14:58.952272 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bqjc2\" (UniqueName: \"kubernetes.io/projected/f1a37491-1665-4349-81cb-fc7268837984-kube-api-access-bqjc2\") pod \"image-registry-66df7c8f76-vltnj\" (UID: \"f1a37491-1665-4349-81cb-fc7268837984\") " pod="openshift-image-registry/image-registry-66df7c8f76-vltnj" Dec 03 14:14:58 crc kubenswrapper[4805]: I1203 14:14:58.963366 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/f1a37491-1665-4349-81cb-fc7268837984-bound-sa-token\") pod \"image-registry-66df7c8f76-vltnj\" (UID: \"f1a37491-1665-4349-81cb-fc7268837984\") " pod="openshift-image-registry/image-registry-66df7c8f76-vltnj" Dec 03 14:14:59 crc kubenswrapper[4805]: I1203 14:14:59.021084 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-66df7c8f76-vltnj" Dec 03 14:14:59 crc kubenswrapper[4805]: I1203 14:14:59.467589 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-66df7c8f76-vltnj"] Dec 03 14:14:59 crc kubenswrapper[4805]: W1203 14:14:59.477358 4805 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf1a37491_1665_4349_81cb_fc7268837984.slice/crio-2b8265efdd6bbd59ddc14f46e68cde976e521bb17c0e4cc2831b943483c907f3 WatchSource:0}: Error finding container 2b8265efdd6bbd59ddc14f46e68cde976e521bb17c0e4cc2831b943483c907f3: Status 404 returned error can't find the container with id 2b8265efdd6bbd59ddc14f46e68cde976e521bb17c0e4cc2831b943483c907f3 Dec 03 14:14:59 crc kubenswrapper[4805]: I1203 14:14:59.820528 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-66df7c8f76-vltnj" event={"ID":"f1a37491-1665-4349-81cb-fc7268837984","Type":"ContainerStarted","Data":"e31b06258d01066cfd3661b468b9d30e42a03ae4cb128085be9146c74f49e357"} Dec 03 14:14:59 crc kubenswrapper[4805]: I1203 14:14:59.820570 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-66df7c8f76-vltnj" event={"ID":"f1a37491-1665-4349-81cb-fc7268837984","Type":"ContainerStarted","Data":"2b8265efdd6bbd59ddc14f46e68cde976e521bb17c0e4cc2831b943483c907f3"} Dec 03 14:14:59 crc kubenswrapper[4805]: I1203 14:14:59.821231 4805 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-image-registry/image-registry-66df7c8f76-vltnj" Dec 03 14:14:59 crc kubenswrapper[4805]: I1203 14:14:59.837137 4805 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/image-registry-66df7c8f76-vltnj" podStartSLOduration=1.837118696 podStartE2EDuration="1.837118696s" podCreationTimestamp="2025-12-03 14:14:58 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 14:14:59.835954434 +0000 UTC m=+329.498871357" watchObservedRunningTime="2025-12-03 14:14:59.837118696 +0000 UTC m=+329.500035619" Dec 03 14:15:00 crc kubenswrapper[4805]: I1203 14:15:00.169903 4805 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29412855-8rxc9"] Dec 03 14:15:00 crc kubenswrapper[4805]: I1203 14:15:00.171129 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29412855-8rxc9" Dec 03 14:15:00 crc kubenswrapper[4805]: I1203 14:15:00.173028 4805 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Dec 03 14:15:00 crc kubenswrapper[4805]: I1203 14:15:00.175048 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Dec 03 14:15:00 crc kubenswrapper[4805]: I1203 14:15:00.179091 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29412855-8rxc9"] Dec 03 14:15:00 crc kubenswrapper[4805]: I1203 14:15:00.360082 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/7c461d93-d743-4981-b160-48e927a4d443-config-volume\") pod \"collect-profiles-29412855-8rxc9\" (UID: \"7c461d93-d743-4981-b160-48e927a4d443\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29412855-8rxc9" Dec 03 14:15:00 crc kubenswrapper[4805]: I1203 14:15:00.360180 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/7c461d93-d743-4981-b160-48e927a4d443-secret-volume\") pod \"collect-profiles-29412855-8rxc9\" (UID: \"7c461d93-d743-4981-b160-48e927a4d443\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29412855-8rxc9" Dec 03 14:15:00 crc kubenswrapper[4805]: I1203 14:15:00.360389 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7fdnh\" (UniqueName: \"kubernetes.io/projected/7c461d93-d743-4981-b160-48e927a4d443-kube-api-access-7fdnh\") pod \"collect-profiles-29412855-8rxc9\" (UID: \"7c461d93-d743-4981-b160-48e927a4d443\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29412855-8rxc9" Dec 03 14:15:00 crc kubenswrapper[4805]: I1203 14:15:00.461886 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/7c461d93-d743-4981-b160-48e927a4d443-secret-volume\") pod \"collect-profiles-29412855-8rxc9\" (UID: \"7c461d93-d743-4981-b160-48e927a4d443\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29412855-8rxc9" Dec 03 14:15:00 crc kubenswrapper[4805]: I1203 14:15:00.462112 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7fdnh\" (UniqueName: \"kubernetes.io/projected/7c461d93-d743-4981-b160-48e927a4d443-kube-api-access-7fdnh\") pod \"collect-profiles-29412855-8rxc9\" (UID: \"7c461d93-d743-4981-b160-48e927a4d443\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29412855-8rxc9" Dec 03 14:15:00 crc kubenswrapper[4805]: I1203 14:15:00.462170 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/7c461d93-d743-4981-b160-48e927a4d443-config-volume\") pod \"collect-profiles-29412855-8rxc9\" (UID: \"7c461d93-d743-4981-b160-48e927a4d443\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29412855-8rxc9" Dec 03 14:15:00 crc kubenswrapper[4805]: I1203 14:15:00.463922 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/7c461d93-d743-4981-b160-48e927a4d443-config-volume\") pod \"collect-profiles-29412855-8rxc9\" (UID: \"7c461d93-d743-4981-b160-48e927a4d443\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29412855-8rxc9" Dec 03 14:15:00 crc kubenswrapper[4805]: I1203 14:15:00.473770 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/7c461d93-d743-4981-b160-48e927a4d443-secret-volume\") pod \"collect-profiles-29412855-8rxc9\" (UID: \"7c461d93-d743-4981-b160-48e927a4d443\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29412855-8rxc9" Dec 03 14:15:00 crc kubenswrapper[4805]: I1203 14:15:00.493275 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7fdnh\" (UniqueName: \"kubernetes.io/projected/7c461d93-d743-4981-b160-48e927a4d443-kube-api-access-7fdnh\") pod \"collect-profiles-29412855-8rxc9\" (UID: \"7c461d93-d743-4981-b160-48e927a4d443\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29412855-8rxc9" Dec 03 14:15:00 crc kubenswrapper[4805]: I1203 14:15:00.497106 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29412855-8rxc9" Dec 03 14:15:00 crc kubenswrapper[4805]: I1203 14:15:00.934499 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29412855-8rxc9"] Dec 03 14:15:00 crc kubenswrapper[4805]: W1203 14:15:00.938235 4805 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod7c461d93_d743_4981_b160_48e927a4d443.slice/crio-72976161ded474cc33328eccdde6e3ff3a7e6930c0134717d28f0b8748596e1a WatchSource:0}: Error finding container 72976161ded474cc33328eccdde6e3ff3a7e6930c0134717d28f0b8748596e1a: Status 404 returned error can't find the container with id 72976161ded474cc33328eccdde6e3ff3a7e6930c0134717d28f0b8748596e1a Dec 03 14:15:01 crc kubenswrapper[4805]: I1203 14:15:01.831159 4805 generic.go:334] "Generic (PLEG): container finished" podID="7c461d93-d743-4981-b160-48e927a4d443" containerID="b5b408b815391ac337fc70c2d75fd6ceee87e7854c2a7696867fcf21ebaf87cb" exitCode=0 Dec 03 14:15:01 crc kubenswrapper[4805]: I1203 14:15:01.831239 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29412855-8rxc9" event={"ID":"7c461d93-d743-4981-b160-48e927a4d443","Type":"ContainerDied","Data":"b5b408b815391ac337fc70c2d75fd6ceee87e7854c2a7696867fcf21ebaf87cb"} Dec 03 14:15:01 crc kubenswrapper[4805]: I1203 14:15:01.831626 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29412855-8rxc9" event={"ID":"7c461d93-d743-4981-b160-48e927a4d443","Type":"ContainerStarted","Data":"72976161ded474cc33328eccdde6e3ff3a7e6930c0134717d28f0b8748596e1a"} Dec 03 14:15:03 crc kubenswrapper[4805]: I1203 14:15:03.185392 4805 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29412855-8rxc9" Dec 03 14:15:03 crc kubenswrapper[4805]: I1203 14:15:03.304190 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7fdnh\" (UniqueName: \"kubernetes.io/projected/7c461d93-d743-4981-b160-48e927a4d443-kube-api-access-7fdnh\") pod \"7c461d93-d743-4981-b160-48e927a4d443\" (UID: \"7c461d93-d743-4981-b160-48e927a4d443\") " Dec 03 14:15:03 crc kubenswrapper[4805]: I1203 14:15:03.304311 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/7c461d93-d743-4981-b160-48e927a4d443-secret-volume\") pod \"7c461d93-d743-4981-b160-48e927a4d443\" (UID: \"7c461d93-d743-4981-b160-48e927a4d443\") " Dec 03 14:15:03 crc kubenswrapper[4805]: I1203 14:15:03.304375 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/7c461d93-d743-4981-b160-48e927a4d443-config-volume\") pod \"7c461d93-d743-4981-b160-48e927a4d443\" (UID: \"7c461d93-d743-4981-b160-48e927a4d443\") " Dec 03 14:15:03 crc kubenswrapper[4805]: I1203 14:15:03.305604 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7c461d93-d743-4981-b160-48e927a4d443-config-volume" (OuterVolumeSpecName: "config-volume") pod "7c461d93-d743-4981-b160-48e927a4d443" (UID: "7c461d93-d743-4981-b160-48e927a4d443"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 14:15:03 crc kubenswrapper[4805]: I1203 14:15:03.309572 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7c461d93-d743-4981-b160-48e927a4d443-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "7c461d93-d743-4981-b160-48e927a4d443" (UID: "7c461d93-d743-4981-b160-48e927a4d443"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 14:15:03 crc kubenswrapper[4805]: I1203 14:15:03.310884 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7c461d93-d743-4981-b160-48e927a4d443-kube-api-access-7fdnh" (OuterVolumeSpecName: "kube-api-access-7fdnh") pod "7c461d93-d743-4981-b160-48e927a4d443" (UID: "7c461d93-d743-4981-b160-48e927a4d443"). InnerVolumeSpecName "kube-api-access-7fdnh". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 14:15:03 crc kubenswrapper[4805]: I1203 14:15:03.405539 4805 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7fdnh\" (UniqueName: \"kubernetes.io/projected/7c461d93-d743-4981-b160-48e927a4d443-kube-api-access-7fdnh\") on node \"crc\" DevicePath \"\"" Dec 03 14:15:03 crc kubenswrapper[4805]: I1203 14:15:03.405591 4805 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/7c461d93-d743-4981-b160-48e927a4d443-secret-volume\") on node \"crc\" DevicePath \"\"" Dec 03 14:15:03 crc kubenswrapper[4805]: I1203 14:15:03.405613 4805 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/7c461d93-d743-4981-b160-48e927a4d443-config-volume\") on node \"crc\" DevicePath \"\"" Dec 03 14:15:03 crc kubenswrapper[4805]: I1203 14:15:03.845600 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29412855-8rxc9" event={"ID":"7c461d93-d743-4981-b160-48e927a4d443","Type":"ContainerDied","Data":"72976161ded474cc33328eccdde6e3ff3a7e6930c0134717d28f0b8748596e1a"} Dec 03 14:15:03 crc kubenswrapper[4805]: I1203 14:15:03.845642 4805 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="72976161ded474cc33328eccdde6e3ff3a7e6930c0134717d28f0b8748596e1a" Dec 03 14:15:03 crc kubenswrapper[4805]: I1203 14:15:03.845739 4805 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29412855-8rxc9" Dec 03 14:15:13 crc kubenswrapper[4805]: I1203 14:15:13.069199 4805 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-794995848b-95m55"] Dec 03 14:15:13 crc kubenswrapper[4805]: I1203 14:15:13.070118 4805 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-controller-manager/controller-manager-794995848b-95m55" podUID="46633c53-ce91-4bbe-a650-b4138cfb8ef1" containerName="controller-manager" containerID="cri-o://f7cac3bc95c2d86e6002ac07d2b1c4a47a19ee1e04d0d49d6214aa3dadf84082" gracePeriod=30 Dec 03 14:15:13 crc kubenswrapper[4805]: I1203 14:15:13.632789 4805 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-794995848b-95m55" Dec 03 14:15:13 crc kubenswrapper[4805]: I1203 14:15:13.808173 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hr9lp\" (UniqueName: \"kubernetes.io/projected/46633c53-ce91-4bbe-a650-b4138cfb8ef1-kube-api-access-hr9lp\") pod \"46633c53-ce91-4bbe-a650-b4138cfb8ef1\" (UID: \"46633c53-ce91-4bbe-a650-b4138cfb8ef1\") " Dec 03 14:15:13 crc kubenswrapper[4805]: I1203 14:15:13.808232 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/46633c53-ce91-4bbe-a650-b4138cfb8ef1-client-ca\") pod \"46633c53-ce91-4bbe-a650-b4138cfb8ef1\" (UID: \"46633c53-ce91-4bbe-a650-b4138cfb8ef1\") " Dec 03 14:15:13 crc kubenswrapper[4805]: I1203 14:15:13.808263 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/46633c53-ce91-4bbe-a650-b4138cfb8ef1-serving-cert\") pod \"46633c53-ce91-4bbe-a650-b4138cfb8ef1\" (UID: \"46633c53-ce91-4bbe-a650-b4138cfb8ef1\") " Dec 03 14:15:13 crc kubenswrapper[4805]: I1203 14:15:13.808290 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/46633c53-ce91-4bbe-a650-b4138cfb8ef1-proxy-ca-bundles\") pod \"46633c53-ce91-4bbe-a650-b4138cfb8ef1\" (UID: \"46633c53-ce91-4bbe-a650-b4138cfb8ef1\") " Dec 03 14:15:13 crc kubenswrapper[4805]: I1203 14:15:13.808314 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/46633c53-ce91-4bbe-a650-b4138cfb8ef1-config\") pod \"46633c53-ce91-4bbe-a650-b4138cfb8ef1\" (UID: \"46633c53-ce91-4bbe-a650-b4138cfb8ef1\") " Dec 03 14:15:13 crc kubenswrapper[4805]: I1203 14:15:13.809005 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/46633c53-ce91-4bbe-a650-b4138cfb8ef1-client-ca" (OuterVolumeSpecName: "client-ca") pod "46633c53-ce91-4bbe-a650-b4138cfb8ef1" (UID: "46633c53-ce91-4bbe-a650-b4138cfb8ef1"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 14:15:13 crc kubenswrapper[4805]: I1203 14:15:13.809471 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/46633c53-ce91-4bbe-a650-b4138cfb8ef1-config" (OuterVolumeSpecName: "config") pod "46633c53-ce91-4bbe-a650-b4138cfb8ef1" (UID: "46633c53-ce91-4bbe-a650-b4138cfb8ef1"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 14:15:13 crc kubenswrapper[4805]: I1203 14:15:13.810532 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/46633c53-ce91-4bbe-a650-b4138cfb8ef1-proxy-ca-bundles" (OuterVolumeSpecName: "proxy-ca-bundles") pod "46633c53-ce91-4bbe-a650-b4138cfb8ef1" (UID: "46633c53-ce91-4bbe-a650-b4138cfb8ef1"). InnerVolumeSpecName "proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 14:15:13 crc kubenswrapper[4805]: I1203 14:15:13.814028 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/46633c53-ce91-4bbe-a650-b4138cfb8ef1-kube-api-access-hr9lp" (OuterVolumeSpecName: "kube-api-access-hr9lp") pod "46633c53-ce91-4bbe-a650-b4138cfb8ef1" (UID: "46633c53-ce91-4bbe-a650-b4138cfb8ef1"). InnerVolumeSpecName "kube-api-access-hr9lp". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 14:15:13 crc kubenswrapper[4805]: I1203 14:15:13.815955 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/46633c53-ce91-4bbe-a650-b4138cfb8ef1-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "46633c53-ce91-4bbe-a650-b4138cfb8ef1" (UID: "46633c53-ce91-4bbe-a650-b4138cfb8ef1"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 14:15:13 crc kubenswrapper[4805]: I1203 14:15:13.904062 4805 generic.go:334] "Generic (PLEG): container finished" podID="46633c53-ce91-4bbe-a650-b4138cfb8ef1" containerID="f7cac3bc95c2d86e6002ac07d2b1c4a47a19ee1e04d0d49d6214aa3dadf84082" exitCode=0 Dec 03 14:15:13 crc kubenswrapper[4805]: I1203 14:15:13.904103 4805 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-794995848b-95m55" Dec 03 14:15:13 crc kubenswrapper[4805]: I1203 14:15:13.904125 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-794995848b-95m55" event={"ID":"46633c53-ce91-4bbe-a650-b4138cfb8ef1","Type":"ContainerDied","Data":"f7cac3bc95c2d86e6002ac07d2b1c4a47a19ee1e04d0d49d6214aa3dadf84082"} Dec 03 14:15:13 crc kubenswrapper[4805]: I1203 14:15:13.904162 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-794995848b-95m55" event={"ID":"46633c53-ce91-4bbe-a650-b4138cfb8ef1","Type":"ContainerDied","Data":"2c699d8733dd380a42b8e7f1babee8c964d71d7e68a13e6d6e02f8cdfecf3fa8"} Dec 03 14:15:13 crc kubenswrapper[4805]: I1203 14:15:13.904190 4805 scope.go:117] "RemoveContainer" containerID="f7cac3bc95c2d86e6002ac07d2b1c4a47a19ee1e04d0d49d6214aa3dadf84082" Dec 03 14:15:13 crc kubenswrapper[4805]: I1203 14:15:13.909726 4805 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/46633c53-ce91-4bbe-a650-b4138cfb8ef1-client-ca\") on node \"crc\" DevicePath \"\"" Dec 03 14:15:13 crc kubenswrapper[4805]: I1203 14:15:13.909763 4805 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/46633c53-ce91-4bbe-a650-b4138cfb8ef1-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 03 14:15:13 crc kubenswrapper[4805]: I1203 14:15:13.909782 4805 reconciler_common.go:293] "Volume detached for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/46633c53-ce91-4bbe-a650-b4138cfb8ef1-proxy-ca-bundles\") on node \"crc\" DevicePath \"\"" Dec 03 14:15:13 crc kubenswrapper[4805]: I1203 14:15:13.909802 4805 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/46633c53-ce91-4bbe-a650-b4138cfb8ef1-config\") on node \"crc\" DevicePath \"\"" Dec 03 14:15:13 crc kubenswrapper[4805]: I1203 14:15:13.910065 4805 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hr9lp\" (UniqueName: \"kubernetes.io/projected/46633c53-ce91-4bbe-a650-b4138cfb8ef1-kube-api-access-hr9lp\") on node \"crc\" DevicePath \"\"" Dec 03 14:15:13 crc kubenswrapper[4805]: I1203 14:15:13.942905 4805 scope.go:117] "RemoveContainer" containerID="f7cac3bc95c2d86e6002ac07d2b1c4a47a19ee1e04d0d49d6214aa3dadf84082" Dec 03 14:15:13 crc kubenswrapper[4805]: E1203 14:15:13.943511 4805 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f7cac3bc95c2d86e6002ac07d2b1c4a47a19ee1e04d0d49d6214aa3dadf84082\": container with ID starting with f7cac3bc95c2d86e6002ac07d2b1c4a47a19ee1e04d0d49d6214aa3dadf84082 not found: ID does not exist" containerID="f7cac3bc95c2d86e6002ac07d2b1c4a47a19ee1e04d0d49d6214aa3dadf84082" Dec 03 14:15:13 crc kubenswrapper[4805]: I1203 14:15:13.943554 4805 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f7cac3bc95c2d86e6002ac07d2b1c4a47a19ee1e04d0d49d6214aa3dadf84082"} err="failed to get container status \"f7cac3bc95c2d86e6002ac07d2b1c4a47a19ee1e04d0d49d6214aa3dadf84082\": rpc error: code = NotFound desc = could not find container \"f7cac3bc95c2d86e6002ac07d2b1c4a47a19ee1e04d0d49d6214aa3dadf84082\": container with ID starting with f7cac3bc95c2d86e6002ac07d2b1c4a47a19ee1e04d0d49d6214aa3dadf84082 not found: ID does not exist" Dec 03 14:15:13 crc kubenswrapper[4805]: I1203 14:15:13.944545 4805 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-794995848b-95m55"] Dec 03 14:15:13 crc kubenswrapper[4805]: I1203 14:15:13.951440 4805 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-controller-manager/controller-manager-794995848b-95m55"] Dec 03 14:15:14 crc kubenswrapper[4805]: I1203 14:15:14.707891 4805 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="46633c53-ce91-4bbe-a650-b4138cfb8ef1" path="/var/lib/kubelet/pods/46633c53-ce91-4bbe-a650-b4138cfb8ef1/volumes" Dec 03 14:15:15 crc kubenswrapper[4805]: I1203 14:15:15.000097 4805 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager/controller-manager-797bfc7f65-k7wn4"] Dec 03 14:15:15 crc kubenswrapper[4805]: E1203 14:15:15.000386 4805 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="46633c53-ce91-4bbe-a650-b4138cfb8ef1" containerName="controller-manager" Dec 03 14:15:15 crc kubenswrapper[4805]: I1203 14:15:15.000399 4805 state_mem.go:107] "Deleted CPUSet assignment" podUID="46633c53-ce91-4bbe-a650-b4138cfb8ef1" containerName="controller-manager" Dec 03 14:15:15 crc kubenswrapper[4805]: E1203 14:15:15.000412 4805 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7c461d93-d743-4981-b160-48e927a4d443" containerName="collect-profiles" Dec 03 14:15:15 crc kubenswrapper[4805]: I1203 14:15:15.000420 4805 state_mem.go:107] "Deleted CPUSet assignment" podUID="7c461d93-d743-4981-b160-48e927a4d443" containerName="collect-profiles" Dec 03 14:15:15 crc kubenswrapper[4805]: I1203 14:15:15.000519 4805 memory_manager.go:354] "RemoveStaleState removing state" podUID="7c461d93-d743-4981-b160-48e927a4d443" containerName="collect-profiles" Dec 03 14:15:15 crc kubenswrapper[4805]: I1203 14:15:15.000532 4805 memory_manager.go:354] "RemoveStaleState removing state" podUID="46633c53-ce91-4bbe-a650-b4138cfb8ef1" containerName="controller-manager" Dec 03 14:15:15 crc kubenswrapper[4805]: I1203 14:15:15.000943 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-797bfc7f65-k7wn4" Dec 03 14:15:15 crc kubenswrapper[4805]: I1203 14:15:15.003570 4805 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-service-ca.crt" Dec 03 14:15:15 crc kubenswrapper[4805]: I1203 14:15:15.004978 4805 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"client-ca" Dec 03 14:15:15 crc kubenswrapper[4805]: I1203 14:15:15.005082 4805 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"kube-root-ca.crt" Dec 03 14:15:15 crc kubenswrapper[4805]: I1203 14:15:15.005806 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"openshift-controller-manager-sa-dockercfg-msq4c" Dec 03 14:15:15 crc kubenswrapper[4805]: I1203 14:15:15.006449 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"serving-cert" Dec 03 14:15:15 crc kubenswrapper[4805]: I1203 14:15:15.006696 4805 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"config" Dec 03 14:15:15 crc kubenswrapper[4805]: I1203 14:15:15.006753 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-797bfc7f65-k7wn4"] Dec 03 14:15:15 crc kubenswrapper[4805]: I1203 14:15:15.024568 4805 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-global-ca" Dec 03 14:15:15 crc kubenswrapper[4805]: I1203 14:15:15.125578 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/98b69e82-9cbe-4d01-9a43-e8d94f911a3f-serving-cert\") pod \"controller-manager-797bfc7f65-k7wn4\" (UID: \"98b69e82-9cbe-4d01-9a43-e8d94f911a3f\") " pod="openshift-controller-manager/controller-manager-797bfc7f65-k7wn4" Dec 03 14:15:15 crc kubenswrapper[4805]: I1203 14:15:15.125975 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/98b69e82-9cbe-4d01-9a43-e8d94f911a3f-proxy-ca-bundles\") pod \"controller-manager-797bfc7f65-k7wn4\" (UID: \"98b69e82-9cbe-4d01-9a43-e8d94f911a3f\") " pod="openshift-controller-manager/controller-manager-797bfc7f65-k7wn4" Dec 03 14:15:15 crc kubenswrapper[4805]: I1203 14:15:15.126042 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/98b69e82-9cbe-4d01-9a43-e8d94f911a3f-config\") pod \"controller-manager-797bfc7f65-k7wn4\" (UID: \"98b69e82-9cbe-4d01-9a43-e8d94f911a3f\") " pod="openshift-controller-manager/controller-manager-797bfc7f65-k7wn4" Dec 03 14:15:15 crc kubenswrapper[4805]: I1203 14:15:15.126106 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/98b69e82-9cbe-4d01-9a43-e8d94f911a3f-client-ca\") pod \"controller-manager-797bfc7f65-k7wn4\" (UID: \"98b69e82-9cbe-4d01-9a43-e8d94f911a3f\") " pod="openshift-controller-manager/controller-manager-797bfc7f65-k7wn4" Dec 03 14:15:15 crc kubenswrapper[4805]: I1203 14:15:15.126173 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-plzh2\" (UniqueName: \"kubernetes.io/projected/98b69e82-9cbe-4d01-9a43-e8d94f911a3f-kube-api-access-plzh2\") pod \"controller-manager-797bfc7f65-k7wn4\" (UID: \"98b69e82-9cbe-4d01-9a43-e8d94f911a3f\") " pod="openshift-controller-manager/controller-manager-797bfc7f65-k7wn4" Dec 03 14:15:15 crc kubenswrapper[4805]: I1203 14:15:15.227530 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-plzh2\" (UniqueName: \"kubernetes.io/projected/98b69e82-9cbe-4d01-9a43-e8d94f911a3f-kube-api-access-plzh2\") pod \"controller-manager-797bfc7f65-k7wn4\" (UID: \"98b69e82-9cbe-4d01-9a43-e8d94f911a3f\") " pod="openshift-controller-manager/controller-manager-797bfc7f65-k7wn4" Dec 03 14:15:15 crc kubenswrapper[4805]: I1203 14:15:15.227612 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/98b69e82-9cbe-4d01-9a43-e8d94f911a3f-serving-cert\") pod \"controller-manager-797bfc7f65-k7wn4\" (UID: \"98b69e82-9cbe-4d01-9a43-e8d94f911a3f\") " pod="openshift-controller-manager/controller-manager-797bfc7f65-k7wn4" Dec 03 14:15:15 crc kubenswrapper[4805]: I1203 14:15:15.227674 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/98b69e82-9cbe-4d01-9a43-e8d94f911a3f-proxy-ca-bundles\") pod \"controller-manager-797bfc7f65-k7wn4\" (UID: \"98b69e82-9cbe-4d01-9a43-e8d94f911a3f\") " pod="openshift-controller-manager/controller-manager-797bfc7f65-k7wn4" Dec 03 14:15:15 crc kubenswrapper[4805]: I1203 14:15:15.235277 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/98b69e82-9cbe-4d01-9a43-e8d94f911a3f-config\") pod \"controller-manager-797bfc7f65-k7wn4\" (UID: \"98b69e82-9cbe-4d01-9a43-e8d94f911a3f\") " pod="openshift-controller-manager/controller-manager-797bfc7f65-k7wn4" Dec 03 14:15:15 crc kubenswrapper[4805]: I1203 14:15:15.235525 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/98b69e82-9cbe-4d01-9a43-e8d94f911a3f-client-ca\") pod \"controller-manager-797bfc7f65-k7wn4\" (UID: \"98b69e82-9cbe-4d01-9a43-e8d94f911a3f\") " pod="openshift-controller-manager/controller-manager-797bfc7f65-k7wn4" Dec 03 14:15:15 crc kubenswrapper[4805]: I1203 14:15:15.236825 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/98b69e82-9cbe-4d01-9a43-e8d94f911a3f-config\") pod \"controller-manager-797bfc7f65-k7wn4\" (UID: \"98b69e82-9cbe-4d01-9a43-e8d94f911a3f\") " pod="openshift-controller-manager/controller-manager-797bfc7f65-k7wn4" Dec 03 14:15:15 crc kubenswrapper[4805]: I1203 14:15:15.237478 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/98b69e82-9cbe-4d01-9a43-e8d94f911a3f-client-ca\") pod \"controller-manager-797bfc7f65-k7wn4\" (UID: \"98b69e82-9cbe-4d01-9a43-e8d94f911a3f\") " pod="openshift-controller-manager/controller-manager-797bfc7f65-k7wn4" Dec 03 14:15:15 crc kubenswrapper[4805]: I1203 14:15:15.244684 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/98b69e82-9cbe-4d01-9a43-e8d94f911a3f-serving-cert\") pod \"controller-manager-797bfc7f65-k7wn4\" (UID: \"98b69e82-9cbe-4d01-9a43-e8d94f911a3f\") " pod="openshift-controller-manager/controller-manager-797bfc7f65-k7wn4" Dec 03 14:15:15 crc kubenswrapper[4805]: I1203 14:15:15.248196 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/98b69e82-9cbe-4d01-9a43-e8d94f911a3f-proxy-ca-bundles\") pod \"controller-manager-797bfc7f65-k7wn4\" (UID: \"98b69e82-9cbe-4d01-9a43-e8d94f911a3f\") " pod="openshift-controller-manager/controller-manager-797bfc7f65-k7wn4" Dec 03 14:15:15 crc kubenswrapper[4805]: I1203 14:15:15.271179 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-plzh2\" (UniqueName: \"kubernetes.io/projected/98b69e82-9cbe-4d01-9a43-e8d94f911a3f-kube-api-access-plzh2\") pod \"controller-manager-797bfc7f65-k7wn4\" (UID: \"98b69e82-9cbe-4d01-9a43-e8d94f911a3f\") " pod="openshift-controller-manager/controller-manager-797bfc7f65-k7wn4" Dec 03 14:15:15 crc kubenswrapper[4805]: I1203 14:15:15.335684 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-797bfc7f65-k7wn4" Dec 03 14:15:15 crc kubenswrapper[4805]: I1203 14:15:15.768939 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-797bfc7f65-k7wn4"] Dec 03 14:15:15 crc kubenswrapper[4805]: W1203 14:15:15.777060 4805 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod98b69e82_9cbe_4d01_9a43_e8d94f911a3f.slice/crio-b955faa5b7d92de133d499b353e6aeca6a4850940a7e6e8ab5893a204e685417 WatchSource:0}: Error finding container b955faa5b7d92de133d499b353e6aeca6a4850940a7e6e8ab5893a204e685417: Status 404 returned error can't find the container with id b955faa5b7d92de133d499b353e6aeca6a4850940a7e6e8ab5893a204e685417 Dec 03 14:15:15 crc kubenswrapper[4805]: I1203 14:15:15.916857 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-797bfc7f65-k7wn4" event={"ID":"98b69e82-9cbe-4d01-9a43-e8d94f911a3f","Type":"ContainerStarted","Data":"b955faa5b7d92de133d499b353e6aeca6a4850940a7e6e8ab5893a204e685417"} Dec 03 14:15:16 crc kubenswrapper[4805]: I1203 14:15:16.399554 4805 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-8gt4x"] Dec 03 14:15:16 crc kubenswrapper[4805]: I1203 14:15:16.399858 4805 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-8gt4x" podUID="b2e4ac3a-e225-46ae-874d-4cf931c13464" containerName="registry-server" containerID="cri-o://ae4620a4d37207bd0317f1311c70528b552e6f40a2fac0b242d743282dadf663" gracePeriod=30 Dec 03 14:15:16 crc kubenswrapper[4805]: I1203 14:15:16.408645 4805 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-w6nm5"] Dec 03 14:15:16 crc kubenswrapper[4805]: I1203 14:15:16.408885 4805 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-w6nm5" podUID="344cb196-3baa-48c0-abcb-7e46cbce614d" containerName="registry-server" containerID="cri-o://d7a629c994143e2afd38ad34222b48e2bbde341dc173df63ac2114c8591c19f7" gracePeriod=30 Dec 03 14:15:16 crc kubenswrapper[4805]: I1203 14:15:16.421115 4805 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-spjlp"] Dec 03 14:15:16 crc kubenswrapper[4805]: I1203 14:15:16.421621 4805 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/marketplace-operator-79b997595-spjlp" podUID="66812e7e-aa85-4045-8e7a-8967b84f8849" containerName="marketplace-operator" containerID="cri-o://a9f2e6afd451ccf4c9cf049256fc598e0cbb0ee7e78cd6ed913ff8afdff9b2bf" gracePeriod=30 Dec 03 14:15:16 crc kubenswrapper[4805]: I1203 14:15:16.432310 4805 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-m7r99"] Dec 03 14:15:16 crc kubenswrapper[4805]: I1203 14:15:16.432608 4805 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-m7r99" podUID="1e412276-9d78-4771-b447-b4e2b388e604" containerName="registry-server" containerID="cri-o://1613fa3f6b59024d57deed3610696ad2566be4c7d5110cec4a2f3f22a3a00311" gracePeriod=30 Dec 03 14:15:16 crc kubenswrapper[4805]: I1203 14:15:16.438374 4805 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-bd7j9"] Dec 03 14:15:16 crc kubenswrapper[4805]: I1203 14:15:16.438593 4805 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-bd7j9" podUID="d0a05061-dab0-4a85-be73-b1d2acbc7b8f" containerName="registry-server" containerID="cri-o://48f11437168216c2fe0dbb6d16be8cd4176c1f7d852e1837bb92258c4b0f67e3" gracePeriod=30 Dec 03 14:15:16 crc kubenswrapper[4805]: I1203 14:15:16.449602 4805 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-6qgcv"] Dec 03 14:15:16 crc kubenswrapper[4805]: I1203 14:15:16.450426 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-6qgcv" Dec 03 14:15:16 crc kubenswrapper[4805]: I1203 14:15:16.461025 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-6qgcv"] Dec 03 14:15:16 crc kubenswrapper[4805]: I1203 14:15:16.553471 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kt5d4\" (UniqueName: \"kubernetes.io/projected/336b0e79-07ec-4899-8cba-a24a65458383-kube-api-access-kt5d4\") pod \"marketplace-operator-79b997595-6qgcv\" (UID: \"336b0e79-07ec-4899-8cba-a24a65458383\") " pod="openshift-marketplace/marketplace-operator-79b997595-6qgcv" Dec 03 14:15:16 crc kubenswrapper[4805]: I1203 14:15:16.553546 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/336b0e79-07ec-4899-8cba-a24a65458383-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-6qgcv\" (UID: \"336b0e79-07ec-4899-8cba-a24a65458383\") " pod="openshift-marketplace/marketplace-operator-79b997595-6qgcv" Dec 03 14:15:16 crc kubenswrapper[4805]: I1203 14:15:16.553578 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/336b0e79-07ec-4899-8cba-a24a65458383-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-6qgcv\" (UID: \"336b0e79-07ec-4899-8cba-a24a65458383\") " pod="openshift-marketplace/marketplace-operator-79b997595-6qgcv" Dec 03 14:15:16 crc kubenswrapper[4805]: I1203 14:15:16.655223 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kt5d4\" (UniqueName: \"kubernetes.io/projected/336b0e79-07ec-4899-8cba-a24a65458383-kube-api-access-kt5d4\") pod \"marketplace-operator-79b997595-6qgcv\" (UID: \"336b0e79-07ec-4899-8cba-a24a65458383\") " pod="openshift-marketplace/marketplace-operator-79b997595-6qgcv" Dec 03 14:15:16 crc kubenswrapper[4805]: I1203 14:15:16.655296 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/336b0e79-07ec-4899-8cba-a24a65458383-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-6qgcv\" (UID: \"336b0e79-07ec-4899-8cba-a24a65458383\") " pod="openshift-marketplace/marketplace-operator-79b997595-6qgcv" Dec 03 14:15:16 crc kubenswrapper[4805]: I1203 14:15:16.655328 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/336b0e79-07ec-4899-8cba-a24a65458383-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-6qgcv\" (UID: \"336b0e79-07ec-4899-8cba-a24a65458383\") " pod="openshift-marketplace/marketplace-operator-79b997595-6qgcv" Dec 03 14:15:16 crc kubenswrapper[4805]: I1203 14:15:16.657381 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/336b0e79-07ec-4899-8cba-a24a65458383-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-6qgcv\" (UID: \"336b0e79-07ec-4899-8cba-a24a65458383\") " pod="openshift-marketplace/marketplace-operator-79b997595-6qgcv" Dec 03 14:15:16 crc kubenswrapper[4805]: I1203 14:15:16.661103 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/336b0e79-07ec-4899-8cba-a24a65458383-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-6qgcv\" (UID: \"336b0e79-07ec-4899-8cba-a24a65458383\") " pod="openshift-marketplace/marketplace-operator-79b997595-6qgcv" Dec 03 14:15:16 crc kubenswrapper[4805]: I1203 14:15:16.672875 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kt5d4\" (UniqueName: \"kubernetes.io/projected/336b0e79-07ec-4899-8cba-a24a65458383-kube-api-access-kt5d4\") pod \"marketplace-operator-79b997595-6qgcv\" (UID: \"336b0e79-07ec-4899-8cba-a24a65458383\") " pod="openshift-marketplace/marketplace-operator-79b997595-6qgcv" Dec 03 14:15:16 crc kubenswrapper[4805]: I1203 14:15:16.802719 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-6qgcv" Dec 03 14:15:16 crc kubenswrapper[4805]: E1203 14:15:16.864134 4805 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 48f11437168216c2fe0dbb6d16be8cd4176c1f7d852e1837bb92258c4b0f67e3 is running failed: container process not found" containerID="48f11437168216c2fe0dbb6d16be8cd4176c1f7d852e1837bb92258c4b0f67e3" cmd=["grpc_health_probe","-addr=:50051"] Dec 03 14:15:16 crc kubenswrapper[4805]: E1203 14:15:16.864467 4805 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 48f11437168216c2fe0dbb6d16be8cd4176c1f7d852e1837bb92258c4b0f67e3 is running failed: container process not found" containerID="48f11437168216c2fe0dbb6d16be8cd4176c1f7d852e1837bb92258c4b0f67e3" cmd=["grpc_health_probe","-addr=:50051"] Dec 03 14:15:16 crc kubenswrapper[4805]: I1203 14:15:16.864913 4805 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-8gt4x" Dec 03 14:15:16 crc kubenswrapper[4805]: E1203 14:15:16.865363 4805 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 48f11437168216c2fe0dbb6d16be8cd4176c1f7d852e1837bb92258c4b0f67e3 is running failed: container process not found" containerID="48f11437168216c2fe0dbb6d16be8cd4176c1f7d852e1837bb92258c4b0f67e3" cmd=["grpc_health_probe","-addr=:50051"] Dec 03 14:15:16 crc kubenswrapper[4805]: E1203 14:15:16.865392 4805 prober.go:104] "Probe errored" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 48f11437168216c2fe0dbb6d16be8cd4176c1f7d852e1837bb92258c4b0f67e3 is running failed: container process not found" probeType="Readiness" pod="openshift-marketplace/redhat-operators-bd7j9" podUID="d0a05061-dab0-4a85-be73-b1d2acbc7b8f" containerName="registry-server" Dec 03 14:15:16 crc kubenswrapper[4805]: I1203 14:15:16.927423 4805 generic.go:334] "Generic (PLEG): container finished" podID="1e412276-9d78-4771-b447-b4e2b388e604" containerID="1613fa3f6b59024d57deed3610696ad2566be4c7d5110cec4a2f3f22a3a00311" exitCode=0 Dec 03 14:15:16 crc kubenswrapper[4805]: I1203 14:15:16.927452 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-m7r99" event={"ID":"1e412276-9d78-4771-b447-b4e2b388e604","Type":"ContainerDied","Data":"1613fa3f6b59024d57deed3610696ad2566be4c7d5110cec4a2f3f22a3a00311"} Dec 03 14:15:16 crc kubenswrapper[4805]: I1203 14:15:16.927508 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-m7r99" event={"ID":"1e412276-9d78-4771-b447-b4e2b388e604","Type":"ContainerDied","Data":"4a3c5188215db263a0646b5984c0ff692c2829ba3f3a403761ac309db9f6b223"} Dec 03 14:15:16 crc kubenswrapper[4805]: I1203 14:15:16.927525 4805 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="4a3c5188215db263a0646b5984c0ff692c2829ba3f3a403761ac309db9f6b223" Dec 03 14:15:16 crc kubenswrapper[4805]: I1203 14:15:16.927434 4805 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-m7r99" Dec 03 14:15:16 crc kubenswrapper[4805]: I1203 14:15:16.933096 4805 generic.go:334] "Generic (PLEG): container finished" podID="66812e7e-aa85-4045-8e7a-8967b84f8849" containerID="a9f2e6afd451ccf4c9cf049256fc598e0cbb0ee7e78cd6ed913ff8afdff9b2bf" exitCode=0 Dec 03 14:15:16 crc kubenswrapper[4805]: I1203 14:15:16.933162 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-spjlp" event={"ID":"66812e7e-aa85-4045-8e7a-8967b84f8849","Type":"ContainerDied","Data":"a9f2e6afd451ccf4c9cf049256fc598e0cbb0ee7e78cd6ed913ff8afdff9b2bf"} Dec 03 14:15:16 crc kubenswrapper[4805]: I1203 14:15:16.933199 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-spjlp" event={"ID":"66812e7e-aa85-4045-8e7a-8967b84f8849","Type":"ContainerDied","Data":"ab75814bb815ba66757f696e18a982e87da1a1e3fa65e8f535b0363cc08e85b8"} Dec 03 14:15:16 crc kubenswrapper[4805]: I1203 14:15:16.933210 4805 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="ab75814bb815ba66757f696e18a982e87da1a1e3fa65e8f535b0363cc08e85b8" Dec 03 14:15:16 crc kubenswrapper[4805]: I1203 14:15:16.933640 4805 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-bd7j9" Dec 03 14:15:16 crc kubenswrapper[4805]: I1203 14:15:16.934953 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-797bfc7f65-k7wn4" event={"ID":"98b69e82-9cbe-4d01-9a43-e8d94f911a3f","Type":"ContainerStarted","Data":"c5fbbe954cc8a568d9939f766e14b5dcb775e51c5e3018e888c3edc54bb09fb9"} Dec 03 14:15:16 crc kubenswrapper[4805]: I1203 14:15:16.935265 4805 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-controller-manager/controller-manager-797bfc7f65-k7wn4" Dec 03 14:15:16 crc kubenswrapper[4805]: I1203 14:15:16.940979 4805 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-controller-manager/controller-manager-797bfc7f65-k7wn4" Dec 03 14:15:16 crc kubenswrapper[4805]: I1203 14:15:16.941237 4805 generic.go:334] "Generic (PLEG): container finished" podID="b2e4ac3a-e225-46ae-874d-4cf931c13464" containerID="ae4620a4d37207bd0317f1311c70528b552e6f40a2fac0b242d743282dadf663" exitCode=0 Dec 03 14:15:16 crc kubenswrapper[4805]: I1203 14:15:16.941293 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-8gt4x" event={"ID":"b2e4ac3a-e225-46ae-874d-4cf931c13464","Type":"ContainerDied","Data":"ae4620a4d37207bd0317f1311c70528b552e6f40a2fac0b242d743282dadf663"} Dec 03 14:15:16 crc kubenswrapper[4805]: I1203 14:15:16.941315 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-8gt4x" event={"ID":"b2e4ac3a-e225-46ae-874d-4cf931c13464","Type":"ContainerDied","Data":"fad96d1f140c18627af015f90094d1e33cfe8bc2801445395bcefd0696ea7cc3"} Dec 03 14:15:16 crc kubenswrapper[4805]: I1203 14:15:16.941337 4805 scope.go:117] "RemoveContainer" containerID="ae4620a4d37207bd0317f1311c70528b552e6f40a2fac0b242d743282dadf663" Dec 03 14:15:16 crc kubenswrapper[4805]: I1203 14:15:16.941299 4805 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-8gt4x" Dec 03 14:15:16 crc kubenswrapper[4805]: I1203 14:15:16.942402 4805 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-w6nm5" Dec 03 14:15:16 crc kubenswrapper[4805]: I1203 14:15:16.945590 4805 generic.go:334] "Generic (PLEG): container finished" podID="d0a05061-dab0-4a85-be73-b1d2acbc7b8f" containerID="48f11437168216c2fe0dbb6d16be8cd4176c1f7d852e1837bb92258c4b0f67e3" exitCode=0 Dec 03 14:15:16 crc kubenswrapper[4805]: I1203 14:15:16.945664 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-bd7j9" event={"ID":"d0a05061-dab0-4a85-be73-b1d2acbc7b8f","Type":"ContainerDied","Data":"48f11437168216c2fe0dbb6d16be8cd4176c1f7d852e1837bb92258c4b0f67e3"} Dec 03 14:15:16 crc kubenswrapper[4805]: I1203 14:15:16.945696 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-bd7j9" event={"ID":"d0a05061-dab0-4a85-be73-b1d2acbc7b8f","Type":"ContainerDied","Data":"77d1a04380488ca0fe6609e31fd04dd77e33e0987bd8b09f722d503c4dcaa265"} Dec 03 14:15:16 crc kubenswrapper[4805]: I1203 14:15:16.946242 4805 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-bd7j9" Dec 03 14:15:16 crc kubenswrapper[4805]: I1203 14:15:16.958990 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b2e4ac3a-e225-46ae-874d-4cf931c13464-catalog-content\") pod \"b2e4ac3a-e225-46ae-874d-4cf931c13464\" (UID: \"b2e4ac3a-e225-46ae-874d-4cf931c13464\") " Dec 03 14:15:16 crc kubenswrapper[4805]: I1203 14:15:16.959038 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b2e4ac3a-e225-46ae-874d-4cf931c13464-utilities\") pod \"b2e4ac3a-e225-46ae-874d-4cf931c13464\" (UID: \"b2e4ac3a-e225-46ae-874d-4cf931c13464\") " Dec 03 14:15:16 crc kubenswrapper[4805]: I1203 14:15:16.959083 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d0a05061-dab0-4a85-be73-b1d2acbc7b8f-catalog-content\") pod \"d0a05061-dab0-4a85-be73-b1d2acbc7b8f\" (UID: \"d0a05061-dab0-4a85-be73-b1d2acbc7b8f\") " Dec 03 14:15:16 crc kubenswrapper[4805]: I1203 14:15:16.959131 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/344cb196-3baa-48c0-abcb-7e46cbce614d-catalog-content\") pod \"344cb196-3baa-48c0-abcb-7e46cbce614d\" (UID: \"344cb196-3baa-48c0-abcb-7e46cbce614d\") " Dec 03 14:15:16 crc kubenswrapper[4805]: I1203 14:15:16.959181 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-n626l\" (UniqueName: \"kubernetes.io/projected/b2e4ac3a-e225-46ae-874d-4cf931c13464-kube-api-access-n626l\") pod \"b2e4ac3a-e225-46ae-874d-4cf931c13464\" (UID: \"b2e4ac3a-e225-46ae-874d-4cf931c13464\") " Dec 03 14:15:16 crc kubenswrapper[4805]: I1203 14:15:16.959208 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-74jhd\" (UniqueName: \"kubernetes.io/projected/d0a05061-dab0-4a85-be73-b1d2acbc7b8f-kube-api-access-74jhd\") pod \"d0a05061-dab0-4a85-be73-b1d2acbc7b8f\" (UID: \"d0a05061-dab0-4a85-be73-b1d2acbc7b8f\") " Dec 03 14:15:16 crc kubenswrapper[4805]: I1203 14:15:16.959232 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1e412276-9d78-4771-b447-b4e2b388e604-utilities\") pod \"1e412276-9d78-4771-b447-b4e2b388e604\" (UID: \"1e412276-9d78-4771-b447-b4e2b388e604\") " Dec 03 14:15:16 crc kubenswrapper[4805]: I1203 14:15:16.959267 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-b5wvv\" (UniqueName: \"kubernetes.io/projected/344cb196-3baa-48c0-abcb-7e46cbce614d-kube-api-access-b5wvv\") pod \"344cb196-3baa-48c0-abcb-7e46cbce614d\" (UID: \"344cb196-3baa-48c0-abcb-7e46cbce614d\") " Dec 03 14:15:16 crc kubenswrapper[4805]: I1203 14:15:16.959308 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/344cb196-3baa-48c0-abcb-7e46cbce614d-utilities\") pod \"344cb196-3baa-48c0-abcb-7e46cbce614d\" (UID: \"344cb196-3baa-48c0-abcb-7e46cbce614d\") " Dec 03 14:15:16 crc kubenswrapper[4805]: I1203 14:15:16.959340 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1e412276-9d78-4771-b447-b4e2b388e604-catalog-content\") pod \"1e412276-9d78-4771-b447-b4e2b388e604\" (UID: \"1e412276-9d78-4771-b447-b4e2b388e604\") " Dec 03 14:15:16 crc kubenswrapper[4805]: I1203 14:15:16.959363 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-h6c8m\" (UniqueName: \"kubernetes.io/projected/1e412276-9d78-4771-b447-b4e2b388e604-kube-api-access-h6c8m\") pod \"1e412276-9d78-4771-b447-b4e2b388e604\" (UID: \"1e412276-9d78-4771-b447-b4e2b388e604\") " Dec 03 14:15:16 crc kubenswrapper[4805]: I1203 14:15:16.959404 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d0a05061-dab0-4a85-be73-b1d2acbc7b8f-utilities\") pod \"d0a05061-dab0-4a85-be73-b1d2acbc7b8f\" (UID: \"d0a05061-dab0-4a85-be73-b1d2acbc7b8f\") " Dec 03 14:15:16 crc kubenswrapper[4805]: I1203 14:15:16.963611 4805 generic.go:334] "Generic (PLEG): container finished" podID="344cb196-3baa-48c0-abcb-7e46cbce614d" containerID="d7a629c994143e2afd38ad34222b48e2bbde341dc173df63ac2114c8591c19f7" exitCode=0 Dec 03 14:15:16 crc kubenswrapper[4805]: I1203 14:15:16.963662 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-w6nm5" event={"ID":"344cb196-3baa-48c0-abcb-7e46cbce614d","Type":"ContainerDied","Data":"d7a629c994143e2afd38ad34222b48e2bbde341dc173df63ac2114c8591c19f7"} Dec 03 14:15:16 crc kubenswrapper[4805]: I1203 14:15:16.963697 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-w6nm5" event={"ID":"344cb196-3baa-48c0-abcb-7e46cbce614d","Type":"ContainerDied","Data":"9c86a4483d9c040cfd4bb2d0e7facf620e95b3e01e98f8356b5085dad93668fe"} Dec 03 14:15:16 crc kubenswrapper[4805]: I1203 14:15:16.963886 4805 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-spjlp" Dec 03 14:15:16 crc kubenswrapper[4805]: I1203 14:15:16.963973 4805 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-w6nm5" Dec 03 14:15:16 crc kubenswrapper[4805]: I1203 14:15:16.965067 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b2e4ac3a-e225-46ae-874d-4cf931c13464-kube-api-access-n626l" (OuterVolumeSpecName: "kube-api-access-n626l") pod "b2e4ac3a-e225-46ae-874d-4cf931c13464" (UID: "b2e4ac3a-e225-46ae-874d-4cf931c13464"). InnerVolumeSpecName "kube-api-access-n626l". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 14:15:16 crc kubenswrapper[4805]: I1203 14:15:16.965564 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1e412276-9d78-4771-b447-b4e2b388e604-kube-api-access-h6c8m" (OuterVolumeSpecName: "kube-api-access-h6c8m") pod "1e412276-9d78-4771-b447-b4e2b388e604" (UID: "1e412276-9d78-4771-b447-b4e2b388e604"). InnerVolumeSpecName "kube-api-access-h6c8m". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 14:15:16 crc kubenswrapper[4805]: I1203 14:15:16.966981 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/344cb196-3baa-48c0-abcb-7e46cbce614d-kube-api-access-b5wvv" (OuterVolumeSpecName: "kube-api-access-b5wvv") pod "344cb196-3baa-48c0-abcb-7e46cbce614d" (UID: "344cb196-3baa-48c0-abcb-7e46cbce614d"). InnerVolumeSpecName "kube-api-access-b5wvv". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 14:15:16 crc kubenswrapper[4805]: I1203 14:15:16.967431 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d0a05061-dab0-4a85-be73-b1d2acbc7b8f-utilities" (OuterVolumeSpecName: "utilities") pod "d0a05061-dab0-4a85-be73-b1d2acbc7b8f" (UID: "d0a05061-dab0-4a85-be73-b1d2acbc7b8f"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 14:15:16 crc kubenswrapper[4805]: I1203 14:15:16.967624 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/344cb196-3baa-48c0-abcb-7e46cbce614d-utilities" (OuterVolumeSpecName: "utilities") pod "344cb196-3baa-48c0-abcb-7e46cbce614d" (UID: "344cb196-3baa-48c0-abcb-7e46cbce614d"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 14:15:16 crc kubenswrapper[4805]: I1203 14:15:16.967798 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1e412276-9d78-4771-b447-b4e2b388e604-utilities" (OuterVolumeSpecName: "utilities") pod "1e412276-9d78-4771-b447-b4e2b388e604" (UID: "1e412276-9d78-4771-b447-b4e2b388e604"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 14:15:16 crc kubenswrapper[4805]: I1203 14:15:16.972525 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b2e4ac3a-e225-46ae-874d-4cf931c13464-utilities" (OuterVolumeSpecName: "utilities") pod "b2e4ac3a-e225-46ae-874d-4cf931c13464" (UID: "b2e4ac3a-e225-46ae-874d-4cf931c13464"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 14:15:16 crc kubenswrapper[4805]: I1203 14:15:16.977041 4805 scope.go:117] "RemoveContainer" containerID="8ab28724d5fa789a469456ce08a8c43bf28fa705458279faad7ea7f330f14a6b" Dec 03 14:15:16 crc kubenswrapper[4805]: I1203 14:15:16.987731 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d0a05061-dab0-4a85-be73-b1d2acbc7b8f-kube-api-access-74jhd" (OuterVolumeSpecName: "kube-api-access-74jhd") pod "d0a05061-dab0-4a85-be73-b1d2acbc7b8f" (UID: "d0a05061-dab0-4a85-be73-b1d2acbc7b8f"). InnerVolumeSpecName "kube-api-access-74jhd". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 14:15:16 crc kubenswrapper[4805]: I1203 14:15:16.994306 4805 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager/controller-manager-797bfc7f65-k7wn4" podStartSLOduration=3.994288573 podStartE2EDuration="3.994288573s" podCreationTimestamp="2025-12-03 14:15:13 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 14:15:16.991084086 +0000 UTC m=+346.654001019" watchObservedRunningTime="2025-12-03 14:15:16.994288573 +0000 UTC m=+346.657205496" Dec 03 14:15:17 crc kubenswrapper[4805]: I1203 14:15:17.000559 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1e412276-9d78-4771-b447-b4e2b388e604-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "1e412276-9d78-4771-b447-b4e2b388e604" (UID: "1e412276-9d78-4771-b447-b4e2b388e604"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 14:15:17 crc kubenswrapper[4805]: I1203 14:15:17.019775 4805 scope.go:117] "RemoveContainer" containerID="397f4b6319d8994e0a243b3377f230736eaef70f2f17cdcde055b94bc9582296" Dec 03 14:15:17 crc kubenswrapper[4805]: I1203 14:15:17.056255 4805 scope.go:117] "RemoveContainer" containerID="ae4620a4d37207bd0317f1311c70528b552e6f40a2fac0b242d743282dadf663" Dec 03 14:15:17 crc kubenswrapper[4805]: E1203 14:15:17.057307 4805 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ae4620a4d37207bd0317f1311c70528b552e6f40a2fac0b242d743282dadf663\": container with ID starting with ae4620a4d37207bd0317f1311c70528b552e6f40a2fac0b242d743282dadf663 not found: ID does not exist" containerID="ae4620a4d37207bd0317f1311c70528b552e6f40a2fac0b242d743282dadf663" Dec 03 14:15:17 crc kubenswrapper[4805]: I1203 14:15:17.057342 4805 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ae4620a4d37207bd0317f1311c70528b552e6f40a2fac0b242d743282dadf663"} err="failed to get container status \"ae4620a4d37207bd0317f1311c70528b552e6f40a2fac0b242d743282dadf663\": rpc error: code = NotFound desc = could not find container \"ae4620a4d37207bd0317f1311c70528b552e6f40a2fac0b242d743282dadf663\": container with ID starting with ae4620a4d37207bd0317f1311c70528b552e6f40a2fac0b242d743282dadf663 not found: ID does not exist" Dec 03 14:15:17 crc kubenswrapper[4805]: I1203 14:15:17.057381 4805 scope.go:117] "RemoveContainer" containerID="8ab28724d5fa789a469456ce08a8c43bf28fa705458279faad7ea7f330f14a6b" Dec 03 14:15:17 crc kubenswrapper[4805]: E1203 14:15:17.058021 4805 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8ab28724d5fa789a469456ce08a8c43bf28fa705458279faad7ea7f330f14a6b\": container with ID starting with 8ab28724d5fa789a469456ce08a8c43bf28fa705458279faad7ea7f330f14a6b not found: ID does not exist" containerID="8ab28724d5fa789a469456ce08a8c43bf28fa705458279faad7ea7f330f14a6b" Dec 03 14:15:17 crc kubenswrapper[4805]: I1203 14:15:17.058072 4805 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8ab28724d5fa789a469456ce08a8c43bf28fa705458279faad7ea7f330f14a6b"} err="failed to get container status \"8ab28724d5fa789a469456ce08a8c43bf28fa705458279faad7ea7f330f14a6b\": rpc error: code = NotFound desc = could not find container \"8ab28724d5fa789a469456ce08a8c43bf28fa705458279faad7ea7f330f14a6b\": container with ID starting with 8ab28724d5fa789a469456ce08a8c43bf28fa705458279faad7ea7f330f14a6b not found: ID does not exist" Dec 03 14:15:17 crc kubenswrapper[4805]: I1203 14:15:17.058087 4805 scope.go:117] "RemoveContainer" containerID="397f4b6319d8994e0a243b3377f230736eaef70f2f17cdcde055b94bc9582296" Dec 03 14:15:17 crc kubenswrapper[4805]: E1203 14:15:17.058336 4805 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"397f4b6319d8994e0a243b3377f230736eaef70f2f17cdcde055b94bc9582296\": container with ID starting with 397f4b6319d8994e0a243b3377f230736eaef70f2f17cdcde055b94bc9582296 not found: ID does not exist" containerID="397f4b6319d8994e0a243b3377f230736eaef70f2f17cdcde055b94bc9582296" Dec 03 14:15:17 crc kubenswrapper[4805]: I1203 14:15:17.058463 4805 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"397f4b6319d8994e0a243b3377f230736eaef70f2f17cdcde055b94bc9582296"} err="failed to get container status \"397f4b6319d8994e0a243b3377f230736eaef70f2f17cdcde055b94bc9582296\": rpc error: code = NotFound desc = could not find container \"397f4b6319d8994e0a243b3377f230736eaef70f2f17cdcde055b94bc9582296\": container with ID starting with 397f4b6319d8994e0a243b3377f230736eaef70f2f17cdcde055b94bc9582296 not found: ID does not exist" Dec 03 14:15:17 crc kubenswrapper[4805]: I1203 14:15:17.058489 4805 scope.go:117] "RemoveContainer" containerID="48f11437168216c2fe0dbb6d16be8cd4176c1f7d852e1837bb92258c4b0f67e3" Dec 03 14:15:17 crc kubenswrapper[4805]: I1203 14:15:17.058550 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b2e4ac3a-e225-46ae-874d-4cf931c13464-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "b2e4ac3a-e225-46ae-874d-4cf931c13464" (UID: "b2e4ac3a-e225-46ae-874d-4cf931c13464"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 14:15:17 crc kubenswrapper[4805]: I1203 14:15:17.060472 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/66812e7e-aa85-4045-8e7a-8967b84f8849-marketplace-trusted-ca\") pod \"66812e7e-aa85-4045-8e7a-8967b84f8849\" (UID: \"66812e7e-aa85-4045-8e7a-8967b84f8849\") " Dec 03 14:15:17 crc kubenswrapper[4805]: I1203 14:15:17.060545 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/66812e7e-aa85-4045-8e7a-8967b84f8849-marketplace-operator-metrics\") pod \"66812e7e-aa85-4045-8e7a-8967b84f8849\" (UID: \"66812e7e-aa85-4045-8e7a-8967b84f8849\") " Dec 03 14:15:17 crc kubenswrapper[4805]: I1203 14:15:17.060646 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-n5f7x\" (UniqueName: \"kubernetes.io/projected/66812e7e-aa85-4045-8e7a-8967b84f8849-kube-api-access-n5f7x\") pod \"66812e7e-aa85-4045-8e7a-8967b84f8849\" (UID: \"66812e7e-aa85-4045-8e7a-8967b84f8849\") " Dec 03 14:15:17 crc kubenswrapper[4805]: I1203 14:15:17.061152 4805 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/344cb196-3baa-48c0-abcb-7e46cbce614d-utilities\") on node \"crc\" DevicePath \"\"" Dec 03 14:15:17 crc kubenswrapper[4805]: I1203 14:15:17.061188 4805 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1e412276-9d78-4771-b447-b4e2b388e604-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 03 14:15:17 crc kubenswrapper[4805]: I1203 14:15:17.061205 4805 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-h6c8m\" (UniqueName: \"kubernetes.io/projected/1e412276-9d78-4771-b447-b4e2b388e604-kube-api-access-h6c8m\") on node \"crc\" DevicePath \"\"" Dec 03 14:15:17 crc kubenswrapper[4805]: I1203 14:15:17.061217 4805 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d0a05061-dab0-4a85-be73-b1d2acbc7b8f-utilities\") on node \"crc\" DevicePath \"\"" Dec 03 14:15:17 crc kubenswrapper[4805]: I1203 14:15:17.061230 4805 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b2e4ac3a-e225-46ae-874d-4cf931c13464-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 03 14:15:17 crc kubenswrapper[4805]: I1203 14:15:17.061246 4805 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b2e4ac3a-e225-46ae-874d-4cf931c13464-utilities\") on node \"crc\" DevicePath \"\"" Dec 03 14:15:17 crc kubenswrapper[4805]: I1203 14:15:17.061258 4805 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-n626l\" (UniqueName: \"kubernetes.io/projected/b2e4ac3a-e225-46ae-874d-4cf931c13464-kube-api-access-n626l\") on node \"crc\" DevicePath \"\"" Dec 03 14:15:17 crc kubenswrapper[4805]: I1203 14:15:17.061270 4805 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-74jhd\" (UniqueName: \"kubernetes.io/projected/d0a05061-dab0-4a85-be73-b1d2acbc7b8f-kube-api-access-74jhd\") on node \"crc\" DevicePath \"\"" Dec 03 14:15:17 crc kubenswrapper[4805]: I1203 14:15:17.061282 4805 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1e412276-9d78-4771-b447-b4e2b388e604-utilities\") on node \"crc\" DevicePath \"\"" Dec 03 14:15:17 crc kubenswrapper[4805]: I1203 14:15:17.061297 4805 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-b5wvv\" (UniqueName: \"kubernetes.io/projected/344cb196-3baa-48c0-abcb-7e46cbce614d-kube-api-access-b5wvv\") on node \"crc\" DevicePath \"\"" Dec 03 14:15:17 crc kubenswrapper[4805]: I1203 14:15:17.061503 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/66812e7e-aa85-4045-8e7a-8967b84f8849-marketplace-trusted-ca" (OuterVolumeSpecName: "marketplace-trusted-ca") pod "66812e7e-aa85-4045-8e7a-8967b84f8849" (UID: "66812e7e-aa85-4045-8e7a-8967b84f8849"). InnerVolumeSpecName "marketplace-trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 14:15:17 crc kubenswrapper[4805]: I1203 14:15:17.065351 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/66812e7e-aa85-4045-8e7a-8967b84f8849-marketplace-operator-metrics" (OuterVolumeSpecName: "marketplace-operator-metrics") pod "66812e7e-aa85-4045-8e7a-8967b84f8849" (UID: "66812e7e-aa85-4045-8e7a-8967b84f8849"). InnerVolumeSpecName "marketplace-operator-metrics". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 14:15:17 crc kubenswrapper[4805]: I1203 14:15:17.066167 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/66812e7e-aa85-4045-8e7a-8967b84f8849-kube-api-access-n5f7x" (OuterVolumeSpecName: "kube-api-access-n5f7x") pod "66812e7e-aa85-4045-8e7a-8967b84f8849" (UID: "66812e7e-aa85-4045-8e7a-8967b84f8849"). InnerVolumeSpecName "kube-api-access-n5f7x". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 14:15:17 crc kubenswrapper[4805]: I1203 14:15:17.081493 4805 scope.go:117] "RemoveContainer" containerID="545b83ba10b4fa150c471ebb69ce19ffc8b1ba8aa337a88331459f13f345068a" Dec 03 14:15:17 crc kubenswrapper[4805]: I1203 14:15:17.091502 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/344cb196-3baa-48c0-abcb-7e46cbce614d-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "344cb196-3baa-48c0-abcb-7e46cbce614d" (UID: "344cb196-3baa-48c0-abcb-7e46cbce614d"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 14:15:17 crc kubenswrapper[4805]: I1203 14:15:17.098212 4805 scope.go:117] "RemoveContainer" containerID="38be2c7eb01fb25fa571a38edfabdf0beede1d902fd2c2f2d3bf004010e331e8" Dec 03 14:15:17 crc kubenswrapper[4805]: I1203 14:15:17.111781 4805 scope.go:117] "RemoveContainer" containerID="48f11437168216c2fe0dbb6d16be8cd4176c1f7d852e1837bb92258c4b0f67e3" Dec 03 14:15:17 crc kubenswrapper[4805]: E1203 14:15:17.112537 4805 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"48f11437168216c2fe0dbb6d16be8cd4176c1f7d852e1837bb92258c4b0f67e3\": container with ID starting with 48f11437168216c2fe0dbb6d16be8cd4176c1f7d852e1837bb92258c4b0f67e3 not found: ID does not exist" containerID="48f11437168216c2fe0dbb6d16be8cd4176c1f7d852e1837bb92258c4b0f67e3" Dec 03 14:15:17 crc kubenswrapper[4805]: I1203 14:15:17.112578 4805 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"48f11437168216c2fe0dbb6d16be8cd4176c1f7d852e1837bb92258c4b0f67e3"} err="failed to get container status \"48f11437168216c2fe0dbb6d16be8cd4176c1f7d852e1837bb92258c4b0f67e3\": rpc error: code = NotFound desc = could not find container \"48f11437168216c2fe0dbb6d16be8cd4176c1f7d852e1837bb92258c4b0f67e3\": container with ID starting with 48f11437168216c2fe0dbb6d16be8cd4176c1f7d852e1837bb92258c4b0f67e3 not found: ID does not exist" Dec 03 14:15:17 crc kubenswrapper[4805]: I1203 14:15:17.112671 4805 scope.go:117] "RemoveContainer" containerID="545b83ba10b4fa150c471ebb69ce19ffc8b1ba8aa337a88331459f13f345068a" Dec 03 14:15:17 crc kubenswrapper[4805]: E1203 14:15:17.113186 4805 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"545b83ba10b4fa150c471ebb69ce19ffc8b1ba8aa337a88331459f13f345068a\": container with ID starting with 545b83ba10b4fa150c471ebb69ce19ffc8b1ba8aa337a88331459f13f345068a not found: ID does not exist" containerID="545b83ba10b4fa150c471ebb69ce19ffc8b1ba8aa337a88331459f13f345068a" Dec 03 14:15:17 crc kubenswrapper[4805]: I1203 14:15:17.113223 4805 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"545b83ba10b4fa150c471ebb69ce19ffc8b1ba8aa337a88331459f13f345068a"} err="failed to get container status \"545b83ba10b4fa150c471ebb69ce19ffc8b1ba8aa337a88331459f13f345068a\": rpc error: code = NotFound desc = could not find container \"545b83ba10b4fa150c471ebb69ce19ffc8b1ba8aa337a88331459f13f345068a\": container with ID starting with 545b83ba10b4fa150c471ebb69ce19ffc8b1ba8aa337a88331459f13f345068a not found: ID does not exist" Dec 03 14:15:17 crc kubenswrapper[4805]: I1203 14:15:17.113240 4805 scope.go:117] "RemoveContainer" containerID="38be2c7eb01fb25fa571a38edfabdf0beede1d902fd2c2f2d3bf004010e331e8" Dec 03 14:15:17 crc kubenswrapper[4805]: E1203 14:15:17.113542 4805 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"38be2c7eb01fb25fa571a38edfabdf0beede1d902fd2c2f2d3bf004010e331e8\": container with ID starting with 38be2c7eb01fb25fa571a38edfabdf0beede1d902fd2c2f2d3bf004010e331e8 not found: ID does not exist" containerID="38be2c7eb01fb25fa571a38edfabdf0beede1d902fd2c2f2d3bf004010e331e8" Dec 03 14:15:17 crc kubenswrapper[4805]: I1203 14:15:17.113570 4805 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"38be2c7eb01fb25fa571a38edfabdf0beede1d902fd2c2f2d3bf004010e331e8"} err="failed to get container status \"38be2c7eb01fb25fa571a38edfabdf0beede1d902fd2c2f2d3bf004010e331e8\": rpc error: code = NotFound desc = could not find container \"38be2c7eb01fb25fa571a38edfabdf0beede1d902fd2c2f2d3bf004010e331e8\": container with ID starting with 38be2c7eb01fb25fa571a38edfabdf0beede1d902fd2c2f2d3bf004010e331e8 not found: ID does not exist" Dec 03 14:15:17 crc kubenswrapper[4805]: I1203 14:15:17.113587 4805 scope.go:117] "RemoveContainer" containerID="d7a629c994143e2afd38ad34222b48e2bbde341dc173df63ac2114c8591c19f7" Dec 03 14:15:17 crc kubenswrapper[4805]: I1203 14:15:17.125710 4805 scope.go:117] "RemoveContainer" containerID="aae5371377082372f9fd92376d8cc23f7bb2d5195e09b55895849e7a7ef4ca6b" Dec 03 14:15:17 crc kubenswrapper[4805]: I1203 14:15:17.150370 4805 scope.go:117] "RemoveContainer" containerID="2fd386dcb356443b4d70345cf5a236e1680e0f2dce2c6a2c7a8ab14803d8d28e" Dec 03 14:15:17 crc kubenswrapper[4805]: I1203 14:15:17.156813 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d0a05061-dab0-4a85-be73-b1d2acbc7b8f-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "d0a05061-dab0-4a85-be73-b1d2acbc7b8f" (UID: "d0a05061-dab0-4a85-be73-b1d2acbc7b8f"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 14:15:17 crc kubenswrapper[4805]: I1203 14:15:17.162583 4805 reconciler_common.go:293] "Volume detached for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/66812e7e-aa85-4045-8e7a-8967b84f8849-marketplace-trusted-ca\") on node \"crc\" DevicePath \"\"" Dec 03 14:15:17 crc kubenswrapper[4805]: I1203 14:15:17.162624 4805 reconciler_common.go:293] "Volume detached for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/66812e7e-aa85-4045-8e7a-8967b84f8849-marketplace-operator-metrics\") on node \"crc\" DevicePath \"\"" Dec 03 14:15:17 crc kubenswrapper[4805]: I1203 14:15:17.162642 4805 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d0a05061-dab0-4a85-be73-b1d2acbc7b8f-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 03 14:15:17 crc kubenswrapper[4805]: I1203 14:15:17.162654 4805 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-n5f7x\" (UniqueName: \"kubernetes.io/projected/66812e7e-aa85-4045-8e7a-8967b84f8849-kube-api-access-n5f7x\") on node \"crc\" DevicePath \"\"" Dec 03 14:15:17 crc kubenswrapper[4805]: I1203 14:15:17.162665 4805 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/344cb196-3baa-48c0-abcb-7e46cbce614d-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 03 14:15:17 crc kubenswrapper[4805]: I1203 14:15:17.164792 4805 scope.go:117] "RemoveContainer" containerID="d7a629c994143e2afd38ad34222b48e2bbde341dc173df63ac2114c8591c19f7" Dec 03 14:15:17 crc kubenswrapper[4805]: E1203 14:15:17.165242 4805 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d7a629c994143e2afd38ad34222b48e2bbde341dc173df63ac2114c8591c19f7\": container with ID starting with d7a629c994143e2afd38ad34222b48e2bbde341dc173df63ac2114c8591c19f7 not found: ID does not exist" containerID="d7a629c994143e2afd38ad34222b48e2bbde341dc173df63ac2114c8591c19f7" Dec 03 14:15:17 crc kubenswrapper[4805]: I1203 14:15:17.165274 4805 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d7a629c994143e2afd38ad34222b48e2bbde341dc173df63ac2114c8591c19f7"} err="failed to get container status \"d7a629c994143e2afd38ad34222b48e2bbde341dc173df63ac2114c8591c19f7\": rpc error: code = NotFound desc = could not find container \"d7a629c994143e2afd38ad34222b48e2bbde341dc173df63ac2114c8591c19f7\": container with ID starting with d7a629c994143e2afd38ad34222b48e2bbde341dc173df63ac2114c8591c19f7 not found: ID does not exist" Dec 03 14:15:17 crc kubenswrapper[4805]: I1203 14:15:17.165301 4805 scope.go:117] "RemoveContainer" containerID="aae5371377082372f9fd92376d8cc23f7bb2d5195e09b55895849e7a7ef4ca6b" Dec 03 14:15:17 crc kubenswrapper[4805]: E1203 14:15:17.165558 4805 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"aae5371377082372f9fd92376d8cc23f7bb2d5195e09b55895849e7a7ef4ca6b\": container with ID starting with aae5371377082372f9fd92376d8cc23f7bb2d5195e09b55895849e7a7ef4ca6b not found: ID does not exist" containerID="aae5371377082372f9fd92376d8cc23f7bb2d5195e09b55895849e7a7ef4ca6b" Dec 03 14:15:17 crc kubenswrapper[4805]: I1203 14:15:17.165581 4805 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"aae5371377082372f9fd92376d8cc23f7bb2d5195e09b55895849e7a7ef4ca6b"} err="failed to get container status \"aae5371377082372f9fd92376d8cc23f7bb2d5195e09b55895849e7a7ef4ca6b\": rpc error: code = NotFound desc = could not find container \"aae5371377082372f9fd92376d8cc23f7bb2d5195e09b55895849e7a7ef4ca6b\": container with ID starting with aae5371377082372f9fd92376d8cc23f7bb2d5195e09b55895849e7a7ef4ca6b not found: ID does not exist" Dec 03 14:15:17 crc kubenswrapper[4805]: I1203 14:15:17.165597 4805 scope.go:117] "RemoveContainer" containerID="2fd386dcb356443b4d70345cf5a236e1680e0f2dce2c6a2c7a8ab14803d8d28e" Dec 03 14:15:17 crc kubenswrapper[4805]: E1203 14:15:17.165876 4805 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2fd386dcb356443b4d70345cf5a236e1680e0f2dce2c6a2c7a8ab14803d8d28e\": container with ID starting with 2fd386dcb356443b4d70345cf5a236e1680e0f2dce2c6a2c7a8ab14803d8d28e not found: ID does not exist" containerID="2fd386dcb356443b4d70345cf5a236e1680e0f2dce2c6a2c7a8ab14803d8d28e" Dec 03 14:15:17 crc kubenswrapper[4805]: I1203 14:15:17.165900 4805 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2fd386dcb356443b4d70345cf5a236e1680e0f2dce2c6a2c7a8ab14803d8d28e"} err="failed to get container status \"2fd386dcb356443b4d70345cf5a236e1680e0f2dce2c6a2c7a8ab14803d8d28e\": rpc error: code = NotFound desc = could not find container \"2fd386dcb356443b4d70345cf5a236e1680e0f2dce2c6a2c7a8ab14803d8d28e\": container with ID starting with 2fd386dcb356443b4d70345cf5a236e1680e0f2dce2c6a2c7a8ab14803d8d28e not found: ID does not exist" Dec 03 14:15:17 crc kubenswrapper[4805]: I1203 14:15:17.307634 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-6qgcv"] Dec 03 14:15:17 crc kubenswrapper[4805]: I1203 14:15:17.315683 4805 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-bd7j9"] Dec 03 14:15:17 crc kubenswrapper[4805]: I1203 14:15:17.319288 4805 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-bd7j9"] Dec 03 14:15:17 crc kubenswrapper[4805]: I1203 14:15:17.332257 4805 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-8gt4x"] Dec 03 14:15:17 crc kubenswrapper[4805]: I1203 14:15:17.339685 4805 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-8gt4x"] Dec 03 14:15:17 crc kubenswrapper[4805]: I1203 14:15:17.343660 4805 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-w6nm5"] Dec 03 14:15:17 crc kubenswrapper[4805]: I1203 14:15:17.348472 4805 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-w6nm5"] Dec 03 14:15:17 crc kubenswrapper[4805]: I1203 14:15:17.971701 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-6qgcv" event={"ID":"336b0e79-07ec-4899-8cba-a24a65458383","Type":"ContainerStarted","Data":"3a4b137bfccee82219d79857e7848e26dc754cacecd82483d2532cb86bbdd072"} Dec 03 14:15:17 crc kubenswrapper[4805]: I1203 14:15:17.972129 4805 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/marketplace-operator-79b997595-6qgcv" Dec 03 14:15:17 crc kubenswrapper[4805]: I1203 14:15:17.972154 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-6qgcv" event={"ID":"336b0e79-07ec-4899-8cba-a24a65458383","Type":"ContainerStarted","Data":"eaea27f8cd138aaa2fe11911a562a15025e34db28c0f94814c73ccd244ef314f"} Dec 03 14:15:17 crc kubenswrapper[4805]: I1203 14:15:17.975276 4805 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-spjlp" Dec 03 14:15:17 crc kubenswrapper[4805]: I1203 14:15:17.977239 4805 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-m7r99" Dec 03 14:15:17 crc kubenswrapper[4805]: I1203 14:15:17.977283 4805 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/marketplace-operator-79b997595-6qgcv" Dec 03 14:15:17 crc kubenswrapper[4805]: I1203 14:15:17.990920 4805 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/marketplace-operator-79b997595-6qgcv" podStartSLOduration=1.9908855490000001 podStartE2EDuration="1.990885549s" podCreationTimestamp="2025-12-03 14:15:16 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 14:15:17.987545638 +0000 UTC m=+347.650462551" watchObservedRunningTime="2025-12-03 14:15:17.990885549 +0000 UTC m=+347.653802492" Dec 03 14:15:18 crc kubenswrapper[4805]: I1203 14:15:18.034339 4805 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-spjlp"] Dec 03 14:15:18 crc kubenswrapper[4805]: I1203 14:15:18.039985 4805 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-spjlp"] Dec 03 14:15:18 crc kubenswrapper[4805]: I1203 14:15:18.046347 4805 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-m7r99"] Dec 03 14:15:18 crc kubenswrapper[4805]: I1203 14:15:18.053123 4805 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-m7r99"] Dec 03 14:15:18 crc kubenswrapper[4805]: I1203 14:15:18.617479 4805 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-jftlh"] Dec 03 14:15:18 crc kubenswrapper[4805]: E1203 14:15:18.617698 4805 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1e412276-9d78-4771-b447-b4e2b388e604" containerName="extract-utilities" Dec 03 14:15:18 crc kubenswrapper[4805]: I1203 14:15:18.617712 4805 state_mem.go:107] "Deleted CPUSet assignment" podUID="1e412276-9d78-4771-b447-b4e2b388e604" containerName="extract-utilities" Dec 03 14:15:18 crc kubenswrapper[4805]: E1203 14:15:18.617722 4805 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b2e4ac3a-e225-46ae-874d-4cf931c13464" containerName="extract-utilities" Dec 03 14:15:18 crc kubenswrapper[4805]: I1203 14:15:18.617732 4805 state_mem.go:107] "Deleted CPUSet assignment" podUID="b2e4ac3a-e225-46ae-874d-4cf931c13464" containerName="extract-utilities" Dec 03 14:15:18 crc kubenswrapper[4805]: E1203 14:15:18.617743 4805 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1e412276-9d78-4771-b447-b4e2b388e604" containerName="registry-server" Dec 03 14:15:18 crc kubenswrapper[4805]: I1203 14:15:18.617751 4805 state_mem.go:107] "Deleted CPUSet assignment" podUID="1e412276-9d78-4771-b447-b4e2b388e604" containerName="registry-server" Dec 03 14:15:18 crc kubenswrapper[4805]: E1203 14:15:18.617763 4805 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="344cb196-3baa-48c0-abcb-7e46cbce614d" containerName="registry-server" Dec 03 14:15:18 crc kubenswrapper[4805]: I1203 14:15:18.617771 4805 state_mem.go:107] "Deleted CPUSet assignment" podUID="344cb196-3baa-48c0-abcb-7e46cbce614d" containerName="registry-server" Dec 03 14:15:18 crc kubenswrapper[4805]: E1203 14:15:18.617785 4805 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b2e4ac3a-e225-46ae-874d-4cf931c13464" containerName="registry-server" Dec 03 14:15:18 crc kubenswrapper[4805]: I1203 14:15:18.617793 4805 state_mem.go:107] "Deleted CPUSet assignment" podUID="b2e4ac3a-e225-46ae-874d-4cf931c13464" containerName="registry-server" Dec 03 14:15:18 crc kubenswrapper[4805]: E1203 14:15:18.617803 4805 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="344cb196-3baa-48c0-abcb-7e46cbce614d" containerName="extract-content" Dec 03 14:15:18 crc kubenswrapper[4805]: I1203 14:15:18.617813 4805 state_mem.go:107] "Deleted CPUSet assignment" podUID="344cb196-3baa-48c0-abcb-7e46cbce614d" containerName="extract-content" Dec 03 14:15:18 crc kubenswrapper[4805]: E1203 14:15:18.617822 4805 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d0a05061-dab0-4a85-be73-b1d2acbc7b8f" containerName="extract-content" Dec 03 14:15:18 crc kubenswrapper[4805]: I1203 14:15:18.617829 4805 state_mem.go:107] "Deleted CPUSet assignment" podUID="d0a05061-dab0-4a85-be73-b1d2acbc7b8f" containerName="extract-content" Dec 03 14:15:18 crc kubenswrapper[4805]: E1203 14:15:18.617927 4805 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1e412276-9d78-4771-b447-b4e2b388e604" containerName="extract-content" Dec 03 14:15:18 crc kubenswrapper[4805]: I1203 14:15:18.617936 4805 state_mem.go:107] "Deleted CPUSet assignment" podUID="1e412276-9d78-4771-b447-b4e2b388e604" containerName="extract-content" Dec 03 14:15:18 crc kubenswrapper[4805]: E1203 14:15:18.617950 4805 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d0a05061-dab0-4a85-be73-b1d2acbc7b8f" containerName="registry-server" Dec 03 14:15:18 crc kubenswrapper[4805]: I1203 14:15:18.617958 4805 state_mem.go:107] "Deleted CPUSet assignment" podUID="d0a05061-dab0-4a85-be73-b1d2acbc7b8f" containerName="registry-server" Dec 03 14:15:18 crc kubenswrapper[4805]: E1203 14:15:18.617984 4805 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="344cb196-3baa-48c0-abcb-7e46cbce614d" containerName="extract-utilities" Dec 03 14:15:18 crc kubenswrapper[4805]: I1203 14:15:18.617992 4805 state_mem.go:107] "Deleted CPUSet assignment" podUID="344cb196-3baa-48c0-abcb-7e46cbce614d" containerName="extract-utilities" Dec 03 14:15:18 crc kubenswrapper[4805]: E1203 14:15:18.618004 4805 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="66812e7e-aa85-4045-8e7a-8967b84f8849" containerName="marketplace-operator" Dec 03 14:15:18 crc kubenswrapper[4805]: I1203 14:15:18.618012 4805 state_mem.go:107] "Deleted CPUSet assignment" podUID="66812e7e-aa85-4045-8e7a-8967b84f8849" containerName="marketplace-operator" Dec 03 14:15:18 crc kubenswrapper[4805]: E1203 14:15:18.618025 4805 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b2e4ac3a-e225-46ae-874d-4cf931c13464" containerName="extract-content" Dec 03 14:15:18 crc kubenswrapper[4805]: I1203 14:15:18.618033 4805 state_mem.go:107] "Deleted CPUSet assignment" podUID="b2e4ac3a-e225-46ae-874d-4cf931c13464" containerName="extract-content" Dec 03 14:15:18 crc kubenswrapper[4805]: E1203 14:15:18.618044 4805 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d0a05061-dab0-4a85-be73-b1d2acbc7b8f" containerName="extract-utilities" Dec 03 14:15:18 crc kubenswrapper[4805]: I1203 14:15:18.618052 4805 state_mem.go:107] "Deleted CPUSet assignment" podUID="d0a05061-dab0-4a85-be73-b1d2acbc7b8f" containerName="extract-utilities" Dec 03 14:15:18 crc kubenswrapper[4805]: I1203 14:15:18.618162 4805 memory_manager.go:354] "RemoveStaleState removing state" podUID="66812e7e-aa85-4045-8e7a-8967b84f8849" containerName="marketplace-operator" Dec 03 14:15:18 crc kubenswrapper[4805]: I1203 14:15:18.618173 4805 memory_manager.go:354] "RemoveStaleState removing state" podUID="1e412276-9d78-4771-b447-b4e2b388e604" containerName="registry-server" Dec 03 14:15:18 crc kubenswrapper[4805]: I1203 14:15:18.618183 4805 memory_manager.go:354] "RemoveStaleState removing state" podUID="b2e4ac3a-e225-46ae-874d-4cf931c13464" containerName="registry-server" Dec 03 14:15:18 crc kubenswrapper[4805]: I1203 14:15:18.618197 4805 memory_manager.go:354] "RemoveStaleState removing state" podUID="344cb196-3baa-48c0-abcb-7e46cbce614d" containerName="registry-server" Dec 03 14:15:18 crc kubenswrapper[4805]: I1203 14:15:18.618210 4805 memory_manager.go:354] "RemoveStaleState removing state" podUID="d0a05061-dab0-4a85-be73-b1d2acbc7b8f" containerName="registry-server" Dec 03 14:15:18 crc kubenswrapper[4805]: I1203 14:15:18.619062 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-jftlh" Dec 03 14:15:18 crc kubenswrapper[4805]: I1203 14:15:18.622047 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-marketplace-dockercfg-x2ctb" Dec 03 14:15:18 crc kubenswrapper[4805]: I1203 14:15:18.632702 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-jftlh"] Dec 03 14:15:18 crc kubenswrapper[4805]: I1203 14:15:18.683687 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jk8sv\" (UniqueName: \"kubernetes.io/projected/0ab10c8b-6fa6-44a5-87af-211ef363ec7f-kube-api-access-jk8sv\") pod \"redhat-marketplace-jftlh\" (UID: \"0ab10c8b-6fa6-44a5-87af-211ef363ec7f\") " pod="openshift-marketplace/redhat-marketplace-jftlh" Dec 03 14:15:18 crc kubenswrapper[4805]: I1203 14:15:18.683741 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0ab10c8b-6fa6-44a5-87af-211ef363ec7f-catalog-content\") pod \"redhat-marketplace-jftlh\" (UID: \"0ab10c8b-6fa6-44a5-87af-211ef363ec7f\") " pod="openshift-marketplace/redhat-marketplace-jftlh" Dec 03 14:15:18 crc kubenswrapper[4805]: I1203 14:15:18.683768 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0ab10c8b-6fa6-44a5-87af-211ef363ec7f-utilities\") pod \"redhat-marketplace-jftlh\" (UID: \"0ab10c8b-6fa6-44a5-87af-211ef363ec7f\") " pod="openshift-marketplace/redhat-marketplace-jftlh" Dec 03 14:15:18 crc kubenswrapper[4805]: I1203 14:15:18.701306 4805 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1e412276-9d78-4771-b447-b4e2b388e604" path="/var/lib/kubelet/pods/1e412276-9d78-4771-b447-b4e2b388e604/volumes" Dec 03 14:15:18 crc kubenswrapper[4805]: I1203 14:15:18.703662 4805 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="344cb196-3baa-48c0-abcb-7e46cbce614d" path="/var/lib/kubelet/pods/344cb196-3baa-48c0-abcb-7e46cbce614d/volumes" Dec 03 14:15:18 crc kubenswrapper[4805]: I1203 14:15:18.704372 4805 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="66812e7e-aa85-4045-8e7a-8967b84f8849" path="/var/lib/kubelet/pods/66812e7e-aa85-4045-8e7a-8967b84f8849/volumes" Dec 03 14:15:18 crc kubenswrapper[4805]: I1203 14:15:18.705283 4805 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b2e4ac3a-e225-46ae-874d-4cf931c13464" path="/var/lib/kubelet/pods/b2e4ac3a-e225-46ae-874d-4cf931c13464/volumes" Dec 03 14:15:18 crc kubenswrapper[4805]: I1203 14:15:18.705831 4805 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d0a05061-dab0-4a85-be73-b1d2acbc7b8f" path="/var/lib/kubelet/pods/d0a05061-dab0-4a85-be73-b1d2acbc7b8f/volumes" Dec 03 14:15:18 crc kubenswrapper[4805]: I1203 14:15:18.785209 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jk8sv\" (UniqueName: \"kubernetes.io/projected/0ab10c8b-6fa6-44a5-87af-211ef363ec7f-kube-api-access-jk8sv\") pod \"redhat-marketplace-jftlh\" (UID: \"0ab10c8b-6fa6-44a5-87af-211ef363ec7f\") " pod="openshift-marketplace/redhat-marketplace-jftlh" Dec 03 14:15:18 crc kubenswrapper[4805]: I1203 14:15:18.785296 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0ab10c8b-6fa6-44a5-87af-211ef363ec7f-catalog-content\") pod \"redhat-marketplace-jftlh\" (UID: \"0ab10c8b-6fa6-44a5-87af-211ef363ec7f\") " pod="openshift-marketplace/redhat-marketplace-jftlh" Dec 03 14:15:18 crc kubenswrapper[4805]: I1203 14:15:18.785327 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0ab10c8b-6fa6-44a5-87af-211ef363ec7f-utilities\") pod \"redhat-marketplace-jftlh\" (UID: \"0ab10c8b-6fa6-44a5-87af-211ef363ec7f\") " pod="openshift-marketplace/redhat-marketplace-jftlh" Dec 03 14:15:18 crc kubenswrapper[4805]: I1203 14:15:18.786109 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0ab10c8b-6fa6-44a5-87af-211ef363ec7f-catalog-content\") pod \"redhat-marketplace-jftlh\" (UID: \"0ab10c8b-6fa6-44a5-87af-211ef363ec7f\") " pod="openshift-marketplace/redhat-marketplace-jftlh" Dec 03 14:15:18 crc kubenswrapper[4805]: I1203 14:15:18.786264 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0ab10c8b-6fa6-44a5-87af-211ef363ec7f-utilities\") pod \"redhat-marketplace-jftlh\" (UID: \"0ab10c8b-6fa6-44a5-87af-211ef363ec7f\") " pod="openshift-marketplace/redhat-marketplace-jftlh" Dec 03 14:15:18 crc kubenswrapper[4805]: I1203 14:15:18.812380 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jk8sv\" (UniqueName: \"kubernetes.io/projected/0ab10c8b-6fa6-44a5-87af-211ef363ec7f-kube-api-access-jk8sv\") pod \"redhat-marketplace-jftlh\" (UID: \"0ab10c8b-6fa6-44a5-87af-211ef363ec7f\") " pod="openshift-marketplace/redhat-marketplace-jftlh" Dec 03 14:15:18 crc kubenswrapper[4805]: I1203 14:15:18.823306 4805 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-v8czz"] Dec 03 14:15:18 crc kubenswrapper[4805]: I1203 14:15:18.828635 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-v8czz" Dec 03 14:15:18 crc kubenswrapper[4805]: I1203 14:15:18.839142 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-operators-dockercfg-ct8rh" Dec 03 14:15:18 crc kubenswrapper[4805]: I1203 14:15:18.850896 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-v8czz"] Dec 03 14:15:18 crc kubenswrapper[4805]: I1203 14:15:18.886025 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/33e30caf-a8b4-4472-9f81-20a67a0ef486-catalog-content\") pod \"redhat-operators-v8czz\" (UID: \"33e30caf-a8b4-4472-9f81-20a67a0ef486\") " pod="openshift-marketplace/redhat-operators-v8czz" Dec 03 14:15:18 crc kubenswrapper[4805]: I1203 14:15:18.886144 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-f2sb7\" (UniqueName: \"kubernetes.io/projected/33e30caf-a8b4-4472-9f81-20a67a0ef486-kube-api-access-f2sb7\") pod \"redhat-operators-v8czz\" (UID: \"33e30caf-a8b4-4472-9f81-20a67a0ef486\") " pod="openshift-marketplace/redhat-operators-v8czz" Dec 03 14:15:18 crc kubenswrapper[4805]: I1203 14:15:18.886175 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/33e30caf-a8b4-4472-9f81-20a67a0ef486-utilities\") pod \"redhat-operators-v8czz\" (UID: \"33e30caf-a8b4-4472-9f81-20a67a0ef486\") " pod="openshift-marketplace/redhat-operators-v8czz" Dec 03 14:15:18 crc kubenswrapper[4805]: I1203 14:15:18.936567 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-jftlh" Dec 03 14:15:18 crc kubenswrapper[4805]: I1203 14:15:18.987115 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-f2sb7\" (UniqueName: \"kubernetes.io/projected/33e30caf-a8b4-4472-9f81-20a67a0ef486-kube-api-access-f2sb7\") pod \"redhat-operators-v8czz\" (UID: \"33e30caf-a8b4-4472-9f81-20a67a0ef486\") " pod="openshift-marketplace/redhat-operators-v8czz" Dec 03 14:15:18 crc kubenswrapper[4805]: I1203 14:15:18.987168 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/33e30caf-a8b4-4472-9f81-20a67a0ef486-utilities\") pod \"redhat-operators-v8czz\" (UID: \"33e30caf-a8b4-4472-9f81-20a67a0ef486\") " pod="openshift-marketplace/redhat-operators-v8czz" Dec 03 14:15:18 crc kubenswrapper[4805]: I1203 14:15:18.987202 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/33e30caf-a8b4-4472-9f81-20a67a0ef486-catalog-content\") pod \"redhat-operators-v8czz\" (UID: \"33e30caf-a8b4-4472-9f81-20a67a0ef486\") " pod="openshift-marketplace/redhat-operators-v8czz" Dec 03 14:15:18 crc kubenswrapper[4805]: I1203 14:15:18.987773 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/33e30caf-a8b4-4472-9f81-20a67a0ef486-catalog-content\") pod \"redhat-operators-v8czz\" (UID: \"33e30caf-a8b4-4472-9f81-20a67a0ef486\") " pod="openshift-marketplace/redhat-operators-v8czz" Dec 03 14:15:18 crc kubenswrapper[4805]: I1203 14:15:18.987870 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/33e30caf-a8b4-4472-9f81-20a67a0ef486-utilities\") pod \"redhat-operators-v8czz\" (UID: \"33e30caf-a8b4-4472-9f81-20a67a0ef486\") " pod="openshift-marketplace/redhat-operators-v8czz" Dec 03 14:15:19 crc kubenswrapper[4805]: I1203 14:15:19.013517 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-f2sb7\" (UniqueName: \"kubernetes.io/projected/33e30caf-a8b4-4472-9f81-20a67a0ef486-kube-api-access-f2sb7\") pod \"redhat-operators-v8czz\" (UID: \"33e30caf-a8b4-4472-9f81-20a67a0ef486\") " pod="openshift-marketplace/redhat-operators-v8czz" Dec 03 14:15:19 crc kubenswrapper[4805]: I1203 14:15:19.027164 4805 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-image-registry/image-registry-66df7c8f76-vltnj" Dec 03 14:15:19 crc kubenswrapper[4805]: I1203 14:15:19.088405 4805 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-jjzct"] Dec 03 14:15:19 crc kubenswrapper[4805]: I1203 14:15:19.159247 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-v8czz" Dec 03 14:15:19 crc kubenswrapper[4805]: I1203 14:15:19.549425 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-jftlh"] Dec 03 14:15:19 crc kubenswrapper[4805]: I1203 14:15:19.613688 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-v8czz"] Dec 03 14:15:19 crc kubenswrapper[4805]: W1203 14:15:19.618061 4805 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod33e30caf_a8b4_4472_9f81_20a67a0ef486.slice/crio-70f6238a36c11035104c0eaf80c9f04ce92ca3fb48e313bee4b43f6ca408156f WatchSource:0}: Error finding container 70f6238a36c11035104c0eaf80c9f04ce92ca3fb48e313bee4b43f6ca408156f: Status 404 returned error can't find the container with id 70f6238a36c11035104c0eaf80c9f04ce92ca3fb48e313bee4b43f6ca408156f Dec 03 14:15:19 crc kubenswrapper[4805]: I1203 14:15:19.987814 4805 generic.go:334] "Generic (PLEG): container finished" podID="33e30caf-a8b4-4472-9f81-20a67a0ef486" containerID="eb5ea4afe6fcba2906e2da9e9b48e6c33218e0e0d29ea8217c0fc04884dacb8d" exitCode=0 Dec 03 14:15:19 crc kubenswrapper[4805]: I1203 14:15:19.987870 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-v8czz" event={"ID":"33e30caf-a8b4-4472-9f81-20a67a0ef486","Type":"ContainerDied","Data":"eb5ea4afe6fcba2906e2da9e9b48e6c33218e0e0d29ea8217c0fc04884dacb8d"} Dec 03 14:15:19 crc kubenswrapper[4805]: I1203 14:15:19.988286 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-v8czz" event={"ID":"33e30caf-a8b4-4472-9f81-20a67a0ef486","Type":"ContainerStarted","Data":"70f6238a36c11035104c0eaf80c9f04ce92ca3fb48e313bee4b43f6ca408156f"} Dec 03 14:15:19 crc kubenswrapper[4805]: I1203 14:15:19.990207 4805 generic.go:334] "Generic (PLEG): container finished" podID="0ab10c8b-6fa6-44a5-87af-211ef363ec7f" containerID="b53d7513ecbc6ff28e3a0740f49f6273acd9bd885e1a664744a5f8464641e79d" exitCode=0 Dec 03 14:15:19 crc kubenswrapper[4805]: I1203 14:15:19.990271 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-jftlh" event={"ID":"0ab10c8b-6fa6-44a5-87af-211ef363ec7f","Type":"ContainerDied","Data":"b53d7513ecbc6ff28e3a0740f49f6273acd9bd885e1a664744a5f8464641e79d"} Dec 03 14:15:19 crc kubenswrapper[4805]: I1203 14:15:19.990300 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-jftlh" event={"ID":"0ab10c8b-6fa6-44a5-87af-211ef363ec7f","Type":"ContainerStarted","Data":"7946485d1fed99af40da1dc4b6a147c9e1908b8fc7d2789483c24f2b5f436e19"} Dec 03 14:15:21 crc kubenswrapper[4805]: I1203 14:15:21.019401 4805 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-c4fdw"] Dec 03 14:15:21 crc kubenswrapper[4805]: I1203 14:15:21.027810 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-c4fdw" Dec 03 14:15:21 crc kubenswrapper[4805]: I1203 14:15:21.039397 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"certified-operators-dockercfg-4rs5g" Dec 03 14:15:21 crc kubenswrapper[4805]: I1203 14:15:21.048958 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-c4fdw"] Dec 03 14:15:21 crc kubenswrapper[4805]: I1203 14:15:21.214135 4805 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-xfg6c"] Dec 03 14:15:21 crc kubenswrapper[4805]: I1203 14:15:21.215046 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-xfg6c" Dec 03 14:15:21 crc kubenswrapper[4805]: I1203 14:15:21.217279 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"community-operators-dockercfg-dmngl" Dec 03 14:15:21 crc kubenswrapper[4805]: I1203 14:15:21.226119 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6fcdd680-c71f-4978-992a-65378f16e0a1-catalog-content\") pod \"certified-operators-c4fdw\" (UID: \"6fcdd680-c71f-4978-992a-65378f16e0a1\") " pod="openshift-marketplace/certified-operators-c4fdw" Dec 03 14:15:21 crc kubenswrapper[4805]: I1203 14:15:21.226152 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mt5wj\" (UniqueName: \"kubernetes.io/projected/6fcdd680-c71f-4978-992a-65378f16e0a1-kube-api-access-mt5wj\") pod \"certified-operators-c4fdw\" (UID: \"6fcdd680-c71f-4978-992a-65378f16e0a1\") " pod="openshift-marketplace/certified-operators-c4fdw" Dec 03 14:15:21 crc kubenswrapper[4805]: I1203 14:15:21.226204 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6fcdd680-c71f-4978-992a-65378f16e0a1-utilities\") pod \"certified-operators-c4fdw\" (UID: \"6fcdd680-c71f-4978-992a-65378f16e0a1\") " pod="openshift-marketplace/certified-operators-c4fdw" Dec 03 14:15:21 crc kubenswrapper[4805]: I1203 14:15:21.227196 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-xfg6c"] Dec 03 14:15:21 crc kubenswrapper[4805]: I1203 14:15:21.328288 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a93d60c0-f172-49ae-af40-0d5bbda4877f-catalog-content\") pod \"community-operators-xfg6c\" (UID: \"a93d60c0-f172-49ae-af40-0d5bbda4877f\") " pod="openshift-marketplace/community-operators-xfg6c" Dec 03 14:15:21 crc kubenswrapper[4805]: I1203 14:15:21.328382 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6fcdd680-c71f-4978-992a-65378f16e0a1-catalog-content\") pod \"certified-operators-c4fdw\" (UID: \"6fcdd680-c71f-4978-992a-65378f16e0a1\") " pod="openshift-marketplace/certified-operators-c4fdw" Dec 03 14:15:21 crc kubenswrapper[4805]: I1203 14:15:21.328492 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mt5wj\" (UniqueName: \"kubernetes.io/projected/6fcdd680-c71f-4978-992a-65378f16e0a1-kube-api-access-mt5wj\") pod \"certified-operators-c4fdw\" (UID: \"6fcdd680-c71f-4978-992a-65378f16e0a1\") " pod="openshift-marketplace/certified-operators-c4fdw" Dec 03 14:15:21 crc kubenswrapper[4805]: I1203 14:15:21.328661 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-c6p7x\" (UniqueName: \"kubernetes.io/projected/a93d60c0-f172-49ae-af40-0d5bbda4877f-kube-api-access-c6p7x\") pod \"community-operators-xfg6c\" (UID: \"a93d60c0-f172-49ae-af40-0d5bbda4877f\") " pod="openshift-marketplace/community-operators-xfg6c" Dec 03 14:15:21 crc kubenswrapper[4805]: I1203 14:15:21.328733 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a93d60c0-f172-49ae-af40-0d5bbda4877f-utilities\") pod \"community-operators-xfg6c\" (UID: \"a93d60c0-f172-49ae-af40-0d5bbda4877f\") " pod="openshift-marketplace/community-operators-xfg6c" Dec 03 14:15:21 crc kubenswrapper[4805]: I1203 14:15:21.328772 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6fcdd680-c71f-4978-992a-65378f16e0a1-utilities\") pod \"certified-operators-c4fdw\" (UID: \"6fcdd680-c71f-4978-992a-65378f16e0a1\") " pod="openshift-marketplace/certified-operators-c4fdw" Dec 03 14:15:21 crc kubenswrapper[4805]: I1203 14:15:21.329296 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6fcdd680-c71f-4978-992a-65378f16e0a1-catalog-content\") pod \"certified-operators-c4fdw\" (UID: \"6fcdd680-c71f-4978-992a-65378f16e0a1\") " pod="openshift-marketplace/certified-operators-c4fdw" Dec 03 14:15:21 crc kubenswrapper[4805]: I1203 14:15:21.329327 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6fcdd680-c71f-4978-992a-65378f16e0a1-utilities\") pod \"certified-operators-c4fdw\" (UID: \"6fcdd680-c71f-4978-992a-65378f16e0a1\") " pod="openshift-marketplace/certified-operators-c4fdw" Dec 03 14:15:21 crc kubenswrapper[4805]: I1203 14:15:21.372570 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mt5wj\" (UniqueName: \"kubernetes.io/projected/6fcdd680-c71f-4978-992a-65378f16e0a1-kube-api-access-mt5wj\") pod \"certified-operators-c4fdw\" (UID: \"6fcdd680-c71f-4978-992a-65378f16e0a1\") " pod="openshift-marketplace/certified-operators-c4fdw" Dec 03 14:15:21 crc kubenswrapper[4805]: I1203 14:15:21.430615 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-c6p7x\" (UniqueName: \"kubernetes.io/projected/a93d60c0-f172-49ae-af40-0d5bbda4877f-kube-api-access-c6p7x\") pod \"community-operators-xfg6c\" (UID: \"a93d60c0-f172-49ae-af40-0d5bbda4877f\") " pod="openshift-marketplace/community-operators-xfg6c" Dec 03 14:15:21 crc kubenswrapper[4805]: I1203 14:15:21.430662 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a93d60c0-f172-49ae-af40-0d5bbda4877f-utilities\") pod \"community-operators-xfg6c\" (UID: \"a93d60c0-f172-49ae-af40-0d5bbda4877f\") " pod="openshift-marketplace/community-operators-xfg6c" Dec 03 14:15:21 crc kubenswrapper[4805]: I1203 14:15:21.430710 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a93d60c0-f172-49ae-af40-0d5bbda4877f-catalog-content\") pod \"community-operators-xfg6c\" (UID: \"a93d60c0-f172-49ae-af40-0d5bbda4877f\") " pod="openshift-marketplace/community-operators-xfg6c" Dec 03 14:15:21 crc kubenswrapper[4805]: I1203 14:15:21.431177 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a93d60c0-f172-49ae-af40-0d5bbda4877f-catalog-content\") pod \"community-operators-xfg6c\" (UID: \"a93d60c0-f172-49ae-af40-0d5bbda4877f\") " pod="openshift-marketplace/community-operators-xfg6c" Dec 03 14:15:21 crc kubenswrapper[4805]: I1203 14:15:21.431318 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a93d60c0-f172-49ae-af40-0d5bbda4877f-utilities\") pod \"community-operators-xfg6c\" (UID: \"a93d60c0-f172-49ae-af40-0d5bbda4877f\") " pod="openshift-marketplace/community-operators-xfg6c" Dec 03 14:15:21 crc kubenswrapper[4805]: I1203 14:15:21.450128 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-c6p7x\" (UniqueName: \"kubernetes.io/projected/a93d60c0-f172-49ae-af40-0d5bbda4877f-kube-api-access-c6p7x\") pod \"community-operators-xfg6c\" (UID: \"a93d60c0-f172-49ae-af40-0d5bbda4877f\") " pod="openshift-marketplace/community-operators-xfg6c" Dec 03 14:15:21 crc kubenswrapper[4805]: I1203 14:15:21.530158 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-xfg6c" Dec 03 14:15:21 crc kubenswrapper[4805]: I1203 14:15:21.670296 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-c4fdw" Dec 03 14:15:21 crc kubenswrapper[4805]: I1203 14:15:21.943978 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-xfg6c"] Dec 03 14:15:22 crc kubenswrapper[4805]: I1203 14:15:22.003130 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-xfg6c" event={"ID":"a93d60c0-f172-49ae-af40-0d5bbda4877f","Type":"ContainerStarted","Data":"3170d0b5b1f9ba7c9e5a8e619fc73bc1ed5c8069923796aab0fb4338cd29f50d"} Dec 03 14:15:22 crc kubenswrapper[4805]: I1203 14:15:22.005039 4805 generic.go:334] "Generic (PLEG): container finished" podID="0ab10c8b-6fa6-44a5-87af-211ef363ec7f" containerID="6f6f25946f317dc523b4025de0bd775fb2728537dd6e4db98f8db20c166e0da1" exitCode=0 Dec 03 14:15:22 crc kubenswrapper[4805]: I1203 14:15:22.005091 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-jftlh" event={"ID":"0ab10c8b-6fa6-44a5-87af-211ef363ec7f","Type":"ContainerDied","Data":"6f6f25946f317dc523b4025de0bd775fb2728537dd6e4db98f8db20c166e0da1"} Dec 03 14:15:22 crc kubenswrapper[4805]: I1203 14:15:22.018340 4805 generic.go:334] "Generic (PLEG): container finished" podID="33e30caf-a8b4-4472-9f81-20a67a0ef486" containerID="81690fa00b58d5532e4b40427ff12823c125c56924c4c41c5ea3b26a1ab16039" exitCode=0 Dec 03 14:15:22 crc kubenswrapper[4805]: I1203 14:15:22.018394 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-v8czz" event={"ID":"33e30caf-a8b4-4472-9f81-20a67a0ef486","Type":"ContainerDied","Data":"81690fa00b58d5532e4b40427ff12823c125c56924c4c41c5ea3b26a1ab16039"} Dec 03 14:15:22 crc kubenswrapper[4805]: I1203 14:15:22.099730 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-c4fdw"] Dec 03 14:15:23 crc kubenswrapper[4805]: I1203 14:15:23.025640 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-v8czz" event={"ID":"33e30caf-a8b4-4472-9f81-20a67a0ef486","Type":"ContainerStarted","Data":"dcabbc03c6bc0c020bcaf3bdac0e38a0ca7d30827b07c019473d511646280e3c"} Dec 03 14:15:23 crc kubenswrapper[4805]: I1203 14:15:23.027025 4805 generic.go:334] "Generic (PLEG): container finished" podID="a93d60c0-f172-49ae-af40-0d5bbda4877f" containerID="75a23bffb5af21cba6e13d685626f4efe044a4a344d428f346a294a1d0b08ec1" exitCode=0 Dec 03 14:15:23 crc kubenswrapper[4805]: I1203 14:15:23.027080 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-xfg6c" event={"ID":"a93d60c0-f172-49ae-af40-0d5bbda4877f","Type":"ContainerDied","Data":"75a23bffb5af21cba6e13d685626f4efe044a4a344d428f346a294a1d0b08ec1"} Dec 03 14:15:23 crc kubenswrapper[4805]: I1203 14:15:23.031616 4805 generic.go:334] "Generic (PLEG): container finished" podID="6fcdd680-c71f-4978-992a-65378f16e0a1" containerID="c11c891142c6362746ca73289c53182b3702296a0f6c384f69e717371e3374d4" exitCode=0 Dec 03 14:15:23 crc kubenswrapper[4805]: I1203 14:15:23.031684 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-c4fdw" event={"ID":"6fcdd680-c71f-4978-992a-65378f16e0a1","Type":"ContainerDied","Data":"c11c891142c6362746ca73289c53182b3702296a0f6c384f69e717371e3374d4"} Dec 03 14:15:23 crc kubenswrapper[4805]: I1203 14:15:23.031714 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-c4fdw" event={"ID":"6fcdd680-c71f-4978-992a-65378f16e0a1","Type":"ContainerStarted","Data":"1136b258d8171bdddf2c28b9ca78a545a0df4044ba14407f25da985539ce93b1"} Dec 03 14:15:23 crc kubenswrapper[4805]: I1203 14:15:23.036528 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-jftlh" event={"ID":"0ab10c8b-6fa6-44a5-87af-211ef363ec7f","Type":"ContainerStarted","Data":"d3c5a24fa5f6827568f392f923f1b616792f8e32bd0fa5c0609273c36e911811"} Dec 03 14:15:23 crc kubenswrapper[4805]: I1203 14:15:23.063614 4805 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-jftlh" podStartSLOduration=2.343647073 podStartE2EDuration="5.063594411s" podCreationTimestamp="2025-12-03 14:15:18 +0000 UTC" firstStartedPulling="2025-12-03 14:15:19.991198941 +0000 UTC m=+349.654115864" lastFinishedPulling="2025-12-03 14:15:22.711146279 +0000 UTC m=+352.374063202" observedRunningTime="2025-12-03 14:15:23.062009958 +0000 UTC m=+352.724926921" watchObservedRunningTime="2025-12-03 14:15:23.063594411 +0000 UTC m=+352.726511374" Dec 03 14:15:23 crc kubenswrapper[4805]: I1203 14:15:23.064681 4805 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-v8czz" podStartSLOduration=2.331444031 podStartE2EDuration="5.06467032s" podCreationTimestamp="2025-12-03 14:15:18 +0000 UTC" firstStartedPulling="2025-12-03 14:15:19.988890718 +0000 UTC m=+349.651807641" lastFinishedPulling="2025-12-03 14:15:22.722117007 +0000 UTC m=+352.385033930" observedRunningTime="2025-12-03 14:15:23.047145624 +0000 UTC m=+352.710062547" watchObservedRunningTime="2025-12-03 14:15:23.06467032 +0000 UTC m=+352.727587283" Dec 03 14:15:24 crc kubenswrapper[4805]: I1203 14:15:24.042792 4805 generic.go:334] "Generic (PLEG): container finished" podID="a93d60c0-f172-49ae-af40-0d5bbda4877f" containerID="aea354e18556fb7d2644769dc2856caf665cf7254e4500c22baeb83b7dc890d4" exitCode=0 Dec 03 14:15:24 crc kubenswrapper[4805]: I1203 14:15:24.043072 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-xfg6c" event={"ID":"a93d60c0-f172-49ae-af40-0d5bbda4877f","Type":"ContainerDied","Data":"aea354e18556fb7d2644769dc2856caf665cf7254e4500c22baeb83b7dc890d4"} Dec 03 14:15:24 crc kubenswrapper[4805]: I1203 14:15:24.044626 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-c4fdw" event={"ID":"6fcdd680-c71f-4978-992a-65378f16e0a1","Type":"ContainerStarted","Data":"95a8267c98281a194dca4270336b0290176ae000bd3a52cc7324fcf221f62576"} Dec 03 14:15:25 crc kubenswrapper[4805]: I1203 14:15:25.051415 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-xfg6c" event={"ID":"a93d60c0-f172-49ae-af40-0d5bbda4877f","Type":"ContainerStarted","Data":"fa82287a64c2d6ed9b8e414be4b21767592b29b109e1bc6bb60d7667454f9b5a"} Dec 03 14:15:25 crc kubenswrapper[4805]: I1203 14:15:25.053315 4805 generic.go:334] "Generic (PLEG): container finished" podID="6fcdd680-c71f-4978-992a-65378f16e0a1" containerID="95a8267c98281a194dca4270336b0290176ae000bd3a52cc7324fcf221f62576" exitCode=0 Dec 03 14:15:25 crc kubenswrapper[4805]: I1203 14:15:25.053365 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-c4fdw" event={"ID":"6fcdd680-c71f-4978-992a-65378f16e0a1","Type":"ContainerDied","Data":"95a8267c98281a194dca4270336b0290176ae000bd3a52cc7324fcf221f62576"} Dec 03 14:15:25 crc kubenswrapper[4805]: I1203 14:15:25.072100 4805 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-xfg6c" podStartSLOduration=2.619457453 podStartE2EDuration="4.072081037s" podCreationTimestamp="2025-12-03 14:15:21 +0000 UTC" firstStartedPulling="2025-12-03 14:15:23.028327922 +0000 UTC m=+352.691244835" lastFinishedPulling="2025-12-03 14:15:24.480951496 +0000 UTC m=+354.143868419" observedRunningTime="2025-12-03 14:15:25.070289408 +0000 UTC m=+354.733206341" watchObservedRunningTime="2025-12-03 14:15:25.072081037 +0000 UTC m=+354.734997980" Dec 03 14:15:27 crc kubenswrapper[4805]: I1203 14:15:27.066412 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-c4fdw" event={"ID":"6fcdd680-c71f-4978-992a-65378f16e0a1","Type":"ContainerStarted","Data":"db5bd15236267f42182c4eb5cc93a48bb3f011d99babef6ba4e0e797994496e4"} Dec 03 14:15:27 crc kubenswrapper[4805]: I1203 14:15:27.085466 4805 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-c4fdw" podStartSLOduration=4.589083555 podStartE2EDuration="7.085448885s" podCreationTimestamp="2025-12-03 14:15:20 +0000 UTC" firstStartedPulling="2025-12-03 14:15:23.034714966 +0000 UTC m=+352.697631889" lastFinishedPulling="2025-12-03 14:15:25.531080296 +0000 UTC m=+355.193997219" observedRunningTime="2025-12-03 14:15:27.083188423 +0000 UTC m=+356.746105346" watchObservedRunningTime="2025-12-03 14:15:27.085448885 +0000 UTC m=+356.748365808" Dec 03 14:15:28 crc kubenswrapper[4805]: I1203 14:15:28.937901 4805 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-jftlh" Dec 03 14:15:28 crc kubenswrapper[4805]: I1203 14:15:28.937978 4805 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-jftlh" Dec 03 14:15:29 crc kubenswrapper[4805]: I1203 14:15:29.001456 4805 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-jftlh" Dec 03 14:15:29 crc kubenswrapper[4805]: I1203 14:15:29.114299 4805 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-jftlh" Dec 03 14:15:29 crc kubenswrapper[4805]: I1203 14:15:29.178990 4805 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-v8czz" Dec 03 14:15:29 crc kubenswrapper[4805]: I1203 14:15:29.179061 4805 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-v8czz" Dec 03 14:15:29 crc kubenswrapper[4805]: I1203 14:15:29.215866 4805 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-v8czz" Dec 03 14:15:30 crc kubenswrapper[4805]: I1203 14:15:30.125606 4805 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-v8czz" Dec 03 14:15:31 crc kubenswrapper[4805]: I1203 14:15:31.530878 4805 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-xfg6c" Dec 03 14:15:31 crc kubenswrapper[4805]: I1203 14:15:31.531033 4805 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-xfg6c" Dec 03 14:15:31 crc kubenswrapper[4805]: I1203 14:15:31.578814 4805 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-xfg6c" Dec 03 14:15:31 crc kubenswrapper[4805]: I1203 14:15:31.670715 4805 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-c4fdw" Dec 03 14:15:31 crc kubenswrapper[4805]: I1203 14:15:31.671059 4805 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-c4fdw" Dec 03 14:15:31 crc kubenswrapper[4805]: I1203 14:15:31.721073 4805 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-c4fdw" Dec 03 14:15:32 crc kubenswrapper[4805]: I1203 14:15:32.148950 4805 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-c4fdw" Dec 03 14:15:32 crc kubenswrapper[4805]: I1203 14:15:32.152006 4805 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-xfg6c" Dec 03 14:15:43 crc kubenswrapper[4805]: I1203 14:15:43.917785 4805 patch_prober.go:28] interesting pod/machine-config-daemon-gskh4 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 03 14:15:43 crc kubenswrapper[4805]: I1203 14:15:43.918876 4805 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-gskh4" podUID="6098937f-e3f6-45e8-a647-4994a79cd711" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 03 14:15:44 crc kubenswrapper[4805]: I1203 14:15:44.125471 4805 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-image-registry/image-registry-697d97f7c8-jjzct" podUID="4655a83e-4980-4a5c-9f61-e305ae418822" containerName="registry" containerID="cri-o://2eb46bb7d9082e503a63d58a515289f423d646a019ee0ec11dd33ae69c6562a5" gracePeriod=30 Dec 03 14:15:44 crc kubenswrapper[4805]: I1203 14:15:44.605055 4805 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-jjzct" Dec 03 14:15:44 crc kubenswrapper[4805]: I1203 14:15:44.678662 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/4655a83e-4980-4a5c-9f61-e305ae418822-ca-trust-extracted\") pod \"4655a83e-4980-4a5c-9f61-e305ae418822\" (UID: \"4655a83e-4980-4a5c-9f61-e305ae418822\") " Dec 03 14:15:44 crc kubenswrapper[4805]: I1203 14:15:44.678723 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/4655a83e-4980-4a5c-9f61-e305ae418822-bound-sa-token\") pod \"4655a83e-4980-4a5c-9f61-e305ae418822\" (UID: \"4655a83e-4980-4a5c-9f61-e305ae418822\") " Dec 03 14:15:44 crc kubenswrapper[4805]: I1203 14:15:44.678742 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-m7vtp\" (UniqueName: \"kubernetes.io/projected/4655a83e-4980-4a5c-9f61-e305ae418822-kube-api-access-m7vtp\") pod \"4655a83e-4980-4a5c-9f61-e305ae418822\" (UID: \"4655a83e-4980-4a5c-9f61-e305ae418822\") " Dec 03 14:15:44 crc kubenswrapper[4805]: I1203 14:15:44.678795 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/4655a83e-4980-4a5c-9f61-e305ae418822-registry-certificates\") pod \"4655a83e-4980-4a5c-9f61-e305ae418822\" (UID: \"4655a83e-4980-4a5c-9f61-e305ae418822\") " Dec 03 14:15:44 crc kubenswrapper[4805]: I1203 14:15:44.678891 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/4655a83e-4980-4a5c-9f61-e305ae418822-trusted-ca\") pod \"4655a83e-4980-4a5c-9f61-e305ae418822\" (UID: \"4655a83e-4980-4a5c-9f61-e305ae418822\") " Dec 03 14:15:44 crc kubenswrapper[4805]: I1203 14:15:44.679061 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-storage\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"4655a83e-4980-4a5c-9f61-e305ae418822\" (UID: \"4655a83e-4980-4a5c-9f61-e305ae418822\") " Dec 03 14:15:44 crc kubenswrapper[4805]: I1203 14:15:44.679099 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/4655a83e-4980-4a5c-9f61-e305ae418822-registry-tls\") pod \"4655a83e-4980-4a5c-9f61-e305ae418822\" (UID: \"4655a83e-4980-4a5c-9f61-e305ae418822\") " Dec 03 14:15:44 crc kubenswrapper[4805]: I1203 14:15:44.679126 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/4655a83e-4980-4a5c-9f61-e305ae418822-installation-pull-secrets\") pod \"4655a83e-4980-4a5c-9f61-e305ae418822\" (UID: \"4655a83e-4980-4a5c-9f61-e305ae418822\") " Dec 03 14:15:44 crc kubenswrapper[4805]: I1203 14:15:44.680957 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4655a83e-4980-4a5c-9f61-e305ae418822-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "4655a83e-4980-4a5c-9f61-e305ae418822" (UID: "4655a83e-4980-4a5c-9f61-e305ae418822"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 14:15:44 crc kubenswrapper[4805]: I1203 14:15:44.681831 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4655a83e-4980-4a5c-9f61-e305ae418822-registry-certificates" (OuterVolumeSpecName: "registry-certificates") pod "4655a83e-4980-4a5c-9f61-e305ae418822" (UID: "4655a83e-4980-4a5c-9f61-e305ae418822"). InnerVolumeSpecName "registry-certificates". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 14:15:44 crc kubenswrapper[4805]: I1203 14:15:44.688380 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4655a83e-4980-4a5c-9f61-e305ae418822-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "4655a83e-4980-4a5c-9f61-e305ae418822" (UID: "4655a83e-4980-4a5c-9f61-e305ae418822"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 14:15:44 crc kubenswrapper[4805]: I1203 14:15:44.696064 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4655a83e-4980-4a5c-9f61-e305ae418822-installation-pull-secrets" (OuterVolumeSpecName: "installation-pull-secrets") pod "4655a83e-4980-4a5c-9f61-e305ae418822" (UID: "4655a83e-4980-4a5c-9f61-e305ae418822"). InnerVolumeSpecName "installation-pull-secrets". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 14:15:44 crc kubenswrapper[4805]: I1203 14:15:44.697277 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/4655a83e-4980-4a5c-9f61-e305ae418822-ca-trust-extracted" (OuterVolumeSpecName: "ca-trust-extracted") pod "4655a83e-4980-4a5c-9f61-e305ae418822" (UID: "4655a83e-4980-4a5c-9f61-e305ae418822"). InnerVolumeSpecName "ca-trust-extracted". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 14:15:44 crc kubenswrapper[4805]: I1203 14:15:44.700140 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4655a83e-4980-4a5c-9f61-e305ae418822-registry-tls" (OuterVolumeSpecName: "registry-tls") pod "4655a83e-4980-4a5c-9f61-e305ae418822" (UID: "4655a83e-4980-4a5c-9f61-e305ae418822"). InnerVolumeSpecName "registry-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 14:15:44 crc kubenswrapper[4805]: I1203 14:15:44.710225 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (OuterVolumeSpecName: "registry-storage") pod "4655a83e-4980-4a5c-9f61-e305ae418822" (UID: "4655a83e-4980-4a5c-9f61-e305ae418822"). InnerVolumeSpecName "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8". PluginName "kubernetes.io/csi", VolumeGidValue "" Dec 03 14:15:44 crc kubenswrapper[4805]: I1203 14:15:44.710269 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4655a83e-4980-4a5c-9f61-e305ae418822-kube-api-access-m7vtp" (OuterVolumeSpecName: "kube-api-access-m7vtp") pod "4655a83e-4980-4a5c-9f61-e305ae418822" (UID: "4655a83e-4980-4a5c-9f61-e305ae418822"). InnerVolumeSpecName "kube-api-access-m7vtp". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 14:15:44 crc kubenswrapper[4805]: I1203 14:15:44.780581 4805 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/4655a83e-4980-4a5c-9f61-e305ae418822-trusted-ca\") on node \"crc\" DevicePath \"\"" Dec 03 14:15:44 crc kubenswrapper[4805]: I1203 14:15:44.780716 4805 reconciler_common.go:293] "Volume detached for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/4655a83e-4980-4a5c-9f61-e305ae418822-registry-tls\") on node \"crc\" DevicePath \"\"" Dec 03 14:15:44 crc kubenswrapper[4805]: I1203 14:15:44.780860 4805 reconciler_common.go:293] "Volume detached for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/4655a83e-4980-4a5c-9f61-e305ae418822-installation-pull-secrets\") on node \"crc\" DevicePath \"\"" Dec 03 14:15:44 crc kubenswrapper[4805]: I1203 14:15:44.780895 4805 reconciler_common.go:293] "Volume detached for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/4655a83e-4980-4a5c-9f61-e305ae418822-ca-trust-extracted\") on node \"crc\" DevicePath \"\"" Dec 03 14:15:44 crc kubenswrapper[4805]: I1203 14:15:44.780906 4805 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/4655a83e-4980-4a5c-9f61-e305ae418822-bound-sa-token\") on node \"crc\" DevicePath \"\"" Dec 03 14:15:44 crc kubenswrapper[4805]: I1203 14:15:44.780915 4805 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-m7vtp\" (UniqueName: \"kubernetes.io/projected/4655a83e-4980-4a5c-9f61-e305ae418822-kube-api-access-m7vtp\") on node \"crc\" DevicePath \"\"" Dec 03 14:15:44 crc kubenswrapper[4805]: I1203 14:15:44.780924 4805 reconciler_common.go:293] "Volume detached for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/4655a83e-4980-4a5c-9f61-e305ae418822-registry-certificates\") on node \"crc\" DevicePath \"\"" Dec 03 14:15:45 crc kubenswrapper[4805]: I1203 14:15:45.167259 4805 generic.go:334] "Generic (PLEG): container finished" podID="4655a83e-4980-4a5c-9f61-e305ae418822" containerID="2eb46bb7d9082e503a63d58a515289f423d646a019ee0ec11dd33ae69c6562a5" exitCode=0 Dec 03 14:15:45 crc kubenswrapper[4805]: I1203 14:15:45.167323 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-jjzct" event={"ID":"4655a83e-4980-4a5c-9f61-e305ae418822","Type":"ContainerDied","Data":"2eb46bb7d9082e503a63d58a515289f423d646a019ee0ec11dd33ae69c6562a5"} Dec 03 14:15:45 crc kubenswrapper[4805]: I1203 14:15:45.167339 4805 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-jjzct" Dec 03 14:15:45 crc kubenswrapper[4805]: I1203 14:15:45.167601 4805 scope.go:117] "RemoveContainer" containerID="2eb46bb7d9082e503a63d58a515289f423d646a019ee0ec11dd33ae69c6562a5" Dec 03 14:15:45 crc kubenswrapper[4805]: I1203 14:15:45.167583 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-jjzct" event={"ID":"4655a83e-4980-4a5c-9f61-e305ae418822","Type":"ContainerDied","Data":"45e95adcdd0b0349f91467461f335ee50bd61445aaf7d6948daf31c5482d4d35"} Dec 03 14:15:45 crc kubenswrapper[4805]: I1203 14:15:45.190032 4805 scope.go:117] "RemoveContainer" containerID="2eb46bb7d9082e503a63d58a515289f423d646a019ee0ec11dd33ae69c6562a5" Dec 03 14:15:45 crc kubenswrapper[4805]: E1203 14:15:45.190419 4805 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2eb46bb7d9082e503a63d58a515289f423d646a019ee0ec11dd33ae69c6562a5\": container with ID starting with 2eb46bb7d9082e503a63d58a515289f423d646a019ee0ec11dd33ae69c6562a5 not found: ID does not exist" containerID="2eb46bb7d9082e503a63d58a515289f423d646a019ee0ec11dd33ae69c6562a5" Dec 03 14:15:45 crc kubenswrapper[4805]: I1203 14:15:45.190451 4805 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2eb46bb7d9082e503a63d58a515289f423d646a019ee0ec11dd33ae69c6562a5"} err="failed to get container status \"2eb46bb7d9082e503a63d58a515289f423d646a019ee0ec11dd33ae69c6562a5\": rpc error: code = NotFound desc = could not find container \"2eb46bb7d9082e503a63d58a515289f423d646a019ee0ec11dd33ae69c6562a5\": container with ID starting with 2eb46bb7d9082e503a63d58a515289f423d646a019ee0ec11dd33ae69c6562a5 not found: ID does not exist" Dec 03 14:15:45 crc kubenswrapper[4805]: I1203 14:15:45.210703 4805 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-jjzct"] Dec 03 14:15:45 crc kubenswrapper[4805]: I1203 14:15:45.219256 4805 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-jjzct"] Dec 03 14:15:46 crc kubenswrapper[4805]: I1203 14:15:46.703085 4805 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4655a83e-4980-4a5c-9f61-e305ae418822" path="/var/lib/kubelet/pods/4655a83e-4980-4a5c-9f61-e305ae418822/volumes" Dec 03 14:16:13 crc kubenswrapper[4805]: I1203 14:16:13.917241 4805 patch_prober.go:28] interesting pod/machine-config-daemon-gskh4 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 03 14:16:13 crc kubenswrapper[4805]: I1203 14:16:13.919443 4805 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-gskh4" podUID="6098937f-e3f6-45e8-a647-4994a79cd711" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 03 14:16:43 crc kubenswrapper[4805]: I1203 14:16:43.917221 4805 patch_prober.go:28] interesting pod/machine-config-daemon-gskh4 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 03 14:16:43 crc kubenswrapper[4805]: I1203 14:16:43.917915 4805 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-gskh4" podUID="6098937f-e3f6-45e8-a647-4994a79cd711" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 03 14:16:43 crc kubenswrapper[4805]: I1203 14:16:43.917973 4805 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-gskh4" Dec 03 14:16:44 crc kubenswrapper[4805]: I1203 14:16:44.539768 4805 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"550f8bfd5120af37ce2c9f018ac5ef6ec16519c28f68395f823626d542873ea8"} pod="openshift-machine-config-operator/machine-config-daemon-gskh4" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 03 14:16:44 crc kubenswrapper[4805]: I1203 14:16:44.539926 4805 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-gskh4" podUID="6098937f-e3f6-45e8-a647-4994a79cd711" containerName="machine-config-daemon" containerID="cri-o://550f8bfd5120af37ce2c9f018ac5ef6ec16519c28f68395f823626d542873ea8" gracePeriod=600 Dec 03 14:16:45 crc kubenswrapper[4805]: I1203 14:16:45.548581 4805 generic.go:334] "Generic (PLEG): container finished" podID="6098937f-e3f6-45e8-a647-4994a79cd711" containerID="550f8bfd5120af37ce2c9f018ac5ef6ec16519c28f68395f823626d542873ea8" exitCode=0 Dec 03 14:16:45 crc kubenswrapper[4805]: I1203 14:16:45.548721 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-gskh4" event={"ID":"6098937f-e3f6-45e8-a647-4994a79cd711","Type":"ContainerDied","Data":"550f8bfd5120af37ce2c9f018ac5ef6ec16519c28f68395f823626d542873ea8"} Dec 03 14:16:45 crc kubenswrapper[4805]: I1203 14:16:45.548880 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-gskh4" event={"ID":"6098937f-e3f6-45e8-a647-4994a79cd711","Type":"ContainerStarted","Data":"af1b8fb24b94658040c784c6bffe5034acb7fffeeb5210dc5f214b62d02ab400"} Dec 03 14:16:45 crc kubenswrapper[4805]: I1203 14:16:45.548903 4805 scope.go:117] "RemoveContainer" containerID="9ad6b68a42fd33ee167e67873c6eed3f11cdf6dc71d22e952c021436eb967dd9" Dec 03 14:18:30 crc kubenswrapper[4805]: I1203 14:18:30.901953 4805 scope.go:117] "RemoveContainer" containerID="01c7edcfff6ab9f5d32726a538449aba4ba91c0b84f5c45ed5b86a66f255feb5" Dec 03 14:18:30 crc kubenswrapper[4805]: I1203 14:18:30.928240 4805 scope.go:117] "RemoveContainer" containerID="a9f2e6afd451ccf4c9cf049256fc598e0cbb0ee7e78cd6ed913ff8afdff9b2bf" Dec 03 14:19:13 crc kubenswrapper[4805]: I1203 14:19:13.917271 4805 patch_prober.go:28] interesting pod/machine-config-daemon-gskh4 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 03 14:19:13 crc kubenswrapper[4805]: I1203 14:19:13.917871 4805 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-gskh4" podUID="6098937f-e3f6-45e8-a647-4994a79cd711" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 03 14:19:30 crc kubenswrapper[4805]: I1203 14:19:30.969443 4805 scope.go:117] "RemoveContainer" containerID="1613fa3f6b59024d57deed3610696ad2566be4c7d5110cec4a2f3f22a3a00311" Dec 03 14:19:30 crc kubenswrapper[4805]: I1203 14:19:30.987515 4805 scope.go:117] "RemoveContainer" containerID="d3a6280f659aef72bcb760b2b10ced59c5868f476392736b6f3c84c58c46fa06" Dec 03 14:19:43 crc kubenswrapper[4805]: I1203 14:19:43.917439 4805 patch_prober.go:28] interesting pod/machine-config-daemon-gskh4 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 03 14:19:43 crc kubenswrapper[4805]: I1203 14:19:43.918093 4805 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-gskh4" podUID="6098937f-e3f6-45e8-a647-4994a79cd711" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 03 14:20:13 crc kubenswrapper[4805]: I1203 14:20:13.917536 4805 patch_prober.go:28] interesting pod/machine-config-daemon-gskh4 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 03 14:20:13 crc kubenswrapper[4805]: I1203 14:20:13.918175 4805 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-gskh4" podUID="6098937f-e3f6-45e8-a647-4994a79cd711" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 03 14:20:13 crc kubenswrapper[4805]: I1203 14:20:13.918228 4805 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-gskh4" Dec 03 14:20:13 crc kubenswrapper[4805]: I1203 14:20:13.918873 4805 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"af1b8fb24b94658040c784c6bffe5034acb7fffeeb5210dc5f214b62d02ab400"} pod="openshift-machine-config-operator/machine-config-daemon-gskh4" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 03 14:20:13 crc kubenswrapper[4805]: I1203 14:20:13.918953 4805 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-gskh4" podUID="6098937f-e3f6-45e8-a647-4994a79cd711" containerName="machine-config-daemon" containerID="cri-o://af1b8fb24b94658040c784c6bffe5034acb7fffeeb5210dc5f214b62d02ab400" gracePeriod=600 Dec 03 14:20:14 crc kubenswrapper[4805]: I1203 14:20:14.858939 4805 generic.go:334] "Generic (PLEG): container finished" podID="6098937f-e3f6-45e8-a647-4994a79cd711" containerID="af1b8fb24b94658040c784c6bffe5034acb7fffeeb5210dc5f214b62d02ab400" exitCode=0 Dec 03 14:20:14 crc kubenswrapper[4805]: I1203 14:20:14.859007 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-gskh4" event={"ID":"6098937f-e3f6-45e8-a647-4994a79cd711","Type":"ContainerDied","Data":"af1b8fb24b94658040c784c6bffe5034acb7fffeeb5210dc5f214b62d02ab400"} Dec 03 14:20:14 crc kubenswrapper[4805]: I1203 14:20:14.859510 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-gskh4" event={"ID":"6098937f-e3f6-45e8-a647-4994a79cd711","Type":"ContainerStarted","Data":"7e81e89ac6c4e0378da77d246d467c4c9c45debada7dcf3d097d3c2112a7ea0f"} Dec 03 14:20:14 crc kubenswrapper[4805]: I1203 14:20:14.859547 4805 scope.go:117] "RemoveContainer" containerID="550f8bfd5120af37ce2c9f018ac5ef6ec16519c28f68395f823626d542873ea8" Dec 03 14:22:03 crc kubenswrapper[4805]: I1203 14:22:03.155127 4805 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["cert-manager/cert-manager-5b446d88c5-vj7mm"] Dec 03 14:22:03 crc kubenswrapper[4805]: E1203 14:22:03.156485 4805 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4655a83e-4980-4a5c-9f61-e305ae418822" containerName="registry" Dec 03 14:22:03 crc kubenswrapper[4805]: I1203 14:22:03.156506 4805 state_mem.go:107] "Deleted CPUSet assignment" podUID="4655a83e-4980-4a5c-9f61-e305ae418822" containerName="registry" Dec 03 14:22:03 crc kubenswrapper[4805]: I1203 14:22:03.156647 4805 memory_manager.go:354] "RemoveStaleState removing state" podUID="4655a83e-4980-4a5c-9f61-e305ae418822" containerName="registry" Dec 03 14:22:03 crc kubenswrapper[4805]: I1203 14:22:03.157250 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-5b446d88c5-vj7mm" Dec 03 14:22:03 crc kubenswrapper[4805]: W1203 14:22:03.159089 4805 reflector.go:561] object-"cert-manager"/"kube-root-ca.crt": failed to list *v1.ConfigMap: configmaps "kube-root-ca.crt" is forbidden: User "system:node:crc" cannot list resource "configmaps" in API group "" in the namespace "cert-manager": no relationship found between node 'crc' and this object Dec 03 14:22:03 crc kubenswrapper[4805]: W1203 14:22:03.159145 4805 reflector.go:561] object-"cert-manager"/"cert-manager-dockercfg-97rn9": failed to list *v1.Secret: secrets "cert-manager-dockercfg-97rn9" is forbidden: User "system:node:crc" cannot list resource "secrets" in API group "" in the namespace "cert-manager": no relationship found between node 'crc' and this object Dec 03 14:22:03 crc kubenswrapper[4805]: E1203 14:22:03.159198 4805 reflector.go:158] "Unhandled Error" err="object-\"cert-manager\"/\"cert-manager-dockercfg-97rn9\": Failed to watch *v1.Secret: failed to list *v1.Secret: secrets \"cert-manager-dockercfg-97rn9\" is forbidden: User \"system:node:crc\" cannot list resource \"secrets\" in API group \"\" in the namespace \"cert-manager\": no relationship found between node 'crc' and this object" logger="UnhandledError" Dec 03 14:22:03 crc kubenswrapper[4805]: E1203 14:22:03.159143 4805 reflector.go:158] "Unhandled Error" err="object-\"cert-manager\"/\"kube-root-ca.crt\": Failed to watch *v1.ConfigMap: failed to list *v1.ConfigMap: configmaps \"kube-root-ca.crt\" is forbidden: User \"system:node:crc\" cannot list resource \"configmaps\" in API group \"\" in the namespace \"cert-manager\": no relationship found between node 'crc' and this object" logger="UnhandledError" Dec 03 14:22:03 crc kubenswrapper[4805]: W1203 14:22:03.159553 4805 reflector.go:561] object-"cert-manager"/"openshift-service-ca.crt": failed to list *v1.ConfigMap: configmaps "openshift-service-ca.crt" is forbidden: User "system:node:crc" cannot list resource "configmaps" in API group "" in the namespace "cert-manager": no relationship found between node 'crc' and this object Dec 03 14:22:03 crc kubenswrapper[4805]: E1203 14:22:03.159575 4805 reflector.go:158] "Unhandled Error" err="object-\"cert-manager\"/\"openshift-service-ca.crt\": Failed to watch *v1.ConfigMap: failed to list *v1.ConfigMap: configmaps \"openshift-service-ca.crt\" is forbidden: User \"system:node:crc\" cannot list resource \"configmaps\" in API group \"\" in the namespace \"cert-manager\": no relationship found between node 'crc' and this object" logger="UnhandledError" Dec 03 14:22:03 crc kubenswrapper[4805]: I1203 14:22:03.159739 4805 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["cert-manager/cert-manager-cainjector-7f985d654d-ll5ml"] Dec 03 14:22:03 crc kubenswrapper[4805]: I1203 14:22:03.160738 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-cainjector-7f985d654d-ll5ml" Dec 03 14:22:03 crc kubenswrapper[4805]: W1203 14:22:03.161906 4805 reflector.go:561] object-"cert-manager"/"cert-manager-cainjector-dockercfg-vjcsj": failed to list *v1.Secret: secrets "cert-manager-cainjector-dockercfg-vjcsj" is forbidden: User "system:node:crc" cannot list resource "secrets" in API group "" in the namespace "cert-manager": no relationship found between node 'crc' and this object Dec 03 14:22:03 crc kubenswrapper[4805]: E1203 14:22:03.161973 4805 reflector.go:158] "Unhandled Error" err="object-\"cert-manager\"/\"cert-manager-cainjector-dockercfg-vjcsj\": Failed to watch *v1.Secret: failed to list *v1.Secret: secrets \"cert-manager-cainjector-dockercfg-vjcsj\" is forbidden: User \"system:node:crc\" cannot list resource \"secrets\" in API group \"\" in the namespace \"cert-manager\": no relationship found between node 'crc' and this object" logger="UnhandledError" Dec 03 14:22:03 crc kubenswrapper[4805]: I1203 14:22:03.171460 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lf7dd\" (UniqueName: \"kubernetes.io/projected/70f90fc2-9a18-4737-8025-850782848e31-kube-api-access-lf7dd\") pod \"cert-manager-cainjector-7f985d654d-ll5ml\" (UID: \"70f90fc2-9a18-4737-8025-850782848e31\") " pod="cert-manager/cert-manager-cainjector-7f985d654d-ll5ml" Dec 03 14:22:03 crc kubenswrapper[4805]: I1203 14:22:03.171495 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fb5rn\" (UniqueName: \"kubernetes.io/projected/c7baa569-d061-4226-a134-2b54cf463f90-kube-api-access-fb5rn\") pod \"cert-manager-5b446d88c5-vj7mm\" (UID: \"c7baa569-d061-4226-a134-2b54cf463f90\") " pod="cert-manager/cert-manager-5b446d88c5-vj7mm" Dec 03 14:22:03 crc kubenswrapper[4805]: I1203 14:22:03.181459 4805 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["cert-manager/cert-manager-webhook-5655c58dd6-mt9mr"] Dec 03 14:22:03 crc kubenswrapper[4805]: I1203 14:22:03.184373 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-webhook-5655c58dd6-mt9mr" Dec 03 14:22:03 crc kubenswrapper[4805]: I1203 14:22:03.188176 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-cainjector-7f985d654d-ll5ml"] Dec 03 14:22:03 crc kubenswrapper[4805]: I1203 14:22:03.197145 4805 reflector.go:368] Caches populated for *v1.Secret from object-"cert-manager"/"cert-manager-webhook-dockercfg-j45dj" Dec 03 14:22:03 crc kubenswrapper[4805]: I1203 14:22:03.207771 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-webhook-5655c58dd6-mt9mr"] Dec 03 14:22:03 crc kubenswrapper[4805]: I1203 14:22:03.224214 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-5b446d88c5-vj7mm"] Dec 03 14:22:03 crc kubenswrapper[4805]: I1203 14:22:03.273235 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pxrzb\" (UniqueName: \"kubernetes.io/projected/55307bc8-2e9c-4ea9-b007-c40c05403707-kube-api-access-pxrzb\") pod \"cert-manager-webhook-5655c58dd6-mt9mr\" (UID: \"55307bc8-2e9c-4ea9-b007-c40c05403707\") " pod="cert-manager/cert-manager-webhook-5655c58dd6-mt9mr" Dec 03 14:22:03 crc kubenswrapper[4805]: I1203 14:22:03.273326 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lf7dd\" (UniqueName: \"kubernetes.io/projected/70f90fc2-9a18-4737-8025-850782848e31-kube-api-access-lf7dd\") pod \"cert-manager-cainjector-7f985d654d-ll5ml\" (UID: \"70f90fc2-9a18-4737-8025-850782848e31\") " pod="cert-manager/cert-manager-cainjector-7f985d654d-ll5ml" Dec 03 14:22:03 crc kubenswrapper[4805]: I1203 14:22:03.273358 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fb5rn\" (UniqueName: \"kubernetes.io/projected/c7baa569-d061-4226-a134-2b54cf463f90-kube-api-access-fb5rn\") pod \"cert-manager-5b446d88c5-vj7mm\" (UID: \"c7baa569-d061-4226-a134-2b54cf463f90\") " pod="cert-manager/cert-manager-5b446d88c5-vj7mm" Dec 03 14:22:03 crc kubenswrapper[4805]: I1203 14:22:03.374765 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pxrzb\" (UniqueName: \"kubernetes.io/projected/55307bc8-2e9c-4ea9-b007-c40c05403707-kube-api-access-pxrzb\") pod \"cert-manager-webhook-5655c58dd6-mt9mr\" (UID: \"55307bc8-2e9c-4ea9-b007-c40c05403707\") " pod="cert-manager/cert-manager-webhook-5655c58dd6-mt9mr" Dec 03 14:22:04 crc kubenswrapper[4805]: I1203 14:22:04.203339 4805 reflector.go:368] Caches populated for *v1.Secret from object-"cert-manager"/"cert-manager-dockercfg-97rn9" Dec 03 14:22:04 crc kubenswrapper[4805]: E1203 14:22:04.287013 4805 projected.go:288] Couldn't get configMap cert-manager/kube-root-ca.crt: failed to sync configmap cache: timed out waiting for the condition Dec 03 14:22:04 crc kubenswrapper[4805]: E1203 14:22:04.287271 4805 projected.go:288] Couldn't get configMap cert-manager/kube-root-ca.crt: failed to sync configmap cache: timed out waiting for the condition Dec 03 14:22:04 crc kubenswrapper[4805]: E1203 14:22:04.389082 4805 projected.go:288] Couldn't get configMap cert-manager/kube-root-ca.crt: failed to sync configmap cache: timed out waiting for the condition Dec 03 14:22:04 crc kubenswrapper[4805]: I1203 14:22:04.421245 4805 reflector.go:368] Caches populated for *v1.ConfigMap from object-"cert-manager"/"kube-root-ca.crt" Dec 03 14:22:04 crc kubenswrapper[4805]: I1203 14:22:04.448306 4805 reflector.go:368] Caches populated for *v1.Secret from object-"cert-manager"/"cert-manager-cainjector-dockercfg-vjcsj" Dec 03 14:22:04 crc kubenswrapper[4805]: I1203 14:22:04.706930 4805 reflector.go:368] Caches populated for *v1.ConfigMap from object-"cert-manager"/"openshift-service-ca.crt" Dec 03 14:22:04 crc kubenswrapper[4805]: E1203 14:22:04.707149 4805 projected.go:194] Error preparing data for projected volume kube-api-access-fb5rn for pod cert-manager/cert-manager-5b446d88c5-vj7mm: failed to sync configmap cache: timed out waiting for the condition Dec 03 14:22:04 crc kubenswrapper[4805]: E1203 14:22:04.707255 4805 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/c7baa569-d061-4226-a134-2b54cf463f90-kube-api-access-fb5rn podName:c7baa569-d061-4226-a134-2b54cf463f90 nodeName:}" failed. No retries permitted until 2025-12-03 14:22:05.207228403 +0000 UTC m=+754.870145326 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "kube-api-access-fb5rn" (UniqueName: "kubernetes.io/projected/c7baa569-d061-4226-a134-2b54cf463f90-kube-api-access-fb5rn") pod "cert-manager-5b446d88c5-vj7mm" (UID: "c7baa569-d061-4226-a134-2b54cf463f90") : failed to sync configmap cache: timed out waiting for the condition Dec 03 14:22:04 crc kubenswrapper[4805]: E1203 14:22:04.707501 4805 projected.go:194] Error preparing data for projected volume kube-api-access-lf7dd for pod cert-manager/cert-manager-cainjector-7f985d654d-ll5ml: failed to sync configmap cache: timed out waiting for the condition Dec 03 14:22:04 crc kubenswrapper[4805]: E1203 14:22:04.707593 4805 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/70f90fc2-9a18-4737-8025-850782848e31-kube-api-access-lf7dd podName:70f90fc2-9a18-4737-8025-850782848e31 nodeName:}" failed. No retries permitted until 2025-12-03 14:22:05.207564192 +0000 UTC m=+754.870481125 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "kube-api-access-lf7dd" (UniqueName: "kubernetes.io/projected/70f90fc2-9a18-4737-8025-850782848e31-kube-api-access-lf7dd") pod "cert-manager-cainjector-7f985d654d-ll5ml" (UID: "70f90fc2-9a18-4737-8025-850782848e31") : failed to sync configmap cache: timed out waiting for the condition Dec 03 14:22:04 crc kubenswrapper[4805]: E1203 14:22:04.709579 4805 projected.go:194] Error preparing data for projected volume kube-api-access-pxrzb for pod cert-manager/cert-manager-webhook-5655c58dd6-mt9mr: failed to sync configmap cache: timed out waiting for the condition Dec 03 14:22:04 crc kubenswrapper[4805]: E1203 14:22:04.709644 4805 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/55307bc8-2e9c-4ea9-b007-c40c05403707-kube-api-access-pxrzb podName:55307bc8-2e9c-4ea9-b007-c40c05403707 nodeName:}" failed. No retries permitted until 2025-12-03 14:22:05.209631467 +0000 UTC m=+754.872548390 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "kube-api-access-pxrzb" (UniqueName: "kubernetes.io/projected/55307bc8-2e9c-4ea9-b007-c40c05403707-kube-api-access-pxrzb") pod "cert-manager-webhook-5655c58dd6-mt9mr" (UID: "55307bc8-2e9c-4ea9-b007-c40c05403707") : failed to sync configmap cache: timed out waiting for the condition Dec 03 14:22:05 crc kubenswrapper[4805]: I1203 14:22:05.298318 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pxrzb\" (UniqueName: \"kubernetes.io/projected/55307bc8-2e9c-4ea9-b007-c40c05403707-kube-api-access-pxrzb\") pod \"cert-manager-webhook-5655c58dd6-mt9mr\" (UID: \"55307bc8-2e9c-4ea9-b007-c40c05403707\") " pod="cert-manager/cert-manager-webhook-5655c58dd6-mt9mr" Dec 03 14:22:05 crc kubenswrapper[4805]: I1203 14:22:05.298502 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lf7dd\" (UniqueName: \"kubernetes.io/projected/70f90fc2-9a18-4737-8025-850782848e31-kube-api-access-lf7dd\") pod \"cert-manager-cainjector-7f985d654d-ll5ml\" (UID: \"70f90fc2-9a18-4737-8025-850782848e31\") " pod="cert-manager/cert-manager-cainjector-7f985d654d-ll5ml" Dec 03 14:22:05 crc kubenswrapper[4805]: I1203 14:22:05.298564 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fb5rn\" (UniqueName: \"kubernetes.io/projected/c7baa569-d061-4226-a134-2b54cf463f90-kube-api-access-fb5rn\") pod \"cert-manager-5b446d88c5-vj7mm\" (UID: \"c7baa569-d061-4226-a134-2b54cf463f90\") " pod="cert-manager/cert-manager-5b446d88c5-vj7mm" Dec 03 14:22:05 crc kubenswrapper[4805]: I1203 14:22:05.305508 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pxrzb\" (UniqueName: \"kubernetes.io/projected/55307bc8-2e9c-4ea9-b007-c40c05403707-kube-api-access-pxrzb\") pod \"cert-manager-webhook-5655c58dd6-mt9mr\" (UID: \"55307bc8-2e9c-4ea9-b007-c40c05403707\") " pod="cert-manager/cert-manager-webhook-5655c58dd6-mt9mr" Dec 03 14:22:05 crc kubenswrapper[4805]: I1203 14:22:05.306236 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fb5rn\" (UniqueName: \"kubernetes.io/projected/c7baa569-d061-4226-a134-2b54cf463f90-kube-api-access-fb5rn\") pod \"cert-manager-5b446d88c5-vj7mm\" (UID: \"c7baa569-d061-4226-a134-2b54cf463f90\") " pod="cert-manager/cert-manager-5b446d88c5-vj7mm" Dec 03 14:22:05 crc kubenswrapper[4805]: I1203 14:22:05.313575 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lf7dd\" (UniqueName: \"kubernetes.io/projected/70f90fc2-9a18-4737-8025-850782848e31-kube-api-access-lf7dd\") pod \"cert-manager-cainjector-7f985d654d-ll5ml\" (UID: \"70f90fc2-9a18-4737-8025-850782848e31\") " pod="cert-manager/cert-manager-cainjector-7f985d654d-ll5ml" Dec 03 14:22:05 crc kubenswrapper[4805]: I1203 14:22:05.575778 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-5b446d88c5-vj7mm" Dec 03 14:22:05 crc kubenswrapper[4805]: I1203 14:22:05.581280 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-cainjector-7f985d654d-ll5ml" Dec 03 14:22:05 crc kubenswrapper[4805]: I1203 14:22:05.595572 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-webhook-5655c58dd6-mt9mr" Dec 03 14:22:06 crc kubenswrapper[4805]: I1203 14:22:06.017001 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-5b446d88c5-vj7mm"] Dec 03 14:22:06 crc kubenswrapper[4805]: I1203 14:22:06.032545 4805 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 03 14:22:06 crc kubenswrapper[4805]: I1203 14:22:06.076585 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-webhook-5655c58dd6-mt9mr"] Dec 03 14:22:06 crc kubenswrapper[4805]: I1203 14:22:06.082159 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-cainjector-7f985d654d-ll5ml"] Dec 03 14:22:06 crc kubenswrapper[4805]: I1203 14:22:06.628862 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-cainjector-7f985d654d-ll5ml" event={"ID":"70f90fc2-9a18-4737-8025-850782848e31","Type":"ContainerStarted","Data":"62c0b7cbd975997748150b2f15dd1be14f29846c405e005fe138b66d50357ca7"} Dec 03 14:22:06 crc kubenswrapper[4805]: I1203 14:22:06.630328 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-webhook-5655c58dd6-mt9mr" event={"ID":"55307bc8-2e9c-4ea9-b007-c40c05403707","Type":"ContainerStarted","Data":"c40142008c51adfe833025c1e26f13804de131e9e930396c3aa698ee8e8769f6"} Dec 03 14:22:06 crc kubenswrapper[4805]: I1203 14:22:06.633034 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-5b446d88c5-vj7mm" event={"ID":"c7baa569-d061-4226-a134-2b54cf463f90","Type":"ContainerStarted","Data":"0552a26e650b4c0a92b488ee005d094ce2c94c4b372a772f318028505ca51858"} Dec 03 14:22:08 crc kubenswrapper[4805]: I1203 14:22:08.280603 4805 dynamic_cafile_content.go:123] "Loaded a new CA Bundle and Verifier" name="client-ca-bundle::/etc/kubernetes/kubelet-ca.crt" Dec 03 14:22:13 crc kubenswrapper[4805]: I1203 14:22:13.692939 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-5b446d88c5-vj7mm" event={"ID":"c7baa569-d061-4226-a134-2b54cf463f90","Type":"ContainerStarted","Data":"583f56090883cc6b82091accc99bdb597837d20c3c92b94c0c853bbf5e7eec2d"} Dec 03 14:22:13 crc kubenswrapper[4805]: I1203 14:22:13.696081 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-cainjector-7f985d654d-ll5ml" event={"ID":"70f90fc2-9a18-4737-8025-850782848e31","Type":"ContainerStarted","Data":"e4ebe940f5cab4f76eeb789b3cc48a27f293cba86fd32f95dee7e545049c378b"} Dec 03 14:22:13 crc kubenswrapper[4805]: I1203 14:22:13.697741 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-webhook-5655c58dd6-mt9mr" event={"ID":"55307bc8-2e9c-4ea9-b007-c40c05403707","Type":"ContainerStarted","Data":"3ea18f387462040af9d8f2afd1a117c25faff76adac2fb49c62a341761f17d3f"} Dec 03 14:22:13 crc kubenswrapper[4805]: I1203 14:22:13.698041 4805 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="cert-manager/cert-manager-webhook-5655c58dd6-mt9mr" Dec 03 14:22:13 crc kubenswrapper[4805]: I1203 14:22:13.707428 4805 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="cert-manager/cert-manager-5b446d88c5-vj7mm" podStartSLOduration=3.721915017 podStartE2EDuration="10.707404441s" podCreationTimestamp="2025-12-03 14:22:03 +0000 UTC" firstStartedPulling="2025-12-03 14:22:06.032291987 +0000 UTC m=+755.695208920" lastFinishedPulling="2025-12-03 14:22:13.017781391 +0000 UTC m=+762.680698344" observedRunningTime="2025-12-03 14:22:13.706533048 +0000 UTC m=+763.369449971" watchObservedRunningTime="2025-12-03 14:22:13.707404441 +0000 UTC m=+763.370321374" Dec 03 14:22:13 crc kubenswrapper[4805]: I1203 14:22:13.735229 4805 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="cert-manager/cert-manager-cainjector-7f985d654d-ll5ml" podStartSLOduration=4.100522867 podStartE2EDuration="10.735211152s" podCreationTimestamp="2025-12-03 14:22:03 +0000 UTC" firstStartedPulling="2025-12-03 14:22:06.088011332 +0000 UTC m=+755.750928265" lastFinishedPulling="2025-12-03 14:22:12.722699617 +0000 UTC m=+762.385616550" observedRunningTime="2025-12-03 14:22:13.72951138 +0000 UTC m=+763.392428313" watchObservedRunningTime="2025-12-03 14:22:13.735211152 +0000 UTC m=+763.398128075" Dec 03 14:22:13 crc kubenswrapper[4805]: I1203 14:22:13.751725 4805 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-psjpm"] Dec 03 14:22:13 crc kubenswrapper[4805]: I1203 14:22:13.752231 4805 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-psjpm" podUID="deeec30d-cad2-4dfd-adb6-64f6646581be" containerName="ovn-controller" containerID="cri-o://47d76962beb8c0f196d50af96fb3ec611097823ca80a1c8bd718f2d18840128c" gracePeriod=30 Dec 03 14:22:13 crc kubenswrapper[4805]: I1203 14:22:13.752324 4805 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-psjpm" podUID="deeec30d-cad2-4dfd-adb6-64f6646581be" containerName="kube-rbac-proxy-ovn-metrics" containerID="cri-o://bd67c6de01a8d196a1f51d4d90640c9e69de6f05d5e201b25e6eaa5c33440ea8" gracePeriod=30 Dec 03 14:22:13 crc kubenswrapper[4805]: I1203 14:22:13.752382 4805 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-psjpm" podUID="deeec30d-cad2-4dfd-adb6-64f6646581be" containerName="nbdb" containerID="cri-o://cb8bf05ef093eb0727a066d4bbbdf9af9422867cd4dadacafc6eeb016867a61e" gracePeriod=30 Dec 03 14:22:13 crc kubenswrapper[4805]: I1203 14:22:13.752348 4805 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-psjpm" podUID="deeec30d-cad2-4dfd-adb6-64f6646581be" containerName="ovn-acl-logging" containerID="cri-o://6ce80ccafb3262a11fbe9923317a9d38b61e5dc5b3e7e7ae136f143d6b643fd7" gracePeriod=30 Dec 03 14:22:13 crc kubenswrapper[4805]: I1203 14:22:13.752370 4805 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-psjpm" podUID="deeec30d-cad2-4dfd-adb6-64f6646581be" containerName="sbdb" containerID="cri-o://ab98e365c1a4242a47b8cbe2dd85109d366d6002c4de1f18ddc950a46bea6c12" gracePeriod=30 Dec 03 14:22:13 crc kubenswrapper[4805]: I1203 14:22:13.752270 4805 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-psjpm" podUID="deeec30d-cad2-4dfd-adb6-64f6646581be" containerName="northd" containerID="cri-o://0304467102ed053628c0d3e06548c74b6492336b6fd044246508d1e8bdb17bc3" gracePeriod=30 Dec 03 14:22:13 crc kubenswrapper[4805]: I1203 14:22:13.752313 4805 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-psjpm" podUID="deeec30d-cad2-4dfd-adb6-64f6646581be" containerName="kube-rbac-proxy-node" containerID="cri-o://4b8d00cdf0f37f2e5cd4e952cf85d3de8d72c43fdf786088fd9f26ed17b1e21f" gracePeriod=30 Dec 03 14:22:13 crc kubenswrapper[4805]: I1203 14:22:13.763161 4805 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="cert-manager/cert-manager-webhook-5655c58dd6-mt9mr" podStartSLOduration=4.123460019 podStartE2EDuration="10.763142226s" podCreationTimestamp="2025-12-03 14:22:03 +0000 UTC" firstStartedPulling="2025-12-03 14:22:06.082748432 +0000 UTC m=+755.745665375" lastFinishedPulling="2025-12-03 14:22:12.722430649 +0000 UTC m=+762.385347582" observedRunningTime="2025-12-03 14:22:13.760673911 +0000 UTC m=+763.423590824" watchObservedRunningTime="2025-12-03 14:22:13.763142226 +0000 UTC m=+763.426059149" Dec 03 14:22:13 crc kubenswrapper[4805]: I1203 14:22:13.789710 4805 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-psjpm" podUID="deeec30d-cad2-4dfd-adb6-64f6646581be" containerName="ovnkube-controller" containerID="cri-o://cb029c1dbe665e82f006ab5f681796fa66f6ff814a3988c3c4b5863923b4a44f" gracePeriod=30 Dec 03 14:22:14 crc kubenswrapper[4805]: I1203 14:22:14.714280 4805 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-psjpm_deeec30d-cad2-4dfd-adb6-64f6646581be/ovnkube-controller/3.log" Dec 03 14:22:14 crc kubenswrapper[4805]: I1203 14:22:14.718520 4805 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-psjpm_deeec30d-cad2-4dfd-adb6-64f6646581be/ovn-acl-logging/0.log" Dec 03 14:22:14 crc kubenswrapper[4805]: I1203 14:22:14.719572 4805 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-psjpm_deeec30d-cad2-4dfd-adb6-64f6646581be/ovn-controller/0.log" Dec 03 14:22:14 crc kubenswrapper[4805]: I1203 14:22:14.720379 4805 generic.go:334] "Generic (PLEG): container finished" podID="deeec30d-cad2-4dfd-adb6-64f6646581be" containerID="cb029c1dbe665e82f006ab5f681796fa66f6ff814a3988c3c4b5863923b4a44f" exitCode=0 Dec 03 14:22:14 crc kubenswrapper[4805]: I1203 14:22:14.720465 4805 generic.go:334] "Generic (PLEG): container finished" podID="deeec30d-cad2-4dfd-adb6-64f6646581be" containerID="ab98e365c1a4242a47b8cbe2dd85109d366d6002c4de1f18ddc950a46bea6c12" exitCode=0 Dec 03 14:22:14 crc kubenswrapper[4805]: I1203 14:22:14.720528 4805 generic.go:334] "Generic (PLEG): container finished" podID="deeec30d-cad2-4dfd-adb6-64f6646581be" containerID="cb8bf05ef093eb0727a066d4bbbdf9af9422867cd4dadacafc6eeb016867a61e" exitCode=0 Dec 03 14:22:14 crc kubenswrapper[4805]: I1203 14:22:14.720582 4805 generic.go:334] "Generic (PLEG): container finished" podID="deeec30d-cad2-4dfd-adb6-64f6646581be" containerID="0304467102ed053628c0d3e06548c74b6492336b6fd044246508d1e8bdb17bc3" exitCode=0 Dec 03 14:22:14 crc kubenswrapper[4805]: I1203 14:22:14.720639 4805 generic.go:334] "Generic (PLEG): container finished" podID="deeec30d-cad2-4dfd-adb6-64f6646581be" containerID="bd67c6de01a8d196a1f51d4d90640c9e69de6f05d5e201b25e6eaa5c33440ea8" exitCode=0 Dec 03 14:22:14 crc kubenswrapper[4805]: I1203 14:22:14.720486 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-psjpm" event={"ID":"deeec30d-cad2-4dfd-adb6-64f6646581be","Type":"ContainerDied","Data":"cb029c1dbe665e82f006ab5f681796fa66f6ff814a3988c3c4b5863923b4a44f"} Dec 03 14:22:14 crc kubenswrapper[4805]: I1203 14:22:14.720773 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-psjpm" event={"ID":"deeec30d-cad2-4dfd-adb6-64f6646581be","Type":"ContainerDied","Data":"ab98e365c1a4242a47b8cbe2dd85109d366d6002c4de1f18ddc950a46bea6c12"} Dec 03 14:22:14 crc kubenswrapper[4805]: I1203 14:22:14.720686 4805 generic.go:334] "Generic (PLEG): container finished" podID="deeec30d-cad2-4dfd-adb6-64f6646581be" containerID="4b8d00cdf0f37f2e5cd4e952cf85d3de8d72c43fdf786088fd9f26ed17b1e21f" exitCode=0 Dec 03 14:22:14 crc kubenswrapper[4805]: I1203 14:22:14.720821 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-psjpm" event={"ID":"deeec30d-cad2-4dfd-adb6-64f6646581be","Type":"ContainerDied","Data":"cb8bf05ef093eb0727a066d4bbbdf9af9422867cd4dadacafc6eeb016867a61e"} Dec 03 14:22:14 crc kubenswrapper[4805]: I1203 14:22:14.720876 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-psjpm" event={"ID":"deeec30d-cad2-4dfd-adb6-64f6646581be","Type":"ContainerDied","Data":"0304467102ed053628c0d3e06548c74b6492336b6fd044246508d1e8bdb17bc3"} Dec 03 14:22:14 crc kubenswrapper[4805]: I1203 14:22:14.720904 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-psjpm" event={"ID":"deeec30d-cad2-4dfd-adb6-64f6646581be","Type":"ContainerDied","Data":"bd67c6de01a8d196a1f51d4d90640c9e69de6f05d5e201b25e6eaa5c33440ea8"} Dec 03 14:22:14 crc kubenswrapper[4805]: I1203 14:22:14.720925 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-psjpm" event={"ID":"deeec30d-cad2-4dfd-adb6-64f6646581be","Type":"ContainerDied","Data":"4b8d00cdf0f37f2e5cd4e952cf85d3de8d72c43fdf786088fd9f26ed17b1e21f"} Dec 03 14:22:14 crc kubenswrapper[4805]: I1203 14:22:14.720920 4805 generic.go:334] "Generic (PLEG): container finished" podID="deeec30d-cad2-4dfd-adb6-64f6646581be" containerID="6ce80ccafb3262a11fbe9923317a9d38b61e5dc5b3e7e7ae136f143d6b643fd7" exitCode=143 Dec 03 14:22:14 crc kubenswrapper[4805]: I1203 14:22:14.720962 4805 generic.go:334] "Generic (PLEG): container finished" podID="deeec30d-cad2-4dfd-adb6-64f6646581be" containerID="47d76962beb8c0f196d50af96fb3ec611097823ca80a1c8bd718f2d18840128c" exitCode=143 Dec 03 14:22:14 crc kubenswrapper[4805]: I1203 14:22:14.720961 4805 scope.go:117] "RemoveContainer" containerID="8f75988e25825e00abdb9e4b2292c5cc8ccab3dd028c882c2a38c796c0fd19c5" Dec 03 14:22:14 crc kubenswrapper[4805]: I1203 14:22:14.720943 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-psjpm" event={"ID":"deeec30d-cad2-4dfd-adb6-64f6646581be","Type":"ContainerDied","Data":"6ce80ccafb3262a11fbe9923317a9d38b61e5dc5b3e7e7ae136f143d6b643fd7"} Dec 03 14:22:14 crc kubenswrapper[4805]: I1203 14:22:14.721104 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-psjpm" event={"ID":"deeec30d-cad2-4dfd-adb6-64f6646581be","Type":"ContainerDied","Data":"47d76962beb8c0f196d50af96fb3ec611097823ca80a1c8bd718f2d18840128c"} Dec 03 14:22:14 crc kubenswrapper[4805]: I1203 14:22:14.725358 4805 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-k4mhs_c09c1cdc-f461-45c9-8444-5b8764bccabc/kube-multus/2.log" Dec 03 14:22:14 crc kubenswrapper[4805]: I1203 14:22:14.726707 4805 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-k4mhs_c09c1cdc-f461-45c9-8444-5b8764bccabc/kube-multus/1.log" Dec 03 14:22:14 crc kubenswrapper[4805]: I1203 14:22:14.726775 4805 generic.go:334] "Generic (PLEG): container finished" podID="c09c1cdc-f461-45c9-8444-5b8764bccabc" containerID="556324f2c4c09f251607fd41d095f4eae1aa27bcb9958f51449b88e6dedcaa86" exitCode=2 Dec 03 14:22:14 crc kubenswrapper[4805]: I1203 14:22:14.726826 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-k4mhs" event={"ID":"c09c1cdc-f461-45c9-8444-5b8764bccabc","Type":"ContainerDied","Data":"556324f2c4c09f251607fd41d095f4eae1aa27bcb9958f51449b88e6dedcaa86"} Dec 03 14:22:14 crc kubenswrapper[4805]: I1203 14:22:14.727668 4805 scope.go:117] "RemoveContainer" containerID="556324f2c4c09f251607fd41d095f4eae1aa27bcb9958f51449b88e6dedcaa86" Dec 03 14:22:14 crc kubenswrapper[4805]: I1203 14:22:14.813517 4805 scope.go:117] "RemoveContainer" containerID="7e077e19b6a5bf7d29b357982e12224d4b7058e0f9ac81b66b5a665c9bbf5014" Dec 03 14:22:14 crc kubenswrapper[4805]: I1203 14:22:14.821751 4805 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-psjpm_deeec30d-cad2-4dfd-adb6-64f6646581be/ovn-acl-logging/0.log" Dec 03 14:22:14 crc kubenswrapper[4805]: I1203 14:22:14.822813 4805 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-psjpm_deeec30d-cad2-4dfd-adb6-64f6646581be/ovn-controller/0.log" Dec 03 14:22:14 crc kubenswrapper[4805]: I1203 14:22:14.823943 4805 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-psjpm" Dec 03 14:22:14 crc kubenswrapper[4805]: I1203 14:22:14.893650 4805 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-6cz9b"] Dec 03 14:22:14 crc kubenswrapper[4805]: E1203 14:22:14.893887 4805 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="deeec30d-cad2-4dfd-adb6-64f6646581be" containerName="ovnkube-controller" Dec 03 14:22:14 crc kubenswrapper[4805]: I1203 14:22:14.893901 4805 state_mem.go:107] "Deleted CPUSet assignment" podUID="deeec30d-cad2-4dfd-adb6-64f6646581be" containerName="ovnkube-controller" Dec 03 14:22:14 crc kubenswrapper[4805]: E1203 14:22:14.893908 4805 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="deeec30d-cad2-4dfd-adb6-64f6646581be" containerName="ovnkube-controller" Dec 03 14:22:14 crc kubenswrapper[4805]: I1203 14:22:14.893913 4805 state_mem.go:107] "Deleted CPUSet assignment" podUID="deeec30d-cad2-4dfd-adb6-64f6646581be" containerName="ovnkube-controller" Dec 03 14:22:14 crc kubenswrapper[4805]: E1203 14:22:14.893921 4805 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="deeec30d-cad2-4dfd-adb6-64f6646581be" containerName="ovnkube-controller" Dec 03 14:22:14 crc kubenswrapper[4805]: I1203 14:22:14.893928 4805 state_mem.go:107] "Deleted CPUSet assignment" podUID="deeec30d-cad2-4dfd-adb6-64f6646581be" containerName="ovnkube-controller" Dec 03 14:22:14 crc kubenswrapper[4805]: E1203 14:22:14.893935 4805 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="deeec30d-cad2-4dfd-adb6-64f6646581be" containerName="ovn-controller" Dec 03 14:22:14 crc kubenswrapper[4805]: I1203 14:22:14.893942 4805 state_mem.go:107] "Deleted CPUSet assignment" podUID="deeec30d-cad2-4dfd-adb6-64f6646581be" containerName="ovn-controller" Dec 03 14:22:14 crc kubenswrapper[4805]: E1203 14:22:14.893953 4805 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="deeec30d-cad2-4dfd-adb6-64f6646581be" containerName="sbdb" Dec 03 14:22:14 crc kubenswrapper[4805]: I1203 14:22:14.893959 4805 state_mem.go:107] "Deleted CPUSet assignment" podUID="deeec30d-cad2-4dfd-adb6-64f6646581be" containerName="sbdb" Dec 03 14:22:14 crc kubenswrapper[4805]: E1203 14:22:14.893968 4805 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="deeec30d-cad2-4dfd-adb6-64f6646581be" containerName="kubecfg-setup" Dec 03 14:22:14 crc kubenswrapper[4805]: I1203 14:22:14.893974 4805 state_mem.go:107] "Deleted CPUSet assignment" podUID="deeec30d-cad2-4dfd-adb6-64f6646581be" containerName="kubecfg-setup" Dec 03 14:22:14 crc kubenswrapper[4805]: E1203 14:22:14.893986 4805 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="deeec30d-cad2-4dfd-adb6-64f6646581be" containerName="ovnkube-controller" Dec 03 14:22:14 crc kubenswrapper[4805]: I1203 14:22:14.893991 4805 state_mem.go:107] "Deleted CPUSet assignment" podUID="deeec30d-cad2-4dfd-adb6-64f6646581be" containerName="ovnkube-controller" Dec 03 14:22:14 crc kubenswrapper[4805]: E1203 14:22:14.894000 4805 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="deeec30d-cad2-4dfd-adb6-64f6646581be" containerName="northd" Dec 03 14:22:14 crc kubenswrapper[4805]: I1203 14:22:14.894006 4805 state_mem.go:107] "Deleted CPUSet assignment" podUID="deeec30d-cad2-4dfd-adb6-64f6646581be" containerName="northd" Dec 03 14:22:14 crc kubenswrapper[4805]: E1203 14:22:14.894016 4805 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="deeec30d-cad2-4dfd-adb6-64f6646581be" containerName="kube-rbac-proxy-node" Dec 03 14:22:14 crc kubenswrapper[4805]: I1203 14:22:14.894021 4805 state_mem.go:107] "Deleted CPUSet assignment" podUID="deeec30d-cad2-4dfd-adb6-64f6646581be" containerName="kube-rbac-proxy-node" Dec 03 14:22:14 crc kubenswrapper[4805]: E1203 14:22:14.894028 4805 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="deeec30d-cad2-4dfd-adb6-64f6646581be" containerName="nbdb" Dec 03 14:22:14 crc kubenswrapper[4805]: I1203 14:22:14.894035 4805 state_mem.go:107] "Deleted CPUSet assignment" podUID="deeec30d-cad2-4dfd-adb6-64f6646581be" containerName="nbdb" Dec 03 14:22:14 crc kubenswrapper[4805]: E1203 14:22:14.894043 4805 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="deeec30d-cad2-4dfd-adb6-64f6646581be" containerName="ovn-acl-logging" Dec 03 14:22:14 crc kubenswrapper[4805]: I1203 14:22:14.894049 4805 state_mem.go:107] "Deleted CPUSet assignment" podUID="deeec30d-cad2-4dfd-adb6-64f6646581be" containerName="ovn-acl-logging" Dec 03 14:22:14 crc kubenswrapper[4805]: E1203 14:22:14.894055 4805 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="deeec30d-cad2-4dfd-adb6-64f6646581be" containerName="kube-rbac-proxy-ovn-metrics" Dec 03 14:22:14 crc kubenswrapper[4805]: I1203 14:22:14.894061 4805 state_mem.go:107] "Deleted CPUSet assignment" podUID="deeec30d-cad2-4dfd-adb6-64f6646581be" containerName="kube-rbac-proxy-ovn-metrics" Dec 03 14:22:14 crc kubenswrapper[4805]: I1203 14:22:14.894157 4805 memory_manager.go:354] "RemoveStaleState removing state" podUID="deeec30d-cad2-4dfd-adb6-64f6646581be" containerName="kube-rbac-proxy-node" Dec 03 14:22:14 crc kubenswrapper[4805]: I1203 14:22:14.894170 4805 memory_manager.go:354] "RemoveStaleState removing state" podUID="deeec30d-cad2-4dfd-adb6-64f6646581be" containerName="ovn-acl-logging" Dec 03 14:22:14 crc kubenswrapper[4805]: I1203 14:22:14.894178 4805 memory_manager.go:354] "RemoveStaleState removing state" podUID="deeec30d-cad2-4dfd-adb6-64f6646581be" containerName="northd" Dec 03 14:22:14 crc kubenswrapper[4805]: I1203 14:22:14.894187 4805 memory_manager.go:354] "RemoveStaleState removing state" podUID="deeec30d-cad2-4dfd-adb6-64f6646581be" containerName="ovnkube-controller" Dec 03 14:22:14 crc kubenswrapper[4805]: I1203 14:22:14.894199 4805 memory_manager.go:354] "RemoveStaleState removing state" podUID="deeec30d-cad2-4dfd-adb6-64f6646581be" containerName="sbdb" Dec 03 14:22:14 crc kubenswrapper[4805]: I1203 14:22:14.894209 4805 memory_manager.go:354] "RemoveStaleState removing state" podUID="deeec30d-cad2-4dfd-adb6-64f6646581be" containerName="kube-rbac-proxy-ovn-metrics" Dec 03 14:22:14 crc kubenswrapper[4805]: I1203 14:22:14.894217 4805 memory_manager.go:354] "RemoveStaleState removing state" podUID="deeec30d-cad2-4dfd-adb6-64f6646581be" containerName="ovnkube-controller" Dec 03 14:22:14 crc kubenswrapper[4805]: I1203 14:22:14.894225 4805 memory_manager.go:354] "RemoveStaleState removing state" podUID="deeec30d-cad2-4dfd-adb6-64f6646581be" containerName="ovnkube-controller" Dec 03 14:22:14 crc kubenswrapper[4805]: I1203 14:22:14.894233 4805 memory_manager.go:354] "RemoveStaleState removing state" podUID="deeec30d-cad2-4dfd-adb6-64f6646581be" containerName="ovn-controller" Dec 03 14:22:14 crc kubenswrapper[4805]: I1203 14:22:14.894241 4805 memory_manager.go:354] "RemoveStaleState removing state" podUID="deeec30d-cad2-4dfd-adb6-64f6646581be" containerName="nbdb" Dec 03 14:22:14 crc kubenswrapper[4805]: E1203 14:22:14.894342 4805 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="deeec30d-cad2-4dfd-adb6-64f6646581be" containerName="ovnkube-controller" Dec 03 14:22:14 crc kubenswrapper[4805]: I1203 14:22:14.894352 4805 state_mem.go:107] "Deleted CPUSet assignment" podUID="deeec30d-cad2-4dfd-adb6-64f6646581be" containerName="ovnkube-controller" Dec 03 14:22:14 crc kubenswrapper[4805]: I1203 14:22:14.894455 4805 memory_manager.go:354] "RemoveStaleState removing state" podUID="deeec30d-cad2-4dfd-adb6-64f6646581be" containerName="ovnkube-controller" Dec 03 14:22:14 crc kubenswrapper[4805]: I1203 14:22:14.894465 4805 memory_manager.go:354] "RemoveStaleState removing state" podUID="deeec30d-cad2-4dfd-adb6-64f6646581be" containerName="ovnkube-controller" Dec 03 14:22:14 crc kubenswrapper[4805]: I1203 14:22:14.896128 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-6cz9b" Dec 03 14:22:14 crc kubenswrapper[4805]: I1203 14:22:14.954310 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/deeec30d-cad2-4dfd-adb6-64f6646581be-ovn-node-metrics-cert\") pod \"deeec30d-cad2-4dfd-adb6-64f6646581be\" (UID: \"deeec30d-cad2-4dfd-adb6-64f6646581be\") " Dec 03 14:22:14 crc kubenswrapper[4805]: I1203 14:22:14.954364 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/deeec30d-cad2-4dfd-adb6-64f6646581be-var-lib-openvswitch\") pod \"deeec30d-cad2-4dfd-adb6-64f6646581be\" (UID: \"deeec30d-cad2-4dfd-adb6-64f6646581be\") " Dec 03 14:22:14 crc kubenswrapper[4805]: I1203 14:22:14.954400 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/deeec30d-cad2-4dfd-adb6-64f6646581be-ovnkube-config\") pod \"deeec30d-cad2-4dfd-adb6-64f6646581be\" (UID: \"deeec30d-cad2-4dfd-adb6-64f6646581be\") " Dec 03 14:22:14 crc kubenswrapper[4805]: I1203 14:22:14.954420 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/deeec30d-cad2-4dfd-adb6-64f6646581be-node-log\") pod \"deeec30d-cad2-4dfd-adb6-64f6646581be\" (UID: \"deeec30d-cad2-4dfd-adb6-64f6646581be\") " Dec 03 14:22:14 crc kubenswrapper[4805]: I1203 14:22:14.954479 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/deeec30d-cad2-4dfd-adb6-64f6646581be-var-lib-openvswitch" (OuterVolumeSpecName: "var-lib-openvswitch") pod "deeec30d-cad2-4dfd-adb6-64f6646581be" (UID: "deeec30d-cad2-4dfd-adb6-64f6646581be"). InnerVolumeSpecName "var-lib-openvswitch". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 03 14:22:14 crc kubenswrapper[4805]: I1203 14:22:14.954498 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/deeec30d-cad2-4dfd-adb6-64f6646581be-log-socket\") pod \"deeec30d-cad2-4dfd-adb6-64f6646581be\" (UID: \"deeec30d-cad2-4dfd-adb6-64f6646581be\") " Dec 03 14:22:14 crc kubenswrapper[4805]: I1203 14:22:14.954484 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/deeec30d-cad2-4dfd-adb6-64f6646581be-node-log" (OuterVolumeSpecName: "node-log") pod "deeec30d-cad2-4dfd-adb6-64f6646581be" (UID: "deeec30d-cad2-4dfd-adb6-64f6646581be"). InnerVolumeSpecName "node-log". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 03 14:22:14 crc kubenswrapper[4805]: I1203 14:22:14.954562 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/deeec30d-cad2-4dfd-adb6-64f6646581be-log-socket" (OuterVolumeSpecName: "log-socket") pod "deeec30d-cad2-4dfd-adb6-64f6646581be" (UID: "deeec30d-cad2-4dfd-adb6-64f6646581be"). InnerVolumeSpecName "log-socket". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 03 14:22:14 crc kubenswrapper[4805]: I1203 14:22:14.954788 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/deeec30d-cad2-4dfd-adb6-64f6646581be-env-overrides\") pod \"deeec30d-cad2-4dfd-adb6-64f6646581be\" (UID: \"deeec30d-cad2-4dfd-adb6-64f6646581be\") " Dec 03 14:22:14 crc kubenswrapper[4805]: I1203 14:22:14.954819 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/deeec30d-cad2-4dfd-adb6-64f6646581be-ovnkube-config" (OuterVolumeSpecName: "ovnkube-config") pod "deeec30d-cad2-4dfd-adb6-64f6646581be" (UID: "deeec30d-cad2-4dfd-adb6-64f6646581be"). InnerVolumeSpecName "ovnkube-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 14:22:14 crc kubenswrapper[4805]: I1203 14:22:14.954884 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/deeec30d-cad2-4dfd-adb6-64f6646581be-host-cni-netd\") pod \"deeec30d-cad2-4dfd-adb6-64f6646581be\" (UID: \"deeec30d-cad2-4dfd-adb6-64f6646581be\") " Dec 03 14:22:14 crc kubenswrapper[4805]: I1203 14:22:14.954906 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/deeec30d-cad2-4dfd-adb6-64f6646581be-host-cni-bin\") pod \"deeec30d-cad2-4dfd-adb6-64f6646581be\" (UID: \"deeec30d-cad2-4dfd-adb6-64f6646581be\") " Dec 03 14:22:14 crc kubenswrapper[4805]: I1203 14:22:14.954936 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/deeec30d-cad2-4dfd-adb6-64f6646581be-run-openvswitch\") pod \"deeec30d-cad2-4dfd-adb6-64f6646581be\" (UID: \"deeec30d-cad2-4dfd-adb6-64f6646581be\") " Dec 03 14:22:14 crc kubenswrapper[4805]: I1203 14:22:14.954939 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/deeec30d-cad2-4dfd-adb6-64f6646581be-host-cni-netd" (OuterVolumeSpecName: "host-cni-netd") pod "deeec30d-cad2-4dfd-adb6-64f6646581be" (UID: "deeec30d-cad2-4dfd-adb6-64f6646581be"). InnerVolumeSpecName "host-cni-netd". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 03 14:22:14 crc kubenswrapper[4805]: I1203 14:22:14.954960 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wdlzx\" (UniqueName: \"kubernetes.io/projected/deeec30d-cad2-4dfd-adb6-64f6646581be-kube-api-access-wdlzx\") pod \"deeec30d-cad2-4dfd-adb6-64f6646581be\" (UID: \"deeec30d-cad2-4dfd-adb6-64f6646581be\") " Dec 03 14:22:14 crc kubenswrapper[4805]: I1203 14:22:14.954964 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/deeec30d-cad2-4dfd-adb6-64f6646581be-host-cni-bin" (OuterVolumeSpecName: "host-cni-bin") pod "deeec30d-cad2-4dfd-adb6-64f6646581be" (UID: "deeec30d-cad2-4dfd-adb6-64f6646581be"). InnerVolumeSpecName "host-cni-bin". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 03 14:22:14 crc kubenswrapper[4805]: I1203 14:22:14.954976 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/deeec30d-cad2-4dfd-adb6-64f6646581be-run-openvswitch" (OuterVolumeSpecName: "run-openvswitch") pod "deeec30d-cad2-4dfd-adb6-64f6646581be" (UID: "deeec30d-cad2-4dfd-adb6-64f6646581be"). InnerVolumeSpecName "run-openvswitch". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 03 14:22:14 crc kubenswrapper[4805]: I1203 14:22:14.955000 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/deeec30d-cad2-4dfd-adb6-64f6646581be-host-run-netns\") pod \"deeec30d-cad2-4dfd-adb6-64f6646581be\" (UID: \"deeec30d-cad2-4dfd-adb6-64f6646581be\") " Dec 03 14:22:14 crc kubenswrapper[4805]: I1203 14:22:14.955020 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/deeec30d-cad2-4dfd-adb6-64f6646581be-host-slash\") pod \"deeec30d-cad2-4dfd-adb6-64f6646581be\" (UID: \"deeec30d-cad2-4dfd-adb6-64f6646581be\") " Dec 03 14:22:14 crc kubenswrapper[4805]: I1203 14:22:14.955047 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/deeec30d-cad2-4dfd-adb6-64f6646581be-host-var-lib-cni-networks-ovn-kubernetes\") pod \"deeec30d-cad2-4dfd-adb6-64f6646581be\" (UID: \"deeec30d-cad2-4dfd-adb6-64f6646581be\") " Dec 03 14:22:14 crc kubenswrapper[4805]: I1203 14:22:14.955073 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/deeec30d-cad2-4dfd-adb6-64f6646581be-run-systemd\") pod \"deeec30d-cad2-4dfd-adb6-64f6646581be\" (UID: \"deeec30d-cad2-4dfd-adb6-64f6646581be\") " Dec 03 14:22:14 crc kubenswrapper[4805]: I1203 14:22:14.955073 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/deeec30d-cad2-4dfd-adb6-64f6646581be-host-run-netns" (OuterVolumeSpecName: "host-run-netns") pod "deeec30d-cad2-4dfd-adb6-64f6646581be" (UID: "deeec30d-cad2-4dfd-adb6-64f6646581be"). InnerVolumeSpecName "host-run-netns". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 03 14:22:14 crc kubenswrapper[4805]: I1203 14:22:14.955094 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/deeec30d-cad2-4dfd-adb6-64f6646581be-host-kubelet\") pod \"deeec30d-cad2-4dfd-adb6-64f6646581be\" (UID: \"deeec30d-cad2-4dfd-adb6-64f6646581be\") " Dec 03 14:22:14 crc kubenswrapper[4805]: I1203 14:22:14.955107 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/deeec30d-cad2-4dfd-adb6-64f6646581be-host-var-lib-cni-networks-ovn-kubernetes" (OuterVolumeSpecName: "host-var-lib-cni-networks-ovn-kubernetes") pod "deeec30d-cad2-4dfd-adb6-64f6646581be" (UID: "deeec30d-cad2-4dfd-adb6-64f6646581be"). InnerVolumeSpecName "host-var-lib-cni-networks-ovn-kubernetes". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 03 14:22:14 crc kubenswrapper[4805]: I1203 14:22:14.955115 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/deeec30d-cad2-4dfd-adb6-64f6646581be-systemd-units\") pod \"deeec30d-cad2-4dfd-adb6-64f6646581be\" (UID: \"deeec30d-cad2-4dfd-adb6-64f6646581be\") " Dec 03 14:22:14 crc kubenswrapper[4805]: I1203 14:22:14.955158 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/deeec30d-cad2-4dfd-adb6-64f6646581be-run-ovn\") pod \"deeec30d-cad2-4dfd-adb6-64f6646581be\" (UID: \"deeec30d-cad2-4dfd-adb6-64f6646581be\") " Dec 03 14:22:14 crc kubenswrapper[4805]: I1203 14:22:14.955134 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/deeec30d-cad2-4dfd-adb6-64f6646581be-host-slash" (OuterVolumeSpecName: "host-slash") pod "deeec30d-cad2-4dfd-adb6-64f6646581be" (UID: "deeec30d-cad2-4dfd-adb6-64f6646581be"). InnerVolumeSpecName "host-slash". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 03 14:22:14 crc kubenswrapper[4805]: I1203 14:22:14.955189 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/deeec30d-cad2-4dfd-adb6-64f6646581be-ovnkube-script-lib\") pod \"deeec30d-cad2-4dfd-adb6-64f6646581be\" (UID: \"deeec30d-cad2-4dfd-adb6-64f6646581be\") " Dec 03 14:22:14 crc kubenswrapper[4805]: I1203 14:22:14.955209 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/deeec30d-cad2-4dfd-adb6-64f6646581be-etc-openvswitch\") pod \"deeec30d-cad2-4dfd-adb6-64f6646581be\" (UID: \"deeec30d-cad2-4dfd-adb6-64f6646581be\") " Dec 03 14:22:14 crc kubenswrapper[4805]: I1203 14:22:14.955212 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/deeec30d-cad2-4dfd-adb6-64f6646581be-systemd-units" (OuterVolumeSpecName: "systemd-units") pod "deeec30d-cad2-4dfd-adb6-64f6646581be" (UID: "deeec30d-cad2-4dfd-adb6-64f6646581be"). InnerVolumeSpecName "systemd-units". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 03 14:22:14 crc kubenswrapper[4805]: I1203 14:22:14.955230 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/deeec30d-cad2-4dfd-adb6-64f6646581be-host-run-ovn-kubernetes\") pod \"deeec30d-cad2-4dfd-adb6-64f6646581be\" (UID: \"deeec30d-cad2-4dfd-adb6-64f6646581be\") " Dec 03 14:22:14 crc kubenswrapper[4805]: I1203 14:22:14.955240 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/deeec30d-cad2-4dfd-adb6-64f6646581be-host-kubelet" (OuterVolumeSpecName: "host-kubelet") pod "deeec30d-cad2-4dfd-adb6-64f6646581be" (UID: "deeec30d-cad2-4dfd-adb6-64f6646581be"). InnerVolumeSpecName "host-kubelet". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 03 14:22:14 crc kubenswrapper[4805]: I1203 14:22:14.955266 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/deeec30d-cad2-4dfd-adb6-64f6646581be-env-overrides" (OuterVolumeSpecName: "env-overrides") pod "deeec30d-cad2-4dfd-adb6-64f6646581be" (UID: "deeec30d-cad2-4dfd-adb6-64f6646581be"). InnerVolumeSpecName "env-overrides". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 14:22:14 crc kubenswrapper[4805]: I1203 14:22:14.955312 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/deeec30d-cad2-4dfd-adb6-64f6646581be-etc-openvswitch" (OuterVolumeSpecName: "etc-openvswitch") pod "deeec30d-cad2-4dfd-adb6-64f6646581be" (UID: "deeec30d-cad2-4dfd-adb6-64f6646581be"). InnerVolumeSpecName "etc-openvswitch". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 03 14:22:14 crc kubenswrapper[4805]: I1203 14:22:14.955314 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/deeec30d-cad2-4dfd-adb6-64f6646581be-run-ovn" (OuterVolumeSpecName: "run-ovn") pod "deeec30d-cad2-4dfd-adb6-64f6646581be" (UID: "deeec30d-cad2-4dfd-adb6-64f6646581be"). InnerVolumeSpecName "run-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 03 14:22:14 crc kubenswrapper[4805]: I1203 14:22:14.955403 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/ea4cce30-4d9a-4384-87d5-8ec365db3247-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-6cz9b\" (UID: \"ea4cce30-4d9a-4384-87d5-8ec365db3247\") " pod="openshift-ovn-kubernetes/ovnkube-node-6cz9b" Dec 03 14:22:14 crc kubenswrapper[4805]: I1203 14:22:14.955434 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/ea4cce30-4d9a-4384-87d5-8ec365db3247-host-kubelet\") pod \"ovnkube-node-6cz9b\" (UID: \"ea4cce30-4d9a-4384-87d5-8ec365db3247\") " pod="openshift-ovn-kubernetes/ovnkube-node-6cz9b" Dec 03 14:22:14 crc kubenswrapper[4805]: I1203 14:22:14.955430 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/deeec30d-cad2-4dfd-adb6-64f6646581be-host-run-ovn-kubernetes" (OuterVolumeSpecName: "host-run-ovn-kubernetes") pod "deeec30d-cad2-4dfd-adb6-64f6646581be" (UID: "deeec30d-cad2-4dfd-adb6-64f6646581be"). InnerVolumeSpecName "host-run-ovn-kubernetes". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 03 14:22:14 crc kubenswrapper[4805]: I1203 14:22:14.955462 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/ea4cce30-4d9a-4384-87d5-8ec365db3247-ovnkube-config\") pod \"ovnkube-node-6cz9b\" (UID: \"ea4cce30-4d9a-4384-87d5-8ec365db3247\") " pod="openshift-ovn-kubernetes/ovnkube-node-6cz9b" Dec 03 14:22:14 crc kubenswrapper[4805]: I1203 14:22:14.955505 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/ea4cce30-4d9a-4384-87d5-8ec365db3247-ovn-node-metrics-cert\") pod \"ovnkube-node-6cz9b\" (UID: \"ea4cce30-4d9a-4384-87d5-8ec365db3247\") " pod="openshift-ovn-kubernetes/ovnkube-node-6cz9b" Dec 03 14:22:14 crc kubenswrapper[4805]: I1203 14:22:14.955529 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/ea4cce30-4d9a-4384-87d5-8ec365db3247-host-cni-netd\") pod \"ovnkube-node-6cz9b\" (UID: \"ea4cce30-4d9a-4384-87d5-8ec365db3247\") " pod="openshift-ovn-kubernetes/ovnkube-node-6cz9b" Dec 03 14:22:14 crc kubenswrapper[4805]: I1203 14:22:14.955549 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/ea4cce30-4d9a-4384-87d5-8ec365db3247-run-systemd\") pod \"ovnkube-node-6cz9b\" (UID: \"ea4cce30-4d9a-4384-87d5-8ec365db3247\") " pod="openshift-ovn-kubernetes/ovnkube-node-6cz9b" Dec 03 14:22:14 crc kubenswrapper[4805]: I1203 14:22:14.955574 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/ea4cce30-4d9a-4384-87d5-8ec365db3247-host-slash\") pod \"ovnkube-node-6cz9b\" (UID: \"ea4cce30-4d9a-4384-87d5-8ec365db3247\") " pod="openshift-ovn-kubernetes/ovnkube-node-6cz9b" Dec 03 14:22:14 crc kubenswrapper[4805]: I1203 14:22:14.955579 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/deeec30d-cad2-4dfd-adb6-64f6646581be-ovnkube-script-lib" (OuterVolumeSpecName: "ovnkube-script-lib") pod "deeec30d-cad2-4dfd-adb6-64f6646581be" (UID: "deeec30d-cad2-4dfd-adb6-64f6646581be"). InnerVolumeSpecName "ovnkube-script-lib". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 14:22:14 crc kubenswrapper[4805]: I1203 14:22:14.955672 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/ea4cce30-4d9a-4384-87d5-8ec365db3247-run-ovn\") pod \"ovnkube-node-6cz9b\" (UID: \"ea4cce30-4d9a-4384-87d5-8ec365db3247\") " pod="openshift-ovn-kubernetes/ovnkube-node-6cz9b" Dec 03 14:22:14 crc kubenswrapper[4805]: I1203 14:22:14.955701 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/ea4cce30-4d9a-4384-87d5-8ec365db3247-log-socket\") pod \"ovnkube-node-6cz9b\" (UID: \"ea4cce30-4d9a-4384-87d5-8ec365db3247\") " pod="openshift-ovn-kubernetes/ovnkube-node-6cz9b" Dec 03 14:22:14 crc kubenswrapper[4805]: I1203 14:22:14.955909 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/ea4cce30-4d9a-4384-87d5-8ec365db3247-ovnkube-script-lib\") pod \"ovnkube-node-6cz9b\" (UID: \"ea4cce30-4d9a-4384-87d5-8ec365db3247\") " pod="openshift-ovn-kubernetes/ovnkube-node-6cz9b" Dec 03 14:22:14 crc kubenswrapper[4805]: I1203 14:22:14.956005 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/ea4cce30-4d9a-4384-87d5-8ec365db3247-var-lib-openvswitch\") pod \"ovnkube-node-6cz9b\" (UID: \"ea4cce30-4d9a-4384-87d5-8ec365db3247\") " pod="openshift-ovn-kubernetes/ovnkube-node-6cz9b" Dec 03 14:22:14 crc kubenswrapper[4805]: I1203 14:22:14.956044 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/ea4cce30-4d9a-4384-87d5-8ec365db3247-host-run-ovn-kubernetes\") pod \"ovnkube-node-6cz9b\" (UID: \"ea4cce30-4d9a-4384-87d5-8ec365db3247\") " pod="openshift-ovn-kubernetes/ovnkube-node-6cz9b" Dec 03 14:22:14 crc kubenswrapper[4805]: I1203 14:22:14.956067 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/ea4cce30-4d9a-4384-87d5-8ec365db3247-systemd-units\") pod \"ovnkube-node-6cz9b\" (UID: \"ea4cce30-4d9a-4384-87d5-8ec365db3247\") " pod="openshift-ovn-kubernetes/ovnkube-node-6cz9b" Dec 03 14:22:14 crc kubenswrapper[4805]: I1203 14:22:14.956091 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/ea4cce30-4d9a-4384-87d5-8ec365db3247-host-cni-bin\") pod \"ovnkube-node-6cz9b\" (UID: \"ea4cce30-4d9a-4384-87d5-8ec365db3247\") " pod="openshift-ovn-kubernetes/ovnkube-node-6cz9b" Dec 03 14:22:14 crc kubenswrapper[4805]: I1203 14:22:14.956133 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/ea4cce30-4d9a-4384-87d5-8ec365db3247-node-log\") pod \"ovnkube-node-6cz9b\" (UID: \"ea4cce30-4d9a-4384-87d5-8ec365db3247\") " pod="openshift-ovn-kubernetes/ovnkube-node-6cz9b" Dec 03 14:22:14 crc kubenswrapper[4805]: I1203 14:22:14.956170 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/ea4cce30-4d9a-4384-87d5-8ec365db3247-host-run-netns\") pod \"ovnkube-node-6cz9b\" (UID: \"ea4cce30-4d9a-4384-87d5-8ec365db3247\") " pod="openshift-ovn-kubernetes/ovnkube-node-6cz9b" Dec 03 14:22:14 crc kubenswrapper[4805]: I1203 14:22:14.956209 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-plnxr\" (UniqueName: \"kubernetes.io/projected/ea4cce30-4d9a-4384-87d5-8ec365db3247-kube-api-access-plnxr\") pod \"ovnkube-node-6cz9b\" (UID: \"ea4cce30-4d9a-4384-87d5-8ec365db3247\") " pod="openshift-ovn-kubernetes/ovnkube-node-6cz9b" Dec 03 14:22:14 crc kubenswrapper[4805]: I1203 14:22:14.956243 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/ea4cce30-4d9a-4384-87d5-8ec365db3247-run-openvswitch\") pod \"ovnkube-node-6cz9b\" (UID: \"ea4cce30-4d9a-4384-87d5-8ec365db3247\") " pod="openshift-ovn-kubernetes/ovnkube-node-6cz9b" Dec 03 14:22:14 crc kubenswrapper[4805]: I1203 14:22:14.956265 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/ea4cce30-4d9a-4384-87d5-8ec365db3247-env-overrides\") pod \"ovnkube-node-6cz9b\" (UID: \"ea4cce30-4d9a-4384-87d5-8ec365db3247\") " pod="openshift-ovn-kubernetes/ovnkube-node-6cz9b" Dec 03 14:22:14 crc kubenswrapper[4805]: I1203 14:22:14.956341 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/ea4cce30-4d9a-4384-87d5-8ec365db3247-etc-openvswitch\") pod \"ovnkube-node-6cz9b\" (UID: \"ea4cce30-4d9a-4384-87d5-8ec365db3247\") " pod="openshift-ovn-kubernetes/ovnkube-node-6cz9b" Dec 03 14:22:14 crc kubenswrapper[4805]: I1203 14:22:14.956431 4805 reconciler_common.go:293] "Volume detached for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/deeec30d-cad2-4dfd-adb6-64f6646581be-host-run-netns\") on node \"crc\" DevicePath \"\"" Dec 03 14:22:14 crc kubenswrapper[4805]: I1203 14:22:14.956452 4805 reconciler_common.go:293] "Volume detached for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/deeec30d-cad2-4dfd-adb6-64f6646581be-host-slash\") on node \"crc\" DevicePath \"\"" Dec 03 14:22:14 crc kubenswrapper[4805]: I1203 14:22:14.956467 4805 reconciler_common.go:293] "Volume detached for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/deeec30d-cad2-4dfd-adb6-64f6646581be-host-var-lib-cni-networks-ovn-kubernetes\") on node \"crc\" DevicePath \"\"" Dec 03 14:22:14 crc kubenswrapper[4805]: I1203 14:22:14.956481 4805 reconciler_common.go:293] "Volume detached for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/deeec30d-cad2-4dfd-adb6-64f6646581be-host-kubelet\") on node \"crc\" DevicePath \"\"" Dec 03 14:22:14 crc kubenswrapper[4805]: I1203 14:22:14.956493 4805 reconciler_common.go:293] "Volume detached for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/deeec30d-cad2-4dfd-adb6-64f6646581be-systemd-units\") on node \"crc\" DevicePath \"\"" Dec 03 14:22:14 crc kubenswrapper[4805]: I1203 14:22:14.956504 4805 reconciler_common.go:293] "Volume detached for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/deeec30d-cad2-4dfd-adb6-64f6646581be-run-ovn\") on node \"crc\" DevicePath \"\"" Dec 03 14:22:14 crc kubenswrapper[4805]: I1203 14:22:14.956516 4805 reconciler_common.go:293] "Volume detached for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/deeec30d-cad2-4dfd-adb6-64f6646581be-ovnkube-script-lib\") on node \"crc\" DevicePath \"\"" Dec 03 14:22:14 crc kubenswrapper[4805]: I1203 14:22:14.956529 4805 reconciler_common.go:293] "Volume detached for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/deeec30d-cad2-4dfd-adb6-64f6646581be-etc-openvswitch\") on node \"crc\" DevicePath \"\"" Dec 03 14:22:14 crc kubenswrapper[4805]: I1203 14:22:14.956540 4805 reconciler_common.go:293] "Volume detached for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/deeec30d-cad2-4dfd-adb6-64f6646581be-host-run-ovn-kubernetes\") on node \"crc\" DevicePath \"\"" Dec 03 14:22:14 crc kubenswrapper[4805]: I1203 14:22:14.956552 4805 reconciler_common.go:293] "Volume detached for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/deeec30d-cad2-4dfd-adb6-64f6646581be-var-lib-openvswitch\") on node \"crc\" DevicePath \"\"" Dec 03 14:22:14 crc kubenswrapper[4805]: I1203 14:22:14.956563 4805 reconciler_common.go:293] "Volume detached for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/deeec30d-cad2-4dfd-adb6-64f6646581be-ovnkube-config\") on node \"crc\" DevicePath \"\"" Dec 03 14:22:14 crc kubenswrapper[4805]: I1203 14:22:14.956573 4805 reconciler_common.go:293] "Volume detached for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/deeec30d-cad2-4dfd-adb6-64f6646581be-node-log\") on node \"crc\" DevicePath \"\"" Dec 03 14:22:14 crc kubenswrapper[4805]: I1203 14:22:14.956583 4805 reconciler_common.go:293] "Volume detached for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/deeec30d-cad2-4dfd-adb6-64f6646581be-log-socket\") on node \"crc\" DevicePath \"\"" Dec 03 14:22:14 crc kubenswrapper[4805]: I1203 14:22:14.956595 4805 reconciler_common.go:293] "Volume detached for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/deeec30d-cad2-4dfd-adb6-64f6646581be-env-overrides\") on node \"crc\" DevicePath \"\"" Dec 03 14:22:14 crc kubenswrapper[4805]: I1203 14:22:14.956606 4805 reconciler_common.go:293] "Volume detached for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/deeec30d-cad2-4dfd-adb6-64f6646581be-host-cni-bin\") on node \"crc\" DevicePath \"\"" Dec 03 14:22:14 crc kubenswrapper[4805]: I1203 14:22:14.956616 4805 reconciler_common.go:293] "Volume detached for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/deeec30d-cad2-4dfd-adb6-64f6646581be-host-cni-netd\") on node \"crc\" DevicePath \"\"" Dec 03 14:22:14 crc kubenswrapper[4805]: I1203 14:22:14.956628 4805 reconciler_common.go:293] "Volume detached for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/deeec30d-cad2-4dfd-adb6-64f6646581be-run-openvswitch\") on node \"crc\" DevicePath \"\"" Dec 03 14:22:14 crc kubenswrapper[4805]: I1203 14:22:14.960683 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/deeec30d-cad2-4dfd-adb6-64f6646581be-ovn-node-metrics-cert" (OuterVolumeSpecName: "ovn-node-metrics-cert") pod "deeec30d-cad2-4dfd-adb6-64f6646581be" (UID: "deeec30d-cad2-4dfd-adb6-64f6646581be"). InnerVolumeSpecName "ovn-node-metrics-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 14:22:14 crc kubenswrapper[4805]: I1203 14:22:14.961964 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/deeec30d-cad2-4dfd-adb6-64f6646581be-kube-api-access-wdlzx" (OuterVolumeSpecName: "kube-api-access-wdlzx") pod "deeec30d-cad2-4dfd-adb6-64f6646581be" (UID: "deeec30d-cad2-4dfd-adb6-64f6646581be"). InnerVolumeSpecName "kube-api-access-wdlzx". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 14:22:14 crc kubenswrapper[4805]: I1203 14:22:14.971338 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/deeec30d-cad2-4dfd-adb6-64f6646581be-run-systemd" (OuterVolumeSpecName: "run-systemd") pod "deeec30d-cad2-4dfd-adb6-64f6646581be" (UID: "deeec30d-cad2-4dfd-adb6-64f6646581be"). InnerVolumeSpecName "run-systemd". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 03 14:22:15 crc kubenswrapper[4805]: I1203 14:22:15.057947 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/ea4cce30-4d9a-4384-87d5-8ec365db3247-etc-openvswitch\") pod \"ovnkube-node-6cz9b\" (UID: \"ea4cce30-4d9a-4384-87d5-8ec365db3247\") " pod="openshift-ovn-kubernetes/ovnkube-node-6cz9b" Dec 03 14:22:15 crc kubenswrapper[4805]: I1203 14:22:15.058504 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/ea4cce30-4d9a-4384-87d5-8ec365db3247-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-6cz9b\" (UID: \"ea4cce30-4d9a-4384-87d5-8ec365db3247\") " pod="openshift-ovn-kubernetes/ovnkube-node-6cz9b" Dec 03 14:22:15 crc kubenswrapper[4805]: I1203 14:22:15.058539 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/ea4cce30-4d9a-4384-87d5-8ec365db3247-host-kubelet\") pod \"ovnkube-node-6cz9b\" (UID: \"ea4cce30-4d9a-4384-87d5-8ec365db3247\") " pod="openshift-ovn-kubernetes/ovnkube-node-6cz9b" Dec 03 14:22:15 crc kubenswrapper[4805]: I1203 14:22:15.058179 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/ea4cce30-4d9a-4384-87d5-8ec365db3247-etc-openvswitch\") pod \"ovnkube-node-6cz9b\" (UID: \"ea4cce30-4d9a-4384-87d5-8ec365db3247\") " pod="openshift-ovn-kubernetes/ovnkube-node-6cz9b" Dec 03 14:22:15 crc kubenswrapper[4805]: I1203 14:22:15.058570 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/ea4cce30-4d9a-4384-87d5-8ec365db3247-ovnkube-config\") pod \"ovnkube-node-6cz9b\" (UID: \"ea4cce30-4d9a-4384-87d5-8ec365db3247\") " pod="openshift-ovn-kubernetes/ovnkube-node-6cz9b" Dec 03 14:22:15 crc kubenswrapper[4805]: I1203 14:22:15.058778 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/ea4cce30-4d9a-4384-87d5-8ec365db3247-ovn-node-metrics-cert\") pod \"ovnkube-node-6cz9b\" (UID: \"ea4cce30-4d9a-4384-87d5-8ec365db3247\") " pod="openshift-ovn-kubernetes/ovnkube-node-6cz9b" Dec 03 14:22:15 crc kubenswrapper[4805]: I1203 14:22:15.058810 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/ea4cce30-4d9a-4384-87d5-8ec365db3247-host-cni-netd\") pod \"ovnkube-node-6cz9b\" (UID: \"ea4cce30-4d9a-4384-87d5-8ec365db3247\") " pod="openshift-ovn-kubernetes/ovnkube-node-6cz9b" Dec 03 14:22:15 crc kubenswrapper[4805]: I1203 14:22:15.058855 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/ea4cce30-4d9a-4384-87d5-8ec365db3247-run-systemd\") pod \"ovnkube-node-6cz9b\" (UID: \"ea4cce30-4d9a-4384-87d5-8ec365db3247\") " pod="openshift-ovn-kubernetes/ovnkube-node-6cz9b" Dec 03 14:22:15 crc kubenswrapper[4805]: I1203 14:22:15.058893 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/ea4cce30-4d9a-4384-87d5-8ec365db3247-host-slash\") pod \"ovnkube-node-6cz9b\" (UID: \"ea4cce30-4d9a-4384-87d5-8ec365db3247\") " pod="openshift-ovn-kubernetes/ovnkube-node-6cz9b" Dec 03 14:22:15 crc kubenswrapper[4805]: I1203 14:22:15.058969 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/ea4cce30-4d9a-4384-87d5-8ec365db3247-run-ovn\") pod \"ovnkube-node-6cz9b\" (UID: \"ea4cce30-4d9a-4384-87d5-8ec365db3247\") " pod="openshift-ovn-kubernetes/ovnkube-node-6cz9b" Dec 03 14:22:15 crc kubenswrapper[4805]: I1203 14:22:15.058990 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/ea4cce30-4d9a-4384-87d5-8ec365db3247-log-socket\") pod \"ovnkube-node-6cz9b\" (UID: \"ea4cce30-4d9a-4384-87d5-8ec365db3247\") " pod="openshift-ovn-kubernetes/ovnkube-node-6cz9b" Dec 03 14:22:15 crc kubenswrapper[4805]: I1203 14:22:15.059034 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/ea4cce30-4d9a-4384-87d5-8ec365db3247-ovnkube-script-lib\") pod \"ovnkube-node-6cz9b\" (UID: \"ea4cce30-4d9a-4384-87d5-8ec365db3247\") " pod="openshift-ovn-kubernetes/ovnkube-node-6cz9b" Dec 03 14:22:15 crc kubenswrapper[4805]: I1203 14:22:15.059031 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/ea4cce30-4d9a-4384-87d5-8ec365db3247-run-systemd\") pod \"ovnkube-node-6cz9b\" (UID: \"ea4cce30-4d9a-4384-87d5-8ec365db3247\") " pod="openshift-ovn-kubernetes/ovnkube-node-6cz9b" Dec 03 14:22:15 crc kubenswrapper[4805]: I1203 14:22:15.059055 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/ea4cce30-4d9a-4384-87d5-8ec365db3247-var-lib-openvswitch\") pod \"ovnkube-node-6cz9b\" (UID: \"ea4cce30-4d9a-4384-87d5-8ec365db3247\") " pod="openshift-ovn-kubernetes/ovnkube-node-6cz9b" Dec 03 14:22:15 crc kubenswrapper[4805]: I1203 14:22:15.059078 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/ea4cce30-4d9a-4384-87d5-8ec365db3247-host-run-ovn-kubernetes\") pod \"ovnkube-node-6cz9b\" (UID: \"ea4cce30-4d9a-4384-87d5-8ec365db3247\") " pod="openshift-ovn-kubernetes/ovnkube-node-6cz9b" Dec 03 14:22:15 crc kubenswrapper[4805]: I1203 14:22:15.059087 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/ea4cce30-4d9a-4384-87d5-8ec365db3247-host-cni-netd\") pod \"ovnkube-node-6cz9b\" (UID: \"ea4cce30-4d9a-4384-87d5-8ec365db3247\") " pod="openshift-ovn-kubernetes/ovnkube-node-6cz9b" Dec 03 14:22:15 crc kubenswrapper[4805]: I1203 14:22:15.058580 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/ea4cce30-4d9a-4384-87d5-8ec365db3247-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-6cz9b\" (UID: \"ea4cce30-4d9a-4384-87d5-8ec365db3247\") " pod="openshift-ovn-kubernetes/ovnkube-node-6cz9b" Dec 03 14:22:15 crc kubenswrapper[4805]: I1203 14:22:15.059101 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/ea4cce30-4d9a-4384-87d5-8ec365db3247-systemd-units\") pod \"ovnkube-node-6cz9b\" (UID: \"ea4cce30-4d9a-4384-87d5-8ec365db3247\") " pod="openshift-ovn-kubernetes/ovnkube-node-6cz9b" Dec 03 14:22:15 crc kubenswrapper[4805]: I1203 14:22:15.059162 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/ea4cce30-4d9a-4384-87d5-8ec365db3247-host-cni-bin\") pod \"ovnkube-node-6cz9b\" (UID: \"ea4cce30-4d9a-4384-87d5-8ec365db3247\") " pod="openshift-ovn-kubernetes/ovnkube-node-6cz9b" Dec 03 14:22:15 crc kubenswrapper[4805]: I1203 14:22:15.059194 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/ea4cce30-4d9a-4384-87d5-8ec365db3247-node-log\") pod \"ovnkube-node-6cz9b\" (UID: \"ea4cce30-4d9a-4384-87d5-8ec365db3247\") " pod="openshift-ovn-kubernetes/ovnkube-node-6cz9b" Dec 03 14:22:15 crc kubenswrapper[4805]: I1203 14:22:15.059225 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/ea4cce30-4d9a-4384-87d5-8ec365db3247-host-run-netns\") pod \"ovnkube-node-6cz9b\" (UID: \"ea4cce30-4d9a-4384-87d5-8ec365db3247\") " pod="openshift-ovn-kubernetes/ovnkube-node-6cz9b" Dec 03 14:22:15 crc kubenswrapper[4805]: I1203 14:22:15.059248 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-plnxr\" (UniqueName: \"kubernetes.io/projected/ea4cce30-4d9a-4384-87d5-8ec365db3247-kube-api-access-plnxr\") pod \"ovnkube-node-6cz9b\" (UID: \"ea4cce30-4d9a-4384-87d5-8ec365db3247\") " pod="openshift-ovn-kubernetes/ovnkube-node-6cz9b" Dec 03 14:22:15 crc kubenswrapper[4805]: I1203 14:22:15.059266 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/ea4cce30-4d9a-4384-87d5-8ec365db3247-env-overrides\") pod \"ovnkube-node-6cz9b\" (UID: \"ea4cce30-4d9a-4384-87d5-8ec365db3247\") " pod="openshift-ovn-kubernetes/ovnkube-node-6cz9b" Dec 03 14:22:15 crc kubenswrapper[4805]: I1203 14:22:15.059287 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/ea4cce30-4d9a-4384-87d5-8ec365db3247-run-openvswitch\") pod \"ovnkube-node-6cz9b\" (UID: \"ea4cce30-4d9a-4384-87d5-8ec365db3247\") " pod="openshift-ovn-kubernetes/ovnkube-node-6cz9b" Dec 03 14:22:15 crc kubenswrapper[4805]: I1203 14:22:15.059379 4805 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wdlzx\" (UniqueName: \"kubernetes.io/projected/deeec30d-cad2-4dfd-adb6-64f6646581be-kube-api-access-wdlzx\") on node \"crc\" DevicePath \"\"" Dec 03 14:22:15 crc kubenswrapper[4805]: I1203 14:22:15.059405 4805 reconciler_common.go:293] "Volume detached for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/deeec30d-cad2-4dfd-adb6-64f6646581be-run-systemd\") on node \"crc\" DevicePath \"\"" Dec 03 14:22:15 crc kubenswrapper[4805]: I1203 14:22:15.059420 4805 reconciler_common.go:293] "Volume detached for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/deeec30d-cad2-4dfd-adb6-64f6646581be-ovn-node-metrics-cert\") on node \"crc\" DevicePath \"\"" Dec 03 14:22:15 crc kubenswrapper[4805]: I1203 14:22:15.059450 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/ea4cce30-4d9a-4384-87d5-8ec365db3247-run-openvswitch\") pod \"ovnkube-node-6cz9b\" (UID: \"ea4cce30-4d9a-4384-87d5-8ec365db3247\") " pod="openshift-ovn-kubernetes/ovnkube-node-6cz9b" Dec 03 14:22:15 crc kubenswrapper[4805]: I1203 14:22:15.059478 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/ea4cce30-4d9a-4384-87d5-8ec365db3247-var-lib-openvswitch\") pod \"ovnkube-node-6cz9b\" (UID: \"ea4cce30-4d9a-4384-87d5-8ec365db3247\") " pod="openshift-ovn-kubernetes/ovnkube-node-6cz9b" Dec 03 14:22:15 crc kubenswrapper[4805]: I1203 14:22:15.058696 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/ea4cce30-4d9a-4384-87d5-8ec365db3247-host-kubelet\") pod \"ovnkube-node-6cz9b\" (UID: \"ea4cce30-4d9a-4384-87d5-8ec365db3247\") " pod="openshift-ovn-kubernetes/ovnkube-node-6cz9b" Dec 03 14:22:15 crc kubenswrapper[4805]: I1203 14:22:15.059515 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/ea4cce30-4d9a-4384-87d5-8ec365db3247-host-run-ovn-kubernetes\") pod \"ovnkube-node-6cz9b\" (UID: \"ea4cce30-4d9a-4384-87d5-8ec365db3247\") " pod="openshift-ovn-kubernetes/ovnkube-node-6cz9b" Dec 03 14:22:15 crc kubenswrapper[4805]: I1203 14:22:15.059544 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/ea4cce30-4d9a-4384-87d5-8ec365db3247-systemd-units\") pod \"ovnkube-node-6cz9b\" (UID: \"ea4cce30-4d9a-4384-87d5-8ec365db3247\") " pod="openshift-ovn-kubernetes/ovnkube-node-6cz9b" Dec 03 14:22:15 crc kubenswrapper[4805]: I1203 14:22:15.059572 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/ea4cce30-4d9a-4384-87d5-8ec365db3247-host-cni-bin\") pod \"ovnkube-node-6cz9b\" (UID: \"ea4cce30-4d9a-4384-87d5-8ec365db3247\") " pod="openshift-ovn-kubernetes/ovnkube-node-6cz9b" Dec 03 14:22:15 crc kubenswrapper[4805]: I1203 14:22:15.059598 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/ea4cce30-4d9a-4384-87d5-8ec365db3247-node-log\") pod \"ovnkube-node-6cz9b\" (UID: \"ea4cce30-4d9a-4384-87d5-8ec365db3247\") " pod="openshift-ovn-kubernetes/ovnkube-node-6cz9b" Dec 03 14:22:15 crc kubenswrapper[4805]: I1203 14:22:15.059628 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/ea4cce30-4d9a-4384-87d5-8ec365db3247-host-run-netns\") pod \"ovnkube-node-6cz9b\" (UID: \"ea4cce30-4d9a-4384-87d5-8ec365db3247\") " pod="openshift-ovn-kubernetes/ovnkube-node-6cz9b" Dec 03 14:22:15 crc kubenswrapper[4805]: I1203 14:22:15.059625 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/ea4cce30-4d9a-4384-87d5-8ec365db3247-ovnkube-config\") pod \"ovnkube-node-6cz9b\" (UID: \"ea4cce30-4d9a-4384-87d5-8ec365db3247\") " pod="openshift-ovn-kubernetes/ovnkube-node-6cz9b" Dec 03 14:22:15 crc kubenswrapper[4805]: I1203 14:22:15.059665 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/ea4cce30-4d9a-4384-87d5-8ec365db3247-run-ovn\") pod \"ovnkube-node-6cz9b\" (UID: \"ea4cce30-4d9a-4384-87d5-8ec365db3247\") " pod="openshift-ovn-kubernetes/ovnkube-node-6cz9b" Dec 03 14:22:15 crc kubenswrapper[4805]: I1203 14:22:15.059720 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/ea4cce30-4d9a-4384-87d5-8ec365db3247-host-slash\") pod \"ovnkube-node-6cz9b\" (UID: \"ea4cce30-4d9a-4384-87d5-8ec365db3247\") " pod="openshift-ovn-kubernetes/ovnkube-node-6cz9b" Dec 03 14:22:15 crc kubenswrapper[4805]: I1203 14:22:15.059755 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/ea4cce30-4d9a-4384-87d5-8ec365db3247-log-socket\") pod \"ovnkube-node-6cz9b\" (UID: \"ea4cce30-4d9a-4384-87d5-8ec365db3247\") " pod="openshift-ovn-kubernetes/ovnkube-node-6cz9b" Dec 03 14:22:15 crc kubenswrapper[4805]: I1203 14:22:15.060086 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/ea4cce30-4d9a-4384-87d5-8ec365db3247-ovnkube-script-lib\") pod \"ovnkube-node-6cz9b\" (UID: \"ea4cce30-4d9a-4384-87d5-8ec365db3247\") " pod="openshift-ovn-kubernetes/ovnkube-node-6cz9b" Dec 03 14:22:15 crc kubenswrapper[4805]: I1203 14:22:15.060244 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/ea4cce30-4d9a-4384-87d5-8ec365db3247-env-overrides\") pod \"ovnkube-node-6cz9b\" (UID: \"ea4cce30-4d9a-4384-87d5-8ec365db3247\") " pod="openshift-ovn-kubernetes/ovnkube-node-6cz9b" Dec 03 14:22:15 crc kubenswrapper[4805]: I1203 14:22:15.066443 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/ea4cce30-4d9a-4384-87d5-8ec365db3247-ovn-node-metrics-cert\") pod \"ovnkube-node-6cz9b\" (UID: \"ea4cce30-4d9a-4384-87d5-8ec365db3247\") " pod="openshift-ovn-kubernetes/ovnkube-node-6cz9b" Dec 03 14:22:15 crc kubenswrapper[4805]: I1203 14:22:15.086576 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-plnxr\" (UniqueName: \"kubernetes.io/projected/ea4cce30-4d9a-4384-87d5-8ec365db3247-kube-api-access-plnxr\") pod \"ovnkube-node-6cz9b\" (UID: \"ea4cce30-4d9a-4384-87d5-8ec365db3247\") " pod="openshift-ovn-kubernetes/ovnkube-node-6cz9b" Dec 03 14:22:15 crc kubenswrapper[4805]: I1203 14:22:15.209646 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-6cz9b" Dec 03 14:22:15 crc kubenswrapper[4805]: W1203 14:22:15.226131 4805 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podea4cce30_4d9a_4384_87d5_8ec365db3247.slice/crio-f0ed24736216f7fe8f8ad0f47a77dbd654e194327d8e1b9c8a1a24bdb61e31c9 WatchSource:0}: Error finding container f0ed24736216f7fe8f8ad0f47a77dbd654e194327d8e1b9c8a1a24bdb61e31c9: Status 404 returned error can't find the container with id f0ed24736216f7fe8f8ad0f47a77dbd654e194327d8e1b9c8a1a24bdb61e31c9 Dec 03 14:22:15 crc kubenswrapper[4805]: I1203 14:22:15.742744 4805 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-psjpm_deeec30d-cad2-4dfd-adb6-64f6646581be/ovn-acl-logging/0.log" Dec 03 14:22:15 crc kubenswrapper[4805]: I1203 14:22:15.743337 4805 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-psjpm_deeec30d-cad2-4dfd-adb6-64f6646581be/ovn-controller/0.log" Dec 03 14:22:15 crc kubenswrapper[4805]: I1203 14:22:15.743754 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-psjpm" event={"ID":"deeec30d-cad2-4dfd-adb6-64f6646581be","Type":"ContainerDied","Data":"fd34043a1f52c16209aad4766079aff91088a34a1a1e9a27a022a8c8083c8d86"} Dec 03 14:22:15 crc kubenswrapper[4805]: I1203 14:22:15.743787 4805 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-psjpm" Dec 03 14:22:15 crc kubenswrapper[4805]: I1203 14:22:15.743792 4805 scope.go:117] "RemoveContainer" containerID="cb029c1dbe665e82f006ab5f681796fa66f6ff814a3988c3c4b5863923b4a44f" Dec 03 14:22:15 crc kubenswrapper[4805]: I1203 14:22:15.749082 4805 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-k4mhs_c09c1cdc-f461-45c9-8444-5b8764bccabc/kube-multus/2.log" Dec 03 14:22:15 crc kubenswrapper[4805]: I1203 14:22:15.749298 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-k4mhs" event={"ID":"c09c1cdc-f461-45c9-8444-5b8764bccabc","Type":"ContainerStarted","Data":"84d205519af43b3d80b78703faa9f5908d0b1a56ee80722d462a752e5d4dbaa9"} Dec 03 14:22:15 crc kubenswrapper[4805]: I1203 14:22:15.751703 4805 generic.go:334] "Generic (PLEG): container finished" podID="ea4cce30-4d9a-4384-87d5-8ec365db3247" containerID="7c633ae4766b3d85a5c028f898e44852b3ba9f2c118da8e51999f75a944bd3b0" exitCode=0 Dec 03 14:22:15 crc kubenswrapper[4805]: I1203 14:22:15.751815 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-6cz9b" event={"ID":"ea4cce30-4d9a-4384-87d5-8ec365db3247","Type":"ContainerDied","Data":"7c633ae4766b3d85a5c028f898e44852b3ba9f2c118da8e51999f75a944bd3b0"} Dec 03 14:22:15 crc kubenswrapper[4805]: I1203 14:22:15.751933 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-6cz9b" event={"ID":"ea4cce30-4d9a-4384-87d5-8ec365db3247","Type":"ContainerStarted","Data":"f0ed24736216f7fe8f8ad0f47a77dbd654e194327d8e1b9c8a1a24bdb61e31c9"} Dec 03 14:22:15 crc kubenswrapper[4805]: I1203 14:22:15.766338 4805 scope.go:117] "RemoveContainer" containerID="ab98e365c1a4242a47b8cbe2dd85109d366d6002c4de1f18ddc950a46bea6c12" Dec 03 14:22:15 crc kubenswrapper[4805]: I1203 14:22:15.801601 4805 scope.go:117] "RemoveContainer" containerID="cb8bf05ef093eb0727a066d4bbbdf9af9422867cd4dadacafc6eeb016867a61e" Dec 03 14:22:15 crc kubenswrapper[4805]: I1203 14:22:15.823826 4805 scope.go:117] "RemoveContainer" containerID="0304467102ed053628c0d3e06548c74b6492336b6fd044246508d1e8bdb17bc3" Dec 03 14:22:15 crc kubenswrapper[4805]: I1203 14:22:15.841059 4805 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-psjpm"] Dec 03 14:22:15 crc kubenswrapper[4805]: I1203 14:22:15.854214 4805 scope.go:117] "RemoveContainer" containerID="bd67c6de01a8d196a1f51d4d90640c9e69de6f05d5e201b25e6eaa5c33440ea8" Dec 03 14:22:15 crc kubenswrapper[4805]: I1203 14:22:15.857578 4805 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-psjpm"] Dec 03 14:22:15 crc kubenswrapper[4805]: I1203 14:22:15.883530 4805 scope.go:117] "RemoveContainer" containerID="4b8d00cdf0f37f2e5cd4e952cf85d3de8d72c43fdf786088fd9f26ed17b1e21f" Dec 03 14:22:15 crc kubenswrapper[4805]: I1203 14:22:15.899061 4805 scope.go:117] "RemoveContainer" containerID="6ce80ccafb3262a11fbe9923317a9d38b61e5dc5b3e7e7ae136f143d6b643fd7" Dec 03 14:22:15 crc kubenswrapper[4805]: I1203 14:22:15.936139 4805 scope.go:117] "RemoveContainer" containerID="47d76962beb8c0f196d50af96fb3ec611097823ca80a1c8bd718f2d18840128c" Dec 03 14:22:15 crc kubenswrapper[4805]: I1203 14:22:15.963173 4805 scope.go:117] "RemoveContainer" containerID="2f875252fbe6a673a72fc9c56bb96899f62732305ad77b3108e3b244bfa91ecb" Dec 03 14:22:16 crc kubenswrapper[4805]: I1203 14:22:16.704329 4805 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="deeec30d-cad2-4dfd-adb6-64f6646581be" path="/var/lib/kubelet/pods/deeec30d-cad2-4dfd-adb6-64f6646581be/volumes" Dec 03 14:22:16 crc kubenswrapper[4805]: I1203 14:22:16.760823 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-6cz9b" event={"ID":"ea4cce30-4d9a-4384-87d5-8ec365db3247","Type":"ContainerStarted","Data":"a362bbb63ac44cc637615dbc197f1628e3ecf918b54b2ad5ba19f4c065c7dc2d"} Dec 03 14:22:16 crc kubenswrapper[4805]: I1203 14:22:16.760906 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-6cz9b" event={"ID":"ea4cce30-4d9a-4384-87d5-8ec365db3247","Type":"ContainerStarted","Data":"81e7e8c9a67851ea74a0ea5648c76dd7a24d5998c47bc6fc2393b573d3469564"} Dec 03 14:22:16 crc kubenswrapper[4805]: I1203 14:22:16.760925 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-6cz9b" event={"ID":"ea4cce30-4d9a-4384-87d5-8ec365db3247","Type":"ContainerStarted","Data":"3a60204256985c52e003574e2747f4e21e06722564478ba92c62d423cfd9019f"} Dec 03 14:22:16 crc kubenswrapper[4805]: I1203 14:22:16.760943 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-6cz9b" event={"ID":"ea4cce30-4d9a-4384-87d5-8ec365db3247","Type":"ContainerStarted","Data":"c06ccfcab0db13a1daf7d4ee8e11584f50cbea3612bd8ab6ac01e6e747f81e59"} Dec 03 14:22:16 crc kubenswrapper[4805]: I1203 14:22:16.760959 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-6cz9b" event={"ID":"ea4cce30-4d9a-4384-87d5-8ec365db3247","Type":"ContainerStarted","Data":"35a5aeff9e0aa98b633c5adaccb5e7e4205f48558774f51d45f61375ebf063c2"} Dec 03 14:22:16 crc kubenswrapper[4805]: I1203 14:22:16.760973 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-6cz9b" event={"ID":"ea4cce30-4d9a-4384-87d5-8ec365db3247","Type":"ContainerStarted","Data":"929c53b454e6cf81a104134ff510f89386f64cef40853a7c89eba48283199bf7"} Dec 03 14:22:19 crc kubenswrapper[4805]: I1203 14:22:19.790100 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-6cz9b" event={"ID":"ea4cce30-4d9a-4384-87d5-8ec365db3247","Type":"ContainerStarted","Data":"2975b588c9caa9d3f7fd6db43f8782832231aa949cb5f40eeca221588442ef6d"} Dec 03 14:22:20 crc kubenswrapper[4805]: I1203 14:22:20.599919 4805 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="cert-manager/cert-manager-webhook-5655c58dd6-mt9mr" Dec 03 14:22:22 crc kubenswrapper[4805]: I1203 14:22:22.820090 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-6cz9b" event={"ID":"ea4cce30-4d9a-4384-87d5-8ec365db3247","Type":"ContainerStarted","Data":"2c608bed6f597d5c36e6df7d461de2215ee02f0ed710eb294bdf27140ab7fedb"} Dec 03 14:22:23 crc kubenswrapper[4805]: I1203 14:22:23.825749 4805 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-6cz9b" Dec 03 14:22:23 crc kubenswrapper[4805]: I1203 14:22:23.826331 4805 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-6cz9b" Dec 03 14:22:23 crc kubenswrapper[4805]: I1203 14:22:23.826350 4805 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-6cz9b" Dec 03 14:22:23 crc kubenswrapper[4805]: I1203 14:22:23.862726 4805 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-6cz9b" Dec 03 14:22:23 crc kubenswrapper[4805]: I1203 14:22:23.868229 4805 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ovn-kubernetes/ovnkube-node-6cz9b" podStartSLOduration=9.86821206 podStartE2EDuration="9.86821206s" podCreationTimestamp="2025-12-03 14:22:14 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 14:22:23.865543849 +0000 UTC m=+773.528460772" watchObservedRunningTime="2025-12-03 14:22:23.86821206 +0000 UTC m=+773.531128983" Dec 03 14:22:23 crc kubenswrapper[4805]: I1203 14:22:23.879458 4805 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-6cz9b" Dec 03 14:22:30 crc kubenswrapper[4805]: I1203 14:22:30.046263 4805 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-rlgrd"] Dec 03 14:22:30 crc kubenswrapper[4805]: I1203 14:22:30.053878 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-rlgrd" Dec 03 14:22:30 crc kubenswrapper[4805]: I1203 14:22:30.063162 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-rlgrd"] Dec 03 14:22:30 crc kubenswrapper[4805]: I1203 14:22:30.098136 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xvm9d\" (UniqueName: \"kubernetes.io/projected/8d213a92-95be-4101-9317-980ce72cf5d1-kube-api-access-xvm9d\") pod \"redhat-marketplace-rlgrd\" (UID: \"8d213a92-95be-4101-9317-980ce72cf5d1\") " pod="openshift-marketplace/redhat-marketplace-rlgrd" Dec 03 14:22:30 crc kubenswrapper[4805]: I1203 14:22:30.098313 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8d213a92-95be-4101-9317-980ce72cf5d1-catalog-content\") pod \"redhat-marketplace-rlgrd\" (UID: \"8d213a92-95be-4101-9317-980ce72cf5d1\") " pod="openshift-marketplace/redhat-marketplace-rlgrd" Dec 03 14:22:30 crc kubenswrapper[4805]: I1203 14:22:30.098410 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8d213a92-95be-4101-9317-980ce72cf5d1-utilities\") pod \"redhat-marketplace-rlgrd\" (UID: \"8d213a92-95be-4101-9317-980ce72cf5d1\") " pod="openshift-marketplace/redhat-marketplace-rlgrd" Dec 03 14:22:30 crc kubenswrapper[4805]: I1203 14:22:30.199775 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8d213a92-95be-4101-9317-980ce72cf5d1-utilities\") pod \"redhat-marketplace-rlgrd\" (UID: \"8d213a92-95be-4101-9317-980ce72cf5d1\") " pod="openshift-marketplace/redhat-marketplace-rlgrd" Dec 03 14:22:30 crc kubenswrapper[4805]: I1203 14:22:30.199880 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xvm9d\" (UniqueName: \"kubernetes.io/projected/8d213a92-95be-4101-9317-980ce72cf5d1-kube-api-access-xvm9d\") pod \"redhat-marketplace-rlgrd\" (UID: \"8d213a92-95be-4101-9317-980ce72cf5d1\") " pod="openshift-marketplace/redhat-marketplace-rlgrd" Dec 03 14:22:30 crc kubenswrapper[4805]: I1203 14:22:30.199904 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8d213a92-95be-4101-9317-980ce72cf5d1-catalog-content\") pod \"redhat-marketplace-rlgrd\" (UID: \"8d213a92-95be-4101-9317-980ce72cf5d1\") " pod="openshift-marketplace/redhat-marketplace-rlgrd" Dec 03 14:22:30 crc kubenswrapper[4805]: I1203 14:22:30.200387 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8d213a92-95be-4101-9317-980ce72cf5d1-catalog-content\") pod \"redhat-marketplace-rlgrd\" (UID: \"8d213a92-95be-4101-9317-980ce72cf5d1\") " pod="openshift-marketplace/redhat-marketplace-rlgrd" Dec 03 14:22:30 crc kubenswrapper[4805]: I1203 14:22:30.200702 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8d213a92-95be-4101-9317-980ce72cf5d1-utilities\") pod \"redhat-marketplace-rlgrd\" (UID: \"8d213a92-95be-4101-9317-980ce72cf5d1\") " pod="openshift-marketplace/redhat-marketplace-rlgrd" Dec 03 14:22:30 crc kubenswrapper[4805]: I1203 14:22:30.225777 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xvm9d\" (UniqueName: \"kubernetes.io/projected/8d213a92-95be-4101-9317-980ce72cf5d1-kube-api-access-xvm9d\") pod \"redhat-marketplace-rlgrd\" (UID: \"8d213a92-95be-4101-9317-980ce72cf5d1\") " pod="openshift-marketplace/redhat-marketplace-rlgrd" Dec 03 14:22:30 crc kubenswrapper[4805]: I1203 14:22:30.381968 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-rlgrd" Dec 03 14:22:30 crc kubenswrapper[4805]: I1203 14:22:30.810132 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-rlgrd"] Dec 03 14:22:30 crc kubenswrapper[4805]: W1203 14:22:30.810443 4805 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod8d213a92_95be_4101_9317_980ce72cf5d1.slice/crio-95116eb1ee650d798cfff77a3e75dba4caf5ac67d391920106f38f1c9f8cc5a4 WatchSource:0}: Error finding container 95116eb1ee650d798cfff77a3e75dba4caf5ac67d391920106f38f1c9f8cc5a4: Status 404 returned error can't find the container with id 95116eb1ee650d798cfff77a3e75dba4caf5ac67d391920106f38f1c9f8cc5a4 Dec 03 14:22:30 crc kubenswrapper[4805]: I1203 14:22:30.884974 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-rlgrd" event={"ID":"8d213a92-95be-4101-9317-980ce72cf5d1","Type":"ContainerStarted","Data":"95116eb1ee650d798cfff77a3e75dba4caf5ac67d391920106f38f1c9f8cc5a4"} Dec 03 14:22:32 crc kubenswrapper[4805]: I1203 14:22:32.896757 4805 generic.go:334] "Generic (PLEG): container finished" podID="8d213a92-95be-4101-9317-980ce72cf5d1" containerID="eedef859d870ca72638cbacc07d0f9c4ebffb350ef0d130ffcd5ecab0f16f595" exitCode=0 Dec 03 14:22:32 crc kubenswrapper[4805]: I1203 14:22:32.896808 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-rlgrd" event={"ID":"8d213a92-95be-4101-9317-980ce72cf5d1","Type":"ContainerDied","Data":"eedef859d870ca72638cbacc07d0f9c4ebffb350ef0d130ffcd5ecab0f16f595"} Dec 03 14:22:33 crc kubenswrapper[4805]: I1203 14:22:33.907363 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-rlgrd" event={"ID":"8d213a92-95be-4101-9317-980ce72cf5d1","Type":"ContainerStarted","Data":"6c318cbda03ccc8640f5d2ada59f42a2b763f11bb34ef29d896b1cfcb228bce3"} Dec 03 14:22:34 crc kubenswrapper[4805]: I1203 14:22:34.916885 4805 generic.go:334] "Generic (PLEG): container finished" podID="8d213a92-95be-4101-9317-980ce72cf5d1" containerID="6c318cbda03ccc8640f5d2ada59f42a2b763f11bb34ef29d896b1cfcb228bce3" exitCode=0 Dec 03 14:22:34 crc kubenswrapper[4805]: I1203 14:22:34.916956 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-rlgrd" event={"ID":"8d213a92-95be-4101-9317-980ce72cf5d1","Type":"ContainerDied","Data":"6c318cbda03ccc8640f5d2ada59f42a2b763f11bb34ef29d896b1cfcb228bce3"} Dec 03 14:22:36 crc kubenswrapper[4805]: I1203 14:22:36.932186 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-rlgrd" event={"ID":"8d213a92-95be-4101-9317-980ce72cf5d1","Type":"ContainerStarted","Data":"4621c2ef911320a4d5fe9191ab74404413c4081e4dea9df1c9aad838489b65e0"} Dec 03 14:22:40 crc kubenswrapper[4805]: I1203 14:22:40.382561 4805 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-rlgrd" Dec 03 14:22:40 crc kubenswrapper[4805]: I1203 14:22:40.382632 4805 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-rlgrd" Dec 03 14:22:40 crc kubenswrapper[4805]: I1203 14:22:40.437511 4805 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-rlgrd" Dec 03 14:22:40 crc kubenswrapper[4805]: I1203 14:22:40.462619 4805 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-rlgrd" podStartSLOduration=7.776914651 podStartE2EDuration="10.462593649s" podCreationTimestamp="2025-12-03 14:22:30 +0000 UTC" firstStartedPulling="2025-12-03 14:22:32.899026847 +0000 UTC m=+782.561943770" lastFinishedPulling="2025-12-03 14:22:35.584705805 +0000 UTC m=+785.247622768" observedRunningTime="2025-12-03 14:22:36.95440732 +0000 UTC m=+786.617324243" watchObservedRunningTime="2025-12-03 14:22:40.462593649 +0000 UTC m=+790.125510612" Dec 03 14:22:43 crc kubenswrapper[4805]: I1203 14:22:43.917492 4805 patch_prober.go:28] interesting pod/machine-config-daemon-gskh4 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 03 14:22:43 crc kubenswrapper[4805]: I1203 14:22:43.917799 4805 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-gskh4" podUID="6098937f-e3f6-45e8-a647-4994a79cd711" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 03 14:22:45 crc kubenswrapper[4805]: I1203 14:22:45.249898 4805 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-6cz9b" Dec 03 14:22:50 crc kubenswrapper[4805]: I1203 14:22:50.422077 4805 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-rlgrd" Dec 03 14:22:50 crc kubenswrapper[4805]: I1203 14:22:50.463538 4805 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-rlgrd"] Dec 03 14:22:51 crc kubenswrapper[4805]: I1203 14:22:51.020940 4805 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-rlgrd" podUID="8d213a92-95be-4101-9317-980ce72cf5d1" containerName="registry-server" containerID="cri-o://4621c2ef911320a4d5fe9191ab74404413c4081e4dea9df1c9aad838489b65e0" gracePeriod=2 Dec 03 14:22:51 crc kubenswrapper[4805]: I1203 14:22:51.945623 4805 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-rlgrd" Dec 03 14:22:52 crc kubenswrapper[4805]: I1203 14:22:52.011109 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8d213a92-95be-4101-9317-980ce72cf5d1-utilities\") pod \"8d213a92-95be-4101-9317-980ce72cf5d1\" (UID: \"8d213a92-95be-4101-9317-980ce72cf5d1\") " Dec 03 14:22:52 crc kubenswrapper[4805]: I1203 14:22:52.011360 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xvm9d\" (UniqueName: \"kubernetes.io/projected/8d213a92-95be-4101-9317-980ce72cf5d1-kube-api-access-xvm9d\") pod \"8d213a92-95be-4101-9317-980ce72cf5d1\" (UID: \"8d213a92-95be-4101-9317-980ce72cf5d1\") " Dec 03 14:22:52 crc kubenswrapper[4805]: I1203 14:22:52.011437 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8d213a92-95be-4101-9317-980ce72cf5d1-catalog-content\") pod \"8d213a92-95be-4101-9317-980ce72cf5d1\" (UID: \"8d213a92-95be-4101-9317-980ce72cf5d1\") " Dec 03 14:22:52 crc kubenswrapper[4805]: I1203 14:22:52.012609 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8d213a92-95be-4101-9317-980ce72cf5d1-utilities" (OuterVolumeSpecName: "utilities") pod "8d213a92-95be-4101-9317-980ce72cf5d1" (UID: "8d213a92-95be-4101-9317-980ce72cf5d1"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 14:22:52 crc kubenswrapper[4805]: I1203 14:22:52.019778 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8d213a92-95be-4101-9317-980ce72cf5d1-kube-api-access-xvm9d" (OuterVolumeSpecName: "kube-api-access-xvm9d") pod "8d213a92-95be-4101-9317-980ce72cf5d1" (UID: "8d213a92-95be-4101-9317-980ce72cf5d1"). InnerVolumeSpecName "kube-api-access-xvm9d". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 14:22:52 crc kubenswrapper[4805]: I1203 14:22:52.028004 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8d213a92-95be-4101-9317-980ce72cf5d1-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "8d213a92-95be-4101-9317-980ce72cf5d1" (UID: "8d213a92-95be-4101-9317-980ce72cf5d1"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 14:22:52 crc kubenswrapper[4805]: I1203 14:22:52.029175 4805 generic.go:334] "Generic (PLEG): container finished" podID="8d213a92-95be-4101-9317-980ce72cf5d1" containerID="4621c2ef911320a4d5fe9191ab74404413c4081e4dea9df1c9aad838489b65e0" exitCode=0 Dec 03 14:22:52 crc kubenswrapper[4805]: I1203 14:22:52.029220 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-rlgrd" event={"ID":"8d213a92-95be-4101-9317-980ce72cf5d1","Type":"ContainerDied","Data":"4621c2ef911320a4d5fe9191ab74404413c4081e4dea9df1c9aad838489b65e0"} Dec 03 14:22:52 crc kubenswrapper[4805]: I1203 14:22:52.029267 4805 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-rlgrd" Dec 03 14:22:52 crc kubenswrapper[4805]: I1203 14:22:52.029298 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-rlgrd" event={"ID":"8d213a92-95be-4101-9317-980ce72cf5d1","Type":"ContainerDied","Data":"95116eb1ee650d798cfff77a3e75dba4caf5ac67d391920106f38f1c9f8cc5a4"} Dec 03 14:22:52 crc kubenswrapper[4805]: I1203 14:22:52.029565 4805 scope.go:117] "RemoveContainer" containerID="4621c2ef911320a4d5fe9191ab74404413c4081e4dea9df1c9aad838489b65e0" Dec 03 14:22:52 crc kubenswrapper[4805]: I1203 14:22:52.050528 4805 scope.go:117] "RemoveContainer" containerID="6c318cbda03ccc8640f5d2ada59f42a2b763f11bb34ef29d896b1cfcb228bce3" Dec 03 14:22:52 crc kubenswrapper[4805]: I1203 14:22:52.061939 4805 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-rlgrd"] Dec 03 14:22:52 crc kubenswrapper[4805]: I1203 14:22:52.065766 4805 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-rlgrd"] Dec 03 14:22:52 crc kubenswrapper[4805]: I1203 14:22:52.094300 4805 scope.go:117] "RemoveContainer" containerID="eedef859d870ca72638cbacc07d0f9c4ebffb350ef0d130ffcd5ecab0f16f595" Dec 03 14:22:52 crc kubenswrapper[4805]: I1203 14:22:52.110275 4805 scope.go:117] "RemoveContainer" containerID="4621c2ef911320a4d5fe9191ab74404413c4081e4dea9df1c9aad838489b65e0" Dec 03 14:22:52 crc kubenswrapper[4805]: E1203 14:22:52.111029 4805 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4621c2ef911320a4d5fe9191ab74404413c4081e4dea9df1c9aad838489b65e0\": container with ID starting with 4621c2ef911320a4d5fe9191ab74404413c4081e4dea9df1c9aad838489b65e0 not found: ID does not exist" containerID="4621c2ef911320a4d5fe9191ab74404413c4081e4dea9df1c9aad838489b65e0" Dec 03 14:22:52 crc kubenswrapper[4805]: I1203 14:22:52.111096 4805 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4621c2ef911320a4d5fe9191ab74404413c4081e4dea9df1c9aad838489b65e0"} err="failed to get container status \"4621c2ef911320a4d5fe9191ab74404413c4081e4dea9df1c9aad838489b65e0\": rpc error: code = NotFound desc = could not find container \"4621c2ef911320a4d5fe9191ab74404413c4081e4dea9df1c9aad838489b65e0\": container with ID starting with 4621c2ef911320a4d5fe9191ab74404413c4081e4dea9df1c9aad838489b65e0 not found: ID does not exist" Dec 03 14:22:52 crc kubenswrapper[4805]: I1203 14:22:52.111140 4805 scope.go:117] "RemoveContainer" containerID="6c318cbda03ccc8640f5d2ada59f42a2b763f11bb34ef29d896b1cfcb228bce3" Dec 03 14:22:52 crc kubenswrapper[4805]: E1203 14:22:52.111682 4805 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6c318cbda03ccc8640f5d2ada59f42a2b763f11bb34ef29d896b1cfcb228bce3\": container with ID starting with 6c318cbda03ccc8640f5d2ada59f42a2b763f11bb34ef29d896b1cfcb228bce3 not found: ID does not exist" containerID="6c318cbda03ccc8640f5d2ada59f42a2b763f11bb34ef29d896b1cfcb228bce3" Dec 03 14:22:52 crc kubenswrapper[4805]: I1203 14:22:52.111745 4805 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6c318cbda03ccc8640f5d2ada59f42a2b763f11bb34ef29d896b1cfcb228bce3"} err="failed to get container status \"6c318cbda03ccc8640f5d2ada59f42a2b763f11bb34ef29d896b1cfcb228bce3\": rpc error: code = NotFound desc = could not find container \"6c318cbda03ccc8640f5d2ada59f42a2b763f11bb34ef29d896b1cfcb228bce3\": container with ID starting with 6c318cbda03ccc8640f5d2ada59f42a2b763f11bb34ef29d896b1cfcb228bce3 not found: ID does not exist" Dec 03 14:22:52 crc kubenswrapper[4805]: I1203 14:22:52.111776 4805 scope.go:117] "RemoveContainer" containerID="eedef859d870ca72638cbacc07d0f9c4ebffb350ef0d130ffcd5ecab0f16f595" Dec 03 14:22:52 crc kubenswrapper[4805]: E1203 14:22:52.112183 4805 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"eedef859d870ca72638cbacc07d0f9c4ebffb350ef0d130ffcd5ecab0f16f595\": container with ID starting with eedef859d870ca72638cbacc07d0f9c4ebffb350ef0d130ffcd5ecab0f16f595 not found: ID does not exist" containerID="eedef859d870ca72638cbacc07d0f9c4ebffb350ef0d130ffcd5ecab0f16f595" Dec 03 14:22:52 crc kubenswrapper[4805]: I1203 14:22:52.112224 4805 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"eedef859d870ca72638cbacc07d0f9c4ebffb350ef0d130ffcd5ecab0f16f595"} err="failed to get container status \"eedef859d870ca72638cbacc07d0f9c4ebffb350ef0d130ffcd5ecab0f16f595\": rpc error: code = NotFound desc = could not find container \"eedef859d870ca72638cbacc07d0f9c4ebffb350ef0d130ffcd5ecab0f16f595\": container with ID starting with eedef859d870ca72638cbacc07d0f9c4ebffb350ef0d130ffcd5ecab0f16f595 not found: ID does not exist" Dec 03 14:22:52 crc kubenswrapper[4805]: I1203 14:22:52.113136 4805 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xvm9d\" (UniqueName: \"kubernetes.io/projected/8d213a92-95be-4101-9317-980ce72cf5d1-kube-api-access-xvm9d\") on node \"crc\" DevicePath \"\"" Dec 03 14:22:52 crc kubenswrapper[4805]: I1203 14:22:52.113201 4805 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8d213a92-95be-4101-9317-980ce72cf5d1-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 03 14:22:52 crc kubenswrapper[4805]: I1203 14:22:52.113219 4805 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8d213a92-95be-4101-9317-980ce72cf5d1-utilities\") on node \"crc\" DevicePath \"\"" Dec 03 14:22:52 crc kubenswrapper[4805]: I1203 14:22:52.705724 4805 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8d213a92-95be-4101-9317-980ce72cf5d1" path="/var/lib/kubelet/pods/8d213a92-95be-4101-9317-980ce72cf5d1/volumes" Dec 03 14:23:13 crc kubenswrapper[4805]: I1203 14:23:13.917363 4805 patch_prober.go:28] interesting pod/machine-config-daemon-gskh4 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 03 14:23:13 crc kubenswrapper[4805]: I1203 14:23:13.917903 4805 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-gskh4" podUID="6098937f-e3f6-45e8-a647-4994a79cd711" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 03 14:23:14 crc kubenswrapper[4805]: I1203 14:23:14.537629 4805 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fsxzq6"] Dec 03 14:23:14 crc kubenswrapper[4805]: E1203 14:23:14.537926 4805 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8d213a92-95be-4101-9317-980ce72cf5d1" containerName="registry-server" Dec 03 14:23:14 crc kubenswrapper[4805]: I1203 14:23:14.537943 4805 state_mem.go:107] "Deleted CPUSet assignment" podUID="8d213a92-95be-4101-9317-980ce72cf5d1" containerName="registry-server" Dec 03 14:23:14 crc kubenswrapper[4805]: E1203 14:23:14.537955 4805 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8d213a92-95be-4101-9317-980ce72cf5d1" containerName="extract-utilities" Dec 03 14:23:14 crc kubenswrapper[4805]: I1203 14:23:14.537963 4805 state_mem.go:107] "Deleted CPUSet assignment" podUID="8d213a92-95be-4101-9317-980ce72cf5d1" containerName="extract-utilities" Dec 03 14:23:14 crc kubenswrapper[4805]: E1203 14:23:14.537980 4805 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8d213a92-95be-4101-9317-980ce72cf5d1" containerName="extract-content" Dec 03 14:23:14 crc kubenswrapper[4805]: I1203 14:23:14.537987 4805 state_mem.go:107] "Deleted CPUSet assignment" podUID="8d213a92-95be-4101-9317-980ce72cf5d1" containerName="extract-content" Dec 03 14:23:14 crc kubenswrapper[4805]: I1203 14:23:14.538089 4805 memory_manager.go:354] "RemoveStaleState removing state" podUID="8d213a92-95be-4101-9317-980ce72cf5d1" containerName="registry-server" Dec 03 14:23:14 crc kubenswrapper[4805]: I1203 14:23:14.538799 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fsxzq6" Dec 03 14:23:14 crc kubenswrapper[4805]: I1203 14:23:14.540719 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"default-dockercfg-vmwhc" Dec 03 14:23:14 crc kubenswrapper[4805]: I1203 14:23:14.547604 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fsxzq6"] Dec 03 14:23:14 crc kubenswrapper[4805]: I1203 14:23:14.632998 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/abc1f29a-e8a4-425d-bbbd-17c106a39c5c-bundle\") pod \"5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fsxzq6\" (UID: \"abc1f29a-e8a4-425d-bbbd-17c106a39c5c\") " pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fsxzq6" Dec 03 14:23:14 crc kubenswrapper[4805]: I1203 14:23:14.633053 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/abc1f29a-e8a4-425d-bbbd-17c106a39c5c-util\") pod \"5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fsxzq6\" (UID: \"abc1f29a-e8a4-425d-bbbd-17c106a39c5c\") " pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fsxzq6" Dec 03 14:23:14 crc kubenswrapper[4805]: I1203 14:23:14.633144 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qccbf\" (UniqueName: \"kubernetes.io/projected/abc1f29a-e8a4-425d-bbbd-17c106a39c5c-kube-api-access-qccbf\") pod \"5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fsxzq6\" (UID: \"abc1f29a-e8a4-425d-bbbd-17c106a39c5c\") " pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fsxzq6" Dec 03 14:23:14 crc kubenswrapper[4805]: I1203 14:23:14.734107 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/abc1f29a-e8a4-425d-bbbd-17c106a39c5c-bundle\") pod \"5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fsxzq6\" (UID: \"abc1f29a-e8a4-425d-bbbd-17c106a39c5c\") " pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fsxzq6" Dec 03 14:23:14 crc kubenswrapper[4805]: I1203 14:23:14.734155 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/abc1f29a-e8a4-425d-bbbd-17c106a39c5c-util\") pod \"5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fsxzq6\" (UID: \"abc1f29a-e8a4-425d-bbbd-17c106a39c5c\") " pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fsxzq6" Dec 03 14:23:14 crc kubenswrapper[4805]: I1203 14:23:14.734189 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qccbf\" (UniqueName: \"kubernetes.io/projected/abc1f29a-e8a4-425d-bbbd-17c106a39c5c-kube-api-access-qccbf\") pod \"5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fsxzq6\" (UID: \"abc1f29a-e8a4-425d-bbbd-17c106a39c5c\") " pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fsxzq6" Dec 03 14:23:14 crc kubenswrapper[4805]: I1203 14:23:14.734695 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/abc1f29a-e8a4-425d-bbbd-17c106a39c5c-bundle\") pod \"5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fsxzq6\" (UID: \"abc1f29a-e8a4-425d-bbbd-17c106a39c5c\") " pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fsxzq6" Dec 03 14:23:14 crc kubenswrapper[4805]: I1203 14:23:14.734817 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/abc1f29a-e8a4-425d-bbbd-17c106a39c5c-util\") pod \"5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fsxzq6\" (UID: \"abc1f29a-e8a4-425d-bbbd-17c106a39c5c\") " pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fsxzq6" Dec 03 14:23:14 crc kubenswrapper[4805]: I1203 14:23:14.756540 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qccbf\" (UniqueName: \"kubernetes.io/projected/abc1f29a-e8a4-425d-bbbd-17c106a39c5c-kube-api-access-qccbf\") pod \"5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fsxzq6\" (UID: \"abc1f29a-e8a4-425d-bbbd-17c106a39c5c\") " pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fsxzq6" Dec 03 14:23:14 crc kubenswrapper[4805]: I1203 14:23:14.857177 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fsxzq6" Dec 03 14:23:15 crc kubenswrapper[4805]: I1203 14:23:15.326277 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fsxzq6"] Dec 03 14:23:15 crc kubenswrapper[4805]: W1203 14:23:15.333668 4805 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podabc1f29a_e8a4_425d_bbbd_17c106a39c5c.slice/crio-0a495b4dd3fdcacc3727879153f6b95ab021e4b48e50208a006bd4e1a6beba1c WatchSource:0}: Error finding container 0a495b4dd3fdcacc3727879153f6b95ab021e4b48e50208a006bd4e1a6beba1c: Status 404 returned error can't find the container with id 0a495b4dd3fdcacc3727879153f6b95ab021e4b48e50208a006bd4e1a6beba1c Dec 03 14:23:16 crc kubenswrapper[4805]: I1203 14:23:16.188344 4805 generic.go:334] "Generic (PLEG): container finished" podID="abc1f29a-e8a4-425d-bbbd-17c106a39c5c" containerID="879b173e442461cb945eba12b29973cd52649a892f80bdd43cac9d2f5d9629e6" exitCode=0 Dec 03 14:23:16 crc kubenswrapper[4805]: I1203 14:23:16.188671 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fsxzq6" event={"ID":"abc1f29a-e8a4-425d-bbbd-17c106a39c5c","Type":"ContainerDied","Data":"879b173e442461cb945eba12b29973cd52649a892f80bdd43cac9d2f5d9629e6"} Dec 03 14:23:16 crc kubenswrapper[4805]: I1203 14:23:16.188703 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fsxzq6" event={"ID":"abc1f29a-e8a4-425d-bbbd-17c106a39c5c","Type":"ContainerStarted","Data":"0a495b4dd3fdcacc3727879153f6b95ab021e4b48e50208a006bd4e1a6beba1c"} Dec 03 14:23:16 crc kubenswrapper[4805]: I1203 14:23:16.264880 4805 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-vljmk"] Dec 03 14:23:16 crc kubenswrapper[4805]: I1203 14:23:16.280524 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-vljmk" Dec 03 14:23:16 crc kubenswrapper[4805]: I1203 14:23:16.293304 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-vljmk"] Dec 03 14:23:16 crc kubenswrapper[4805]: I1203 14:23:16.456965 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9n6l8\" (UniqueName: \"kubernetes.io/projected/9cd4ed75-301f-42cf-b5d6-629eb4497892-kube-api-access-9n6l8\") pod \"redhat-operators-vljmk\" (UID: \"9cd4ed75-301f-42cf-b5d6-629eb4497892\") " pod="openshift-marketplace/redhat-operators-vljmk" Dec 03 14:23:16 crc kubenswrapper[4805]: I1203 14:23:16.457069 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9cd4ed75-301f-42cf-b5d6-629eb4497892-catalog-content\") pod \"redhat-operators-vljmk\" (UID: \"9cd4ed75-301f-42cf-b5d6-629eb4497892\") " pod="openshift-marketplace/redhat-operators-vljmk" Dec 03 14:23:16 crc kubenswrapper[4805]: I1203 14:23:16.457299 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9cd4ed75-301f-42cf-b5d6-629eb4497892-utilities\") pod \"redhat-operators-vljmk\" (UID: \"9cd4ed75-301f-42cf-b5d6-629eb4497892\") " pod="openshift-marketplace/redhat-operators-vljmk" Dec 03 14:23:16 crc kubenswrapper[4805]: I1203 14:23:16.558133 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9cd4ed75-301f-42cf-b5d6-629eb4497892-catalog-content\") pod \"redhat-operators-vljmk\" (UID: \"9cd4ed75-301f-42cf-b5d6-629eb4497892\") " pod="openshift-marketplace/redhat-operators-vljmk" Dec 03 14:23:16 crc kubenswrapper[4805]: I1203 14:23:16.558196 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9cd4ed75-301f-42cf-b5d6-629eb4497892-utilities\") pod \"redhat-operators-vljmk\" (UID: \"9cd4ed75-301f-42cf-b5d6-629eb4497892\") " pod="openshift-marketplace/redhat-operators-vljmk" Dec 03 14:23:16 crc kubenswrapper[4805]: I1203 14:23:16.558232 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9n6l8\" (UniqueName: \"kubernetes.io/projected/9cd4ed75-301f-42cf-b5d6-629eb4497892-kube-api-access-9n6l8\") pod \"redhat-operators-vljmk\" (UID: \"9cd4ed75-301f-42cf-b5d6-629eb4497892\") " pod="openshift-marketplace/redhat-operators-vljmk" Dec 03 14:23:16 crc kubenswrapper[4805]: I1203 14:23:16.558717 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9cd4ed75-301f-42cf-b5d6-629eb4497892-catalog-content\") pod \"redhat-operators-vljmk\" (UID: \"9cd4ed75-301f-42cf-b5d6-629eb4497892\") " pod="openshift-marketplace/redhat-operators-vljmk" Dec 03 14:23:16 crc kubenswrapper[4805]: I1203 14:23:16.558804 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9cd4ed75-301f-42cf-b5d6-629eb4497892-utilities\") pod \"redhat-operators-vljmk\" (UID: \"9cd4ed75-301f-42cf-b5d6-629eb4497892\") " pod="openshift-marketplace/redhat-operators-vljmk" Dec 03 14:23:16 crc kubenswrapper[4805]: I1203 14:23:16.583262 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9n6l8\" (UniqueName: \"kubernetes.io/projected/9cd4ed75-301f-42cf-b5d6-629eb4497892-kube-api-access-9n6l8\") pod \"redhat-operators-vljmk\" (UID: \"9cd4ed75-301f-42cf-b5d6-629eb4497892\") " pod="openshift-marketplace/redhat-operators-vljmk" Dec 03 14:23:16 crc kubenswrapper[4805]: I1203 14:23:16.671682 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-vljmk" Dec 03 14:23:16 crc kubenswrapper[4805]: I1203 14:23:16.873503 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-vljmk"] Dec 03 14:23:17 crc kubenswrapper[4805]: I1203 14:23:17.196517 4805 generic.go:334] "Generic (PLEG): container finished" podID="9cd4ed75-301f-42cf-b5d6-629eb4497892" containerID="f144eafd040f138842332f6935fa8b4bdea4da55167921b80b5ba4b9a871523f" exitCode=0 Dec 03 14:23:17 crc kubenswrapper[4805]: I1203 14:23:17.196558 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-vljmk" event={"ID":"9cd4ed75-301f-42cf-b5d6-629eb4497892","Type":"ContainerDied","Data":"f144eafd040f138842332f6935fa8b4bdea4da55167921b80b5ba4b9a871523f"} Dec 03 14:23:17 crc kubenswrapper[4805]: I1203 14:23:17.196583 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-vljmk" event={"ID":"9cd4ed75-301f-42cf-b5d6-629eb4497892","Type":"ContainerStarted","Data":"952e31494068c03ae79558e9cd865112c76c3379f6271034df07ea6e41d76bd9"} Dec 03 14:23:18 crc kubenswrapper[4805]: I1203 14:23:18.205702 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-vljmk" event={"ID":"9cd4ed75-301f-42cf-b5d6-629eb4497892","Type":"ContainerStarted","Data":"42ca20e4846cffb850e29e9fc238b2928340fd3a8e89b827acf0b046a33c0312"} Dec 03 14:23:19 crc kubenswrapper[4805]: I1203 14:23:19.218062 4805 generic.go:334] "Generic (PLEG): container finished" podID="9cd4ed75-301f-42cf-b5d6-629eb4497892" containerID="42ca20e4846cffb850e29e9fc238b2928340fd3a8e89b827acf0b046a33c0312" exitCode=0 Dec 03 14:23:19 crc kubenswrapper[4805]: I1203 14:23:19.218180 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-vljmk" event={"ID":"9cd4ed75-301f-42cf-b5d6-629eb4497892","Type":"ContainerDied","Data":"42ca20e4846cffb850e29e9fc238b2928340fd3a8e89b827acf0b046a33c0312"} Dec 03 14:23:19 crc kubenswrapper[4805]: I1203 14:23:19.221400 4805 generic.go:334] "Generic (PLEG): container finished" podID="abc1f29a-e8a4-425d-bbbd-17c106a39c5c" containerID="ed69a1b66fefae5517bd5d19a8ff97e98866b224044e19034a986844fc42d8ca" exitCode=0 Dec 03 14:23:19 crc kubenswrapper[4805]: I1203 14:23:19.221461 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fsxzq6" event={"ID":"abc1f29a-e8a4-425d-bbbd-17c106a39c5c","Type":"ContainerDied","Data":"ed69a1b66fefae5517bd5d19a8ff97e98866b224044e19034a986844fc42d8ca"} Dec 03 14:23:20 crc kubenswrapper[4805]: I1203 14:23:20.239529 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-vljmk" event={"ID":"9cd4ed75-301f-42cf-b5d6-629eb4497892","Type":"ContainerStarted","Data":"811878ae45c02073bacce56ca70f5afd123f00942b7f196877994ecede2a809a"} Dec 03 14:23:20 crc kubenswrapper[4805]: I1203 14:23:20.245743 4805 generic.go:334] "Generic (PLEG): container finished" podID="abc1f29a-e8a4-425d-bbbd-17c106a39c5c" containerID="68daf551f55e3a9c558bb71e0fe7b8dbd2aed008b059e54ec79fc96e570af2bb" exitCode=0 Dec 03 14:23:20 crc kubenswrapper[4805]: I1203 14:23:20.245789 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fsxzq6" event={"ID":"abc1f29a-e8a4-425d-bbbd-17c106a39c5c","Type":"ContainerDied","Data":"68daf551f55e3a9c558bb71e0fe7b8dbd2aed008b059e54ec79fc96e570af2bb"} Dec 03 14:23:20 crc kubenswrapper[4805]: I1203 14:23:20.270389 4805 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-vljmk" podStartSLOduration=1.39113873 podStartE2EDuration="4.270373303s" podCreationTimestamp="2025-12-03 14:23:16 +0000 UTC" firstStartedPulling="2025-12-03 14:23:17.198485386 +0000 UTC m=+826.861402309" lastFinishedPulling="2025-12-03 14:23:20.077719959 +0000 UTC m=+829.740636882" observedRunningTime="2025-12-03 14:23:20.262695167 +0000 UTC m=+829.925612090" watchObservedRunningTime="2025-12-03 14:23:20.270373303 +0000 UTC m=+829.933290226" Dec 03 14:23:21 crc kubenswrapper[4805]: I1203 14:23:21.604470 4805 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fsxzq6" Dec 03 14:23:21 crc kubenswrapper[4805]: I1203 14:23:21.770989 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/abc1f29a-e8a4-425d-bbbd-17c106a39c5c-bundle\") pod \"abc1f29a-e8a4-425d-bbbd-17c106a39c5c\" (UID: \"abc1f29a-e8a4-425d-bbbd-17c106a39c5c\") " Dec 03 14:23:21 crc kubenswrapper[4805]: I1203 14:23:21.771267 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/abc1f29a-e8a4-425d-bbbd-17c106a39c5c-util\") pod \"abc1f29a-e8a4-425d-bbbd-17c106a39c5c\" (UID: \"abc1f29a-e8a4-425d-bbbd-17c106a39c5c\") " Dec 03 14:23:21 crc kubenswrapper[4805]: I1203 14:23:21.771374 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qccbf\" (UniqueName: \"kubernetes.io/projected/abc1f29a-e8a4-425d-bbbd-17c106a39c5c-kube-api-access-qccbf\") pod \"abc1f29a-e8a4-425d-bbbd-17c106a39c5c\" (UID: \"abc1f29a-e8a4-425d-bbbd-17c106a39c5c\") " Dec 03 14:23:21 crc kubenswrapper[4805]: I1203 14:23:21.771515 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/abc1f29a-e8a4-425d-bbbd-17c106a39c5c-bundle" (OuterVolumeSpecName: "bundle") pod "abc1f29a-e8a4-425d-bbbd-17c106a39c5c" (UID: "abc1f29a-e8a4-425d-bbbd-17c106a39c5c"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 14:23:21 crc kubenswrapper[4805]: I1203 14:23:21.772174 4805 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/abc1f29a-e8a4-425d-bbbd-17c106a39c5c-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 14:23:21 crc kubenswrapper[4805]: I1203 14:23:21.777671 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/abc1f29a-e8a4-425d-bbbd-17c106a39c5c-kube-api-access-qccbf" (OuterVolumeSpecName: "kube-api-access-qccbf") pod "abc1f29a-e8a4-425d-bbbd-17c106a39c5c" (UID: "abc1f29a-e8a4-425d-bbbd-17c106a39c5c"). InnerVolumeSpecName "kube-api-access-qccbf". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 14:23:21 crc kubenswrapper[4805]: I1203 14:23:21.782093 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/abc1f29a-e8a4-425d-bbbd-17c106a39c5c-util" (OuterVolumeSpecName: "util") pod "abc1f29a-e8a4-425d-bbbd-17c106a39c5c" (UID: "abc1f29a-e8a4-425d-bbbd-17c106a39c5c"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 14:23:21 crc kubenswrapper[4805]: I1203 14:23:21.872974 4805 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/abc1f29a-e8a4-425d-bbbd-17c106a39c5c-util\") on node \"crc\" DevicePath \"\"" Dec 03 14:23:21 crc kubenswrapper[4805]: I1203 14:23:21.873020 4805 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qccbf\" (UniqueName: \"kubernetes.io/projected/abc1f29a-e8a4-425d-bbbd-17c106a39c5c-kube-api-access-qccbf\") on node \"crc\" DevicePath \"\"" Dec 03 14:23:22 crc kubenswrapper[4805]: I1203 14:23:22.260660 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fsxzq6" event={"ID":"abc1f29a-e8a4-425d-bbbd-17c106a39c5c","Type":"ContainerDied","Data":"0a495b4dd3fdcacc3727879153f6b95ab021e4b48e50208a006bd4e1a6beba1c"} Dec 03 14:23:22 crc kubenswrapper[4805]: I1203 14:23:22.260698 4805 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="0a495b4dd3fdcacc3727879153f6b95ab021e4b48e50208a006bd4e1a6beba1c" Dec 03 14:23:22 crc kubenswrapper[4805]: I1203 14:23:22.261054 4805 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fsxzq6" Dec 03 14:23:25 crc kubenswrapper[4805]: I1203 14:23:25.522080 4805 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-operator-5b5b58f5c8-tstgp"] Dec 03 14:23:25 crc kubenswrapper[4805]: E1203 14:23:25.522768 4805 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="abc1f29a-e8a4-425d-bbbd-17c106a39c5c" containerName="util" Dec 03 14:23:25 crc kubenswrapper[4805]: I1203 14:23:25.522784 4805 state_mem.go:107] "Deleted CPUSet assignment" podUID="abc1f29a-e8a4-425d-bbbd-17c106a39c5c" containerName="util" Dec 03 14:23:25 crc kubenswrapper[4805]: E1203 14:23:25.522794 4805 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="abc1f29a-e8a4-425d-bbbd-17c106a39c5c" containerName="extract" Dec 03 14:23:25 crc kubenswrapper[4805]: I1203 14:23:25.522800 4805 state_mem.go:107] "Deleted CPUSet assignment" podUID="abc1f29a-e8a4-425d-bbbd-17c106a39c5c" containerName="extract" Dec 03 14:23:25 crc kubenswrapper[4805]: E1203 14:23:25.522811 4805 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="abc1f29a-e8a4-425d-bbbd-17c106a39c5c" containerName="pull" Dec 03 14:23:25 crc kubenswrapper[4805]: I1203 14:23:25.522820 4805 state_mem.go:107] "Deleted CPUSet assignment" podUID="abc1f29a-e8a4-425d-bbbd-17c106a39c5c" containerName="pull" Dec 03 14:23:25 crc kubenswrapper[4805]: I1203 14:23:25.522939 4805 memory_manager.go:354] "RemoveStaleState removing state" podUID="abc1f29a-e8a4-425d-bbbd-17c106a39c5c" containerName="extract" Dec 03 14:23:25 crc kubenswrapper[4805]: I1203 14:23:25.523438 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-operator-5b5b58f5c8-tstgp" Dec 03 14:23:25 crc kubenswrapper[4805]: I1203 14:23:25.526652 4805 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-nmstate"/"kube-root-ca.crt" Dec 03 14:23:25 crc kubenswrapper[4805]: I1203 14:23:25.526670 4805 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-nmstate"/"openshift-service-ca.crt" Dec 03 14:23:25 crc kubenswrapper[4805]: I1203 14:23:25.527072 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"nmstate-operator-dockercfg-jb5db" Dec 03 14:23:25 crc kubenswrapper[4805]: I1203 14:23:25.540758 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-operator-5b5b58f5c8-tstgp"] Dec 03 14:23:25 crc kubenswrapper[4805]: I1203 14:23:25.622476 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rz9bb\" (UniqueName: \"kubernetes.io/projected/7f35b1a1-05c4-453a-9101-d43cf4fa3739-kube-api-access-rz9bb\") pod \"nmstate-operator-5b5b58f5c8-tstgp\" (UID: \"7f35b1a1-05c4-453a-9101-d43cf4fa3739\") " pod="openshift-nmstate/nmstate-operator-5b5b58f5c8-tstgp" Dec 03 14:23:25 crc kubenswrapper[4805]: I1203 14:23:25.724662 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rz9bb\" (UniqueName: \"kubernetes.io/projected/7f35b1a1-05c4-453a-9101-d43cf4fa3739-kube-api-access-rz9bb\") pod \"nmstate-operator-5b5b58f5c8-tstgp\" (UID: \"7f35b1a1-05c4-453a-9101-d43cf4fa3739\") " pod="openshift-nmstate/nmstate-operator-5b5b58f5c8-tstgp" Dec 03 14:23:25 crc kubenswrapper[4805]: I1203 14:23:25.760204 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rz9bb\" (UniqueName: \"kubernetes.io/projected/7f35b1a1-05c4-453a-9101-d43cf4fa3739-kube-api-access-rz9bb\") pod \"nmstate-operator-5b5b58f5c8-tstgp\" (UID: \"7f35b1a1-05c4-453a-9101-d43cf4fa3739\") " pod="openshift-nmstate/nmstate-operator-5b5b58f5c8-tstgp" Dec 03 14:23:25 crc kubenswrapper[4805]: I1203 14:23:25.839876 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-operator-5b5b58f5c8-tstgp" Dec 03 14:23:26 crc kubenswrapper[4805]: I1203 14:23:26.339310 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-operator-5b5b58f5c8-tstgp"] Dec 03 14:23:26 crc kubenswrapper[4805]: I1203 14:23:26.672023 4805 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-vljmk" Dec 03 14:23:26 crc kubenswrapper[4805]: I1203 14:23:26.672116 4805 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-vljmk" Dec 03 14:23:27 crc kubenswrapper[4805]: I1203 14:23:27.292077 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-operator-5b5b58f5c8-tstgp" event={"ID":"7f35b1a1-05c4-453a-9101-d43cf4fa3739","Type":"ContainerStarted","Data":"c4b55f581b29517717d1d0d24890bdc22d271ba83ad29ad019093c9115c42793"} Dec 03 14:23:27 crc kubenswrapper[4805]: I1203 14:23:27.720218 4805 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-vljmk" podUID="9cd4ed75-301f-42cf-b5d6-629eb4497892" containerName="registry-server" probeResult="failure" output=< Dec 03 14:23:27 crc kubenswrapper[4805]: timeout: failed to connect service ":50051" within 1s Dec 03 14:23:27 crc kubenswrapper[4805]: > Dec 03 14:23:29 crc kubenswrapper[4805]: I1203 14:23:29.307458 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-operator-5b5b58f5c8-tstgp" event={"ID":"7f35b1a1-05c4-453a-9101-d43cf4fa3739","Type":"ContainerStarted","Data":"46ffd900a43dd1a6909457b7a0044a725a5ecc43674a198e07c72d0c0b244282"} Dec 03 14:23:29 crc kubenswrapper[4805]: I1203 14:23:29.329880 4805 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-operator-5b5b58f5c8-tstgp" podStartSLOduration=2.23615191 podStartE2EDuration="4.329830287s" podCreationTimestamp="2025-12-03 14:23:25 +0000 UTC" firstStartedPulling="2025-12-03 14:23:26.34832206 +0000 UTC m=+836.011238993" lastFinishedPulling="2025-12-03 14:23:28.442000447 +0000 UTC m=+838.104917370" observedRunningTime="2025-12-03 14:23:29.327864654 +0000 UTC m=+838.990781587" watchObservedRunningTime="2025-12-03 14:23:29.329830287 +0000 UTC m=+838.992747220" Dec 03 14:23:35 crc kubenswrapper[4805]: I1203 14:23:35.064805 4805 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-metrics-7f946cbc9-6r8jm"] Dec 03 14:23:35 crc kubenswrapper[4805]: I1203 14:23:35.066176 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-metrics-7f946cbc9-6r8jm" Dec 03 14:23:35 crc kubenswrapper[4805]: I1203 14:23:35.072270 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"nmstate-handler-dockercfg-px68p" Dec 03 14:23:35 crc kubenswrapper[4805]: I1203 14:23:35.072641 4805 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-webhook-5f6d4c5ccb-2zszs"] Dec 03 14:23:35 crc kubenswrapper[4805]: I1203 14:23:35.073418 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-webhook-5f6d4c5ccb-2zszs" Dec 03 14:23:35 crc kubenswrapper[4805]: I1203 14:23:35.075794 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"openshift-nmstate-webhook" Dec 03 14:23:35 crc kubenswrapper[4805]: I1203 14:23:35.083264 4805 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-handler-xqhj5"] Dec 03 14:23:35 crc kubenswrapper[4805]: I1203 14:23:35.084122 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-handler-xqhj5" Dec 03 14:23:35 crc kubenswrapper[4805]: I1203 14:23:35.092178 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-metrics-7f946cbc9-6r8jm"] Dec 03 14:23:35 crc kubenswrapper[4805]: I1203 14:23:35.100530 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-webhook-5f6d4c5ccb-2zszs"] Dec 03 14:23:35 crc kubenswrapper[4805]: I1203 14:23:35.207872 4805 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-console-plugin-7fbb5f6569-mbhzq"] Dec 03 14:23:35 crc kubenswrapper[4805]: I1203 14:23:35.208506 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-mbhzq" Dec 03 14:23:35 crc kubenswrapper[4805]: I1203 14:23:35.211883 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"default-dockercfg-n7ndj" Dec 03 14:23:35 crc kubenswrapper[4805]: I1203 14:23:35.212110 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"plugin-serving-cert" Dec 03 14:23:35 crc kubenswrapper[4805]: I1203 14:23:35.212279 4805 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-nmstate"/"nginx-conf" Dec 03 14:23:35 crc kubenswrapper[4805]: I1203 14:23:35.216160 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-console-plugin-7fbb5f6569-mbhzq"] Dec 03 14:23:35 crc kubenswrapper[4805]: I1203 14:23:35.250147 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tls-key-pair\" (UniqueName: \"kubernetes.io/secret/9654448a-2df5-4cfc-b6e5-401ad15bc52e-tls-key-pair\") pod \"nmstate-webhook-5f6d4c5ccb-2zszs\" (UID: \"9654448a-2df5-4cfc-b6e5-401ad15bc52e\") " pod="openshift-nmstate/nmstate-webhook-5f6d4c5ccb-2zszs" Dec 03 14:23:35 crc kubenswrapper[4805]: I1203 14:23:35.250187 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nmstate-lock\" (UniqueName: \"kubernetes.io/host-path/4f623c1b-94c0-43cd-ae9f-56597f4542b4-nmstate-lock\") pod \"nmstate-handler-xqhj5\" (UID: \"4f623c1b-94c0-43cd-ae9f-56597f4542b4\") " pod="openshift-nmstate/nmstate-handler-xqhj5" Dec 03 14:23:35 crc kubenswrapper[4805]: I1203 14:23:35.250209 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovs-socket\" (UniqueName: \"kubernetes.io/host-path/4f623c1b-94c0-43cd-ae9f-56597f4542b4-ovs-socket\") pod \"nmstate-handler-xqhj5\" (UID: \"4f623c1b-94c0-43cd-ae9f-56597f4542b4\") " pod="openshift-nmstate/nmstate-handler-xqhj5" Dec 03 14:23:35 crc kubenswrapper[4805]: I1203 14:23:35.250228 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hmm6x\" (UniqueName: \"kubernetes.io/projected/9654448a-2df5-4cfc-b6e5-401ad15bc52e-kube-api-access-hmm6x\") pod \"nmstate-webhook-5f6d4c5ccb-2zszs\" (UID: \"9654448a-2df5-4cfc-b6e5-401ad15bc52e\") " pod="openshift-nmstate/nmstate-webhook-5f6d4c5ccb-2zszs" Dec 03 14:23:35 crc kubenswrapper[4805]: I1203 14:23:35.250250 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4hjz7\" (UniqueName: \"kubernetes.io/projected/4f623c1b-94c0-43cd-ae9f-56597f4542b4-kube-api-access-4hjz7\") pod \"nmstate-handler-xqhj5\" (UID: \"4f623c1b-94c0-43cd-ae9f-56597f4542b4\") " pod="openshift-nmstate/nmstate-handler-xqhj5" Dec 03 14:23:35 crc kubenswrapper[4805]: I1203 14:23:35.250274 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zckbt\" (UniqueName: \"kubernetes.io/projected/9d39f3b1-337a-4aeb-b1cf-9ca9290f7791-kube-api-access-zckbt\") pod \"nmstate-metrics-7f946cbc9-6r8jm\" (UID: \"9d39f3b1-337a-4aeb-b1cf-9ca9290f7791\") " pod="openshift-nmstate/nmstate-metrics-7f946cbc9-6r8jm" Dec 03 14:23:35 crc kubenswrapper[4805]: I1203 14:23:35.250305 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dbus-socket\" (UniqueName: \"kubernetes.io/host-path/4f623c1b-94c0-43cd-ae9f-56597f4542b4-dbus-socket\") pod \"nmstate-handler-xqhj5\" (UID: \"4f623c1b-94c0-43cd-ae9f-56597f4542b4\") " pod="openshift-nmstate/nmstate-handler-xqhj5" Dec 03 14:23:35 crc kubenswrapper[4805]: I1203 14:23:35.351007 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dbus-socket\" (UniqueName: \"kubernetes.io/host-path/4f623c1b-94c0-43cd-ae9f-56597f4542b4-dbus-socket\") pod \"nmstate-handler-xqhj5\" (UID: \"4f623c1b-94c0-43cd-ae9f-56597f4542b4\") " pod="openshift-nmstate/nmstate-handler-xqhj5" Dec 03 14:23:35 crc kubenswrapper[4805]: I1203 14:23:35.351061 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-r8vs6\" (UniqueName: \"kubernetes.io/projected/5d4ab6ea-87ca-46dc-a9ad-f315faf0a9e2-kube-api-access-r8vs6\") pod \"nmstate-console-plugin-7fbb5f6569-mbhzq\" (UID: \"5d4ab6ea-87ca-46dc-a9ad-f315faf0a9e2\") " pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-mbhzq" Dec 03 14:23:35 crc kubenswrapper[4805]: I1203 14:23:35.351125 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugin-serving-cert\" (UniqueName: \"kubernetes.io/secret/5d4ab6ea-87ca-46dc-a9ad-f315faf0a9e2-plugin-serving-cert\") pod \"nmstate-console-plugin-7fbb5f6569-mbhzq\" (UID: \"5d4ab6ea-87ca-46dc-a9ad-f315faf0a9e2\") " pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-mbhzq" Dec 03 14:23:35 crc kubenswrapper[4805]: I1203 14:23:35.351152 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tls-key-pair\" (UniqueName: \"kubernetes.io/secret/9654448a-2df5-4cfc-b6e5-401ad15bc52e-tls-key-pair\") pod \"nmstate-webhook-5f6d4c5ccb-2zszs\" (UID: \"9654448a-2df5-4cfc-b6e5-401ad15bc52e\") " pod="openshift-nmstate/nmstate-webhook-5f6d4c5ccb-2zszs" Dec 03 14:23:35 crc kubenswrapper[4805]: I1203 14:23:35.351173 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nmstate-lock\" (UniqueName: \"kubernetes.io/host-path/4f623c1b-94c0-43cd-ae9f-56597f4542b4-nmstate-lock\") pod \"nmstate-handler-xqhj5\" (UID: \"4f623c1b-94c0-43cd-ae9f-56597f4542b4\") " pod="openshift-nmstate/nmstate-handler-xqhj5" Dec 03 14:23:35 crc kubenswrapper[4805]: I1203 14:23:35.351193 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovs-socket\" (UniqueName: \"kubernetes.io/host-path/4f623c1b-94c0-43cd-ae9f-56597f4542b4-ovs-socket\") pod \"nmstate-handler-xqhj5\" (UID: \"4f623c1b-94c0-43cd-ae9f-56597f4542b4\") " pod="openshift-nmstate/nmstate-handler-xqhj5" Dec 03 14:23:35 crc kubenswrapper[4805]: I1203 14:23:35.351215 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hmm6x\" (UniqueName: \"kubernetes.io/projected/9654448a-2df5-4cfc-b6e5-401ad15bc52e-kube-api-access-hmm6x\") pod \"nmstate-webhook-5f6d4c5ccb-2zszs\" (UID: \"9654448a-2df5-4cfc-b6e5-401ad15bc52e\") " pod="openshift-nmstate/nmstate-webhook-5f6d4c5ccb-2zszs" Dec 03 14:23:35 crc kubenswrapper[4805]: I1203 14:23:35.351236 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4hjz7\" (UniqueName: \"kubernetes.io/projected/4f623c1b-94c0-43cd-ae9f-56597f4542b4-kube-api-access-4hjz7\") pod \"nmstate-handler-xqhj5\" (UID: \"4f623c1b-94c0-43cd-ae9f-56597f4542b4\") " pod="openshift-nmstate/nmstate-handler-xqhj5" Dec 03 14:23:35 crc kubenswrapper[4805]: I1203 14:23:35.351262 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5d4ab6ea-87ca-46dc-a9ad-f315faf0a9e2-nginx-conf\") pod \"nmstate-console-plugin-7fbb5f6569-mbhzq\" (UID: \"5d4ab6ea-87ca-46dc-a9ad-f315faf0a9e2\") " pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-mbhzq" Dec 03 14:23:35 crc kubenswrapper[4805]: I1203 14:23:35.351290 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zckbt\" (UniqueName: \"kubernetes.io/projected/9d39f3b1-337a-4aeb-b1cf-9ca9290f7791-kube-api-access-zckbt\") pod \"nmstate-metrics-7f946cbc9-6r8jm\" (UID: \"9d39f3b1-337a-4aeb-b1cf-9ca9290f7791\") " pod="openshift-nmstate/nmstate-metrics-7f946cbc9-6r8jm" Dec 03 14:23:35 crc kubenswrapper[4805]: I1203 14:23:35.351890 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dbus-socket\" (UniqueName: \"kubernetes.io/host-path/4f623c1b-94c0-43cd-ae9f-56597f4542b4-dbus-socket\") pod \"nmstate-handler-xqhj5\" (UID: \"4f623c1b-94c0-43cd-ae9f-56597f4542b4\") " pod="openshift-nmstate/nmstate-handler-xqhj5" Dec 03 14:23:35 crc kubenswrapper[4805]: I1203 14:23:35.352102 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nmstate-lock\" (UniqueName: \"kubernetes.io/host-path/4f623c1b-94c0-43cd-ae9f-56597f4542b4-nmstate-lock\") pod \"nmstate-handler-xqhj5\" (UID: \"4f623c1b-94c0-43cd-ae9f-56597f4542b4\") " pod="openshift-nmstate/nmstate-handler-xqhj5" Dec 03 14:23:35 crc kubenswrapper[4805]: I1203 14:23:35.352140 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovs-socket\" (UniqueName: \"kubernetes.io/host-path/4f623c1b-94c0-43cd-ae9f-56597f4542b4-ovs-socket\") pod \"nmstate-handler-xqhj5\" (UID: \"4f623c1b-94c0-43cd-ae9f-56597f4542b4\") " pod="openshift-nmstate/nmstate-handler-xqhj5" Dec 03 14:23:35 crc kubenswrapper[4805]: I1203 14:23:35.360427 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tls-key-pair\" (UniqueName: \"kubernetes.io/secret/9654448a-2df5-4cfc-b6e5-401ad15bc52e-tls-key-pair\") pod \"nmstate-webhook-5f6d4c5ccb-2zszs\" (UID: \"9654448a-2df5-4cfc-b6e5-401ad15bc52e\") " pod="openshift-nmstate/nmstate-webhook-5f6d4c5ccb-2zszs" Dec 03 14:23:35 crc kubenswrapper[4805]: I1203 14:23:35.367982 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zckbt\" (UniqueName: \"kubernetes.io/projected/9d39f3b1-337a-4aeb-b1cf-9ca9290f7791-kube-api-access-zckbt\") pod \"nmstate-metrics-7f946cbc9-6r8jm\" (UID: \"9d39f3b1-337a-4aeb-b1cf-9ca9290f7791\") " pod="openshift-nmstate/nmstate-metrics-7f946cbc9-6r8jm" Dec 03 14:23:35 crc kubenswrapper[4805]: I1203 14:23:35.376330 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hmm6x\" (UniqueName: \"kubernetes.io/projected/9654448a-2df5-4cfc-b6e5-401ad15bc52e-kube-api-access-hmm6x\") pod \"nmstate-webhook-5f6d4c5ccb-2zszs\" (UID: \"9654448a-2df5-4cfc-b6e5-401ad15bc52e\") " pod="openshift-nmstate/nmstate-webhook-5f6d4c5ccb-2zszs" Dec 03 14:23:35 crc kubenswrapper[4805]: I1203 14:23:35.381495 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4hjz7\" (UniqueName: \"kubernetes.io/projected/4f623c1b-94c0-43cd-ae9f-56597f4542b4-kube-api-access-4hjz7\") pod \"nmstate-handler-xqhj5\" (UID: \"4f623c1b-94c0-43cd-ae9f-56597f4542b4\") " pod="openshift-nmstate/nmstate-handler-xqhj5" Dec 03 14:23:35 crc kubenswrapper[4805]: I1203 14:23:35.388648 4805 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console/console-68845dcc8-7ksxw"] Dec 03 14:23:35 crc kubenswrapper[4805]: I1203 14:23:35.389484 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-68845dcc8-7ksxw" Dec 03 14:23:35 crc kubenswrapper[4805]: I1203 14:23:35.392675 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-metrics-7f946cbc9-6r8jm" Dec 03 14:23:35 crc kubenswrapper[4805]: I1203 14:23:35.400345 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-68845dcc8-7ksxw"] Dec 03 14:23:35 crc kubenswrapper[4805]: I1203 14:23:35.413422 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-webhook-5f6d4c5ccb-2zszs" Dec 03 14:23:35 crc kubenswrapper[4805]: I1203 14:23:35.416063 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-handler-xqhj5" Dec 03 14:23:35 crc kubenswrapper[4805]: I1203 14:23:35.452191 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugin-serving-cert\" (UniqueName: \"kubernetes.io/secret/5d4ab6ea-87ca-46dc-a9ad-f315faf0a9e2-plugin-serving-cert\") pod \"nmstate-console-plugin-7fbb5f6569-mbhzq\" (UID: \"5d4ab6ea-87ca-46dc-a9ad-f315faf0a9e2\") " pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-mbhzq" Dec 03 14:23:35 crc kubenswrapper[4805]: I1203 14:23:35.452575 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5d4ab6ea-87ca-46dc-a9ad-f315faf0a9e2-nginx-conf\") pod \"nmstate-console-plugin-7fbb5f6569-mbhzq\" (UID: \"5d4ab6ea-87ca-46dc-a9ad-f315faf0a9e2\") " pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-mbhzq" Dec 03 14:23:35 crc kubenswrapper[4805]: I1203 14:23:35.453334 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-r8vs6\" (UniqueName: \"kubernetes.io/projected/5d4ab6ea-87ca-46dc-a9ad-f315faf0a9e2-kube-api-access-r8vs6\") pod \"nmstate-console-plugin-7fbb5f6569-mbhzq\" (UID: \"5d4ab6ea-87ca-46dc-a9ad-f315faf0a9e2\") " pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-mbhzq" Dec 03 14:23:35 crc kubenswrapper[4805]: I1203 14:23:35.453280 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5d4ab6ea-87ca-46dc-a9ad-f315faf0a9e2-nginx-conf\") pod \"nmstate-console-plugin-7fbb5f6569-mbhzq\" (UID: \"5d4ab6ea-87ca-46dc-a9ad-f315faf0a9e2\") " pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-mbhzq" Dec 03 14:23:35 crc kubenswrapper[4805]: E1203 14:23:35.452366 4805 secret.go:188] Couldn't get secret openshift-nmstate/plugin-serving-cert: secret "plugin-serving-cert" not found Dec 03 14:23:35 crc kubenswrapper[4805]: E1203 14:23:35.453440 4805 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5d4ab6ea-87ca-46dc-a9ad-f315faf0a9e2-plugin-serving-cert podName:5d4ab6ea-87ca-46dc-a9ad-f315faf0a9e2 nodeName:}" failed. No retries permitted until 2025-12-03 14:23:35.95342519 +0000 UTC m=+845.616342103 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "plugin-serving-cert" (UniqueName: "kubernetes.io/secret/5d4ab6ea-87ca-46dc-a9ad-f315faf0a9e2-plugin-serving-cert") pod "nmstate-console-plugin-7fbb5f6569-mbhzq" (UID: "5d4ab6ea-87ca-46dc-a9ad-f315faf0a9e2") : secret "plugin-serving-cert" not found Dec 03 14:23:35 crc kubenswrapper[4805]: I1203 14:23:35.469485 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-r8vs6\" (UniqueName: \"kubernetes.io/projected/5d4ab6ea-87ca-46dc-a9ad-f315faf0a9e2-kube-api-access-r8vs6\") pod \"nmstate-console-plugin-7fbb5f6569-mbhzq\" (UID: \"5d4ab6ea-87ca-46dc-a9ad-f315faf0a9e2\") " pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-mbhzq" Dec 03 14:23:35 crc kubenswrapper[4805]: I1203 14:23:35.554636 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/12f1f398-bfbc-4e05-bfc6-c654d7b9e97b-oauth-serving-cert\") pod \"console-68845dcc8-7ksxw\" (UID: \"12f1f398-bfbc-4e05-bfc6-c654d7b9e97b\") " pod="openshift-console/console-68845dcc8-7ksxw" Dec 03 14:23:35 crc kubenswrapper[4805]: I1203 14:23:35.554686 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/12f1f398-bfbc-4e05-bfc6-c654d7b9e97b-trusted-ca-bundle\") pod \"console-68845dcc8-7ksxw\" (UID: \"12f1f398-bfbc-4e05-bfc6-c654d7b9e97b\") " pod="openshift-console/console-68845dcc8-7ksxw" Dec 03 14:23:35 crc kubenswrapper[4805]: I1203 14:23:35.555814 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/12f1f398-bfbc-4e05-bfc6-c654d7b9e97b-service-ca\") pod \"console-68845dcc8-7ksxw\" (UID: \"12f1f398-bfbc-4e05-bfc6-c654d7b9e97b\") " pod="openshift-console/console-68845dcc8-7ksxw" Dec 03 14:23:35 crc kubenswrapper[4805]: I1203 14:23:35.557395 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/12f1f398-bfbc-4e05-bfc6-c654d7b9e97b-console-serving-cert\") pod \"console-68845dcc8-7ksxw\" (UID: \"12f1f398-bfbc-4e05-bfc6-c654d7b9e97b\") " pod="openshift-console/console-68845dcc8-7ksxw" Dec 03 14:23:35 crc kubenswrapper[4805]: I1203 14:23:35.557457 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cp6bk\" (UniqueName: \"kubernetes.io/projected/12f1f398-bfbc-4e05-bfc6-c654d7b9e97b-kube-api-access-cp6bk\") pod \"console-68845dcc8-7ksxw\" (UID: \"12f1f398-bfbc-4e05-bfc6-c654d7b9e97b\") " pod="openshift-console/console-68845dcc8-7ksxw" Dec 03 14:23:35 crc kubenswrapper[4805]: I1203 14:23:35.557497 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/12f1f398-bfbc-4e05-bfc6-c654d7b9e97b-console-oauth-config\") pod \"console-68845dcc8-7ksxw\" (UID: \"12f1f398-bfbc-4e05-bfc6-c654d7b9e97b\") " pod="openshift-console/console-68845dcc8-7ksxw" Dec 03 14:23:35 crc kubenswrapper[4805]: I1203 14:23:35.557526 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/12f1f398-bfbc-4e05-bfc6-c654d7b9e97b-console-config\") pod \"console-68845dcc8-7ksxw\" (UID: \"12f1f398-bfbc-4e05-bfc6-c654d7b9e97b\") " pod="openshift-console/console-68845dcc8-7ksxw" Dec 03 14:23:35 crc kubenswrapper[4805]: I1203 14:23:35.607186 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-metrics-7f946cbc9-6r8jm"] Dec 03 14:23:35 crc kubenswrapper[4805]: I1203 14:23:35.651348 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-webhook-5f6d4c5ccb-2zszs"] Dec 03 14:23:35 crc kubenswrapper[4805]: W1203 14:23:35.655084 4805 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod9654448a_2df5_4cfc_b6e5_401ad15bc52e.slice/crio-5697c7db08639bbe3a1d3de3d19a6c7f6066bf96ecae16baefe5d340b4d90e32 WatchSource:0}: Error finding container 5697c7db08639bbe3a1d3de3d19a6c7f6066bf96ecae16baefe5d340b4d90e32: Status 404 returned error can't find the container with id 5697c7db08639bbe3a1d3de3d19a6c7f6066bf96ecae16baefe5d340b4d90e32 Dec 03 14:23:35 crc kubenswrapper[4805]: I1203 14:23:35.658154 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/12f1f398-bfbc-4e05-bfc6-c654d7b9e97b-trusted-ca-bundle\") pod \"console-68845dcc8-7ksxw\" (UID: \"12f1f398-bfbc-4e05-bfc6-c654d7b9e97b\") " pod="openshift-console/console-68845dcc8-7ksxw" Dec 03 14:23:35 crc kubenswrapper[4805]: I1203 14:23:35.658202 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/12f1f398-bfbc-4e05-bfc6-c654d7b9e97b-service-ca\") pod \"console-68845dcc8-7ksxw\" (UID: \"12f1f398-bfbc-4e05-bfc6-c654d7b9e97b\") " pod="openshift-console/console-68845dcc8-7ksxw" Dec 03 14:23:35 crc kubenswrapper[4805]: I1203 14:23:35.658245 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/12f1f398-bfbc-4e05-bfc6-c654d7b9e97b-console-serving-cert\") pod \"console-68845dcc8-7ksxw\" (UID: \"12f1f398-bfbc-4e05-bfc6-c654d7b9e97b\") " pod="openshift-console/console-68845dcc8-7ksxw" Dec 03 14:23:35 crc kubenswrapper[4805]: I1203 14:23:35.658268 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cp6bk\" (UniqueName: \"kubernetes.io/projected/12f1f398-bfbc-4e05-bfc6-c654d7b9e97b-kube-api-access-cp6bk\") pod \"console-68845dcc8-7ksxw\" (UID: \"12f1f398-bfbc-4e05-bfc6-c654d7b9e97b\") " pod="openshift-console/console-68845dcc8-7ksxw" Dec 03 14:23:35 crc kubenswrapper[4805]: I1203 14:23:35.658286 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/12f1f398-bfbc-4e05-bfc6-c654d7b9e97b-console-oauth-config\") pod \"console-68845dcc8-7ksxw\" (UID: \"12f1f398-bfbc-4e05-bfc6-c654d7b9e97b\") " pod="openshift-console/console-68845dcc8-7ksxw" Dec 03 14:23:35 crc kubenswrapper[4805]: I1203 14:23:35.658303 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/12f1f398-bfbc-4e05-bfc6-c654d7b9e97b-console-config\") pod \"console-68845dcc8-7ksxw\" (UID: \"12f1f398-bfbc-4e05-bfc6-c654d7b9e97b\") " pod="openshift-console/console-68845dcc8-7ksxw" Dec 03 14:23:35 crc kubenswrapper[4805]: I1203 14:23:35.658328 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/12f1f398-bfbc-4e05-bfc6-c654d7b9e97b-oauth-serving-cert\") pod \"console-68845dcc8-7ksxw\" (UID: \"12f1f398-bfbc-4e05-bfc6-c654d7b9e97b\") " pod="openshift-console/console-68845dcc8-7ksxw" Dec 03 14:23:35 crc kubenswrapper[4805]: I1203 14:23:35.659324 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/12f1f398-bfbc-4e05-bfc6-c654d7b9e97b-service-ca\") pod \"console-68845dcc8-7ksxw\" (UID: \"12f1f398-bfbc-4e05-bfc6-c654d7b9e97b\") " pod="openshift-console/console-68845dcc8-7ksxw" Dec 03 14:23:35 crc kubenswrapper[4805]: I1203 14:23:35.659455 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/12f1f398-bfbc-4e05-bfc6-c654d7b9e97b-console-config\") pod \"console-68845dcc8-7ksxw\" (UID: \"12f1f398-bfbc-4e05-bfc6-c654d7b9e97b\") " pod="openshift-console/console-68845dcc8-7ksxw" Dec 03 14:23:35 crc kubenswrapper[4805]: I1203 14:23:35.660081 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/12f1f398-bfbc-4e05-bfc6-c654d7b9e97b-trusted-ca-bundle\") pod \"console-68845dcc8-7ksxw\" (UID: \"12f1f398-bfbc-4e05-bfc6-c654d7b9e97b\") " pod="openshift-console/console-68845dcc8-7ksxw" Dec 03 14:23:35 crc kubenswrapper[4805]: I1203 14:23:35.660355 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/12f1f398-bfbc-4e05-bfc6-c654d7b9e97b-oauth-serving-cert\") pod \"console-68845dcc8-7ksxw\" (UID: \"12f1f398-bfbc-4e05-bfc6-c654d7b9e97b\") " pod="openshift-console/console-68845dcc8-7ksxw" Dec 03 14:23:35 crc kubenswrapper[4805]: I1203 14:23:35.666405 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/12f1f398-bfbc-4e05-bfc6-c654d7b9e97b-console-oauth-config\") pod \"console-68845dcc8-7ksxw\" (UID: \"12f1f398-bfbc-4e05-bfc6-c654d7b9e97b\") " pod="openshift-console/console-68845dcc8-7ksxw" Dec 03 14:23:35 crc kubenswrapper[4805]: I1203 14:23:35.666522 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/12f1f398-bfbc-4e05-bfc6-c654d7b9e97b-console-serving-cert\") pod \"console-68845dcc8-7ksxw\" (UID: \"12f1f398-bfbc-4e05-bfc6-c654d7b9e97b\") " pod="openshift-console/console-68845dcc8-7ksxw" Dec 03 14:23:35 crc kubenswrapper[4805]: I1203 14:23:35.674326 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cp6bk\" (UniqueName: \"kubernetes.io/projected/12f1f398-bfbc-4e05-bfc6-c654d7b9e97b-kube-api-access-cp6bk\") pod \"console-68845dcc8-7ksxw\" (UID: \"12f1f398-bfbc-4e05-bfc6-c654d7b9e97b\") " pod="openshift-console/console-68845dcc8-7ksxw" Dec 03 14:23:35 crc kubenswrapper[4805]: I1203 14:23:35.757663 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-68845dcc8-7ksxw" Dec 03 14:23:35 crc kubenswrapper[4805]: I1203 14:23:35.961859 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugin-serving-cert\" (UniqueName: \"kubernetes.io/secret/5d4ab6ea-87ca-46dc-a9ad-f315faf0a9e2-plugin-serving-cert\") pod \"nmstate-console-plugin-7fbb5f6569-mbhzq\" (UID: \"5d4ab6ea-87ca-46dc-a9ad-f315faf0a9e2\") " pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-mbhzq" Dec 03 14:23:35 crc kubenswrapper[4805]: I1203 14:23:35.968720 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugin-serving-cert\" (UniqueName: \"kubernetes.io/secret/5d4ab6ea-87ca-46dc-a9ad-f315faf0a9e2-plugin-serving-cert\") pod \"nmstate-console-plugin-7fbb5f6569-mbhzq\" (UID: \"5d4ab6ea-87ca-46dc-a9ad-f315faf0a9e2\") " pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-mbhzq" Dec 03 14:23:35 crc kubenswrapper[4805]: I1203 14:23:35.973754 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-68845dcc8-7ksxw"] Dec 03 14:23:35 crc kubenswrapper[4805]: W1203 14:23:35.983210 4805 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod12f1f398_bfbc_4e05_bfc6_c654d7b9e97b.slice/crio-624df663e702d6c48a99e2144e715288a1acb9e9e6ecf23b8ae38a7d3b5e719a WatchSource:0}: Error finding container 624df663e702d6c48a99e2144e715288a1acb9e9e6ecf23b8ae38a7d3b5e719a: Status 404 returned error can't find the container with id 624df663e702d6c48a99e2144e715288a1acb9e9e6ecf23b8ae38a7d3b5e719a Dec 03 14:23:36 crc kubenswrapper[4805]: I1203 14:23:36.142120 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-mbhzq" Dec 03 14:23:36 crc kubenswrapper[4805]: I1203 14:23:36.343312 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-console-plugin-7fbb5f6569-mbhzq"] Dec 03 14:23:36 crc kubenswrapper[4805]: I1203 14:23:36.346554 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-68845dcc8-7ksxw" event={"ID":"12f1f398-bfbc-4e05-bfc6-c654d7b9e97b","Type":"ContainerStarted","Data":"3cfe56d53c6bf4666394552beac7e0cf79a6477375e898839b199ddca59b8064"} Dec 03 14:23:36 crc kubenswrapper[4805]: I1203 14:23:36.346608 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-68845dcc8-7ksxw" event={"ID":"12f1f398-bfbc-4e05-bfc6-c654d7b9e97b","Type":"ContainerStarted","Data":"624df663e702d6c48a99e2144e715288a1acb9e9e6ecf23b8ae38a7d3b5e719a"} Dec 03 14:23:36 crc kubenswrapper[4805]: I1203 14:23:36.347640 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-metrics-7f946cbc9-6r8jm" event={"ID":"9d39f3b1-337a-4aeb-b1cf-9ca9290f7791","Type":"ContainerStarted","Data":"be158bf4f302667b5552fe46dd1de4928b80a67adf0cada2de795c41e440f173"} Dec 03 14:23:36 crc kubenswrapper[4805]: I1203 14:23:36.348305 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-handler-xqhj5" event={"ID":"4f623c1b-94c0-43cd-ae9f-56597f4542b4","Type":"ContainerStarted","Data":"f058adc9bcd8e1b39515c0e04938d485e9060144c43ab04ca67d7c82dd92ab23"} Dec 03 14:23:36 crc kubenswrapper[4805]: I1203 14:23:36.349040 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-webhook-5f6d4c5ccb-2zszs" event={"ID":"9654448a-2df5-4cfc-b6e5-401ad15bc52e","Type":"ContainerStarted","Data":"5697c7db08639bbe3a1d3de3d19a6c7f6066bf96ecae16baefe5d340b4d90e32"} Dec 03 14:23:36 crc kubenswrapper[4805]: W1203 14:23:36.349408 4805 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod5d4ab6ea_87ca_46dc_a9ad_f315faf0a9e2.slice/crio-4a2ddf48a4b6c1f95a8b512848fb2f154f75f52d47df8db8077ad5270fa69767 WatchSource:0}: Error finding container 4a2ddf48a4b6c1f95a8b512848fb2f154f75f52d47df8db8077ad5270fa69767: Status 404 returned error can't find the container with id 4a2ddf48a4b6c1f95a8b512848fb2f154f75f52d47df8db8077ad5270fa69767 Dec 03 14:23:36 crc kubenswrapper[4805]: I1203 14:23:36.711104 4805 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-vljmk" Dec 03 14:23:36 crc kubenswrapper[4805]: I1203 14:23:36.756735 4805 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-vljmk" Dec 03 14:23:37 crc kubenswrapper[4805]: I1203 14:23:37.355568 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-mbhzq" event={"ID":"5d4ab6ea-87ca-46dc-a9ad-f315faf0a9e2","Type":"ContainerStarted","Data":"4a2ddf48a4b6c1f95a8b512848fb2f154f75f52d47df8db8077ad5270fa69767"} Dec 03 14:23:37 crc kubenswrapper[4805]: I1203 14:23:37.382712 4805 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console/console-68845dcc8-7ksxw" podStartSLOduration=2.382693632 podStartE2EDuration="2.382693632s" podCreationTimestamp="2025-12-03 14:23:35 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 14:23:37.372613671 +0000 UTC m=+847.035530594" watchObservedRunningTime="2025-12-03 14:23:37.382693632 +0000 UTC m=+847.045610545" Dec 03 14:23:39 crc kubenswrapper[4805]: I1203 14:23:39.049405 4805 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-vljmk"] Dec 03 14:23:39 crc kubenswrapper[4805]: I1203 14:23:39.049934 4805 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-vljmk" podUID="9cd4ed75-301f-42cf-b5d6-629eb4497892" containerName="registry-server" containerID="cri-o://811878ae45c02073bacce56ca70f5afd123f00942b7f196877994ecede2a809a" gracePeriod=2 Dec 03 14:23:39 crc kubenswrapper[4805]: I1203 14:23:39.368739 4805 generic.go:334] "Generic (PLEG): container finished" podID="9cd4ed75-301f-42cf-b5d6-629eb4497892" containerID="811878ae45c02073bacce56ca70f5afd123f00942b7f196877994ecede2a809a" exitCode=0 Dec 03 14:23:39 crc kubenswrapper[4805]: I1203 14:23:39.368765 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-vljmk" event={"ID":"9cd4ed75-301f-42cf-b5d6-629eb4497892","Type":"ContainerDied","Data":"811878ae45c02073bacce56ca70f5afd123f00942b7f196877994ecede2a809a"} Dec 03 14:23:40 crc kubenswrapper[4805]: I1203 14:23:40.048799 4805 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-vljmk" Dec 03 14:23:40 crc kubenswrapper[4805]: I1203 14:23:40.148089 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9n6l8\" (UniqueName: \"kubernetes.io/projected/9cd4ed75-301f-42cf-b5d6-629eb4497892-kube-api-access-9n6l8\") pod \"9cd4ed75-301f-42cf-b5d6-629eb4497892\" (UID: \"9cd4ed75-301f-42cf-b5d6-629eb4497892\") " Dec 03 14:23:40 crc kubenswrapper[4805]: I1203 14:23:40.148162 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9cd4ed75-301f-42cf-b5d6-629eb4497892-utilities\") pod \"9cd4ed75-301f-42cf-b5d6-629eb4497892\" (UID: \"9cd4ed75-301f-42cf-b5d6-629eb4497892\") " Dec 03 14:23:40 crc kubenswrapper[4805]: I1203 14:23:40.148458 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9cd4ed75-301f-42cf-b5d6-629eb4497892-catalog-content\") pod \"9cd4ed75-301f-42cf-b5d6-629eb4497892\" (UID: \"9cd4ed75-301f-42cf-b5d6-629eb4497892\") " Dec 03 14:23:40 crc kubenswrapper[4805]: I1203 14:23:40.152273 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9cd4ed75-301f-42cf-b5d6-629eb4497892-kube-api-access-9n6l8" (OuterVolumeSpecName: "kube-api-access-9n6l8") pod "9cd4ed75-301f-42cf-b5d6-629eb4497892" (UID: "9cd4ed75-301f-42cf-b5d6-629eb4497892"). InnerVolumeSpecName "kube-api-access-9n6l8". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 14:23:40 crc kubenswrapper[4805]: I1203 14:23:40.161370 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9cd4ed75-301f-42cf-b5d6-629eb4497892-utilities" (OuterVolumeSpecName: "utilities") pod "9cd4ed75-301f-42cf-b5d6-629eb4497892" (UID: "9cd4ed75-301f-42cf-b5d6-629eb4497892"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 14:23:40 crc kubenswrapper[4805]: I1203 14:23:40.250456 4805 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9n6l8\" (UniqueName: \"kubernetes.io/projected/9cd4ed75-301f-42cf-b5d6-629eb4497892-kube-api-access-9n6l8\") on node \"crc\" DevicePath \"\"" Dec 03 14:23:40 crc kubenswrapper[4805]: I1203 14:23:40.250491 4805 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9cd4ed75-301f-42cf-b5d6-629eb4497892-utilities\") on node \"crc\" DevicePath \"\"" Dec 03 14:23:40 crc kubenswrapper[4805]: I1203 14:23:40.268885 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9cd4ed75-301f-42cf-b5d6-629eb4497892-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "9cd4ed75-301f-42cf-b5d6-629eb4497892" (UID: "9cd4ed75-301f-42cf-b5d6-629eb4497892"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 14:23:40 crc kubenswrapper[4805]: I1203 14:23:40.351043 4805 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9cd4ed75-301f-42cf-b5d6-629eb4497892-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 03 14:23:40 crc kubenswrapper[4805]: I1203 14:23:40.377587 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-vljmk" event={"ID":"9cd4ed75-301f-42cf-b5d6-629eb4497892","Type":"ContainerDied","Data":"952e31494068c03ae79558e9cd865112c76c3379f6271034df07ea6e41d76bd9"} Dec 03 14:23:40 crc kubenswrapper[4805]: I1203 14:23:40.377634 4805 scope.go:117] "RemoveContainer" containerID="811878ae45c02073bacce56ca70f5afd123f00942b7f196877994ecede2a809a" Dec 03 14:23:40 crc kubenswrapper[4805]: I1203 14:23:40.377733 4805 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-vljmk" Dec 03 14:23:40 crc kubenswrapper[4805]: I1203 14:23:40.379704 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-metrics-7f946cbc9-6r8jm" event={"ID":"9d39f3b1-337a-4aeb-b1cf-9ca9290f7791","Type":"ContainerStarted","Data":"2ed3052eddddb5396d8815590f6ff6f3a27212b64c899bce8695676b4435490a"} Dec 03 14:23:40 crc kubenswrapper[4805]: I1203 14:23:40.381667 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-handler-xqhj5" event={"ID":"4f623c1b-94c0-43cd-ae9f-56597f4542b4","Type":"ContainerStarted","Data":"dee76e1ddd6fe68342c0fc2fe987ff6e0736851264e746d055947c79de96f102"} Dec 03 14:23:40 crc kubenswrapper[4805]: I1203 14:23:40.381962 4805 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-nmstate/nmstate-handler-xqhj5" Dec 03 14:23:40 crc kubenswrapper[4805]: I1203 14:23:40.384225 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-mbhzq" event={"ID":"5d4ab6ea-87ca-46dc-a9ad-f315faf0a9e2","Type":"ContainerStarted","Data":"22db18f093751dd19e28c3a0e6c67fc1c2c03dd03dcf0027fede930d7bd0789e"} Dec 03 14:23:40 crc kubenswrapper[4805]: I1203 14:23:40.386468 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-webhook-5f6d4c5ccb-2zszs" event={"ID":"9654448a-2df5-4cfc-b6e5-401ad15bc52e","Type":"ContainerStarted","Data":"a339bebbc64c830cbc8fb44d5f1aecfd4e31cfb5683d3581681e09cf1ad04057"} Dec 03 14:23:40 crc kubenswrapper[4805]: I1203 14:23:40.386920 4805 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-nmstate/nmstate-webhook-5f6d4c5ccb-2zszs" Dec 03 14:23:40 crc kubenswrapper[4805]: I1203 14:23:40.401766 4805 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-handler-xqhj5" podStartSLOduration=1.242438999 podStartE2EDuration="5.401749569s" podCreationTimestamp="2025-12-03 14:23:35 +0000 UTC" firstStartedPulling="2025-12-03 14:23:35.460578402 +0000 UTC m=+845.123495325" lastFinishedPulling="2025-12-03 14:23:39.619888932 +0000 UTC m=+849.282805895" observedRunningTime="2025-12-03 14:23:40.39733538 +0000 UTC m=+850.060252363" watchObservedRunningTime="2025-12-03 14:23:40.401749569 +0000 UTC m=+850.064666492" Dec 03 14:23:40 crc kubenswrapper[4805]: I1203 14:23:40.403130 4805 scope.go:117] "RemoveContainer" containerID="42ca20e4846cffb850e29e9fc238b2928340fd3a8e89b827acf0b046a33c0312" Dec 03 14:23:40 crc kubenswrapper[4805]: I1203 14:23:40.419497 4805 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-mbhzq" podStartSLOduration=1.729125556 podStartE2EDuration="5.419475326s" podCreationTimestamp="2025-12-03 14:23:35 +0000 UTC" firstStartedPulling="2025-12-03 14:23:36.351368112 +0000 UTC m=+846.014285035" lastFinishedPulling="2025-12-03 14:23:40.041717872 +0000 UTC m=+849.704634805" observedRunningTime="2025-12-03 14:23:40.412026376 +0000 UTC m=+850.074943349" watchObservedRunningTime="2025-12-03 14:23:40.419475326 +0000 UTC m=+850.082392259" Dec 03 14:23:40 crc kubenswrapper[4805]: I1203 14:23:40.429544 4805 scope.go:117] "RemoveContainer" containerID="f144eafd040f138842332f6935fa8b4bdea4da55167921b80b5ba4b9a871523f" Dec 03 14:23:40 crc kubenswrapper[4805]: I1203 14:23:40.429659 4805 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-vljmk"] Dec 03 14:23:40 crc kubenswrapper[4805]: I1203 14:23:40.435187 4805 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-vljmk"] Dec 03 14:23:40 crc kubenswrapper[4805]: I1203 14:23:40.449966 4805 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-webhook-5f6d4c5ccb-2zszs" podStartSLOduration=1.4602297 podStartE2EDuration="5.449940756s" podCreationTimestamp="2025-12-03 14:23:35 +0000 UTC" firstStartedPulling="2025-12-03 14:23:35.657293485 +0000 UTC m=+845.320210408" lastFinishedPulling="2025-12-03 14:23:39.647004501 +0000 UTC m=+849.309921464" observedRunningTime="2025-12-03 14:23:40.446043761 +0000 UTC m=+850.108960684" watchObservedRunningTime="2025-12-03 14:23:40.449940756 +0000 UTC m=+850.112857679" Dec 03 14:23:40 crc kubenswrapper[4805]: I1203 14:23:40.710897 4805 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9cd4ed75-301f-42cf-b5d6-629eb4497892" path="/var/lib/kubelet/pods/9cd4ed75-301f-42cf-b5d6-629eb4497892/volumes" Dec 03 14:23:43 crc kubenswrapper[4805]: I1203 14:23:43.917897 4805 patch_prober.go:28] interesting pod/machine-config-daemon-gskh4 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 03 14:23:43 crc kubenswrapper[4805]: I1203 14:23:43.918235 4805 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-gskh4" podUID="6098937f-e3f6-45e8-a647-4994a79cd711" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 03 14:23:43 crc kubenswrapper[4805]: I1203 14:23:43.918279 4805 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-gskh4" Dec 03 14:23:43 crc kubenswrapper[4805]: I1203 14:23:43.918952 4805 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"7e81e89ac6c4e0378da77d246d467c4c9c45debada7dcf3d097d3c2112a7ea0f"} pod="openshift-machine-config-operator/machine-config-daemon-gskh4" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 03 14:23:43 crc kubenswrapper[4805]: I1203 14:23:43.919003 4805 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-gskh4" podUID="6098937f-e3f6-45e8-a647-4994a79cd711" containerName="machine-config-daemon" containerID="cri-o://7e81e89ac6c4e0378da77d246d467c4c9c45debada7dcf3d097d3c2112a7ea0f" gracePeriod=600 Dec 03 14:23:44 crc kubenswrapper[4805]: I1203 14:23:44.423108 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-metrics-7f946cbc9-6r8jm" event={"ID":"9d39f3b1-337a-4aeb-b1cf-9ca9290f7791","Type":"ContainerStarted","Data":"debc31195475572b98e9b5f54a4544a5a866ba24bbc1c5e01394594af16316b2"} Dec 03 14:23:44 crc kubenswrapper[4805]: I1203 14:23:44.427173 4805 generic.go:334] "Generic (PLEG): container finished" podID="6098937f-e3f6-45e8-a647-4994a79cd711" containerID="7e81e89ac6c4e0378da77d246d467c4c9c45debada7dcf3d097d3c2112a7ea0f" exitCode=0 Dec 03 14:23:44 crc kubenswrapper[4805]: I1203 14:23:44.427223 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-gskh4" event={"ID":"6098937f-e3f6-45e8-a647-4994a79cd711","Type":"ContainerDied","Data":"7e81e89ac6c4e0378da77d246d467c4c9c45debada7dcf3d097d3c2112a7ea0f"} Dec 03 14:23:44 crc kubenswrapper[4805]: I1203 14:23:44.427249 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-gskh4" event={"ID":"6098937f-e3f6-45e8-a647-4994a79cd711","Type":"ContainerStarted","Data":"5079b541103d89d1ef42b4f7093c19ac19ef0baacd14c63b2eb7fb16b384fba4"} Dec 03 14:23:44 crc kubenswrapper[4805]: I1203 14:23:44.427288 4805 scope.go:117] "RemoveContainer" containerID="af1b8fb24b94658040c784c6bffe5034acb7fffeeb5210dc5f214b62d02ab400" Dec 03 14:23:44 crc kubenswrapper[4805]: I1203 14:23:44.447144 4805 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-metrics-7f946cbc9-6r8jm" podStartSLOduration=1.5295764059999999 podStartE2EDuration="9.447128983s" podCreationTimestamp="2025-12-03 14:23:35 +0000 UTC" firstStartedPulling="2025-12-03 14:23:35.615895641 +0000 UTC m=+845.278812564" lastFinishedPulling="2025-12-03 14:23:43.533448178 +0000 UTC m=+853.196365141" observedRunningTime="2025-12-03 14:23:44.445549021 +0000 UTC m=+854.108465944" watchObservedRunningTime="2025-12-03 14:23:44.447128983 +0000 UTC m=+854.110045906" Dec 03 14:23:45 crc kubenswrapper[4805]: I1203 14:23:45.446807 4805 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-nmstate/nmstate-handler-xqhj5" Dec 03 14:23:45 crc kubenswrapper[4805]: I1203 14:23:45.758145 4805 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-console/console-68845dcc8-7ksxw" Dec 03 14:23:45 crc kubenswrapper[4805]: I1203 14:23:45.758206 4805 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console/console-68845dcc8-7ksxw" Dec 03 14:23:45 crc kubenswrapper[4805]: I1203 14:23:45.765106 4805 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-console/console-68845dcc8-7ksxw" Dec 03 14:23:46 crc kubenswrapper[4805]: I1203 14:23:46.450752 4805 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console/console-68845dcc8-7ksxw" Dec 03 14:23:46 crc kubenswrapper[4805]: I1203 14:23:46.509016 4805 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-console/console-f9d7485db-pp9tv"] Dec 03 14:23:55 crc kubenswrapper[4805]: I1203 14:23:55.422357 4805 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-nmstate/nmstate-webhook-5f6d4c5ccb-2zszs" Dec 03 14:23:57 crc kubenswrapper[4805]: I1203 14:23:57.666697 4805 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-zrnp6"] Dec 03 14:23:57 crc kubenswrapper[4805]: E1203 14:23:57.667010 4805 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9cd4ed75-301f-42cf-b5d6-629eb4497892" containerName="extract-utilities" Dec 03 14:23:57 crc kubenswrapper[4805]: I1203 14:23:57.667029 4805 state_mem.go:107] "Deleted CPUSet assignment" podUID="9cd4ed75-301f-42cf-b5d6-629eb4497892" containerName="extract-utilities" Dec 03 14:23:57 crc kubenswrapper[4805]: E1203 14:23:57.667046 4805 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9cd4ed75-301f-42cf-b5d6-629eb4497892" containerName="extract-content" Dec 03 14:23:57 crc kubenswrapper[4805]: I1203 14:23:57.667054 4805 state_mem.go:107] "Deleted CPUSet assignment" podUID="9cd4ed75-301f-42cf-b5d6-629eb4497892" containerName="extract-content" Dec 03 14:23:57 crc kubenswrapper[4805]: E1203 14:23:57.667073 4805 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9cd4ed75-301f-42cf-b5d6-629eb4497892" containerName="registry-server" Dec 03 14:23:57 crc kubenswrapper[4805]: I1203 14:23:57.667081 4805 state_mem.go:107] "Deleted CPUSet assignment" podUID="9cd4ed75-301f-42cf-b5d6-629eb4497892" containerName="registry-server" Dec 03 14:23:57 crc kubenswrapper[4805]: I1203 14:23:57.667230 4805 memory_manager.go:354] "RemoveStaleState removing state" podUID="9cd4ed75-301f-42cf-b5d6-629eb4497892" containerName="registry-server" Dec 03 14:23:57 crc kubenswrapper[4805]: I1203 14:23:57.668197 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-zrnp6" Dec 03 14:23:57 crc kubenswrapper[4805]: I1203 14:23:57.675723 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-zrnp6"] Dec 03 14:23:57 crc kubenswrapper[4805]: I1203 14:23:57.806536 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-m5qmf\" (UniqueName: \"kubernetes.io/projected/e2055b14-d7a1-4ac4-9c2e-ef12644113b2-kube-api-access-m5qmf\") pod \"certified-operators-zrnp6\" (UID: \"e2055b14-d7a1-4ac4-9c2e-ef12644113b2\") " pod="openshift-marketplace/certified-operators-zrnp6" Dec 03 14:23:57 crc kubenswrapper[4805]: I1203 14:23:57.806945 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e2055b14-d7a1-4ac4-9c2e-ef12644113b2-catalog-content\") pod \"certified-operators-zrnp6\" (UID: \"e2055b14-d7a1-4ac4-9c2e-ef12644113b2\") " pod="openshift-marketplace/certified-operators-zrnp6" Dec 03 14:23:57 crc kubenswrapper[4805]: I1203 14:23:57.807000 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e2055b14-d7a1-4ac4-9c2e-ef12644113b2-utilities\") pod \"certified-operators-zrnp6\" (UID: \"e2055b14-d7a1-4ac4-9c2e-ef12644113b2\") " pod="openshift-marketplace/certified-operators-zrnp6" Dec 03 14:23:57 crc kubenswrapper[4805]: I1203 14:23:57.907642 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e2055b14-d7a1-4ac4-9c2e-ef12644113b2-catalog-content\") pod \"certified-operators-zrnp6\" (UID: \"e2055b14-d7a1-4ac4-9c2e-ef12644113b2\") " pod="openshift-marketplace/certified-operators-zrnp6" Dec 03 14:23:57 crc kubenswrapper[4805]: I1203 14:23:57.907712 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e2055b14-d7a1-4ac4-9c2e-ef12644113b2-utilities\") pod \"certified-operators-zrnp6\" (UID: \"e2055b14-d7a1-4ac4-9c2e-ef12644113b2\") " pod="openshift-marketplace/certified-operators-zrnp6" Dec 03 14:23:57 crc kubenswrapper[4805]: I1203 14:23:57.907755 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-m5qmf\" (UniqueName: \"kubernetes.io/projected/e2055b14-d7a1-4ac4-9c2e-ef12644113b2-kube-api-access-m5qmf\") pod \"certified-operators-zrnp6\" (UID: \"e2055b14-d7a1-4ac4-9c2e-ef12644113b2\") " pod="openshift-marketplace/certified-operators-zrnp6" Dec 03 14:23:57 crc kubenswrapper[4805]: I1203 14:23:57.908345 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e2055b14-d7a1-4ac4-9c2e-ef12644113b2-catalog-content\") pod \"certified-operators-zrnp6\" (UID: \"e2055b14-d7a1-4ac4-9c2e-ef12644113b2\") " pod="openshift-marketplace/certified-operators-zrnp6" Dec 03 14:23:57 crc kubenswrapper[4805]: I1203 14:23:57.908406 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e2055b14-d7a1-4ac4-9c2e-ef12644113b2-utilities\") pod \"certified-operators-zrnp6\" (UID: \"e2055b14-d7a1-4ac4-9c2e-ef12644113b2\") " pod="openshift-marketplace/certified-operators-zrnp6" Dec 03 14:23:57 crc kubenswrapper[4805]: I1203 14:23:57.940108 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-m5qmf\" (UniqueName: \"kubernetes.io/projected/e2055b14-d7a1-4ac4-9c2e-ef12644113b2-kube-api-access-m5qmf\") pod \"certified-operators-zrnp6\" (UID: \"e2055b14-d7a1-4ac4-9c2e-ef12644113b2\") " pod="openshift-marketplace/certified-operators-zrnp6" Dec 03 14:23:57 crc kubenswrapper[4805]: I1203 14:23:57.992294 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-zrnp6" Dec 03 14:23:58 crc kubenswrapper[4805]: I1203 14:23:58.251152 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-zrnp6"] Dec 03 14:23:58 crc kubenswrapper[4805]: W1203 14:23:58.279964 4805 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pode2055b14_d7a1_4ac4_9c2e_ef12644113b2.slice/crio-3b08d74b972880645b3ca2a9bc41c2b564b111189d7d951d03809856ca52b51f WatchSource:0}: Error finding container 3b08d74b972880645b3ca2a9bc41c2b564b111189d7d951d03809856ca52b51f: Status 404 returned error can't find the container with id 3b08d74b972880645b3ca2a9bc41c2b564b111189d7d951d03809856ca52b51f Dec 03 14:23:58 crc kubenswrapper[4805]: I1203 14:23:58.533235 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-zrnp6" event={"ID":"e2055b14-d7a1-4ac4-9c2e-ef12644113b2","Type":"ContainerStarted","Data":"3b08d74b972880645b3ca2a9bc41c2b564b111189d7d951d03809856ca52b51f"} Dec 03 14:23:59 crc kubenswrapper[4805]: I1203 14:23:59.542412 4805 generic.go:334] "Generic (PLEG): container finished" podID="e2055b14-d7a1-4ac4-9c2e-ef12644113b2" containerID="8259535f3eecadad90f99e2ad8437469bd973664b02575624c5239287cb63465" exitCode=0 Dec 03 14:23:59 crc kubenswrapper[4805]: I1203 14:23:59.542501 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-zrnp6" event={"ID":"e2055b14-d7a1-4ac4-9c2e-ef12644113b2","Type":"ContainerDied","Data":"8259535f3eecadad90f99e2ad8437469bd973664b02575624c5239287cb63465"} Dec 03 14:24:00 crc kubenswrapper[4805]: I1203 14:24:00.549759 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-zrnp6" event={"ID":"e2055b14-d7a1-4ac4-9c2e-ef12644113b2","Type":"ContainerStarted","Data":"1c88df15d36b59aae4a17ce193c12ebcfcbbed517f968f721e2dd45166c9edb8"} Dec 03 14:24:01 crc kubenswrapper[4805]: I1203 14:24:01.559353 4805 generic.go:334] "Generic (PLEG): container finished" podID="e2055b14-d7a1-4ac4-9c2e-ef12644113b2" containerID="1c88df15d36b59aae4a17ce193c12ebcfcbbed517f968f721e2dd45166c9edb8" exitCode=0 Dec 03 14:24:01 crc kubenswrapper[4805]: I1203 14:24:01.559417 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-zrnp6" event={"ID":"e2055b14-d7a1-4ac4-9c2e-ef12644113b2","Type":"ContainerDied","Data":"1c88df15d36b59aae4a17ce193c12ebcfcbbed517f968f721e2dd45166c9edb8"} Dec 03 14:24:03 crc kubenswrapper[4805]: I1203 14:24:03.575921 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-zrnp6" event={"ID":"e2055b14-d7a1-4ac4-9c2e-ef12644113b2","Type":"ContainerStarted","Data":"d921eb746733a7032565457067aef5acf11a18adedcd0c5d20cfe3f751be49aa"} Dec 03 14:24:03 crc kubenswrapper[4805]: I1203 14:24:03.612556 4805 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-zrnp6" podStartSLOduration=3.017401379 podStartE2EDuration="6.612541268s" podCreationTimestamp="2025-12-03 14:23:57 +0000 UTC" firstStartedPulling="2025-12-03 14:23:59.545207743 +0000 UTC m=+869.208124706" lastFinishedPulling="2025-12-03 14:24:03.140347642 +0000 UTC m=+872.803264595" observedRunningTime="2025-12-03 14:24:03.609494666 +0000 UTC m=+873.272411589" watchObservedRunningTime="2025-12-03 14:24:03.612541268 +0000 UTC m=+873.275458191" Dec 03 14:24:07 crc kubenswrapper[4805]: I1203 14:24:07.993318 4805 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-zrnp6" Dec 03 14:24:07 crc kubenswrapper[4805]: I1203 14:24:07.993982 4805 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-zrnp6" Dec 03 14:24:08 crc kubenswrapper[4805]: I1203 14:24:08.050596 4805 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-zrnp6" Dec 03 14:24:08 crc kubenswrapper[4805]: I1203 14:24:08.672768 4805 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-zrnp6" Dec 03 14:24:10 crc kubenswrapper[4805]: I1203 14:24:10.727697 4805 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83glh2z"] Dec 03 14:24:10 crc kubenswrapper[4805]: I1203 14:24:10.737753 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83glh2z" Dec 03 14:24:10 crc kubenswrapper[4805]: I1203 14:24:10.741136 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"default-dockercfg-vmwhc" Dec 03 14:24:10 crc kubenswrapper[4805]: I1203 14:24:10.752571 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83glh2z"] Dec 03 14:24:10 crc kubenswrapper[4805]: I1203 14:24:10.847949 4805 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-zrnp6"] Dec 03 14:24:10 crc kubenswrapper[4805]: I1203 14:24:10.848209 4805 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-zrnp6" podUID="e2055b14-d7a1-4ac4-9c2e-ef12644113b2" containerName="registry-server" containerID="cri-o://d921eb746733a7032565457067aef5acf11a18adedcd0c5d20cfe3f751be49aa" gracePeriod=2 Dec 03 14:24:10 crc kubenswrapper[4805]: I1203 14:24:10.890817 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kl88t\" (UniqueName: \"kubernetes.io/projected/449e7507-4d25-4399-b1bd-b174d65a5041-kube-api-access-kl88t\") pod \"af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83glh2z\" (UID: \"449e7507-4d25-4399-b1bd-b174d65a5041\") " pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83glh2z" Dec 03 14:24:10 crc kubenswrapper[4805]: I1203 14:24:10.890977 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/449e7507-4d25-4399-b1bd-b174d65a5041-bundle\") pod \"af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83glh2z\" (UID: \"449e7507-4d25-4399-b1bd-b174d65a5041\") " pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83glh2z" Dec 03 14:24:10 crc kubenswrapper[4805]: I1203 14:24:10.891014 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/449e7507-4d25-4399-b1bd-b174d65a5041-util\") pod \"af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83glh2z\" (UID: \"449e7507-4d25-4399-b1bd-b174d65a5041\") " pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83glh2z" Dec 03 14:24:10 crc kubenswrapper[4805]: I1203 14:24:10.992535 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kl88t\" (UniqueName: \"kubernetes.io/projected/449e7507-4d25-4399-b1bd-b174d65a5041-kube-api-access-kl88t\") pod \"af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83glh2z\" (UID: \"449e7507-4d25-4399-b1bd-b174d65a5041\") " pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83glh2z" Dec 03 14:24:10 crc kubenswrapper[4805]: I1203 14:24:10.992611 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/449e7507-4d25-4399-b1bd-b174d65a5041-bundle\") pod \"af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83glh2z\" (UID: \"449e7507-4d25-4399-b1bd-b174d65a5041\") " pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83glh2z" Dec 03 14:24:10 crc kubenswrapper[4805]: I1203 14:24:10.992638 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/449e7507-4d25-4399-b1bd-b174d65a5041-util\") pod \"af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83glh2z\" (UID: \"449e7507-4d25-4399-b1bd-b174d65a5041\") " pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83glh2z" Dec 03 14:24:10 crc kubenswrapper[4805]: I1203 14:24:10.993461 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/449e7507-4d25-4399-b1bd-b174d65a5041-util\") pod \"af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83glh2z\" (UID: \"449e7507-4d25-4399-b1bd-b174d65a5041\") " pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83glh2z" Dec 03 14:24:10 crc kubenswrapper[4805]: I1203 14:24:10.993780 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/449e7507-4d25-4399-b1bd-b174d65a5041-bundle\") pod \"af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83glh2z\" (UID: \"449e7507-4d25-4399-b1bd-b174d65a5041\") " pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83glh2z" Dec 03 14:24:11 crc kubenswrapper[4805]: I1203 14:24:11.018822 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kl88t\" (UniqueName: \"kubernetes.io/projected/449e7507-4d25-4399-b1bd-b174d65a5041-kube-api-access-kl88t\") pod \"af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83glh2z\" (UID: \"449e7507-4d25-4399-b1bd-b174d65a5041\") " pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83glh2z" Dec 03 14:24:11 crc kubenswrapper[4805]: I1203 14:24:11.065286 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83glh2z" Dec 03 14:24:11 crc kubenswrapper[4805]: I1203 14:24:11.364073 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83glh2z"] Dec 03 14:24:11 crc kubenswrapper[4805]: I1203 14:24:11.368787 4805 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-zrnp6" Dec 03 14:24:11 crc kubenswrapper[4805]: I1203 14:24:11.500585 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e2055b14-d7a1-4ac4-9c2e-ef12644113b2-catalog-content\") pod \"e2055b14-d7a1-4ac4-9c2e-ef12644113b2\" (UID: \"e2055b14-d7a1-4ac4-9c2e-ef12644113b2\") " Dec 03 14:24:11 crc kubenswrapper[4805]: I1203 14:24:11.500986 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-m5qmf\" (UniqueName: \"kubernetes.io/projected/e2055b14-d7a1-4ac4-9c2e-ef12644113b2-kube-api-access-m5qmf\") pod \"e2055b14-d7a1-4ac4-9c2e-ef12644113b2\" (UID: \"e2055b14-d7a1-4ac4-9c2e-ef12644113b2\") " Dec 03 14:24:11 crc kubenswrapper[4805]: I1203 14:24:11.501038 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e2055b14-d7a1-4ac4-9c2e-ef12644113b2-utilities\") pod \"e2055b14-d7a1-4ac4-9c2e-ef12644113b2\" (UID: \"e2055b14-d7a1-4ac4-9c2e-ef12644113b2\") " Dec 03 14:24:11 crc kubenswrapper[4805]: I1203 14:24:11.502162 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e2055b14-d7a1-4ac4-9c2e-ef12644113b2-utilities" (OuterVolumeSpecName: "utilities") pod "e2055b14-d7a1-4ac4-9c2e-ef12644113b2" (UID: "e2055b14-d7a1-4ac4-9c2e-ef12644113b2"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 14:24:11 crc kubenswrapper[4805]: I1203 14:24:11.505776 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e2055b14-d7a1-4ac4-9c2e-ef12644113b2-kube-api-access-m5qmf" (OuterVolumeSpecName: "kube-api-access-m5qmf") pod "e2055b14-d7a1-4ac4-9c2e-ef12644113b2" (UID: "e2055b14-d7a1-4ac4-9c2e-ef12644113b2"). InnerVolumeSpecName "kube-api-access-m5qmf". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 14:24:11 crc kubenswrapper[4805]: I1203 14:24:11.545581 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e2055b14-d7a1-4ac4-9c2e-ef12644113b2-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "e2055b14-d7a1-4ac4-9c2e-ef12644113b2" (UID: "e2055b14-d7a1-4ac4-9c2e-ef12644113b2"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 14:24:11 crc kubenswrapper[4805]: I1203 14:24:11.568583 4805 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-console/console-f9d7485db-pp9tv" podUID="e2034338-8a35-4fd2-9bbf-be733939fc25" containerName="console" containerID="cri-o://2a23c6b4702b54967793e4d45488577f06641ef8826624873c784933f74405a3" gracePeriod=15 Dec 03 14:24:11 crc kubenswrapper[4805]: I1203 14:24:11.602930 4805 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-m5qmf\" (UniqueName: \"kubernetes.io/projected/e2055b14-d7a1-4ac4-9c2e-ef12644113b2-kube-api-access-m5qmf\") on node \"crc\" DevicePath \"\"" Dec 03 14:24:11 crc kubenswrapper[4805]: I1203 14:24:11.602970 4805 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e2055b14-d7a1-4ac4-9c2e-ef12644113b2-utilities\") on node \"crc\" DevicePath \"\"" Dec 03 14:24:11 crc kubenswrapper[4805]: I1203 14:24:11.602984 4805 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e2055b14-d7a1-4ac4-9c2e-ef12644113b2-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 03 14:24:11 crc kubenswrapper[4805]: I1203 14:24:11.624513 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83glh2z" event={"ID":"449e7507-4d25-4399-b1bd-b174d65a5041","Type":"ContainerStarted","Data":"b6cc18fa11d33b5315a32a5fe4285abacc55ac3ab200ba2d2c8f72023f2e8c86"} Dec 03 14:24:11 crc kubenswrapper[4805]: I1203 14:24:11.627478 4805 generic.go:334] "Generic (PLEG): container finished" podID="e2055b14-d7a1-4ac4-9c2e-ef12644113b2" containerID="d921eb746733a7032565457067aef5acf11a18adedcd0c5d20cfe3f751be49aa" exitCode=0 Dec 03 14:24:11 crc kubenswrapper[4805]: I1203 14:24:11.627511 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-zrnp6" event={"ID":"e2055b14-d7a1-4ac4-9c2e-ef12644113b2","Type":"ContainerDied","Data":"d921eb746733a7032565457067aef5acf11a18adedcd0c5d20cfe3f751be49aa"} Dec 03 14:24:11 crc kubenswrapper[4805]: I1203 14:24:11.627531 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-zrnp6" event={"ID":"e2055b14-d7a1-4ac4-9c2e-ef12644113b2","Type":"ContainerDied","Data":"3b08d74b972880645b3ca2a9bc41c2b564b111189d7d951d03809856ca52b51f"} Dec 03 14:24:11 crc kubenswrapper[4805]: I1203 14:24:11.627563 4805 scope.go:117] "RemoveContainer" containerID="d921eb746733a7032565457067aef5acf11a18adedcd0c5d20cfe3f751be49aa" Dec 03 14:24:11 crc kubenswrapper[4805]: I1203 14:24:11.627561 4805 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-zrnp6" Dec 03 14:24:11 crc kubenswrapper[4805]: I1203 14:24:11.648218 4805 scope.go:117] "RemoveContainer" containerID="1c88df15d36b59aae4a17ce193c12ebcfcbbed517f968f721e2dd45166c9edb8" Dec 03 14:24:11 crc kubenswrapper[4805]: I1203 14:24:11.669039 4805 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-zrnp6"] Dec 03 14:24:11 crc kubenswrapper[4805]: I1203 14:24:11.674065 4805 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-zrnp6"] Dec 03 14:24:11 crc kubenswrapper[4805]: I1203 14:24:11.689515 4805 scope.go:117] "RemoveContainer" containerID="8259535f3eecadad90f99e2ad8437469bd973664b02575624c5239287cb63465" Dec 03 14:24:11 crc kubenswrapper[4805]: I1203 14:24:11.713335 4805 scope.go:117] "RemoveContainer" containerID="d921eb746733a7032565457067aef5acf11a18adedcd0c5d20cfe3f751be49aa" Dec 03 14:24:11 crc kubenswrapper[4805]: E1203 14:24:11.713770 4805 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d921eb746733a7032565457067aef5acf11a18adedcd0c5d20cfe3f751be49aa\": container with ID starting with d921eb746733a7032565457067aef5acf11a18adedcd0c5d20cfe3f751be49aa not found: ID does not exist" containerID="d921eb746733a7032565457067aef5acf11a18adedcd0c5d20cfe3f751be49aa" Dec 03 14:24:11 crc kubenswrapper[4805]: I1203 14:24:11.713807 4805 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d921eb746733a7032565457067aef5acf11a18adedcd0c5d20cfe3f751be49aa"} err="failed to get container status \"d921eb746733a7032565457067aef5acf11a18adedcd0c5d20cfe3f751be49aa\": rpc error: code = NotFound desc = could not find container \"d921eb746733a7032565457067aef5acf11a18adedcd0c5d20cfe3f751be49aa\": container with ID starting with d921eb746733a7032565457067aef5acf11a18adedcd0c5d20cfe3f751be49aa not found: ID does not exist" Dec 03 14:24:11 crc kubenswrapper[4805]: I1203 14:24:11.713867 4805 scope.go:117] "RemoveContainer" containerID="1c88df15d36b59aae4a17ce193c12ebcfcbbed517f968f721e2dd45166c9edb8" Dec 03 14:24:11 crc kubenswrapper[4805]: E1203 14:24:11.714445 4805 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1c88df15d36b59aae4a17ce193c12ebcfcbbed517f968f721e2dd45166c9edb8\": container with ID starting with 1c88df15d36b59aae4a17ce193c12ebcfcbbed517f968f721e2dd45166c9edb8 not found: ID does not exist" containerID="1c88df15d36b59aae4a17ce193c12ebcfcbbed517f968f721e2dd45166c9edb8" Dec 03 14:24:11 crc kubenswrapper[4805]: I1203 14:24:11.714487 4805 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1c88df15d36b59aae4a17ce193c12ebcfcbbed517f968f721e2dd45166c9edb8"} err="failed to get container status \"1c88df15d36b59aae4a17ce193c12ebcfcbbed517f968f721e2dd45166c9edb8\": rpc error: code = NotFound desc = could not find container \"1c88df15d36b59aae4a17ce193c12ebcfcbbed517f968f721e2dd45166c9edb8\": container with ID starting with 1c88df15d36b59aae4a17ce193c12ebcfcbbed517f968f721e2dd45166c9edb8 not found: ID does not exist" Dec 03 14:24:11 crc kubenswrapper[4805]: I1203 14:24:11.714506 4805 scope.go:117] "RemoveContainer" containerID="8259535f3eecadad90f99e2ad8437469bd973664b02575624c5239287cb63465" Dec 03 14:24:11 crc kubenswrapper[4805]: E1203 14:24:11.714992 4805 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8259535f3eecadad90f99e2ad8437469bd973664b02575624c5239287cb63465\": container with ID starting with 8259535f3eecadad90f99e2ad8437469bd973664b02575624c5239287cb63465 not found: ID does not exist" containerID="8259535f3eecadad90f99e2ad8437469bd973664b02575624c5239287cb63465" Dec 03 14:24:11 crc kubenswrapper[4805]: I1203 14:24:11.715082 4805 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8259535f3eecadad90f99e2ad8437469bd973664b02575624c5239287cb63465"} err="failed to get container status \"8259535f3eecadad90f99e2ad8437469bd973664b02575624c5239287cb63465\": rpc error: code = NotFound desc = could not find container \"8259535f3eecadad90f99e2ad8437469bd973664b02575624c5239287cb63465\": container with ID starting with 8259535f3eecadad90f99e2ad8437469bd973664b02575624c5239287cb63465 not found: ID does not exist" Dec 03 14:24:12 crc kubenswrapper[4805]: I1203 14:24:12.525132 4805 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-console_console-f9d7485db-pp9tv_e2034338-8a35-4fd2-9bbf-be733939fc25/console/0.log" Dec 03 14:24:12 crc kubenswrapper[4805]: I1203 14:24:12.525447 4805 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-f9d7485db-pp9tv" Dec 03 14:24:12 crc kubenswrapper[4805]: I1203 14:24:12.636994 4805 generic.go:334] "Generic (PLEG): container finished" podID="449e7507-4d25-4399-b1bd-b174d65a5041" containerID="b5e33c25923565409241ecf7df222eb4f502f0db1e4f72be087a1c106b745f9e" exitCode=0 Dec 03 14:24:12 crc kubenswrapper[4805]: I1203 14:24:12.637086 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83glh2z" event={"ID":"449e7507-4d25-4399-b1bd-b174d65a5041","Type":"ContainerDied","Data":"b5e33c25923565409241ecf7df222eb4f502f0db1e4f72be087a1c106b745f9e"} Dec 03 14:24:12 crc kubenswrapper[4805]: I1203 14:24:12.640517 4805 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-console_console-f9d7485db-pp9tv_e2034338-8a35-4fd2-9bbf-be733939fc25/console/0.log" Dec 03 14:24:12 crc kubenswrapper[4805]: I1203 14:24:12.640564 4805 generic.go:334] "Generic (PLEG): container finished" podID="e2034338-8a35-4fd2-9bbf-be733939fc25" containerID="2a23c6b4702b54967793e4d45488577f06641ef8826624873c784933f74405a3" exitCode=2 Dec 03 14:24:12 crc kubenswrapper[4805]: I1203 14:24:12.640596 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-f9d7485db-pp9tv" event={"ID":"e2034338-8a35-4fd2-9bbf-be733939fc25","Type":"ContainerDied","Data":"2a23c6b4702b54967793e4d45488577f06641ef8826624873c784933f74405a3"} Dec 03 14:24:12 crc kubenswrapper[4805]: I1203 14:24:12.640623 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-f9d7485db-pp9tv" event={"ID":"e2034338-8a35-4fd2-9bbf-be733939fc25","Type":"ContainerDied","Data":"b956cee4baf3251b5e35ca263db8e1fbff64366ec9922a8a0edb94f3320e271f"} Dec 03 14:24:12 crc kubenswrapper[4805]: I1203 14:24:12.640638 4805 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-f9d7485db-pp9tv" Dec 03 14:24:12 crc kubenswrapper[4805]: I1203 14:24:12.640644 4805 scope.go:117] "RemoveContainer" containerID="2a23c6b4702b54967793e4d45488577f06641ef8826624873c784933f74405a3" Dec 03 14:24:12 crc kubenswrapper[4805]: I1203 14:24:12.658746 4805 scope.go:117] "RemoveContainer" containerID="2a23c6b4702b54967793e4d45488577f06641ef8826624873c784933f74405a3" Dec 03 14:24:12 crc kubenswrapper[4805]: E1203 14:24:12.659910 4805 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2a23c6b4702b54967793e4d45488577f06641ef8826624873c784933f74405a3\": container with ID starting with 2a23c6b4702b54967793e4d45488577f06641ef8826624873c784933f74405a3 not found: ID does not exist" containerID="2a23c6b4702b54967793e4d45488577f06641ef8826624873c784933f74405a3" Dec 03 14:24:12 crc kubenswrapper[4805]: I1203 14:24:12.659968 4805 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2a23c6b4702b54967793e4d45488577f06641ef8826624873c784933f74405a3"} err="failed to get container status \"2a23c6b4702b54967793e4d45488577f06641ef8826624873c784933f74405a3\": rpc error: code = NotFound desc = could not find container \"2a23c6b4702b54967793e4d45488577f06641ef8826624873c784933f74405a3\": container with ID starting with 2a23c6b4702b54967793e4d45488577f06641ef8826624873c784933f74405a3 not found: ID does not exist" Dec 03 14:24:12 crc kubenswrapper[4805]: I1203 14:24:12.701030 4805 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e2055b14-d7a1-4ac4-9c2e-ef12644113b2" path="/var/lib/kubelet/pods/e2055b14-d7a1-4ac4-9c2e-ef12644113b2/volumes" Dec 03 14:24:12 crc kubenswrapper[4805]: I1203 14:24:12.716832 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/e2034338-8a35-4fd2-9bbf-be733939fc25-console-oauth-config\") pod \"e2034338-8a35-4fd2-9bbf-be733939fc25\" (UID: \"e2034338-8a35-4fd2-9bbf-be733939fc25\") " Dec 03 14:24:12 crc kubenswrapper[4805]: I1203 14:24:12.716890 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/e2034338-8a35-4fd2-9bbf-be733939fc25-console-config\") pod \"e2034338-8a35-4fd2-9bbf-be733939fc25\" (UID: \"e2034338-8a35-4fd2-9bbf-be733939fc25\") " Dec 03 14:24:12 crc kubenswrapper[4805]: I1203 14:24:12.716947 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/e2034338-8a35-4fd2-9bbf-be733939fc25-service-ca\") pod \"e2034338-8a35-4fd2-9bbf-be733939fc25\" (UID: \"e2034338-8a35-4fd2-9bbf-be733939fc25\") " Dec 03 14:24:12 crc kubenswrapper[4805]: I1203 14:24:12.716982 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/e2034338-8a35-4fd2-9bbf-be733939fc25-oauth-serving-cert\") pod \"e2034338-8a35-4fd2-9bbf-be733939fc25\" (UID: \"e2034338-8a35-4fd2-9bbf-be733939fc25\") " Dec 03 14:24:12 crc kubenswrapper[4805]: I1203 14:24:12.717760 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/e2034338-8a35-4fd2-9bbf-be733939fc25-console-serving-cert\") pod \"e2034338-8a35-4fd2-9bbf-be733939fc25\" (UID: \"e2034338-8a35-4fd2-9bbf-be733939fc25\") " Dec 03 14:24:12 crc kubenswrapper[4805]: I1203 14:24:12.717811 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-46hn5\" (UniqueName: \"kubernetes.io/projected/e2034338-8a35-4fd2-9bbf-be733939fc25-kube-api-access-46hn5\") pod \"e2034338-8a35-4fd2-9bbf-be733939fc25\" (UID: \"e2034338-8a35-4fd2-9bbf-be733939fc25\") " Dec 03 14:24:12 crc kubenswrapper[4805]: I1203 14:24:12.717681 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e2034338-8a35-4fd2-9bbf-be733939fc25-service-ca" (OuterVolumeSpecName: "service-ca") pod "e2034338-8a35-4fd2-9bbf-be733939fc25" (UID: "e2034338-8a35-4fd2-9bbf-be733939fc25"). InnerVolumeSpecName "service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 14:24:12 crc kubenswrapper[4805]: I1203 14:24:12.717672 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e2034338-8a35-4fd2-9bbf-be733939fc25-oauth-serving-cert" (OuterVolumeSpecName: "oauth-serving-cert") pod "e2034338-8a35-4fd2-9bbf-be733939fc25" (UID: "e2034338-8a35-4fd2-9bbf-be733939fc25"). InnerVolumeSpecName "oauth-serving-cert". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 14:24:12 crc kubenswrapper[4805]: I1203 14:24:12.717920 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/e2034338-8a35-4fd2-9bbf-be733939fc25-trusted-ca-bundle\") pod \"e2034338-8a35-4fd2-9bbf-be733939fc25\" (UID: \"e2034338-8a35-4fd2-9bbf-be733939fc25\") " Dec 03 14:24:12 crc kubenswrapper[4805]: I1203 14:24:12.718153 4805 reconciler_common.go:293] "Volume detached for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/e2034338-8a35-4fd2-9bbf-be733939fc25-service-ca\") on node \"crc\" DevicePath \"\"" Dec 03 14:24:12 crc kubenswrapper[4805]: I1203 14:24:12.718163 4805 reconciler_common.go:293] "Volume detached for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/e2034338-8a35-4fd2-9bbf-be733939fc25-oauth-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 03 14:24:12 crc kubenswrapper[4805]: I1203 14:24:12.718605 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e2034338-8a35-4fd2-9bbf-be733939fc25-console-config" (OuterVolumeSpecName: "console-config") pod "e2034338-8a35-4fd2-9bbf-be733939fc25" (UID: "e2034338-8a35-4fd2-9bbf-be733939fc25"). InnerVolumeSpecName "console-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 14:24:12 crc kubenswrapper[4805]: I1203 14:24:12.719202 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e2034338-8a35-4fd2-9bbf-be733939fc25-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "e2034338-8a35-4fd2-9bbf-be733939fc25" (UID: "e2034338-8a35-4fd2-9bbf-be733939fc25"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 14:24:12 crc kubenswrapper[4805]: I1203 14:24:12.723713 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e2034338-8a35-4fd2-9bbf-be733939fc25-kube-api-access-46hn5" (OuterVolumeSpecName: "kube-api-access-46hn5") pod "e2034338-8a35-4fd2-9bbf-be733939fc25" (UID: "e2034338-8a35-4fd2-9bbf-be733939fc25"). InnerVolumeSpecName "kube-api-access-46hn5". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 14:24:12 crc kubenswrapper[4805]: I1203 14:24:12.727856 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e2034338-8a35-4fd2-9bbf-be733939fc25-console-oauth-config" (OuterVolumeSpecName: "console-oauth-config") pod "e2034338-8a35-4fd2-9bbf-be733939fc25" (UID: "e2034338-8a35-4fd2-9bbf-be733939fc25"). InnerVolumeSpecName "console-oauth-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 14:24:12 crc kubenswrapper[4805]: I1203 14:24:12.731897 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e2034338-8a35-4fd2-9bbf-be733939fc25-console-serving-cert" (OuterVolumeSpecName: "console-serving-cert") pod "e2034338-8a35-4fd2-9bbf-be733939fc25" (UID: "e2034338-8a35-4fd2-9bbf-be733939fc25"). InnerVolumeSpecName "console-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 14:24:12 crc kubenswrapper[4805]: I1203 14:24:12.818961 4805 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/e2034338-8a35-4fd2-9bbf-be733939fc25-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 14:24:12 crc kubenswrapper[4805]: I1203 14:24:12.818992 4805 reconciler_common.go:293] "Volume detached for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/e2034338-8a35-4fd2-9bbf-be733939fc25-console-oauth-config\") on node \"crc\" DevicePath \"\"" Dec 03 14:24:12 crc kubenswrapper[4805]: I1203 14:24:12.819002 4805 reconciler_common.go:293] "Volume detached for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/e2034338-8a35-4fd2-9bbf-be733939fc25-console-config\") on node \"crc\" DevicePath \"\"" Dec 03 14:24:12 crc kubenswrapper[4805]: I1203 14:24:12.819011 4805 reconciler_common.go:293] "Volume detached for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/e2034338-8a35-4fd2-9bbf-be733939fc25-console-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 03 14:24:12 crc kubenswrapper[4805]: I1203 14:24:12.819021 4805 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-46hn5\" (UniqueName: \"kubernetes.io/projected/e2034338-8a35-4fd2-9bbf-be733939fc25-kube-api-access-46hn5\") on node \"crc\" DevicePath \"\"" Dec 03 14:24:12 crc kubenswrapper[4805]: I1203 14:24:12.981501 4805 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-console/console-f9d7485db-pp9tv"] Dec 03 14:24:12 crc kubenswrapper[4805]: I1203 14:24:12.986117 4805 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-console/console-f9d7485db-pp9tv"] Dec 03 14:24:14 crc kubenswrapper[4805]: I1203 14:24:14.656987 4805 generic.go:334] "Generic (PLEG): container finished" podID="449e7507-4d25-4399-b1bd-b174d65a5041" containerID="dd3bdc8ebe2699fc19b5817fa93daf8bff0b77e652cf870d6e3c75452eacace5" exitCode=0 Dec 03 14:24:14 crc kubenswrapper[4805]: I1203 14:24:14.657083 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83glh2z" event={"ID":"449e7507-4d25-4399-b1bd-b174d65a5041","Type":"ContainerDied","Data":"dd3bdc8ebe2699fc19b5817fa93daf8bff0b77e652cf870d6e3c75452eacace5"} Dec 03 14:24:14 crc kubenswrapper[4805]: I1203 14:24:14.710661 4805 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e2034338-8a35-4fd2-9bbf-be733939fc25" path="/var/lib/kubelet/pods/e2034338-8a35-4fd2-9bbf-be733939fc25/volumes" Dec 03 14:24:15 crc kubenswrapper[4805]: I1203 14:24:15.666489 4805 generic.go:334] "Generic (PLEG): container finished" podID="449e7507-4d25-4399-b1bd-b174d65a5041" containerID="0295db85e2a8be6fba2d7a293d8cd8a6a4bb99c5cdbd59ea7524a8aa6f301229" exitCode=0 Dec 03 14:24:15 crc kubenswrapper[4805]: I1203 14:24:15.666581 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83glh2z" event={"ID":"449e7507-4d25-4399-b1bd-b174d65a5041","Type":"ContainerDied","Data":"0295db85e2a8be6fba2d7a293d8cd8a6a4bb99c5cdbd59ea7524a8aa6f301229"} Dec 03 14:24:16 crc kubenswrapper[4805]: I1203 14:24:16.985904 4805 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83glh2z" Dec 03 14:24:17 crc kubenswrapper[4805]: I1203 14:24:17.075561 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kl88t\" (UniqueName: \"kubernetes.io/projected/449e7507-4d25-4399-b1bd-b174d65a5041-kube-api-access-kl88t\") pod \"449e7507-4d25-4399-b1bd-b174d65a5041\" (UID: \"449e7507-4d25-4399-b1bd-b174d65a5041\") " Dec 03 14:24:17 crc kubenswrapper[4805]: I1203 14:24:17.084326 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/449e7507-4d25-4399-b1bd-b174d65a5041-kube-api-access-kl88t" (OuterVolumeSpecName: "kube-api-access-kl88t") pod "449e7507-4d25-4399-b1bd-b174d65a5041" (UID: "449e7507-4d25-4399-b1bd-b174d65a5041"). InnerVolumeSpecName "kube-api-access-kl88t". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 14:24:17 crc kubenswrapper[4805]: I1203 14:24:17.176694 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/449e7507-4d25-4399-b1bd-b174d65a5041-bundle\") pod \"449e7507-4d25-4399-b1bd-b174d65a5041\" (UID: \"449e7507-4d25-4399-b1bd-b174d65a5041\") " Dec 03 14:24:17 crc kubenswrapper[4805]: I1203 14:24:17.176804 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/449e7507-4d25-4399-b1bd-b174d65a5041-util\") pod \"449e7507-4d25-4399-b1bd-b174d65a5041\" (UID: \"449e7507-4d25-4399-b1bd-b174d65a5041\") " Dec 03 14:24:17 crc kubenswrapper[4805]: I1203 14:24:17.177307 4805 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kl88t\" (UniqueName: \"kubernetes.io/projected/449e7507-4d25-4399-b1bd-b174d65a5041-kube-api-access-kl88t\") on node \"crc\" DevicePath \"\"" Dec 03 14:24:17 crc kubenswrapper[4805]: I1203 14:24:17.179033 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/449e7507-4d25-4399-b1bd-b174d65a5041-bundle" (OuterVolumeSpecName: "bundle") pod "449e7507-4d25-4399-b1bd-b174d65a5041" (UID: "449e7507-4d25-4399-b1bd-b174d65a5041"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 14:24:17 crc kubenswrapper[4805]: I1203 14:24:17.279287 4805 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/449e7507-4d25-4399-b1bd-b174d65a5041-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 14:24:17 crc kubenswrapper[4805]: I1203 14:24:17.330793 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/449e7507-4d25-4399-b1bd-b174d65a5041-util" (OuterVolumeSpecName: "util") pod "449e7507-4d25-4399-b1bd-b174d65a5041" (UID: "449e7507-4d25-4399-b1bd-b174d65a5041"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 14:24:17 crc kubenswrapper[4805]: I1203 14:24:17.380664 4805 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/449e7507-4d25-4399-b1bd-b174d65a5041-util\") on node \"crc\" DevicePath \"\"" Dec 03 14:24:17 crc kubenswrapper[4805]: I1203 14:24:17.689998 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83glh2z" event={"ID":"449e7507-4d25-4399-b1bd-b174d65a5041","Type":"ContainerDied","Data":"b6cc18fa11d33b5315a32a5fe4285abacc55ac3ab200ba2d2c8f72023f2e8c86"} Dec 03 14:24:17 crc kubenswrapper[4805]: I1203 14:24:17.690476 4805 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="b6cc18fa11d33b5315a32a5fe4285abacc55ac3ab200ba2d2c8f72023f2e8c86" Dec 03 14:24:17 crc kubenswrapper[4805]: I1203 14:24:17.690244 4805 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83glh2z" Dec 03 14:24:31 crc kubenswrapper[4805]: I1203 14:24:31.469817 4805 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/metallb-operator-controller-manager-76979bd97f-kgnj5"] Dec 03 14:24:31 crc kubenswrapper[4805]: E1203 14:24:31.470621 4805 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="449e7507-4d25-4399-b1bd-b174d65a5041" containerName="util" Dec 03 14:24:31 crc kubenswrapper[4805]: I1203 14:24:31.470636 4805 state_mem.go:107] "Deleted CPUSet assignment" podUID="449e7507-4d25-4399-b1bd-b174d65a5041" containerName="util" Dec 03 14:24:31 crc kubenswrapper[4805]: E1203 14:24:31.470649 4805 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e2034338-8a35-4fd2-9bbf-be733939fc25" containerName="console" Dec 03 14:24:31 crc kubenswrapper[4805]: I1203 14:24:31.470657 4805 state_mem.go:107] "Deleted CPUSet assignment" podUID="e2034338-8a35-4fd2-9bbf-be733939fc25" containerName="console" Dec 03 14:24:31 crc kubenswrapper[4805]: E1203 14:24:31.470676 4805 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="449e7507-4d25-4399-b1bd-b174d65a5041" containerName="pull" Dec 03 14:24:31 crc kubenswrapper[4805]: I1203 14:24:31.470684 4805 state_mem.go:107] "Deleted CPUSet assignment" podUID="449e7507-4d25-4399-b1bd-b174d65a5041" containerName="pull" Dec 03 14:24:31 crc kubenswrapper[4805]: E1203 14:24:31.470696 4805 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="449e7507-4d25-4399-b1bd-b174d65a5041" containerName="extract" Dec 03 14:24:31 crc kubenswrapper[4805]: I1203 14:24:31.470704 4805 state_mem.go:107] "Deleted CPUSet assignment" podUID="449e7507-4d25-4399-b1bd-b174d65a5041" containerName="extract" Dec 03 14:24:31 crc kubenswrapper[4805]: E1203 14:24:31.470720 4805 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e2055b14-d7a1-4ac4-9c2e-ef12644113b2" containerName="extract-content" Dec 03 14:24:31 crc kubenswrapper[4805]: I1203 14:24:31.470727 4805 state_mem.go:107] "Deleted CPUSet assignment" podUID="e2055b14-d7a1-4ac4-9c2e-ef12644113b2" containerName="extract-content" Dec 03 14:24:31 crc kubenswrapper[4805]: E1203 14:24:31.470739 4805 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e2055b14-d7a1-4ac4-9c2e-ef12644113b2" containerName="extract-utilities" Dec 03 14:24:31 crc kubenswrapper[4805]: I1203 14:24:31.470747 4805 state_mem.go:107] "Deleted CPUSet assignment" podUID="e2055b14-d7a1-4ac4-9c2e-ef12644113b2" containerName="extract-utilities" Dec 03 14:24:31 crc kubenswrapper[4805]: E1203 14:24:31.470762 4805 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e2055b14-d7a1-4ac4-9c2e-ef12644113b2" containerName="registry-server" Dec 03 14:24:31 crc kubenswrapper[4805]: I1203 14:24:31.470769 4805 state_mem.go:107] "Deleted CPUSet assignment" podUID="e2055b14-d7a1-4ac4-9c2e-ef12644113b2" containerName="registry-server" Dec 03 14:24:31 crc kubenswrapper[4805]: I1203 14:24:31.470932 4805 memory_manager.go:354] "RemoveStaleState removing state" podUID="e2034338-8a35-4fd2-9bbf-be733939fc25" containerName="console" Dec 03 14:24:31 crc kubenswrapper[4805]: I1203 14:24:31.470948 4805 memory_manager.go:354] "RemoveStaleState removing state" podUID="e2055b14-d7a1-4ac4-9c2e-ef12644113b2" containerName="registry-server" Dec 03 14:24:31 crc kubenswrapper[4805]: I1203 14:24:31.470962 4805 memory_manager.go:354] "RemoveStaleState removing state" podUID="449e7507-4d25-4399-b1bd-b174d65a5041" containerName="extract" Dec 03 14:24:31 crc kubenswrapper[4805]: I1203 14:24:31.471417 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-controller-manager-76979bd97f-kgnj5" Dec 03 14:24:31 crc kubenswrapper[4805]: I1203 14:24:31.475528 4805 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-operator-controller-manager-service-cert" Dec 03 14:24:31 crc kubenswrapper[4805]: I1203 14:24:31.475736 4805 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"kube-root-ca.crt" Dec 03 14:24:31 crc kubenswrapper[4805]: I1203 14:24:31.475808 4805 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"openshift-service-ca.crt" Dec 03 14:24:31 crc kubenswrapper[4805]: I1203 14:24:31.475888 4805 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"manager-account-dockercfg-7nzvt" Dec 03 14:24:31 crc kubenswrapper[4805]: I1203 14:24:31.475929 4805 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-operator-webhook-server-cert" Dec 03 14:24:31 crc kubenswrapper[4805]: I1203 14:24:31.488805 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-controller-manager-76979bd97f-kgnj5"] Dec 03 14:24:31 crc kubenswrapper[4805]: I1203 14:24:31.612017 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/33acb320-cfe5-4d01-9543-f56ec8c33633-webhook-cert\") pod \"metallb-operator-controller-manager-76979bd97f-kgnj5\" (UID: \"33acb320-cfe5-4d01-9543-f56ec8c33633\") " pod="metallb-system/metallb-operator-controller-manager-76979bd97f-kgnj5" Dec 03 14:24:31 crc kubenswrapper[4805]: I1203 14:24:31.612078 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-k8x2g\" (UniqueName: \"kubernetes.io/projected/33acb320-cfe5-4d01-9543-f56ec8c33633-kube-api-access-k8x2g\") pod \"metallb-operator-controller-manager-76979bd97f-kgnj5\" (UID: \"33acb320-cfe5-4d01-9543-f56ec8c33633\") " pod="metallb-system/metallb-operator-controller-manager-76979bd97f-kgnj5" Dec 03 14:24:31 crc kubenswrapper[4805]: I1203 14:24:31.612206 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/33acb320-cfe5-4d01-9543-f56ec8c33633-apiservice-cert\") pod \"metallb-operator-controller-manager-76979bd97f-kgnj5\" (UID: \"33acb320-cfe5-4d01-9543-f56ec8c33633\") " pod="metallb-system/metallb-operator-controller-manager-76979bd97f-kgnj5" Dec 03 14:24:31 crc kubenswrapper[4805]: I1203 14:24:31.713462 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/33acb320-cfe5-4d01-9543-f56ec8c33633-webhook-cert\") pod \"metallb-operator-controller-manager-76979bd97f-kgnj5\" (UID: \"33acb320-cfe5-4d01-9543-f56ec8c33633\") " pod="metallb-system/metallb-operator-controller-manager-76979bd97f-kgnj5" Dec 03 14:24:31 crc kubenswrapper[4805]: I1203 14:24:31.713542 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-k8x2g\" (UniqueName: \"kubernetes.io/projected/33acb320-cfe5-4d01-9543-f56ec8c33633-kube-api-access-k8x2g\") pod \"metallb-operator-controller-manager-76979bd97f-kgnj5\" (UID: \"33acb320-cfe5-4d01-9543-f56ec8c33633\") " pod="metallb-system/metallb-operator-controller-manager-76979bd97f-kgnj5" Dec 03 14:24:31 crc kubenswrapper[4805]: I1203 14:24:31.713625 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/33acb320-cfe5-4d01-9543-f56ec8c33633-apiservice-cert\") pod \"metallb-operator-controller-manager-76979bd97f-kgnj5\" (UID: \"33acb320-cfe5-4d01-9543-f56ec8c33633\") " pod="metallb-system/metallb-operator-controller-manager-76979bd97f-kgnj5" Dec 03 14:24:31 crc kubenswrapper[4805]: I1203 14:24:31.721046 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/33acb320-cfe5-4d01-9543-f56ec8c33633-apiservice-cert\") pod \"metallb-operator-controller-manager-76979bd97f-kgnj5\" (UID: \"33acb320-cfe5-4d01-9543-f56ec8c33633\") " pod="metallb-system/metallb-operator-controller-manager-76979bd97f-kgnj5" Dec 03 14:24:31 crc kubenswrapper[4805]: I1203 14:24:31.721786 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/33acb320-cfe5-4d01-9543-f56ec8c33633-webhook-cert\") pod \"metallb-operator-controller-manager-76979bd97f-kgnj5\" (UID: \"33acb320-cfe5-4d01-9543-f56ec8c33633\") " pod="metallb-system/metallb-operator-controller-manager-76979bd97f-kgnj5" Dec 03 14:24:31 crc kubenswrapper[4805]: I1203 14:24:31.723185 4805 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/metallb-operator-webhook-server-5587f68f8b-4vcdf"] Dec 03 14:24:31 crc kubenswrapper[4805]: I1203 14:24:31.724205 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-webhook-server-5587f68f8b-4vcdf" Dec 03 14:24:31 crc kubenswrapper[4805]: I1203 14:24:31.726358 4805 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"controller-dockercfg-tn89z" Dec 03 14:24:31 crc kubenswrapper[4805]: I1203 14:24:31.729213 4805 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-webhook-cert" Dec 03 14:24:31 crc kubenswrapper[4805]: I1203 14:24:31.735553 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-k8x2g\" (UniqueName: \"kubernetes.io/projected/33acb320-cfe5-4d01-9543-f56ec8c33633-kube-api-access-k8x2g\") pod \"metallb-operator-controller-manager-76979bd97f-kgnj5\" (UID: \"33acb320-cfe5-4d01-9543-f56ec8c33633\") " pod="metallb-system/metallb-operator-controller-manager-76979bd97f-kgnj5" Dec 03 14:24:31 crc kubenswrapper[4805]: I1203 14:24:31.736513 4805 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-operator-webhook-server-service-cert" Dec 03 14:24:31 crc kubenswrapper[4805]: I1203 14:24:31.748217 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-webhook-server-5587f68f8b-4vcdf"] Dec 03 14:24:31 crc kubenswrapper[4805]: I1203 14:24:31.786057 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-controller-manager-76979bd97f-kgnj5" Dec 03 14:24:31 crc kubenswrapper[4805]: I1203 14:24:31.814726 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/f2e7792c-c1c1-41b0-8e90-fe082ba1c366-webhook-cert\") pod \"metallb-operator-webhook-server-5587f68f8b-4vcdf\" (UID: \"f2e7792c-c1c1-41b0-8e90-fe082ba1c366\") " pod="metallb-system/metallb-operator-webhook-server-5587f68f8b-4vcdf" Dec 03 14:24:31 crc kubenswrapper[4805]: I1203 14:24:31.815099 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/f2e7792c-c1c1-41b0-8e90-fe082ba1c366-apiservice-cert\") pod \"metallb-operator-webhook-server-5587f68f8b-4vcdf\" (UID: \"f2e7792c-c1c1-41b0-8e90-fe082ba1c366\") " pod="metallb-system/metallb-operator-webhook-server-5587f68f8b-4vcdf" Dec 03 14:24:31 crc kubenswrapper[4805]: I1203 14:24:31.815137 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-69l2r\" (UniqueName: \"kubernetes.io/projected/f2e7792c-c1c1-41b0-8e90-fe082ba1c366-kube-api-access-69l2r\") pod \"metallb-operator-webhook-server-5587f68f8b-4vcdf\" (UID: \"f2e7792c-c1c1-41b0-8e90-fe082ba1c366\") " pod="metallb-system/metallb-operator-webhook-server-5587f68f8b-4vcdf" Dec 03 14:24:31 crc kubenswrapper[4805]: I1203 14:24:31.916525 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/f2e7792c-c1c1-41b0-8e90-fe082ba1c366-apiservice-cert\") pod \"metallb-operator-webhook-server-5587f68f8b-4vcdf\" (UID: \"f2e7792c-c1c1-41b0-8e90-fe082ba1c366\") " pod="metallb-system/metallb-operator-webhook-server-5587f68f8b-4vcdf" Dec 03 14:24:31 crc kubenswrapper[4805]: I1203 14:24:31.916585 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-69l2r\" (UniqueName: \"kubernetes.io/projected/f2e7792c-c1c1-41b0-8e90-fe082ba1c366-kube-api-access-69l2r\") pod \"metallb-operator-webhook-server-5587f68f8b-4vcdf\" (UID: \"f2e7792c-c1c1-41b0-8e90-fe082ba1c366\") " pod="metallb-system/metallb-operator-webhook-server-5587f68f8b-4vcdf" Dec 03 14:24:31 crc kubenswrapper[4805]: I1203 14:24:31.916636 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/f2e7792c-c1c1-41b0-8e90-fe082ba1c366-webhook-cert\") pod \"metallb-operator-webhook-server-5587f68f8b-4vcdf\" (UID: \"f2e7792c-c1c1-41b0-8e90-fe082ba1c366\") " pod="metallb-system/metallb-operator-webhook-server-5587f68f8b-4vcdf" Dec 03 14:24:31 crc kubenswrapper[4805]: I1203 14:24:31.936352 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/f2e7792c-c1c1-41b0-8e90-fe082ba1c366-webhook-cert\") pod \"metallb-operator-webhook-server-5587f68f8b-4vcdf\" (UID: \"f2e7792c-c1c1-41b0-8e90-fe082ba1c366\") " pod="metallb-system/metallb-operator-webhook-server-5587f68f8b-4vcdf" Dec 03 14:24:31 crc kubenswrapper[4805]: I1203 14:24:31.943028 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/f2e7792c-c1c1-41b0-8e90-fe082ba1c366-apiservice-cert\") pod \"metallb-operator-webhook-server-5587f68f8b-4vcdf\" (UID: \"f2e7792c-c1c1-41b0-8e90-fe082ba1c366\") " pod="metallb-system/metallb-operator-webhook-server-5587f68f8b-4vcdf" Dec 03 14:24:31 crc kubenswrapper[4805]: I1203 14:24:31.945165 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-69l2r\" (UniqueName: \"kubernetes.io/projected/f2e7792c-c1c1-41b0-8e90-fe082ba1c366-kube-api-access-69l2r\") pod \"metallb-operator-webhook-server-5587f68f8b-4vcdf\" (UID: \"f2e7792c-c1c1-41b0-8e90-fe082ba1c366\") " pod="metallb-system/metallb-operator-webhook-server-5587f68f8b-4vcdf" Dec 03 14:24:32 crc kubenswrapper[4805]: I1203 14:24:32.082980 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-webhook-server-5587f68f8b-4vcdf" Dec 03 14:24:32 crc kubenswrapper[4805]: I1203 14:24:32.109774 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-controller-manager-76979bd97f-kgnj5"] Dec 03 14:24:32 crc kubenswrapper[4805]: W1203 14:24:32.122396 4805 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod33acb320_cfe5_4d01_9543_f56ec8c33633.slice/crio-f30c303d712686df4b96d548b471b14ce421d7696c827a21eea54b6ea16f63fb WatchSource:0}: Error finding container f30c303d712686df4b96d548b471b14ce421d7696c827a21eea54b6ea16f63fb: Status 404 returned error can't find the container with id f30c303d712686df4b96d548b471b14ce421d7696c827a21eea54b6ea16f63fb Dec 03 14:24:32 crc kubenswrapper[4805]: I1203 14:24:32.300551 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-webhook-server-5587f68f8b-4vcdf"] Dec 03 14:24:32 crc kubenswrapper[4805]: W1203 14:24:32.305415 4805 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podf2e7792c_c1c1_41b0_8e90_fe082ba1c366.slice/crio-cb18da8409f217050cea785b0f3e7b24bcf651840a383debb504a8f5ddc14c20 WatchSource:0}: Error finding container cb18da8409f217050cea785b0f3e7b24bcf651840a383debb504a8f5ddc14c20: Status 404 returned error can't find the container with id cb18da8409f217050cea785b0f3e7b24bcf651840a383debb504a8f5ddc14c20 Dec 03 14:24:32 crc kubenswrapper[4805]: I1203 14:24:32.780827 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-controller-manager-76979bd97f-kgnj5" event={"ID":"33acb320-cfe5-4d01-9543-f56ec8c33633","Type":"ContainerStarted","Data":"f30c303d712686df4b96d548b471b14ce421d7696c827a21eea54b6ea16f63fb"} Dec 03 14:24:32 crc kubenswrapper[4805]: I1203 14:24:32.782354 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-webhook-server-5587f68f8b-4vcdf" event={"ID":"f2e7792c-c1c1-41b0-8e90-fe082ba1c366","Type":"ContainerStarted","Data":"cb18da8409f217050cea785b0f3e7b24bcf651840a383debb504a8f5ddc14c20"} Dec 03 14:24:33 crc kubenswrapper[4805]: I1203 14:24:33.271924 4805 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-kkln5"] Dec 03 14:24:33 crc kubenswrapper[4805]: I1203 14:24:33.274310 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-kkln5" Dec 03 14:24:33 crc kubenswrapper[4805]: I1203 14:24:33.284761 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-kkln5"] Dec 03 14:24:33 crc kubenswrapper[4805]: I1203 14:24:33.338042 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/19095d12-19a7-44bd-a321-e783b685ac57-catalog-content\") pod \"community-operators-kkln5\" (UID: \"19095d12-19a7-44bd-a321-e783b685ac57\") " pod="openshift-marketplace/community-operators-kkln5" Dec 03 14:24:33 crc kubenswrapper[4805]: I1203 14:24:33.338103 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/19095d12-19a7-44bd-a321-e783b685ac57-utilities\") pod \"community-operators-kkln5\" (UID: \"19095d12-19a7-44bd-a321-e783b685ac57\") " pod="openshift-marketplace/community-operators-kkln5" Dec 03 14:24:33 crc kubenswrapper[4805]: I1203 14:24:33.338393 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-b54hj\" (UniqueName: \"kubernetes.io/projected/19095d12-19a7-44bd-a321-e783b685ac57-kube-api-access-b54hj\") pod \"community-operators-kkln5\" (UID: \"19095d12-19a7-44bd-a321-e783b685ac57\") " pod="openshift-marketplace/community-operators-kkln5" Dec 03 14:24:33 crc kubenswrapper[4805]: I1203 14:24:33.439259 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-b54hj\" (UniqueName: \"kubernetes.io/projected/19095d12-19a7-44bd-a321-e783b685ac57-kube-api-access-b54hj\") pod \"community-operators-kkln5\" (UID: \"19095d12-19a7-44bd-a321-e783b685ac57\") " pod="openshift-marketplace/community-operators-kkln5" Dec 03 14:24:33 crc kubenswrapper[4805]: I1203 14:24:33.439406 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/19095d12-19a7-44bd-a321-e783b685ac57-catalog-content\") pod \"community-operators-kkln5\" (UID: \"19095d12-19a7-44bd-a321-e783b685ac57\") " pod="openshift-marketplace/community-operators-kkln5" Dec 03 14:24:33 crc kubenswrapper[4805]: I1203 14:24:33.439460 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/19095d12-19a7-44bd-a321-e783b685ac57-utilities\") pod \"community-operators-kkln5\" (UID: \"19095d12-19a7-44bd-a321-e783b685ac57\") " pod="openshift-marketplace/community-operators-kkln5" Dec 03 14:24:33 crc kubenswrapper[4805]: I1203 14:24:33.440020 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/19095d12-19a7-44bd-a321-e783b685ac57-utilities\") pod \"community-operators-kkln5\" (UID: \"19095d12-19a7-44bd-a321-e783b685ac57\") " pod="openshift-marketplace/community-operators-kkln5" Dec 03 14:24:33 crc kubenswrapper[4805]: I1203 14:24:33.440118 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/19095d12-19a7-44bd-a321-e783b685ac57-catalog-content\") pod \"community-operators-kkln5\" (UID: \"19095d12-19a7-44bd-a321-e783b685ac57\") " pod="openshift-marketplace/community-operators-kkln5" Dec 03 14:24:33 crc kubenswrapper[4805]: I1203 14:24:33.464221 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-b54hj\" (UniqueName: \"kubernetes.io/projected/19095d12-19a7-44bd-a321-e783b685ac57-kube-api-access-b54hj\") pod \"community-operators-kkln5\" (UID: \"19095d12-19a7-44bd-a321-e783b685ac57\") " pod="openshift-marketplace/community-operators-kkln5" Dec 03 14:24:33 crc kubenswrapper[4805]: I1203 14:24:33.605158 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-kkln5" Dec 03 14:24:33 crc kubenswrapper[4805]: I1203 14:24:33.885021 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-kkln5"] Dec 03 14:24:34 crc kubenswrapper[4805]: I1203 14:24:34.806560 4805 generic.go:334] "Generic (PLEG): container finished" podID="19095d12-19a7-44bd-a321-e783b685ac57" containerID="398bb8f0e85bfa7009fd599f7bf6477672fe06593e2f5e0f80a8c3dfe321a250" exitCode=0 Dec 03 14:24:34 crc kubenswrapper[4805]: I1203 14:24:34.806745 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-kkln5" event={"ID":"19095d12-19a7-44bd-a321-e783b685ac57","Type":"ContainerDied","Data":"398bb8f0e85bfa7009fd599f7bf6477672fe06593e2f5e0f80a8c3dfe321a250"} Dec 03 14:24:34 crc kubenswrapper[4805]: I1203 14:24:34.807062 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-kkln5" event={"ID":"19095d12-19a7-44bd-a321-e783b685ac57","Type":"ContainerStarted","Data":"e159eb14a80d6c484e8cb91eed9324e523dd81c1f7deb966f80f2ad8dffbb55e"} Dec 03 14:24:38 crc kubenswrapper[4805]: I1203 14:24:38.832405 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-webhook-server-5587f68f8b-4vcdf" event={"ID":"f2e7792c-c1c1-41b0-8e90-fe082ba1c366","Type":"ContainerStarted","Data":"5c010775959b80b13a900737f7f3f3f5d5f7ec8111c669279e01556a23cfebe0"} Dec 03 14:24:38 crc kubenswrapper[4805]: I1203 14:24:38.833329 4805 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/metallb-operator-webhook-server-5587f68f8b-4vcdf" Dec 03 14:24:38 crc kubenswrapper[4805]: I1203 14:24:38.835390 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-controller-manager-76979bd97f-kgnj5" event={"ID":"33acb320-cfe5-4d01-9543-f56ec8c33633","Type":"ContainerStarted","Data":"3be17b9412825975650e555a47cae4c24a27b9a9eedd7e744ddb3bcd51ad1b20"} Dec 03 14:24:38 crc kubenswrapper[4805]: I1203 14:24:38.835528 4805 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/metallb-operator-controller-manager-76979bd97f-kgnj5" Dec 03 14:24:38 crc kubenswrapper[4805]: I1203 14:24:38.855934 4805 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/metallb-operator-webhook-server-5587f68f8b-4vcdf" podStartSLOduration=1.932207716 podStartE2EDuration="7.855913581s" podCreationTimestamp="2025-12-03 14:24:31 +0000 UTC" firstStartedPulling="2025-12-03 14:24:32.308334688 +0000 UTC m=+901.971251611" lastFinishedPulling="2025-12-03 14:24:38.232040553 +0000 UTC m=+907.894957476" observedRunningTime="2025-12-03 14:24:38.850730571 +0000 UTC m=+908.513647574" watchObservedRunningTime="2025-12-03 14:24:38.855913581 +0000 UTC m=+908.518830504" Dec 03 14:24:38 crc kubenswrapper[4805]: I1203 14:24:38.882030 4805 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/metallb-operator-controller-manager-76979bd97f-kgnj5" podStartSLOduration=1.796051862 podStartE2EDuration="7.882015883s" podCreationTimestamp="2025-12-03 14:24:31 +0000 UTC" firstStartedPulling="2025-12-03 14:24:32.125302273 +0000 UTC m=+901.788219196" lastFinishedPulling="2025-12-03 14:24:38.211266294 +0000 UTC m=+907.874183217" observedRunningTime="2025-12-03 14:24:38.880312147 +0000 UTC m=+908.543229070" watchObservedRunningTime="2025-12-03 14:24:38.882015883 +0000 UTC m=+908.544932796" Dec 03 14:24:39 crc kubenswrapper[4805]: I1203 14:24:39.847879 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-kkln5" event={"ID":"19095d12-19a7-44bd-a321-e783b685ac57","Type":"ContainerStarted","Data":"b9edb601f43b33a3dd1a39386d95ce9471f7581e364c6416d4fe38c9dc3df4b0"} Dec 03 14:24:40 crc kubenswrapper[4805]: I1203 14:24:40.859455 4805 generic.go:334] "Generic (PLEG): container finished" podID="19095d12-19a7-44bd-a321-e783b685ac57" containerID="b9edb601f43b33a3dd1a39386d95ce9471f7581e364c6416d4fe38c9dc3df4b0" exitCode=0 Dec 03 14:24:40 crc kubenswrapper[4805]: I1203 14:24:40.859594 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-kkln5" event={"ID":"19095d12-19a7-44bd-a321-e783b685ac57","Type":"ContainerDied","Data":"b9edb601f43b33a3dd1a39386d95ce9471f7581e364c6416d4fe38c9dc3df4b0"} Dec 03 14:24:41 crc kubenswrapper[4805]: I1203 14:24:41.869293 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-kkln5" event={"ID":"19095d12-19a7-44bd-a321-e783b685ac57","Type":"ContainerStarted","Data":"3f01856fe940ad969f6cad6b31518256b9156f720a789d9e10efb28d1f53b83b"} Dec 03 14:24:41 crc kubenswrapper[4805]: I1203 14:24:41.896021 4805 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-kkln5" podStartSLOduration=2.456338505 podStartE2EDuration="8.896002754s" podCreationTimestamp="2025-12-03 14:24:33 +0000 UTC" firstStartedPulling="2025-12-03 14:24:34.812294114 +0000 UTC m=+904.475211037" lastFinishedPulling="2025-12-03 14:24:41.251958353 +0000 UTC m=+910.914875286" observedRunningTime="2025-12-03 14:24:41.891440381 +0000 UTC m=+911.554357314" watchObservedRunningTime="2025-12-03 14:24:41.896002754 +0000 UTC m=+911.558919687" Dec 03 14:24:43 crc kubenswrapper[4805]: I1203 14:24:43.605740 4805 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-kkln5" Dec 03 14:24:43 crc kubenswrapper[4805]: I1203 14:24:43.606119 4805 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-kkln5" Dec 03 14:24:44 crc kubenswrapper[4805]: I1203 14:24:44.651491 4805 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/community-operators-kkln5" podUID="19095d12-19a7-44bd-a321-e783b685ac57" containerName="registry-server" probeResult="failure" output=< Dec 03 14:24:44 crc kubenswrapper[4805]: timeout: failed to connect service ":50051" within 1s Dec 03 14:24:44 crc kubenswrapper[4805]: > Dec 03 14:24:52 crc kubenswrapper[4805]: I1203 14:24:52.088832 4805 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/metallb-operator-webhook-server-5587f68f8b-4vcdf" Dec 03 14:24:53 crc kubenswrapper[4805]: I1203 14:24:53.664785 4805 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-kkln5" Dec 03 14:24:53 crc kubenswrapper[4805]: I1203 14:24:53.737413 4805 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-kkln5" Dec 03 14:24:53 crc kubenswrapper[4805]: I1203 14:24:53.907956 4805 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-kkln5"] Dec 03 14:24:54 crc kubenswrapper[4805]: I1203 14:24:54.945023 4805 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-kkln5" podUID="19095d12-19a7-44bd-a321-e783b685ac57" containerName="registry-server" containerID="cri-o://3f01856fe940ad969f6cad6b31518256b9156f720a789d9e10efb28d1f53b83b" gracePeriod=2 Dec 03 14:24:55 crc kubenswrapper[4805]: I1203 14:24:55.834589 4805 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-kkln5" Dec 03 14:24:55 crc kubenswrapper[4805]: I1203 14:24:55.951702 4805 generic.go:334] "Generic (PLEG): container finished" podID="19095d12-19a7-44bd-a321-e783b685ac57" containerID="3f01856fe940ad969f6cad6b31518256b9156f720a789d9e10efb28d1f53b83b" exitCode=0 Dec 03 14:24:55 crc kubenswrapper[4805]: I1203 14:24:55.951764 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-kkln5" event={"ID":"19095d12-19a7-44bd-a321-e783b685ac57","Type":"ContainerDied","Data":"3f01856fe940ad969f6cad6b31518256b9156f720a789d9e10efb28d1f53b83b"} Dec 03 14:24:55 crc kubenswrapper[4805]: I1203 14:24:55.951833 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-kkln5" event={"ID":"19095d12-19a7-44bd-a321-e783b685ac57","Type":"ContainerDied","Data":"e159eb14a80d6c484e8cb91eed9324e523dd81c1f7deb966f80f2ad8dffbb55e"} Dec 03 14:24:55 crc kubenswrapper[4805]: I1203 14:24:55.951827 4805 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-kkln5" Dec 03 14:24:55 crc kubenswrapper[4805]: I1203 14:24:55.951888 4805 scope.go:117] "RemoveContainer" containerID="3f01856fe940ad969f6cad6b31518256b9156f720a789d9e10efb28d1f53b83b" Dec 03 14:24:55 crc kubenswrapper[4805]: I1203 14:24:55.969656 4805 scope.go:117] "RemoveContainer" containerID="b9edb601f43b33a3dd1a39386d95ce9471f7581e364c6416d4fe38c9dc3df4b0" Dec 03 14:24:55 crc kubenswrapper[4805]: I1203 14:24:55.972688 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-b54hj\" (UniqueName: \"kubernetes.io/projected/19095d12-19a7-44bd-a321-e783b685ac57-kube-api-access-b54hj\") pod \"19095d12-19a7-44bd-a321-e783b685ac57\" (UID: \"19095d12-19a7-44bd-a321-e783b685ac57\") " Dec 03 14:24:55 crc kubenswrapper[4805]: I1203 14:24:55.972748 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/19095d12-19a7-44bd-a321-e783b685ac57-utilities\") pod \"19095d12-19a7-44bd-a321-e783b685ac57\" (UID: \"19095d12-19a7-44bd-a321-e783b685ac57\") " Dec 03 14:24:55 crc kubenswrapper[4805]: I1203 14:24:55.972782 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/19095d12-19a7-44bd-a321-e783b685ac57-catalog-content\") pod \"19095d12-19a7-44bd-a321-e783b685ac57\" (UID: \"19095d12-19a7-44bd-a321-e783b685ac57\") " Dec 03 14:24:55 crc kubenswrapper[4805]: I1203 14:24:55.974544 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/19095d12-19a7-44bd-a321-e783b685ac57-utilities" (OuterVolumeSpecName: "utilities") pod "19095d12-19a7-44bd-a321-e783b685ac57" (UID: "19095d12-19a7-44bd-a321-e783b685ac57"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 14:24:55 crc kubenswrapper[4805]: I1203 14:24:55.978408 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/19095d12-19a7-44bd-a321-e783b685ac57-kube-api-access-b54hj" (OuterVolumeSpecName: "kube-api-access-b54hj") pod "19095d12-19a7-44bd-a321-e783b685ac57" (UID: "19095d12-19a7-44bd-a321-e783b685ac57"). InnerVolumeSpecName "kube-api-access-b54hj". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 14:24:55 crc kubenswrapper[4805]: I1203 14:24:55.987182 4805 scope.go:117] "RemoveContainer" containerID="398bb8f0e85bfa7009fd599f7bf6477672fe06593e2f5e0f80a8c3dfe321a250" Dec 03 14:24:56 crc kubenswrapper[4805]: I1203 14:24:56.024324 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/19095d12-19a7-44bd-a321-e783b685ac57-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "19095d12-19a7-44bd-a321-e783b685ac57" (UID: "19095d12-19a7-44bd-a321-e783b685ac57"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 14:24:56 crc kubenswrapper[4805]: I1203 14:24:56.033684 4805 scope.go:117] "RemoveContainer" containerID="3f01856fe940ad969f6cad6b31518256b9156f720a789d9e10efb28d1f53b83b" Dec 03 14:24:56 crc kubenswrapper[4805]: E1203 14:24:56.034039 4805 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3f01856fe940ad969f6cad6b31518256b9156f720a789d9e10efb28d1f53b83b\": container with ID starting with 3f01856fe940ad969f6cad6b31518256b9156f720a789d9e10efb28d1f53b83b not found: ID does not exist" containerID="3f01856fe940ad969f6cad6b31518256b9156f720a789d9e10efb28d1f53b83b" Dec 03 14:24:56 crc kubenswrapper[4805]: I1203 14:24:56.034082 4805 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3f01856fe940ad969f6cad6b31518256b9156f720a789d9e10efb28d1f53b83b"} err="failed to get container status \"3f01856fe940ad969f6cad6b31518256b9156f720a789d9e10efb28d1f53b83b\": rpc error: code = NotFound desc = could not find container \"3f01856fe940ad969f6cad6b31518256b9156f720a789d9e10efb28d1f53b83b\": container with ID starting with 3f01856fe940ad969f6cad6b31518256b9156f720a789d9e10efb28d1f53b83b not found: ID does not exist" Dec 03 14:24:56 crc kubenswrapper[4805]: I1203 14:24:56.034109 4805 scope.go:117] "RemoveContainer" containerID="b9edb601f43b33a3dd1a39386d95ce9471f7581e364c6416d4fe38c9dc3df4b0" Dec 03 14:24:56 crc kubenswrapper[4805]: E1203 14:24:56.034432 4805 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b9edb601f43b33a3dd1a39386d95ce9471f7581e364c6416d4fe38c9dc3df4b0\": container with ID starting with b9edb601f43b33a3dd1a39386d95ce9471f7581e364c6416d4fe38c9dc3df4b0 not found: ID does not exist" containerID="b9edb601f43b33a3dd1a39386d95ce9471f7581e364c6416d4fe38c9dc3df4b0" Dec 03 14:24:56 crc kubenswrapper[4805]: I1203 14:24:56.034494 4805 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b9edb601f43b33a3dd1a39386d95ce9471f7581e364c6416d4fe38c9dc3df4b0"} err="failed to get container status \"b9edb601f43b33a3dd1a39386d95ce9471f7581e364c6416d4fe38c9dc3df4b0\": rpc error: code = NotFound desc = could not find container \"b9edb601f43b33a3dd1a39386d95ce9471f7581e364c6416d4fe38c9dc3df4b0\": container with ID starting with b9edb601f43b33a3dd1a39386d95ce9471f7581e364c6416d4fe38c9dc3df4b0 not found: ID does not exist" Dec 03 14:24:56 crc kubenswrapper[4805]: I1203 14:24:56.034526 4805 scope.go:117] "RemoveContainer" containerID="398bb8f0e85bfa7009fd599f7bf6477672fe06593e2f5e0f80a8c3dfe321a250" Dec 03 14:24:56 crc kubenswrapper[4805]: E1203 14:24:56.034829 4805 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"398bb8f0e85bfa7009fd599f7bf6477672fe06593e2f5e0f80a8c3dfe321a250\": container with ID starting with 398bb8f0e85bfa7009fd599f7bf6477672fe06593e2f5e0f80a8c3dfe321a250 not found: ID does not exist" containerID="398bb8f0e85bfa7009fd599f7bf6477672fe06593e2f5e0f80a8c3dfe321a250" Dec 03 14:24:56 crc kubenswrapper[4805]: I1203 14:24:56.035095 4805 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"398bb8f0e85bfa7009fd599f7bf6477672fe06593e2f5e0f80a8c3dfe321a250"} err="failed to get container status \"398bb8f0e85bfa7009fd599f7bf6477672fe06593e2f5e0f80a8c3dfe321a250\": rpc error: code = NotFound desc = could not find container \"398bb8f0e85bfa7009fd599f7bf6477672fe06593e2f5e0f80a8c3dfe321a250\": container with ID starting with 398bb8f0e85bfa7009fd599f7bf6477672fe06593e2f5e0f80a8c3dfe321a250 not found: ID does not exist" Dec 03 14:24:56 crc kubenswrapper[4805]: I1203 14:24:56.074722 4805 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-b54hj\" (UniqueName: \"kubernetes.io/projected/19095d12-19a7-44bd-a321-e783b685ac57-kube-api-access-b54hj\") on node \"crc\" DevicePath \"\"" Dec 03 14:24:56 crc kubenswrapper[4805]: I1203 14:24:56.074760 4805 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/19095d12-19a7-44bd-a321-e783b685ac57-utilities\") on node \"crc\" DevicePath \"\"" Dec 03 14:24:56 crc kubenswrapper[4805]: I1203 14:24:56.074769 4805 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/19095d12-19a7-44bd-a321-e783b685ac57-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 03 14:24:56 crc kubenswrapper[4805]: I1203 14:24:56.279734 4805 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-kkln5"] Dec 03 14:24:56 crc kubenswrapper[4805]: I1203 14:24:56.286080 4805 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-kkln5"] Dec 03 14:24:56 crc kubenswrapper[4805]: I1203 14:24:56.703186 4805 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="19095d12-19a7-44bd-a321-e783b685ac57" path="/var/lib/kubelet/pods/19095d12-19a7-44bd-a321-e783b685ac57/volumes" Dec 03 14:25:11 crc kubenswrapper[4805]: I1203 14:25:11.790782 4805 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/metallb-operator-controller-manager-76979bd97f-kgnj5" Dec 03 14:25:12 crc kubenswrapper[4805]: I1203 14:25:12.528028 4805 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/frr-k8s-jsrqx"] Dec 03 14:25:12 crc kubenswrapper[4805]: E1203 14:25:12.528429 4805 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="19095d12-19a7-44bd-a321-e783b685ac57" containerName="extract-utilities" Dec 03 14:25:12 crc kubenswrapper[4805]: I1203 14:25:12.528443 4805 state_mem.go:107] "Deleted CPUSet assignment" podUID="19095d12-19a7-44bd-a321-e783b685ac57" containerName="extract-utilities" Dec 03 14:25:12 crc kubenswrapper[4805]: E1203 14:25:12.528462 4805 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="19095d12-19a7-44bd-a321-e783b685ac57" containerName="registry-server" Dec 03 14:25:12 crc kubenswrapper[4805]: I1203 14:25:12.528468 4805 state_mem.go:107] "Deleted CPUSet assignment" podUID="19095d12-19a7-44bd-a321-e783b685ac57" containerName="registry-server" Dec 03 14:25:12 crc kubenswrapper[4805]: E1203 14:25:12.528485 4805 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="19095d12-19a7-44bd-a321-e783b685ac57" containerName="extract-content" Dec 03 14:25:12 crc kubenswrapper[4805]: I1203 14:25:12.528491 4805 state_mem.go:107] "Deleted CPUSet assignment" podUID="19095d12-19a7-44bd-a321-e783b685ac57" containerName="extract-content" Dec 03 14:25:12 crc kubenswrapper[4805]: I1203 14:25:12.528726 4805 memory_manager.go:354] "RemoveStaleState removing state" podUID="19095d12-19a7-44bd-a321-e783b685ac57" containerName="registry-server" Dec 03 14:25:12 crc kubenswrapper[4805]: I1203 14:25:12.542308 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-jsrqx" Dec 03 14:25:12 crc kubenswrapper[4805]: I1203 14:25:12.547934 4805 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"frr-k8s-certs-secret" Dec 03 14:25:12 crc kubenswrapper[4805]: I1203 14:25:12.548225 4805 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"frr-startup" Dec 03 14:25:12 crc kubenswrapper[4805]: I1203 14:25:12.548334 4805 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"frr-k8s-daemon-dockercfg-lcm8w" Dec 03 14:25:12 crc kubenswrapper[4805]: I1203 14:25:12.551172 4805 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/frr-k8s-webhook-server-7fcb986d4-lmm42"] Dec 03 14:25:12 crc kubenswrapper[4805]: I1203 14:25:12.552006 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-lmm42" Dec 03 14:25:12 crc kubenswrapper[4805]: I1203 14:25:12.561243 4805 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"frr-k8s-webhook-server-cert" Dec 03 14:25:12 crc kubenswrapper[4805]: I1203 14:25:12.569321 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/frr-k8s-webhook-server-7fcb986d4-lmm42"] Dec 03 14:25:12 crc kubenswrapper[4805]: I1203 14:25:12.608825 4805 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/speaker-jpk6x"] Dec 03 14:25:12 crc kubenswrapper[4805]: I1203 14:25:12.609784 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/speaker-jpk6x" Dec 03 14:25:12 crc kubenswrapper[4805]: I1203 14:25:12.612143 4805 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"speaker-dockercfg-2b9c8" Dec 03 14:25:12 crc kubenswrapper[4805]: I1203 14:25:12.612423 4805 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-memberlist" Dec 03 14:25:12 crc kubenswrapper[4805]: I1203 14:25:12.612553 4805 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"speaker-certs-secret" Dec 03 14:25:12 crc kubenswrapper[4805]: I1203 14:25:12.612756 4805 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"metallb-excludel2" Dec 03 14:25:12 crc kubenswrapper[4805]: I1203 14:25:12.631807 4805 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/controller-f8648f98b-n75gf"] Dec 03 14:25:12 crc kubenswrapper[4805]: I1203 14:25:12.633403 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/controller-f8648f98b-n75gf" Dec 03 14:25:12 crc kubenswrapper[4805]: I1203 14:25:12.641108 4805 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"controller-certs-secret" Dec 03 14:25:12 crc kubenswrapper[4805]: I1203 14:25:12.651055 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/controller-f8648f98b-n75gf"] Dec 03 14:25:12 crc kubenswrapper[4805]: I1203 14:25:12.699672 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ctb5f\" (UniqueName: \"kubernetes.io/projected/06e98f2f-39d1-4899-9a18-7cc27e006de9-kube-api-access-ctb5f\") pod \"speaker-jpk6x\" (UID: \"06e98f2f-39d1-4899-9a18-7cc27e006de9\") " pod="metallb-system/speaker-jpk6x" Dec 03 14:25:12 crc kubenswrapper[4805]: I1203 14:25:12.699714 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/06e98f2f-39d1-4899-9a18-7cc27e006de9-metrics-certs\") pod \"speaker-jpk6x\" (UID: \"06e98f2f-39d1-4899-9a18-7cc27e006de9\") " pod="metallb-system/speaker-jpk6x" Dec 03 14:25:12 crc kubenswrapper[4805]: I1203 14:25:12.699745 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/06e98f2f-39d1-4899-9a18-7cc27e006de9-memberlist\") pod \"speaker-jpk6x\" (UID: \"06e98f2f-39d1-4899-9a18-7cc27e006de9\") " pod="metallb-system/speaker-jpk6x" Dec 03 14:25:12 crc kubenswrapper[4805]: I1203 14:25:12.699771 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"frr-conf\" (UniqueName: \"kubernetes.io/empty-dir/2ba805a2-6c47-4649-b8dc-eb92fb614437-frr-conf\") pod \"frr-k8s-jsrqx\" (UID: \"2ba805a2-6c47-4649-b8dc-eb92fb614437\") " pod="metallb-system/frr-k8s-jsrqx" Dec 03 14:25:12 crc kubenswrapper[4805]: I1203 14:25:12.699795 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"frr-sockets\" (UniqueName: \"kubernetes.io/empty-dir/2ba805a2-6c47-4649-b8dc-eb92fb614437-frr-sockets\") pod \"frr-k8s-jsrqx\" (UID: \"2ba805a2-6c47-4649-b8dc-eb92fb614437\") " pod="metallb-system/frr-k8s-jsrqx" Dec 03 14:25:12 crc kubenswrapper[4805]: I1203 14:25:12.699923 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"reloader\" (UniqueName: \"kubernetes.io/empty-dir/2ba805a2-6c47-4649-b8dc-eb92fb614437-reloader\") pod \"frr-k8s-jsrqx\" (UID: \"2ba805a2-6c47-4649-b8dc-eb92fb614437\") " pod="metallb-system/frr-k8s-jsrqx" Dec 03 14:25:12 crc kubenswrapper[4805]: I1203 14:25:12.699989 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"frr-startup\" (UniqueName: \"kubernetes.io/configmap/2ba805a2-6c47-4649-b8dc-eb92fb614437-frr-startup\") pod \"frr-k8s-jsrqx\" (UID: \"2ba805a2-6c47-4649-b8dc-eb92fb614437\") " pod="metallb-system/frr-k8s-jsrqx" Dec 03 14:25:12 crc kubenswrapper[4805]: I1203 14:25:12.700034 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4z4vh\" (UniqueName: \"kubernetes.io/projected/24cc3c3f-7f72-49c4-b531-093bf5851177-kube-api-access-4z4vh\") pod \"frr-k8s-webhook-server-7fcb986d4-lmm42\" (UID: \"24cc3c3f-7f72-49c4-b531-093bf5851177\") " pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-lmm42" Dec 03 14:25:12 crc kubenswrapper[4805]: I1203 14:25:12.700067 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metallb-excludel2\" (UniqueName: \"kubernetes.io/configmap/06e98f2f-39d1-4899-9a18-7cc27e006de9-metallb-excludel2\") pod \"speaker-jpk6x\" (UID: \"06e98f2f-39d1-4899-9a18-7cc27e006de9\") " pod="metallb-system/speaker-jpk6x" Dec 03 14:25:12 crc kubenswrapper[4805]: I1203 14:25:12.700091 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/24cc3c3f-7f72-49c4-b531-093bf5851177-cert\") pod \"frr-k8s-webhook-server-7fcb986d4-lmm42\" (UID: \"24cc3c3f-7f72-49c4-b531-093bf5851177\") " pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-lmm42" Dec 03 14:25:12 crc kubenswrapper[4805]: I1203 14:25:12.700127 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9bnsw\" (UniqueName: \"kubernetes.io/projected/2ba805a2-6c47-4649-b8dc-eb92fb614437-kube-api-access-9bnsw\") pod \"frr-k8s-jsrqx\" (UID: \"2ba805a2-6c47-4649-b8dc-eb92fb614437\") " pod="metallb-system/frr-k8s-jsrqx" Dec 03 14:25:12 crc kubenswrapper[4805]: I1203 14:25:12.700157 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/2ba805a2-6c47-4649-b8dc-eb92fb614437-metrics-certs\") pod \"frr-k8s-jsrqx\" (UID: \"2ba805a2-6c47-4649-b8dc-eb92fb614437\") " pod="metallb-system/frr-k8s-jsrqx" Dec 03 14:25:12 crc kubenswrapper[4805]: I1203 14:25:12.700190 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics\" (UniqueName: \"kubernetes.io/empty-dir/2ba805a2-6c47-4649-b8dc-eb92fb614437-metrics\") pod \"frr-k8s-jsrqx\" (UID: \"2ba805a2-6c47-4649-b8dc-eb92fb614437\") " pod="metallb-system/frr-k8s-jsrqx" Dec 03 14:25:12 crc kubenswrapper[4805]: I1203 14:25:12.801696 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"frr-startup\" (UniqueName: \"kubernetes.io/configmap/2ba805a2-6c47-4649-b8dc-eb92fb614437-frr-startup\") pod \"frr-k8s-jsrqx\" (UID: \"2ba805a2-6c47-4649-b8dc-eb92fb614437\") " pod="metallb-system/frr-k8s-jsrqx" Dec 03 14:25:12 crc kubenswrapper[4805]: I1203 14:25:12.801757 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4z4vh\" (UniqueName: \"kubernetes.io/projected/24cc3c3f-7f72-49c4-b531-093bf5851177-kube-api-access-4z4vh\") pod \"frr-k8s-webhook-server-7fcb986d4-lmm42\" (UID: \"24cc3c3f-7f72-49c4-b531-093bf5851177\") " pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-lmm42" Dec 03 14:25:12 crc kubenswrapper[4805]: I1203 14:25:12.801808 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metallb-excludel2\" (UniqueName: \"kubernetes.io/configmap/06e98f2f-39d1-4899-9a18-7cc27e006de9-metallb-excludel2\") pod \"speaker-jpk6x\" (UID: \"06e98f2f-39d1-4899-9a18-7cc27e006de9\") " pod="metallb-system/speaker-jpk6x" Dec 03 14:25:12 crc kubenswrapper[4805]: I1203 14:25:12.801835 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/24cc3c3f-7f72-49c4-b531-093bf5851177-cert\") pod \"frr-k8s-webhook-server-7fcb986d4-lmm42\" (UID: \"24cc3c3f-7f72-49c4-b531-093bf5851177\") " pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-lmm42" Dec 03 14:25:12 crc kubenswrapper[4805]: I1203 14:25:12.802231 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lq259\" (UniqueName: \"kubernetes.io/projected/a3361a03-7368-4bb6-b209-485190285e82-kube-api-access-lq259\") pod \"controller-f8648f98b-n75gf\" (UID: \"a3361a03-7368-4bb6-b209-485190285e82\") " pod="metallb-system/controller-f8648f98b-n75gf" Dec 03 14:25:12 crc kubenswrapper[4805]: I1203 14:25:12.802260 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9bnsw\" (UniqueName: \"kubernetes.io/projected/2ba805a2-6c47-4649-b8dc-eb92fb614437-kube-api-access-9bnsw\") pod \"frr-k8s-jsrqx\" (UID: \"2ba805a2-6c47-4649-b8dc-eb92fb614437\") " pod="metallb-system/frr-k8s-jsrqx" Dec 03 14:25:12 crc kubenswrapper[4805]: I1203 14:25:12.802335 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/2ba805a2-6c47-4649-b8dc-eb92fb614437-metrics-certs\") pod \"frr-k8s-jsrqx\" (UID: \"2ba805a2-6c47-4649-b8dc-eb92fb614437\") " pod="metallb-system/frr-k8s-jsrqx" Dec 03 14:25:12 crc kubenswrapper[4805]: I1203 14:25:12.802552 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics\" (UniqueName: \"kubernetes.io/empty-dir/2ba805a2-6c47-4649-b8dc-eb92fb614437-metrics\") pod \"frr-k8s-jsrqx\" (UID: \"2ba805a2-6c47-4649-b8dc-eb92fb614437\") " pod="metallb-system/frr-k8s-jsrqx" Dec 03 14:25:12 crc kubenswrapper[4805]: I1203 14:25:12.802825 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"frr-startup\" (UniqueName: \"kubernetes.io/configmap/2ba805a2-6c47-4649-b8dc-eb92fb614437-frr-startup\") pod \"frr-k8s-jsrqx\" (UID: \"2ba805a2-6c47-4649-b8dc-eb92fb614437\") " pod="metallb-system/frr-k8s-jsrqx" Dec 03 14:25:12 crc kubenswrapper[4805]: I1203 14:25:12.802775 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metallb-excludel2\" (UniqueName: \"kubernetes.io/configmap/06e98f2f-39d1-4899-9a18-7cc27e006de9-metallb-excludel2\") pod \"speaker-jpk6x\" (UID: \"06e98f2f-39d1-4899-9a18-7cc27e006de9\") " pod="metallb-system/speaker-jpk6x" Dec 03 14:25:12 crc kubenswrapper[4805]: I1203 14:25:12.802862 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/a3361a03-7368-4bb6-b209-485190285e82-cert\") pod \"controller-f8648f98b-n75gf\" (UID: \"a3361a03-7368-4bb6-b209-485190285e82\") " pod="metallb-system/controller-f8648f98b-n75gf" Dec 03 14:25:12 crc kubenswrapper[4805]: I1203 14:25:12.802932 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ctb5f\" (UniqueName: \"kubernetes.io/projected/06e98f2f-39d1-4899-9a18-7cc27e006de9-kube-api-access-ctb5f\") pod \"speaker-jpk6x\" (UID: \"06e98f2f-39d1-4899-9a18-7cc27e006de9\") " pod="metallb-system/speaker-jpk6x" Dec 03 14:25:12 crc kubenswrapper[4805]: I1203 14:25:12.802989 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/06e98f2f-39d1-4899-9a18-7cc27e006de9-metrics-certs\") pod \"speaker-jpk6x\" (UID: \"06e98f2f-39d1-4899-9a18-7cc27e006de9\") " pod="metallb-system/speaker-jpk6x" Dec 03 14:25:12 crc kubenswrapper[4805]: I1203 14:25:12.803017 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/a3361a03-7368-4bb6-b209-485190285e82-metrics-certs\") pod \"controller-f8648f98b-n75gf\" (UID: \"a3361a03-7368-4bb6-b209-485190285e82\") " pod="metallb-system/controller-f8648f98b-n75gf" Dec 03 14:25:12 crc kubenswrapper[4805]: I1203 14:25:12.803077 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics\" (UniqueName: \"kubernetes.io/empty-dir/2ba805a2-6c47-4649-b8dc-eb92fb614437-metrics\") pod \"frr-k8s-jsrqx\" (UID: \"2ba805a2-6c47-4649-b8dc-eb92fb614437\") " pod="metallb-system/frr-k8s-jsrqx" Dec 03 14:25:12 crc kubenswrapper[4805]: I1203 14:25:12.803095 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/06e98f2f-39d1-4899-9a18-7cc27e006de9-memberlist\") pod \"speaker-jpk6x\" (UID: \"06e98f2f-39d1-4899-9a18-7cc27e006de9\") " pod="metallb-system/speaker-jpk6x" Dec 03 14:25:12 crc kubenswrapper[4805]: I1203 14:25:12.803154 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"frr-conf\" (UniqueName: \"kubernetes.io/empty-dir/2ba805a2-6c47-4649-b8dc-eb92fb614437-frr-conf\") pod \"frr-k8s-jsrqx\" (UID: \"2ba805a2-6c47-4649-b8dc-eb92fb614437\") " pod="metallb-system/frr-k8s-jsrqx" Dec 03 14:25:12 crc kubenswrapper[4805]: E1203 14:25:12.803184 4805 secret.go:188] Couldn't get secret metallb-system/speaker-certs-secret: secret "speaker-certs-secret" not found Dec 03 14:25:12 crc kubenswrapper[4805]: E1203 14:25:12.803241 4805 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/06e98f2f-39d1-4899-9a18-7cc27e006de9-metrics-certs podName:06e98f2f-39d1-4899-9a18-7cc27e006de9 nodeName:}" failed. No retries permitted until 2025-12-03 14:25:13.303220786 +0000 UTC m=+942.966137719 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/06e98f2f-39d1-4899-9a18-7cc27e006de9-metrics-certs") pod "speaker-jpk6x" (UID: "06e98f2f-39d1-4899-9a18-7cc27e006de9") : secret "speaker-certs-secret" not found Dec 03 14:25:12 crc kubenswrapper[4805]: I1203 14:25:12.803184 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"frr-sockets\" (UniqueName: \"kubernetes.io/empty-dir/2ba805a2-6c47-4649-b8dc-eb92fb614437-frr-sockets\") pod \"frr-k8s-jsrqx\" (UID: \"2ba805a2-6c47-4649-b8dc-eb92fb614437\") " pod="metallb-system/frr-k8s-jsrqx" Dec 03 14:25:12 crc kubenswrapper[4805]: E1203 14:25:12.803460 4805 secret.go:188] Couldn't get secret metallb-system/metallb-memberlist: secret "metallb-memberlist" not found Dec 03 14:25:12 crc kubenswrapper[4805]: I1203 14:25:12.803579 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"reloader\" (UniqueName: \"kubernetes.io/empty-dir/2ba805a2-6c47-4649-b8dc-eb92fb614437-reloader\") pod \"frr-k8s-jsrqx\" (UID: \"2ba805a2-6c47-4649-b8dc-eb92fb614437\") " pod="metallb-system/frr-k8s-jsrqx" Dec 03 14:25:12 crc kubenswrapper[4805]: I1203 14:25:12.803392 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"frr-sockets\" (UniqueName: \"kubernetes.io/empty-dir/2ba805a2-6c47-4649-b8dc-eb92fb614437-frr-sockets\") pod \"frr-k8s-jsrqx\" (UID: \"2ba805a2-6c47-4649-b8dc-eb92fb614437\") " pod="metallb-system/frr-k8s-jsrqx" Dec 03 14:25:12 crc kubenswrapper[4805]: I1203 14:25:12.803741 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"frr-conf\" (UniqueName: \"kubernetes.io/empty-dir/2ba805a2-6c47-4649-b8dc-eb92fb614437-frr-conf\") pod \"frr-k8s-jsrqx\" (UID: \"2ba805a2-6c47-4649-b8dc-eb92fb614437\") " pod="metallb-system/frr-k8s-jsrqx" Dec 03 14:25:12 crc kubenswrapper[4805]: E1203 14:25:12.804017 4805 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/06e98f2f-39d1-4899-9a18-7cc27e006de9-memberlist podName:06e98f2f-39d1-4899-9a18-7cc27e006de9 nodeName:}" failed. No retries permitted until 2025-12-03 14:25:13.303986907 +0000 UTC m=+942.966903930 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "memberlist" (UniqueName: "kubernetes.io/secret/06e98f2f-39d1-4899-9a18-7cc27e006de9-memberlist") pod "speaker-jpk6x" (UID: "06e98f2f-39d1-4899-9a18-7cc27e006de9") : secret "metallb-memberlist" not found Dec 03 14:25:12 crc kubenswrapper[4805]: I1203 14:25:12.804333 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"reloader\" (UniqueName: \"kubernetes.io/empty-dir/2ba805a2-6c47-4649-b8dc-eb92fb614437-reloader\") pod \"frr-k8s-jsrqx\" (UID: \"2ba805a2-6c47-4649-b8dc-eb92fb614437\") " pod="metallb-system/frr-k8s-jsrqx" Dec 03 14:25:12 crc kubenswrapper[4805]: I1203 14:25:12.807725 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/2ba805a2-6c47-4649-b8dc-eb92fb614437-metrics-certs\") pod \"frr-k8s-jsrqx\" (UID: \"2ba805a2-6c47-4649-b8dc-eb92fb614437\") " pod="metallb-system/frr-k8s-jsrqx" Dec 03 14:25:12 crc kubenswrapper[4805]: I1203 14:25:12.825080 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/24cc3c3f-7f72-49c4-b531-093bf5851177-cert\") pod \"frr-k8s-webhook-server-7fcb986d4-lmm42\" (UID: \"24cc3c3f-7f72-49c4-b531-093bf5851177\") " pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-lmm42" Dec 03 14:25:12 crc kubenswrapper[4805]: I1203 14:25:12.830559 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ctb5f\" (UniqueName: \"kubernetes.io/projected/06e98f2f-39d1-4899-9a18-7cc27e006de9-kube-api-access-ctb5f\") pod \"speaker-jpk6x\" (UID: \"06e98f2f-39d1-4899-9a18-7cc27e006de9\") " pod="metallb-system/speaker-jpk6x" Dec 03 14:25:12 crc kubenswrapper[4805]: I1203 14:25:12.831139 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9bnsw\" (UniqueName: \"kubernetes.io/projected/2ba805a2-6c47-4649-b8dc-eb92fb614437-kube-api-access-9bnsw\") pod \"frr-k8s-jsrqx\" (UID: \"2ba805a2-6c47-4649-b8dc-eb92fb614437\") " pod="metallb-system/frr-k8s-jsrqx" Dec 03 14:25:12 crc kubenswrapper[4805]: I1203 14:25:12.848695 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4z4vh\" (UniqueName: \"kubernetes.io/projected/24cc3c3f-7f72-49c4-b531-093bf5851177-kube-api-access-4z4vh\") pod \"frr-k8s-webhook-server-7fcb986d4-lmm42\" (UID: \"24cc3c3f-7f72-49c4-b531-093bf5851177\") " pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-lmm42" Dec 03 14:25:12 crc kubenswrapper[4805]: I1203 14:25:12.873298 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-jsrqx" Dec 03 14:25:12 crc kubenswrapper[4805]: I1203 14:25:12.888012 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-lmm42" Dec 03 14:25:12 crc kubenswrapper[4805]: I1203 14:25:12.905372 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lq259\" (UniqueName: \"kubernetes.io/projected/a3361a03-7368-4bb6-b209-485190285e82-kube-api-access-lq259\") pod \"controller-f8648f98b-n75gf\" (UID: \"a3361a03-7368-4bb6-b209-485190285e82\") " pod="metallb-system/controller-f8648f98b-n75gf" Dec 03 14:25:12 crc kubenswrapper[4805]: I1203 14:25:12.905434 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/a3361a03-7368-4bb6-b209-485190285e82-cert\") pod \"controller-f8648f98b-n75gf\" (UID: \"a3361a03-7368-4bb6-b209-485190285e82\") " pod="metallb-system/controller-f8648f98b-n75gf" Dec 03 14:25:12 crc kubenswrapper[4805]: I1203 14:25:12.905880 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/a3361a03-7368-4bb6-b209-485190285e82-metrics-certs\") pod \"controller-f8648f98b-n75gf\" (UID: \"a3361a03-7368-4bb6-b209-485190285e82\") " pod="metallb-system/controller-f8648f98b-n75gf" Dec 03 14:25:12 crc kubenswrapper[4805]: I1203 14:25:12.908211 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/a3361a03-7368-4bb6-b209-485190285e82-cert\") pod \"controller-f8648f98b-n75gf\" (UID: \"a3361a03-7368-4bb6-b209-485190285e82\") " pod="metallb-system/controller-f8648f98b-n75gf" Dec 03 14:25:12 crc kubenswrapper[4805]: I1203 14:25:12.910923 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/a3361a03-7368-4bb6-b209-485190285e82-metrics-certs\") pod \"controller-f8648f98b-n75gf\" (UID: \"a3361a03-7368-4bb6-b209-485190285e82\") " pod="metallb-system/controller-f8648f98b-n75gf" Dec 03 14:25:12 crc kubenswrapper[4805]: I1203 14:25:12.926616 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lq259\" (UniqueName: \"kubernetes.io/projected/a3361a03-7368-4bb6-b209-485190285e82-kube-api-access-lq259\") pod \"controller-f8648f98b-n75gf\" (UID: \"a3361a03-7368-4bb6-b209-485190285e82\") " pod="metallb-system/controller-f8648f98b-n75gf" Dec 03 14:25:12 crc kubenswrapper[4805]: I1203 14:25:12.957689 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/controller-f8648f98b-n75gf" Dec 03 14:25:13 crc kubenswrapper[4805]: I1203 14:25:13.070094 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-jsrqx" event={"ID":"2ba805a2-6c47-4649-b8dc-eb92fb614437","Type":"ContainerStarted","Data":"613e0de3e00631e75ff969b00cbd1cda0301c5b80d32b06b150fecb3b5fb29f1"} Dec 03 14:25:13 crc kubenswrapper[4805]: I1203 14:25:13.139766 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/frr-k8s-webhook-server-7fcb986d4-lmm42"] Dec 03 14:25:13 crc kubenswrapper[4805]: W1203 14:25:13.150194 4805 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod24cc3c3f_7f72_49c4_b531_093bf5851177.slice/crio-a6e30b49d99bb97d1dd0e853fb97d9f10c763956e929e0fa63cefeced8813165 WatchSource:0}: Error finding container a6e30b49d99bb97d1dd0e853fb97d9f10c763956e929e0fa63cefeced8813165: Status 404 returned error can't find the container with id a6e30b49d99bb97d1dd0e853fb97d9f10c763956e929e0fa63cefeced8813165 Dec 03 14:25:13 crc kubenswrapper[4805]: I1203 14:25:13.178784 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/controller-f8648f98b-n75gf"] Dec 03 14:25:13 crc kubenswrapper[4805]: I1203 14:25:13.320831 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/06e98f2f-39d1-4899-9a18-7cc27e006de9-metrics-certs\") pod \"speaker-jpk6x\" (UID: \"06e98f2f-39d1-4899-9a18-7cc27e006de9\") " pod="metallb-system/speaker-jpk6x" Dec 03 14:25:13 crc kubenswrapper[4805]: I1203 14:25:13.320924 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/06e98f2f-39d1-4899-9a18-7cc27e006de9-memberlist\") pod \"speaker-jpk6x\" (UID: \"06e98f2f-39d1-4899-9a18-7cc27e006de9\") " pod="metallb-system/speaker-jpk6x" Dec 03 14:25:13 crc kubenswrapper[4805]: E1203 14:25:13.321074 4805 secret.go:188] Couldn't get secret metallb-system/metallb-memberlist: secret "metallb-memberlist" not found Dec 03 14:25:13 crc kubenswrapper[4805]: E1203 14:25:13.321146 4805 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/06e98f2f-39d1-4899-9a18-7cc27e006de9-memberlist podName:06e98f2f-39d1-4899-9a18-7cc27e006de9 nodeName:}" failed. No retries permitted until 2025-12-03 14:25:14.3211282 +0000 UTC m=+943.984045143 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "memberlist" (UniqueName: "kubernetes.io/secret/06e98f2f-39d1-4899-9a18-7cc27e006de9-memberlist") pod "speaker-jpk6x" (UID: "06e98f2f-39d1-4899-9a18-7cc27e006de9") : secret "metallb-memberlist" not found Dec 03 14:25:13 crc kubenswrapper[4805]: I1203 14:25:13.335039 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/06e98f2f-39d1-4899-9a18-7cc27e006de9-metrics-certs\") pod \"speaker-jpk6x\" (UID: \"06e98f2f-39d1-4899-9a18-7cc27e006de9\") " pod="metallb-system/speaker-jpk6x" Dec 03 14:25:14 crc kubenswrapper[4805]: I1203 14:25:14.081110 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/controller-f8648f98b-n75gf" event={"ID":"a3361a03-7368-4bb6-b209-485190285e82","Type":"ContainerStarted","Data":"fb4bacbc8fe7542f94de5507bc8c5665586f2c0460c730b5ab4f290d4f942cc5"} Dec 03 14:25:14 crc kubenswrapper[4805]: I1203 14:25:14.081187 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/controller-f8648f98b-n75gf" event={"ID":"a3361a03-7368-4bb6-b209-485190285e82","Type":"ContainerStarted","Data":"e9c799b26a6e9fab4a9c3cce33711874ee00589809fc814d73eadf7086943e3a"} Dec 03 14:25:14 crc kubenswrapper[4805]: I1203 14:25:14.081207 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/controller-f8648f98b-n75gf" event={"ID":"a3361a03-7368-4bb6-b209-485190285e82","Type":"ContainerStarted","Data":"deaa62145b35b20ae9238dd5113c678204f9a912162bf48dfd9634e1af5c755c"} Dec 03 14:25:14 crc kubenswrapper[4805]: I1203 14:25:14.081235 4805 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/controller-f8648f98b-n75gf" Dec 03 14:25:14 crc kubenswrapper[4805]: I1203 14:25:14.082956 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-lmm42" event={"ID":"24cc3c3f-7f72-49c4-b531-093bf5851177","Type":"ContainerStarted","Data":"a6e30b49d99bb97d1dd0e853fb97d9f10c763956e929e0fa63cefeced8813165"} Dec 03 14:25:14 crc kubenswrapper[4805]: I1203 14:25:14.108253 4805 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/controller-f8648f98b-n75gf" podStartSLOduration=2.108231886 podStartE2EDuration="2.108231886s" podCreationTimestamp="2025-12-03 14:25:12 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 14:25:14.10652299 +0000 UTC m=+943.769439953" watchObservedRunningTime="2025-12-03 14:25:14.108231886 +0000 UTC m=+943.771148809" Dec 03 14:25:14 crc kubenswrapper[4805]: I1203 14:25:14.336193 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/06e98f2f-39d1-4899-9a18-7cc27e006de9-memberlist\") pod \"speaker-jpk6x\" (UID: \"06e98f2f-39d1-4899-9a18-7cc27e006de9\") " pod="metallb-system/speaker-jpk6x" Dec 03 14:25:14 crc kubenswrapper[4805]: I1203 14:25:14.339872 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/06e98f2f-39d1-4899-9a18-7cc27e006de9-memberlist\") pod \"speaker-jpk6x\" (UID: \"06e98f2f-39d1-4899-9a18-7cc27e006de9\") " pod="metallb-system/speaker-jpk6x" Dec 03 14:25:14 crc kubenswrapper[4805]: I1203 14:25:14.425881 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/speaker-jpk6x" Dec 03 14:25:15 crc kubenswrapper[4805]: I1203 14:25:15.111186 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/speaker-jpk6x" event={"ID":"06e98f2f-39d1-4899-9a18-7cc27e006de9","Type":"ContainerStarted","Data":"a899c1b3b67d6f5e35b475bd80d398a5e78754ec7b93367283a9abf5fcfc667f"} Dec 03 14:25:15 crc kubenswrapper[4805]: I1203 14:25:15.111582 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/speaker-jpk6x" event={"ID":"06e98f2f-39d1-4899-9a18-7cc27e006de9","Type":"ContainerStarted","Data":"28ede75329357787268983eedfc7343e186d5575d9a978db47b99347743a63cb"} Dec 03 14:25:15 crc kubenswrapper[4805]: I1203 14:25:15.111600 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/speaker-jpk6x" event={"ID":"06e98f2f-39d1-4899-9a18-7cc27e006de9","Type":"ContainerStarted","Data":"3d0e6ac8e22224f8ab79a4e5a33b5d35c024009a2111d08de8811bd25c4099fb"} Dec 03 14:25:15 crc kubenswrapper[4805]: I1203 14:25:15.111791 4805 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/speaker-jpk6x" Dec 03 14:25:15 crc kubenswrapper[4805]: I1203 14:25:15.133406 4805 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/speaker-jpk6x" podStartSLOduration=3.133390277 podStartE2EDuration="3.133390277s" podCreationTimestamp="2025-12-03 14:25:12 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 14:25:15.128283739 +0000 UTC m=+944.791200662" watchObservedRunningTime="2025-12-03 14:25:15.133390277 +0000 UTC m=+944.796307200" Dec 03 14:25:21 crc kubenswrapper[4805]: I1203 14:25:21.146230 4805 generic.go:334] "Generic (PLEG): container finished" podID="2ba805a2-6c47-4649-b8dc-eb92fb614437" containerID="38bfcf5c56ffbc28984a3f08fa13807e76ed0501e82cc130edee3fa776be08cc" exitCode=0 Dec 03 14:25:21 crc kubenswrapper[4805]: I1203 14:25:21.146321 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-jsrqx" event={"ID":"2ba805a2-6c47-4649-b8dc-eb92fb614437","Type":"ContainerDied","Data":"38bfcf5c56ffbc28984a3f08fa13807e76ed0501e82cc130edee3fa776be08cc"} Dec 03 14:25:21 crc kubenswrapper[4805]: I1203 14:25:21.150116 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-lmm42" event={"ID":"24cc3c3f-7f72-49c4-b531-093bf5851177","Type":"ContainerStarted","Data":"b6938fe3ba42a8683b374eadc3c539491509a786ae8f227789e12817048f165a"} Dec 03 14:25:21 crc kubenswrapper[4805]: I1203 14:25:21.150394 4805 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-lmm42" Dec 03 14:25:21 crc kubenswrapper[4805]: I1203 14:25:21.189983 4805 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-lmm42" podStartSLOduration=1.567099692 podStartE2EDuration="9.189952875s" podCreationTimestamp="2025-12-03 14:25:12 +0000 UTC" firstStartedPulling="2025-12-03 14:25:13.152669053 +0000 UTC m=+942.815585976" lastFinishedPulling="2025-12-03 14:25:20.775522246 +0000 UTC m=+950.438439159" observedRunningTime="2025-12-03 14:25:21.184924669 +0000 UTC m=+950.847841612" watchObservedRunningTime="2025-12-03 14:25:21.189952875 +0000 UTC m=+950.852869828" Dec 03 14:25:22 crc kubenswrapper[4805]: I1203 14:25:22.158190 4805 generic.go:334] "Generic (PLEG): container finished" podID="2ba805a2-6c47-4649-b8dc-eb92fb614437" containerID="bed80d057b9ea2a4eab794cc5a83f8a175db23785ccbeb7cb8e2cbf1cc66523e" exitCode=0 Dec 03 14:25:22 crc kubenswrapper[4805]: I1203 14:25:22.158281 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-jsrqx" event={"ID":"2ba805a2-6c47-4649-b8dc-eb92fb614437","Type":"ContainerDied","Data":"bed80d057b9ea2a4eab794cc5a83f8a175db23785ccbeb7cb8e2cbf1cc66523e"} Dec 03 14:25:23 crc kubenswrapper[4805]: I1203 14:25:23.165150 4805 generic.go:334] "Generic (PLEG): container finished" podID="2ba805a2-6c47-4649-b8dc-eb92fb614437" containerID="62e598b63ec01a2b646d93def31daa8400b8d184af4657cc65de1dd426547da0" exitCode=0 Dec 03 14:25:23 crc kubenswrapper[4805]: I1203 14:25:23.165209 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-jsrqx" event={"ID":"2ba805a2-6c47-4649-b8dc-eb92fb614437","Type":"ContainerDied","Data":"62e598b63ec01a2b646d93def31daa8400b8d184af4657cc65de1dd426547da0"} Dec 03 14:25:24 crc kubenswrapper[4805]: I1203 14:25:24.175453 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-jsrqx" event={"ID":"2ba805a2-6c47-4649-b8dc-eb92fb614437","Type":"ContainerStarted","Data":"dfbf721d65dc6727ff5c9858de6fc8baeb33af3422bf36023b72553c2fad8bc9"} Dec 03 14:25:24 crc kubenswrapper[4805]: I1203 14:25:24.175511 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-jsrqx" event={"ID":"2ba805a2-6c47-4649-b8dc-eb92fb614437","Type":"ContainerStarted","Data":"3b466f53172cafd7e37aad199e1d384c37b0e288682d930537825505a4c490ce"} Dec 03 14:25:24 crc kubenswrapper[4805]: I1203 14:25:24.175524 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-jsrqx" event={"ID":"2ba805a2-6c47-4649-b8dc-eb92fb614437","Type":"ContainerStarted","Data":"12007b228b947da81ce644ac20ee91397fe7c29c04d659832e6f79097cb63593"} Dec 03 14:25:24 crc kubenswrapper[4805]: I1203 14:25:24.432101 4805 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/speaker-jpk6x" Dec 03 14:25:25 crc kubenswrapper[4805]: I1203 14:25:25.185641 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-jsrqx" event={"ID":"2ba805a2-6c47-4649-b8dc-eb92fb614437","Type":"ContainerStarted","Data":"2347736e16c63d5bbd15f54d1b107f803891e4a3736b911ff3b67659f243e712"} Dec 03 14:25:25 crc kubenswrapper[4805]: I1203 14:25:25.186047 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-jsrqx" event={"ID":"2ba805a2-6c47-4649-b8dc-eb92fb614437","Type":"ContainerStarted","Data":"826db89c45046d7fdb51c85beef9db0de05525b8e66096cfdaf927032a68cd4d"} Dec 03 14:25:26 crc kubenswrapper[4805]: I1203 14:25:26.202024 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-jsrqx" event={"ID":"2ba805a2-6c47-4649-b8dc-eb92fb614437","Type":"ContainerStarted","Data":"acdb082d64ca1f6a1715cdfabd29b12a3ce05c7875743cc9f3bc917ed3e79859"} Dec 03 14:25:26 crc kubenswrapper[4805]: I1203 14:25:26.202334 4805 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/frr-k8s-jsrqx" Dec 03 14:25:26 crc kubenswrapper[4805]: I1203 14:25:26.246302 4805 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/frr-k8s-jsrqx" podStartSLOduration=6.513222256 podStartE2EDuration="14.246277638s" podCreationTimestamp="2025-12-03 14:25:12 +0000 UTC" firstStartedPulling="2025-12-03 14:25:13.037932151 +0000 UTC m=+942.700849074" lastFinishedPulling="2025-12-03 14:25:20.770987533 +0000 UTC m=+950.433904456" observedRunningTime="2025-12-03 14:25:26.23711473 +0000 UTC m=+955.900031673" watchObservedRunningTime="2025-12-03 14:25:26.246277638 +0000 UTC m=+955.909194601" Dec 03 14:25:27 crc kubenswrapper[4805]: I1203 14:25:27.873951 4805 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="metallb-system/frr-k8s-jsrqx" Dec 03 14:25:27 crc kubenswrapper[4805]: I1203 14:25:27.941260 4805 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="metallb-system/frr-k8s-jsrqx" Dec 03 14:25:28 crc kubenswrapper[4805]: I1203 14:25:28.182152 4805 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-operator-index-pbpk5"] Dec 03 14:25:28 crc kubenswrapper[4805]: I1203 14:25:28.183766 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-pbpk5" Dec 03 14:25:28 crc kubenswrapper[4805]: I1203 14:25:28.191167 4805 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack-operators"/"openshift-service-ca.crt" Dec 03 14:25:28 crc kubenswrapper[4805]: I1203 14:25:28.191353 4805 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack-operators"/"kube-root-ca.crt" Dec 03 14:25:28 crc kubenswrapper[4805]: I1203 14:25:28.191500 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-operator-index-dockercfg-fl5sw" Dec 03 14:25:28 crc kubenswrapper[4805]: I1203 14:25:28.247146 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-index-pbpk5"] Dec 03 14:25:28 crc kubenswrapper[4805]: I1203 14:25:28.275403 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wssl7\" (UniqueName: \"kubernetes.io/projected/682558e2-bc38-4aae-aedf-c54605f12dd6-kube-api-access-wssl7\") pod \"openstack-operator-index-pbpk5\" (UID: \"682558e2-bc38-4aae-aedf-c54605f12dd6\") " pod="openstack-operators/openstack-operator-index-pbpk5" Dec 03 14:25:28 crc kubenswrapper[4805]: I1203 14:25:28.376583 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wssl7\" (UniqueName: \"kubernetes.io/projected/682558e2-bc38-4aae-aedf-c54605f12dd6-kube-api-access-wssl7\") pod \"openstack-operator-index-pbpk5\" (UID: \"682558e2-bc38-4aae-aedf-c54605f12dd6\") " pod="openstack-operators/openstack-operator-index-pbpk5" Dec 03 14:25:28 crc kubenswrapper[4805]: I1203 14:25:28.399740 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wssl7\" (UniqueName: \"kubernetes.io/projected/682558e2-bc38-4aae-aedf-c54605f12dd6-kube-api-access-wssl7\") pod \"openstack-operator-index-pbpk5\" (UID: \"682558e2-bc38-4aae-aedf-c54605f12dd6\") " pod="openstack-operators/openstack-operator-index-pbpk5" Dec 03 14:25:28 crc kubenswrapper[4805]: I1203 14:25:28.500177 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-pbpk5" Dec 03 14:25:28 crc kubenswrapper[4805]: I1203 14:25:28.938193 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-index-pbpk5"] Dec 03 14:25:28 crc kubenswrapper[4805]: W1203 14:25:28.948153 4805 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod682558e2_bc38_4aae_aedf_c54605f12dd6.slice/crio-b7d412c00fe569c39091f7d6ee00605dd12a9bc87a849ada7baaf70e8828c040 WatchSource:0}: Error finding container b7d412c00fe569c39091f7d6ee00605dd12a9bc87a849ada7baaf70e8828c040: Status 404 returned error can't find the container with id b7d412c00fe569c39091f7d6ee00605dd12a9bc87a849ada7baaf70e8828c040 Dec 03 14:25:29 crc kubenswrapper[4805]: I1203 14:25:29.227167 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-pbpk5" event={"ID":"682558e2-bc38-4aae-aedf-c54605f12dd6","Type":"ContainerStarted","Data":"b7d412c00fe569c39091f7d6ee00605dd12a9bc87a849ada7baaf70e8828c040"} Dec 03 14:25:31 crc kubenswrapper[4805]: I1203 14:25:31.563936 4805 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack-operators/openstack-operator-index-pbpk5"] Dec 03 14:25:32 crc kubenswrapper[4805]: I1203 14:25:32.165748 4805 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-operator-index-p6vl8"] Dec 03 14:25:32 crc kubenswrapper[4805]: I1203 14:25:32.168009 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-p6vl8" Dec 03 14:25:32 crc kubenswrapper[4805]: I1203 14:25:32.175422 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-index-p6vl8"] Dec 03 14:25:32 crc kubenswrapper[4805]: I1203 14:25:32.247591 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-pbpk5" event={"ID":"682558e2-bc38-4aae-aedf-c54605f12dd6","Type":"ContainerStarted","Data":"5849ded655a1776fac2e6356f7207035fba4c827079f725e722e852ba9e9bbd1"} Dec 03 14:25:32 crc kubenswrapper[4805]: I1203 14:25:32.247747 4805 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack-operators/openstack-operator-index-pbpk5" podUID="682558e2-bc38-4aae-aedf-c54605f12dd6" containerName="registry-server" containerID="cri-o://5849ded655a1776fac2e6356f7207035fba4c827079f725e722e852ba9e9bbd1" gracePeriod=2 Dec 03 14:25:32 crc kubenswrapper[4805]: I1203 14:25:32.253452 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bnxx5\" (UniqueName: \"kubernetes.io/projected/f6e38dfc-ef6b-4813-8bd4-027bed8dfb81-kube-api-access-bnxx5\") pod \"openstack-operator-index-p6vl8\" (UID: \"f6e38dfc-ef6b-4813-8bd4-027bed8dfb81\") " pod="openstack-operators/openstack-operator-index-p6vl8" Dec 03 14:25:32 crc kubenswrapper[4805]: I1203 14:25:32.268354 4805 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-operator-index-pbpk5" podStartSLOduration=1.8023902870000001 podStartE2EDuration="4.26833299s" podCreationTimestamp="2025-12-03 14:25:28 +0000 UTC" firstStartedPulling="2025-12-03 14:25:28.951269485 +0000 UTC m=+958.614186448" lastFinishedPulling="2025-12-03 14:25:31.417212218 +0000 UTC m=+961.080129151" observedRunningTime="2025-12-03 14:25:32.266078359 +0000 UTC m=+961.928995292" watchObservedRunningTime="2025-12-03 14:25:32.26833299 +0000 UTC m=+961.931249903" Dec 03 14:25:32 crc kubenswrapper[4805]: I1203 14:25:32.354419 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bnxx5\" (UniqueName: \"kubernetes.io/projected/f6e38dfc-ef6b-4813-8bd4-027bed8dfb81-kube-api-access-bnxx5\") pod \"openstack-operator-index-p6vl8\" (UID: \"f6e38dfc-ef6b-4813-8bd4-027bed8dfb81\") " pod="openstack-operators/openstack-operator-index-p6vl8" Dec 03 14:25:32 crc kubenswrapper[4805]: I1203 14:25:32.374035 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bnxx5\" (UniqueName: \"kubernetes.io/projected/f6e38dfc-ef6b-4813-8bd4-027bed8dfb81-kube-api-access-bnxx5\") pod \"openstack-operator-index-p6vl8\" (UID: \"f6e38dfc-ef6b-4813-8bd4-027bed8dfb81\") " pod="openstack-operators/openstack-operator-index-p6vl8" Dec 03 14:25:32 crc kubenswrapper[4805]: I1203 14:25:32.483326 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-p6vl8" Dec 03 14:25:32 crc kubenswrapper[4805]: I1203 14:25:32.625464 4805 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-pbpk5" Dec 03 14:25:32 crc kubenswrapper[4805]: I1203 14:25:32.728570 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-index-p6vl8"] Dec 03 14:25:32 crc kubenswrapper[4805]: W1203 14:25:32.732072 4805 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf6e38dfc_ef6b_4813_8bd4_027bed8dfb81.slice/crio-413293ddce89ff5fe8f39110976c7b862e55d7b29ef180a54ce0f20135acd541 WatchSource:0}: Error finding container 413293ddce89ff5fe8f39110976c7b862e55d7b29ef180a54ce0f20135acd541: Status 404 returned error can't find the container with id 413293ddce89ff5fe8f39110976c7b862e55d7b29ef180a54ce0f20135acd541 Dec 03 14:25:32 crc kubenswrapper[4805]: I1203 14:25:32.760199 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wssl7\" (UniqueName: \"kubernetes.io/projected/682558e2-bc38-4aae-aedf-c54605f12dd6-kube-api-access-wssl7\") pod \"682558e2-bc38-4aae-aedf-c54605f12dd6\" (UID: \"682558e2-bc38-4aae-aedf-c54605f12dd6\") " Dec 03 14:25:32 crc kubenswrapper[4805]: I1203 14:25:32.763888 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/682558e2-bc38-4aae-aedf-c54605f12dd6-kube-api-access-wssl7" (OuterVolumeSpecName: "kube-api-access-wssl7") pod "682558e2-bc38-4aae-aedf-c54605f12dd6" (UID: "682558e2-bc38-4aae-aedf-c54605f12dd6"). InnerVolumeSpecName "kube-api-access-wssl7". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 14:25:32 crc kubenswrapper[4805]: I1203 14:25:32.861291 4805 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wssl7\" (UniqueName: \"kubernetes.io/projected/682558e2-bc38-4aae-aedf-c54605f12dd6-kube-api-access-wssl7\") on node \"crc\" DevicePath \"\"" Dec 03 14:25:32 crc kubenswrapper[4805]: I1203 14:25:32.891560 4805 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-lmm42" Dec 03 14:25:32 crc kubenswrapper[4805]: I1203 14:25:32.964030 4805 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/controller-f8648f98b-n75gf" Dec 03 14:25:33 crc kubenswrapper[4805]: I1203 14:25:33.254974 4805 generic.go:334] "Generic (PLEG): container finished" podID="682558e2-bc38-4aae-aedf-c54605f12dd6" containerID="5849ded655a1776fac2e6356f7207035fba4c827079f725e722e852ba9e9bbd1" exitCode=0 Dec 03 14:25:33 crc kubenswrapper[4805]: I1203 14:25:33.255036 4805 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-pbpk5" Dec 03 14:25:33 crc kubenswrapper[4805]: I1203 14:25:33.255035 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-pbpk5" event={"ID":"682558e2-bc38-4aae-aedf-c54605f12dd6","Type":"ContainerDied","Data":"5849ded655a1776fac2e6356f7207035fba4c827079f725e722e852ba9e9bbd1"} Dec 03 14:25:33 crc kubenswrapper[4805]: I1203 14:25:33.255385 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-pbpk5" event={"ID":"682558e2-bc38-4aae-aedf-c54605f12dd6","Type":"ContainerDied","Data":"b7d412c00fe569c39091f7d6ee00605dd12a9bc87a849ada7baaf70e8828c040"} Dec 03 14:25:33 crc kubenswrapper[4805]: I1203 14:25:33.255414 4805 scope.go:117] "RemoveContainer" containerID="5849ded655a1776fac2e6356f7207035fba4c827079f725e722e852ba9e9bbd1" Dec 03 14:25:33 crc kubenswrapper[4805]: I1203 14:25:33.256729 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-p6vl8" event={"ID":"f6e38dfc-ef6b-4813-8bd4-027bed8dfb81","Type":"ContainerStarted","Data":"7a16e9e724a56e507724323780fdb9f86d4007757ebf023b3b37c26111204e94"} Dec 03 14:25:33 crc kubenswrapper[4805]: I1203 14:25:33.256756 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-p6vl8" event={"ID":"f6e38dfc-ef6b-4813-8bd4-027bed8dfb81","Type":"ContainerStarted","Data":"413293ddce89ff5fe8f39110976c7b862e55d7b29ef180a54ce0f20135acd541"} Dec 03 14:25:33 crc kubenswrapper[4805]: I1203 14:25:33.274581 4805 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-operator-index-p6vl8" podStartSLOduration=1.228456437 podStartE2EDuration="1.274568048s" podCreationTimestamp="2025-12-03 14:25:32 +0000 UTC" firstStartedPulling="2025-12-03 14:25:32.73701204 +0000 UTC m=+962.399928963" lastFinishedPulling="2025-12-03 14:25:32.783123651 +0000 UTC m=+962.446040574" observedRunningTime="2025-12-03 14:25:33.27055922 +0000 UTC m=+962.933476133" watchObservedRunningTime="2025-12-03 14:25:33.274568048 +0000 UTC m=+962.937484971" Dec 03 14:25:33 crc kubenswrapper[4805]: I1203 14:25:33.284618 4805 scope.go:117] "RemoveContainer" containerID="5849ded655a1776fac2e6356f7207035fba4c827079f725e722e852ba9e9bbd1" Dec 03 14:25:33 crc kubenswrapper[4805]: E1203 14:25:33.285108 4805 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5849ded655a1776fac2e6356f7207035fba4c827079f725e722e852ba9e9bbd1\": container with ID starting with 5849ded655a1776fac2e6356f7207035fba4c827079f725e722e852ba9e9bbd1 not found: ID does not exist" containerID="5849ded655a1776fac2e6356f7207035fba4c827079f725e722e852ba9e9bbd1" Dec 03 14:25:33 crc kubenswrapper[4805]: I1203 14:25:33.285192 4805 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5849ded655a1776fac2e6356f7207035fba4c827079f725e722e852ba9e9bbd1"} err="failed to get container status \"5849ded655a1776fac2e6356f7207035fba4c827079f725e722e852ba9e9bbd1\": rpc error: code = NotFound desc = could not find container \"5849ded655a1776fac2e6356f7207035fba4c827079f725e722e852ba9e9bbd1\": container with ID starting with 5849ded655a1776fac2e6356f7207035fba4c827079f725e722e852ba9e9bbd1 not found: ID does not exist" Dec 03 14:25:33 crc kubenswrapper[4805]: I1203 14:25:33.292149 4805 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack-operators/openstack-operator-index-pbpk5"] Dec 03 14:25:33 crc kubenswrapper[4805]: I1203 14:25:33.297285 4805 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack-operators/openstack-operator-index-pbpk5"] Dec 03 14:25:34 crc kubenswrapper[4805]: I1203 14:25:34.709248 4805 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="682558e2-bc38-4aae-aedf-c54605f12dd6" path="/var/lib/kubelet/pods/682558e2-bc38-4aae-aedf-c54605f12dd6/volumes" Dec 03 14:25:42 crc kubenswrapper[4805]: I1203 14:25:42.483769 4805 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/openstack-operator-index-p6vl8" Dec 03 14:25:42 crc kubenswrapper[4805]: I1203 14:25:42.485720 4805 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack-operators/openstack-operator-index-p6vl8" Dec 03 14:25:42 crc kubenswrapper[4805]: I1203 14:25:42.523539 4805 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack-operators/openstack-operator-index-p6vl8" Dec 03 14:25:42 crc kubenswrapper[4805]: I1203 14:25:42.882960 4805 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/frr-k8s-jsrqx" Dec 03 14:25:43 crc kubenswrapper[4805]: I1203 14:25:43.360188 4805 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/openstack-operator-index-p6vl8" Dec 03 14:25:50 crc kubenswrapper[4805]: I1203 14:25:50.090814 4805 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/01b9b0248b373c1370a002b5a413a9a23eeeadf6c66c17ba9e1f19741esbt54"] Dec 03 14:25:50 crc kubenswrapper[4805]: E1203 14:25:50.091631 4805 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="682558e2-bc38-4aae-aedf-c54605f12dd6" containerName="registry-server" Dec 03 14:25:50 crc kubenswrapper[4805]: I1203 14:25:50.091647 4805 state_mem.go:107] "Deleted CPUSet assignment" podUID="682558e2-bc38-4aae-aedf-c54605f12dd6" containerName="registry-server" Dec 03 14:25:50 crc kubenswrapper[4805]: I1203 14:25:50.091767 4805 memory_manager.go:354] "RemoveStaleState removing state" podUID="682558e2-bc38-4aae-aedf-c54605f12dd6" containerName="registry-server" Dec 03 14:25:50 crc kubenswrapper[4805]: I1203 14:25:50.092697 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/01b9b0248b373c1370a002b5a413a9a23eeeadf6c66c17ba9e1f19741esbt54" Dec 03 14:25:50 crc kubenswrapper[4805]: I1203 14:25:50.095557 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"default-dockercfg-lcjqk" Dec 03 14:25:50 crc kubenswrapper[4805]: I1203 14:25:50.105101 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/01b9b0248b373c1370a002b5a413a9a23eeeadf6c66c17ba9e1f19741esbt54"] Dec 03 14:25:50 crc kubenswrapper[4805]: I1203 14:25:50.224285 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/7811f073-78a5-49a2-8b6e-336505b43756-bundle\") pod \"01b9b0248b373c1370a002b5a413a9a23eeeadf6c66c17ba9e1f19741esbt54\" (UID: \"7811f073-78a5-49a2-8b6e-336505b43756\") " pod="openstack-operators/01b9b0248b373c1370a002b5a413a9a23eeeadf6c66c17ba9e1f19741esbt54" Dec 03 14:25:50 crc kubenswrapper[4805]: I1203 14:25:50.224620 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-f4k25\" (UniqueName: \"kubernetes.io/projected/7811f073-78a5-49a2-8b6e-336505b43756-kube-api-access-f4k25\") pod \"01b9b0248b373c1370a002b5a413a9a23eeeadf6c66c17ba9e1f19741esbt54\" (UID: \"7811f073-78a5-49a2-8b6e-336505b43756\") " pod="openstack-operators/01b9b0248b373c1370a002b5a413a9a23eeeadf6c66c17ba9e1f19741esbt54" Dec 03 14:25:50 crc kubenswrapper[4805]: I1203 14:25:50.224785 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/7811f073-78a5-49a2-8b6e-336505b43756-util\") pod \"01b9b0248b373c1370a002b5a413a9a23eeeadf6c66c17ba9e1f19741esbt54\" (UID: \"7811f073-78a5-49a2-8b6e-336505b43756\") " pod="openstack-operators/01b9b0248b373c1370a002b5a413a9a23eeeadf6c66c17ba9e1f19741esbt54" Dec 03 14:25:50 crc kubenswrapper[4805]: I1203 14:25:50.325941 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/7811f073-78a5-49a2-8b6e-336505b43756-bundle\") pod \"01b9b0248b373c1370a002b5a413a9a23eeeadf6c66c17ba9e1f19741esbt54\" (UID: \"7811f073-78a5-49a2-8b6e-336505b43756\") " pod="openstack-operators/01b9b0248b373c1370a002b5a413a9a23eeeadf6c66c17ba9e1f19741esbt54" Dec 03 14:25:50 crc kubenswrapper[4805]: I1203 14:25:50.326010 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-f4k25\" (UniqueName: \"kubernetes.io/projected/7811f073-78a5-49a2-8b6e-336505b43756-kube-api-access-f4k25\") pod \"01b9b0248b373c1370a002b5a413a9a23eeeadf6c66c17ba9e1f19741esbt54\" (UID: \"7811f073-78a5-49a2-8b6e-336505b43756\") " pod="openstack-operators/01b9b0248b373c1370a002b5a413a9a23eeeadf6c66c17ba9e1f19741esbt54" Dec 03 14:25:50 crc kubenswrapper[4805]: I1203 14:25:50.326067 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/7811f073-78a5-49a2-8b6e-336505b43756-util\") pod \"01b9b0248b373c1370a002b5a413a9a23eeeadf6c66c17ba9e1f19741esbt54\" (UID: \"7811f073-78a5-49a2-8b6e-336505b43756\") " pod="openstack-operators/01b9b0248b373c1370a002b5a413a9a23eeeadf6c66c17ba9e1f19741esbt54" Dec 03 14:25:50 crc kubenswrapper[4805]: I1203 14:25:50.326688 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/7811f073-78a5-49a2-8b6e-336505b43756-util\") pod \"01b9b0248b373c1370a002b5a413a9a23eeeadf6c66c17ba9e1f19741esbt54\" (UID: \"7811f073-78a5-49a2-8b6e-336505b43756\") " pod="openstack-operators/01b9b0248b373c1370a002b5a413a9a23eeeadf6c66c17ba9e1f19741esbt54" Dec 03 14:25:50 crc kubenswrapper[4805]: I1203 14:25:50.326948 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/7811f073-78a5-49a2-8b6e-336505b43756-bundle\") pod \"01b9b0248b373c1370a002b5a413a9a23eeeadf6c66c17ba9e1f19741esbt54\" (UID: \"7811f073-78a5-49a2-8b6e-336505b43756\") " pod="openstack-operators/01b9b0248b373c1370a002b5a413a9a23eeeadf6c66c17ba9e1f19741esbt54" Dec 03 14:25:50 crc kubenswrapper[4805]: I1203 14:25:50.351004 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-f4k25\" (UniqueName: \"kubernetes.io/projected/7811f073-78a5-49a2-8b6e-336505b43756-kube-api-access-f4k25\") pod \"01b9b0248b373c1370a002b5a413a9a23eeeadf6c66c17ba9e1f19741esbt54\" (UID: \"7811f073-78a5-49a2-8b6e-336505b43756\") " pod="openstack-operators/01b9b0248b373c1370a002b5a413a9a23eeeadf6c66c17ba9e1f19741esbt54" Dec 03 14:25:50 crc kubenswrapper[4805]: I1203 14:25:50.420833 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/01b9b0248b373c1370a002b5a413a9a23eeeadf6c66c17ba9e1f19741esbt54" Dec 03 14:25:50 crc kubenswrapper[4805]: I1203 14:25:50.858167 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/01b9b0248b373c1370a002b5a413a9a23eeeadf6c66c17ba9e1f19741esbt54"] Dec 03 14:25:50 crc kubenswrapper[4805]: W1203 14:25:50.861347 4805 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod7811f073_78a5_49a2_8b6e_336505b43756.slice/crio-b96ec67bb239acf38221142988631ecb3def8d996c01aaa4b21cd1fb00402f09 WatchSource:0}: Error finding container b96ec67bb239acf38221142988631ecb3def8d996c01aaa4b21cd1fb00402f09: Status 404 returned error can't find the container with id b96ec67bb239acf38221142988631ecb3def8d996c01aaa4b21cd1fb00402f09 Dec 03 14:25:51 crc kubenswrapper[4805]: I1203 14:25:51.378445 4805 generic.go:334] "Generic (PLEG): container finished" podID="7811f073-78a5-49a2-8b6e-336505b43756" containerID="358109620cd577f0c51a1195911ac79bfb219bdb78db569e2d2bdbada95eec58" exitCode=0 Dec 03 14:25:51 crc kubenswrapper[4805]: I1203 14:25:51.378500 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/01b9b0248b373c1370a002b5a413a9a23eeeadf6c66c17ba9e1f19741esbt54" event={"ID":"7811f073-78a5-49a2-8b6e-336505b43756","Type":"ContainerDied","Data":"358109620cd577f0c51a1195911ac79bfb219bdb78db569e2d2bdbada95eec58"} Dec 03 14:25:51 crc kubenswrapper[4805]: I1203 14:25:51.378712 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/01b9b0248b373c1370a002b5a413a9a23eeeadf6c66c17ba9e1f19741esbt54" event={"ID":"7811f073-78a5-49a2-8b6e-336505b43756","Type":"ContainerStarted","Data":"b96ec67bb239acf38221142988631ecb3def8d996c01aaa4b21cd1fb00402f09"} Dec 03 14:25:52 crc kubenswrapper[4805]: I1203 14:25:52.386425 4805 generic.go:334] "Generic (PLEG): container finished" podID="7811f073-78a5-49a2-8b6e-336505b43756" containerID="a56e61a628e52fda087232fb4fc71bcbad1888cbbb130e56934079f058b20855" exitCode=0 Dec 03 14:25:52 crc kubenswrapper[4805]: I1203 14:25:52.386479 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/01b9b0248b373c1370a002b5a413a9a23eeeadf6c66c17ba9e1f19741esbt54" event={"ID":"7811f073-78a5-49a2-8b6e-336505b43756","Type":"ContainerDied","Data":"a56e61a628e52fda087232fb4fc71bcbad1888cbbb130e56934079f058b20855"} Dec 03 14:25:53 crc kubenswrapper[4805]: I1203 14:25:53.400215 4805 generic.go:334] "Generic (PLEG): container finished" podID="7811f073-78a5-49a2-8b6e-336505b43756" containerID="ab2e30a8c6dd773199ee2b993587a7ab183abf70edb4c7ae5ae86325cd0571dc" exitCode=0 Dec 03 14:25:53 crc kubenswrapper[4805]: I1203 14:25:53.400303 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/01b9b0248b373c1370a002b5a413a9a23eeeadf6c66c17ba9e1f19741esbt54" event={"ID":"7811f073-78a5-49a2-8b6e-336505b43756","Type":"ContainerDied","Data":"ab2e30a8c6dd773199ee2b993587a7ab183abf70edb4c7ae5ae86325cd0571dc"} Dec 03 14:25:54 crc kubenswrapper[4805]: I1203 14:25:54.714474 4805 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/01b9b0248b373c1370a002b5a413a9a23eeeadf6c66c17ba9e1f19741esbt54" Dec 03 14:25:54 crc kubenswrapper[4805]: I1203 14:25:54.787400 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/7811f073-78a5-49a2-8b6e-336505b43756-bundle\") pod \"7811f073-78a5-49a2-8b6e-336505b43756\" (UID: \"7811f073-78a5-49a2-8b6e-336505b43756\") " Dec 03 14:25:54 crc kubenswrapper[4805]: I1203 14:25:54.787466 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/7811f073-78a5-49a2-8b6e-336505b43756-util\") pod \"7811f073-78a5-49a2-8b6e-336505b43756\" (UID: \"7811f073-78a5-49a2-8b6e-336505b43756\") " Dec 03 14:25:54 crc kubenswrapper[4805]: I1203 14:25:54.787602 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-f4k25\" (UniqueName: \"kubernetes.io/projected/7811f073-78a5-49a2-8b6e-336505b43756-kube-api-access-f4k25\") pod \"7811f073-78a5-49a2-8b6e-336505b43756\" (UID: \"7811f073-78a5-49a2-8b6e-336505b43756\") " Dec 03 14:25:54 crc kubenswrapper[4805]: I1203 14:25:54.788175 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7811f073-78a5-49a2-8b6e-336505b43756-bundle" (OuterVolumeSpecName: "bundle") pod "7811f073-78a5-49a2-8b6e-336505b43756" (UID: "7811f073-78a5-49a2-8b6e-336505b43756"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 14:25:54 crc kubenswrapper[4805]: I1203 14:25:54.792999 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7811f073-78a5-49a2-8b6e-336505b43756-kube-api-access-f4k25" (OuterVolumeSpecName: "kube-api-access-f4k25") pod "7811f073-78a5-49a2-8b6e-336505b43756" (UID: "7811f073-78a5-49a2-8b6e-336505b43756"). InnerVolumeSpecName "kube-api-access-f4k25". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 14:25:54 crc kubenswrapper[4805]: I1203 14:25:54.803193 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7811f073-78a5-49a2-8b6e-336505b43756-util" (OuterVolumeSpecName: "util") pod "7811f073-78a5-49a2-8b6e-336505b43756" (UID: "7811f073-78a5-49a2-8b6e-336505b43756"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 14:25:54 crc kubenswrapper[4805]: I1203 14:25:54.889090 4805 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-f4k25\" (UniqueName: \"kubernetes.io/projected/7811f073-78a5-49a2-8b6e-336505b43756-kube-api-access-f4k25\") on node \"crc\" DevicePath \"\"" Dec 03 14:25:54 crc kubenswrapper[4805]: I1203 14:25:54.889126 4805 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/7811f073-78a5-49a2-8b6e-336505b43756-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 14:25:54 crc kubenswrapper[4805]: I1203 14:25:54.889136 4805 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/7811f073-78a5-49a2-8b6e-336505b43756-util\") on node \"crc\" DevicePath \"\"" Dec 03 14:25:55 crc kubenswrapper[4805]: I1203 14:25:55.419052 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/01b9b0248b373c1370a002b5a413a9a23eeeadf6c66c17ba9e1f19741esbt54" event={"ID":"7811f073-78a5-49a2-8b6e-336505b43756","Type":"ContainerDied","Data":"b96ec67bb239acf38221142988631ecb3def8d996c01aaa4b21cd1fb00402f09"} Dec 03 14:25:55 crc kubenswrapper[4805]: I1203 14:25:55.419357 4805 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="b96ec67bb239acf38221142988631ecb3def8d996c01aaa4b21cd1fb00402f09" Dec 03 14:25:55 crc kubenswrapper[4805]: I1203 14:25:55.419123 4805 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/01b9b0248b373c1370a002b5a413a9a23eeeadf6c66c17ba9e1f19741esbt54" Dec 03 14:26:02 crc kubenswrapper[4805]: I1203 14:26:02.531162 4805 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-operator-controller-operator-796df97b8b-rmv2c"] Dec 03 14:26:02 crc kubenswrapper[4805]: E1203 14:26:02.532184 4805 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7811f073-78a5-49a2-8b6e-336505b43756" containerName="extract" Dec 03 14:26:02 crc kubenswrapper[4805]: I1203 14:26:02.532208 4805 state_mem.go:107] "Deleted CPUSet assignment" podUID="7811f073-78a5-49a2-8b6e-336505b43756" containerName="extract" Dec 03 14:26:02 crc kubenswrapper[4805]: E1203 14:26:02.532230 4805 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7811f073-78a5-49a2-8b6e-336505b43756" containerName="util" Dec 03 14:26:02 crc kubenswrapper[4805]: I1203 14:26:02.532242 4805 state_mem.go:107] "Deleted CPUSet assignment" podUID="7811f073-78a5-49a2-8b6e-336505b43756" containerName="util" Dec 03 14:26:02 crc kubenswrapper[4805]: E1203 14:26:02.532263 4805 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7811f073-78a5-49a2-8b6e-336505b43756" containerName="pull" Dec 03 14:26:02 crc kubenswrapper[4805]: I1203 14:26:02.532275 4805 state_mem.go:107] "Deleted CPUSet assignment" podUID="7811f073-78a5-49a2-8b6e-336505b43756" containerName="pull" Dec 03 14:26:02 crc kubenswrapper[4805]: I1203 14:26:02.532475 4805 memory_manager.go:354] "RemoveStaleState removing state" podUID="7811f073-78a5-49a2-8b6e-336505b43756" containerName="extract" Dec 03 14:26:02 crc kubenswrapper[4805]: I1203 14:26:02.533149 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-controller-operator-796df97b8b-rmv2c" Dec 03 14:26:02 crc kubenswrapper[4805]: I1203 14:26:02.535327 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-operator-controller-operator-dockercfg-fqsxj" Dec 03 14:26:02 crc kubenswrapper[4805]: I1203 14:26:02.556111 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-controller-operator-796df97b8b-rmv2c"] Dec 03 14:26:02 crc kubenswrapper[4805]: I1203 14:26:02.707529 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8mm8c\" (UniqueName: \"kubernetes.io/projected/9da93959-4eef-466e-aab4-1a36202b9970-kube-api-access-8mm8c\") pod \"openstack-operator-controller-operator-796df97b8b-rmv2c\" (UID: \"9da93959-4eef-466e-aab4-1a36202b9970\") " pod="openstack-operators/openstack-operator-controller-operator-796df97b8b-rmv2c" Dec 03 14:26:02 crc kubenswrapper[4805]: I1203 14:26:02.808531 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8mm8c\" (UniqueName: \"kubernetes.io/projected/9da93959-4eef-466e-aab4-1a36202b9970-kube-api-access-8mm8c\") pod \"openstack-operator-controller-operator-796df97b8b-rmv2c\" (UID: \"9da93959-4eef-466e-aab4-1a36202b9970\") " pod="openstack-operators/openstack-operator-controller-operator-796df97b8b-rmv2c" Dec 03 14:26:02 crc kubenswrapper[4805]: I1203 14:26:02.831826 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8mm8c\" (UniqueName: \"kubernetes.io/projected/9da93959-4eef-466e-aab4-1a36202b9970-kube-api-access-8mm8c\") pod \"openstack-operator-controller-operator-796df97b8b-rmv2c\" (UID: \"9da93959-4eef-466e-aab4-1a36202b9970\") " pod="openstack-operators/openstack-operator-controller-operator-796df97b8b-rmv2c" Dec 03 14:26:02 crc kubenswrapper[4805]: I1203 14:26:02.857656 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-controller-operator-796df97b8b-rmv2c" Dec 03 14:26:03 crc kubenswrapper[4805]: I1203 14:26:03.287216 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-controller-operator-796df97b8b-rmv2c"] Dec 03 14:26:03 crc kubenswrapper[4805]: I1203 14:26:03.479407 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-operator-796df97b8b-rmv2c" event={"ID":"9da93959-4eef-466e-aab4-1a36202b9970","Type":"ContainerStarted","Data":"6f3d09074cb6104fff66a3ef760ea3ce28003c2ec250461153840a55bc62992f"} Dec 03 14:26:09 crc kubenswrapper[4805]: I1203 14:26:09.518997 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-operator-796df97b8b-rmv2c" event={"ID":"9da93959-4eef-466e-aab4-1a36202b9970","Type":"ContainerStarted","Data":"5dd3ecf475008bb30c27fde0ca490a119d78c2ba3c1da6f32b2d699b80649aab"} Dec 03 14:26:09 crc kubenswrapper[4805]: I1203 14:26:09.519509 4805 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/openstack-operator-controller-operator-796df97b8b-rmv2c" Dec 03 14:26:09 crc kubenswrapper[4805]: I1203 14:26:09.568633 4805 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-operator-controller-operator-796df97b8b-rmv2c" podStartSLOduration=2.030976146 podStartE2EDuration="7.568607641s" podCreationTimestamp="2025-12-03 14:26:02 +0000 UTC" firstStartedPulling="2025-12-03 14:26:03.299006265 +0000 UTC m=+992.961923188" lastFinishedPulling="2025-12-03 14:26:08.83663777 +0000 UTC m=+998.499554683" observedRunningTime="2025-12-03 14:26:09.564647353 +0000 UTC m=+999.227564326" watchObservedRunningTime="2025-12-03 14:26:09.568607641 +0000 UTC m=+999.231524594" Dec 03 14:26:13 crc kubenswrapper[4805]: I1203 14:26:13.917683 4805 patch_prober.go:28] interesting pod/machine-config-daemon-gskh4 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 03 14:26:13 crc kubenswrapper[4805]: I1203 14:26:13.918437 4805 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-gskh4" podUID="6098937f-e3f6-45e8-a647-4994a79cd711" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 03 14:26:22 crc kubenswrapper[4805]: I1203 14:26:22.860449 4805 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/openstack-operator-controller-operator-796df97b8b-rmv2c" Dec 03 14:26:43 crc kubenswrapper[4805]: I1203 14:26:43.917739 4805 patch_prober.go:28] interesting pod/machine-config-daemon-gskh4 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 03 14:26:43 crc kubenswrapper[4805]: I1203 14:26:43.919038 4805 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-gskh4" podUID="6098937f-e3f6-45e8-a647-4994a79cd711" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 03 14:26:50 crc kubenswrapper[4805]: I1203 14:26:50.793118 4805 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/barbican-operator-controller-manager-7d9dfd778-jc4dw"] Dec 03 14:26:50 crc kubenswrapper[4805]: I1203 14:26:50.794351 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/barbican-operator-controller-manager-7d9dfd778-jc4dw" Dec 03 14:26:50 crc kubenswrapper[4805]: I1203 14:26:50.796415 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"barbican-operator-controller-manager-dockercfg-2sxt6" Dec 03 14:26:50 crc kubenswrapper[4805]: I1203 14:26:50.798498 4805 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/cinder-operator-controller-manager-859b6ccc6-2nzgx"] Dec 03 14:26:50 crc kubenswrapper[4805]: I1203 14:26:50.799683 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/cinder-operator-controller-manager-859b6ccc6-2nzgx" Dec 03 14:26:50 crc kubenswrapper[4805]: I1203 14:26:50.801765 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"cinder-operator-controller-manager-dockercfg-zd2dn" Dec 03 14:26:50 crc kubenswrapper[4805]: I1203 14:26:50.806561 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/barbican-operator-controller-manager-7d9dfd778-jc4dw"] Dec 03 14:26:50 crc kubenswrapper[4805]: I1203 14:26:50.819214 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/cinder-operator-controller-manager-859b6ccc6-2nzgx"] Dec 03 14:26:50 crc kubenswrapper[4805]: I1203 14:26:50.826559 4805 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/designate-operator-controller-manager-78b4bc895b-w56sr"] Dec 03 14:26:50 crc kubenswrapper[4805]: I1203 14:26:50.828010 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/designate-operator-controller-manager-78b4bc895b-w56sr" Dec 03 14:26:50 crc kubenswrapper[4805]: I1203 14:26:50.830650 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"designate-operator-controller-manager-dockercfg-qszrx" Dec 03 14:26:50 crc kubenswrapper[4805]: I1203 14:26:50.847145 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/designate-operator-controller-manager-78b4bc895b-w56sr"] Dec 03 14:26:50 crc kubenswrapper[4805]: I1203 14:26:50.857148 4805 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/glance-operator-controller-manager-77987cd8cd-pt9fw"] Dec 03 14:26:50 crc kubenswrapper[4805]: I1203 14:26:50.870278 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/glance-operator-controller-manager-77987cd8cd-pt9fw" Dec 03 14:26:50 crc kubenswrapper[4805]: I1203 14:26:50.910533 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"glance-operator-controller-manager-dockercfg-z7dn9" Dec 03 14:26:50 crc kubenswrapper[4805]: I1203 14:26:50.915952 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/glance-operator-controller-manager-77987cd8cd-pt9fw"] Dec 03 14:26:50 crc kubenswrapper[4805]: I1203 14:26:50.943887 4805 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/heat-operator-controller-manager-5f64f6f8bb-86sbc"] Dec 03 14:26:50 crc kubenswrapper[4805]: I1203 14:26:50.953310 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/heat-operator-controller-manager-5f64f6f8bb-86sbc" Dec 03 14:26:50 crc kubenswrapper[4805]: I1203 14:26:50.960546 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"heat-operator-controller-manager-dockercfg-hdbxk" Dec 03 14:26:50 crc kubenswrapper[4805]: I1203 14:26:50.987881 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/heat-operator-controller-manager-5f64f6f8bb-86sbc"] Dec 03 14:26:51 crc kubenswrapper[4805]: I1203 14:26:51.008957 4805 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/horizon-operator-controller-manager-68c6d99b8f-kxbr8"] Dec 03 14:26:51 crc kubenswrapper[4805]: I1203 14:26:51.009869 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/horizon-operator-controller-manager-68c6d99b8f-kxbr8" Dec 03 14:26:51 crc kubenswrapper[4805]: I1203 14:26:51.014633 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-462zf\" (UniqueName: \"kubernetes.io/projected/a748cfe4-2e54-4a7b-a353-e1dfea6c14b3-kube-api-access-462zf\") pod \"heat-operator-controller-manager-5f64f6f8bb-86sbc\" (UID: \"a748cfe4-2e54-4a7b-a353-e1dfea6c14b3\") " pod="openstack-operators/heat-operator-controller-manager-5f64f6f8bb-86sbc" Dec 03 14:26:51 crc kubenswrapper[4805]: I1203 14:26:51.014680 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pxqph\" (UniqueName: \"kubernetes.io/projected/b29e4120-0cbc-4de5-a4b6-97b97761f49a-kube-api-access-pxqph\") pod \"horizon-operator-controller-manager-68c6d99b8f-kxbr8\" (UID: \"b29e4120-0cbc-4de5-a4b6-97b97761f49a\") " pod="openstack-operators/horizon-operator-controller-manager-68c6d99b8f-kxbr8" Dec 03 14:26:51 crc kubenswrapper[4805]: I1203 14:26:51.014705 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-89ppm\" (UniqueName: \"kubernetes.io/projected/f9b6b508-1cf6-40aa-829e-b9cf264921e5-kube-api-access-89ppm\") pod \"cinder-operator-controller-manager-859b6ccc6-2nzgx\" (UID: \"f9b6b508-1cf6-40aa-829e-b9cf264921e5\") " pod="openstack-operators/cinder-operator-controller-manager-859b6ccc6-2nzgx" Dec 03 14:26:51 crc kubenswrapper[4805]: I1203 14:26:51.014721 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-f9wlh\" (UniqueName: \"kubernetes.io/projected/5615ce63-1ec1-4a77-8e44-49e6d114eec6-kube-api-access-f9wlh\") pod \"glance-operator-controller-manager-77987cd8cd-pt9fw\" (UID: \"5615ce63-1ec1-4a77-8e44-49e6d114eec6\") " pod="openstack-operators/glance-operator-controller-manager-77987cd8cd-pt9fw" Dec 03 14:26:51 crc kubenswrapper[4805]: I1203 14:26:51.014749 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4dpg9\" (UniqueName: \"kubernetes.io/projected/326305da-3798-498c-9b47-0aeb71bd2205-kube-api-access-4dpg9\") pod \"designate-operator-controller-manager-78b4bc895b-w56sr\" (UID: \"326305da-3798-498c-9b47-0aeb71bd2205\") " pod="openstack-operators/designate-operator-controller-manager-78b4bc895b-w56sr" Dec 03 14:26:51 crc kubenswrapper[4805]: I1203 14:26:51.014766 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-s9bwv\" (UniqueName: \"kubernetes.io/projected/cbf3813c-fb86-49d2-8fac-25487ff08601-kube-api-access-s9bwv\") pod \"barbican-operator-controller-manager-7d9dfd778-jc4dw\" (UID: \"cbf3813c-fb86-49d2-8fac-25487ff08601\") " pod="openstack-operators/barbican-operator-controller-manager-7d9dfd778-jc4dw" Dec 03 14:26:51 crc kubenswrapper[4805]: I1203 14:26:51.017267 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"horizon-operator-controller-manager-dockercfg-kfbcg" Dec 03 14:26:51 crc kubenswrapper[4805]: I1203 14:26:51.027420 4805 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/infra-operator-controller-manager-57548d458d-77t9s"] Dec 03 14:26:51 crc kubenswrapper[4805]: I1203 14:26:51.028351 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/infra-operator-controller-manager-57548d458d-77t9s" Dec 03 14:26:51 crc kubenswrapper[4805]: I1203 14:26:51.031776 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"infra-operator-controller-manager-dockercfg-l4pt8" Dec 03 14:26:51 crc kubenswrapper[4805]: I1203 14:26:51.035160 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"infra-operator-webhook-server-cert" Dec 03 14:26:51 crc kubenswrapper[4805]: I1203 14:26:51.042027 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/horizon-operator-controller-manager-68c6d99b8f-kxbr8"] Dec 03 14:26:51 crc kubenswrapper[4805]: I1203 14:26:51.071891 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/infra-operator-controller-manager-57548d458d-77t9s"] Dec 03 14:26:51 crc kubenswrapper[4805]: I1203 14:26:51.090911 4805 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/ironic-operator-controller-manager-6c548fd776-mz4pw"] Dec 03 14:26:51 crc kubenswrapper[4805]: I1203 14:26:51.092313 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/ironic-operator-controller-manager-6c548fd776-mz4pw" Dec 03 14:26:51 crc kubenswrapper[4805]: I1203 14:26:51.094829 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"ironic-operator-controller-manager-dockercfg-gkwld" Dec 03 14:26:51 crc kubenswrapper[4805]: I1203 14:26:51.100171 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/ironic-operator-controller-manager-6c548fd776-mz4pw"] Dec 03 14:26:51 crc kubenswrapper[4805]: I1203 14:26:51.113547 4805 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/keystone-operator-controller-manager-7765d96ddf-qzzz9"] Dec 03 14:26:51 crc kubenswrapper[4805]: I1203 14:26:51.114631 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/keystone-operator-controller-manager-7765d96ddf-qzzz9" Dec 03 14:26:51 crc kubenswrapper[4805]: I1203 14:26:51.115937 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-462zf\" (UniqueName: \"kubernetes.io/projected/a748cfe4-2e54-4a7b-a353-e1dfea6c14b3-kube-api-access-462zf\") pod \"heat-operator-controller-manager-5f64f6f8bb-86sbc\" (UID: \"a748cfe4-2e54-4a7b-a353-e1dfea6c14b3\") " pod="openstack-operators/heat-operator-controller-manager-5f64f6f8bb-86sbc" Dec 03 14:26:51 crc kubenswrapper[4805]: I1203 14:26:51.115970 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pxqph\" (UniqueName: \"kubernetes.io/projected/b29e4120-0cbc-4de5-a4b6-97b97761f49a-kube-api-access-pxqph\") pod \"horizon-operator-controller-manager-68c6d99b8f-kxbr8\" (UID: \"b29e4120-0cbc-4de5-a4b6-97b97761f49a\") " pod="openstack-operators/horizon-operator-controller-manager-68c6d99b8f-kxbr8" Dec 03 14:26:51 crc kubenswrapper[4805]: I1203 14:26:51.115993 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-89ppm\" (UniqueName: \"kubernetes.io/projected/f9b6b508-1cf6-40aa-829e-b9cf264921e5-kube-api-access-89ppm\") pod \"cinder-operator-controller-manager-859b6ccc6-2nzgx\" (UID: \"f9b6b508-1cf6-40aa-829e-b9cf264921e5\") " pod="openstack-operators/cinder-operator-controller-manager-859b6ccc6-2nzgx" Dec 03 14:26:51 crc kubenswrapper[4805]: I1203 14:26:51.116010 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-f9wlh\" (UniqueName: \"kubernetes.io/projected/5615ce63-1ec1-4a77-8e44-49e6d114eec6-kube-api-access-f9wlh\") pod \"glance-operator-controller-manager-77987cd8cd-pt9fw\" (UID: \"5615ce63-1ec1-4a77-8e44-49e6d114eec6\") " pod="openstack-operators/glance-operator-controller-manager-77987cd8cd-pt9fw" Dec 03 14:26:51 crc kubenswrapper[4805]: I1203 14:26:51.116041 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4dpg9\" (UniqueName: \"kubernetes.io/projected/326305da-3798-498c-9b47-0aeb71bd2205-kube-api-access-4dpg9\") pod \"designate-operator-controller-manager-78b4bc895b-w56sr\" (UID: \"326305da-3798-498c-9b47-0aeb71bd2205\") " pod="openstack-operators/designate-operator-controller-manager-78b4bc895b-w56sr" Dec 03 14:26:51 crc kubenswrapper[4805]: I1203 14:26:51.116062 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/f21790d8-af47-4be7-8758-f038bda76908-cert\") pod \"infra-operator-controller-manager-57548d458d-77t9s\" (UID: \"f21790d8-af47-4be7-8758-f038bda76908\") " pod="openstack-operators/infra-operator-controller-manager-57548d458d-77t9s" Dec 03 14:26:51 crc kubenswrapper[4805]: I1203 14:26:51.116090 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s9bwv\" (UniqueName: \"kubernetes.io/projected/cbf3813c-fb86-49d2-8fac-25487ff08601-kube-api-access-s9bwv\") pod \"barbican-operator-controller-manager-7d9dfd778-jc4dw\" (UID: \"cbf3813c-fb86-49d2-8fac-25487ff08601\") " pod="openstack-operators/barbican-operator-controller-manager-7d9dfd778-jc4dw" Dec 03 14:26:51 crc kubenswrapper[4805]: I1203 14:26:51.116112 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-scslw\" (UniqueName: \"kubernetes.io/projected/f21790d8-af47-4be7-8758-f038bda76908-kube-api-access-scslw\") pod \"infra-operator-controller-manager-57548d458d-77t9s\" (UID: \"f21790d8-af47-4be7-8758-f038bda76908\") " pod="openstack-operators/infra-operator-controller-manager-57548d458d-77t9s" Dec 03 14:26:51 crc kubenswrapper[4805]: I1203 14:26:51.121684 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/keystone-operator-controller-manager-7765d96ddf-qzzz9"] Dec 03 14:26:51 crc kubenswrapper[4805]: I1203 14:26:51.126192 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"keystone-operator-controller-manager-dockercfg-bnvqs" Dec 03 14:26:51 crc kubenswrapper[4805]: I1203 14:26:51.136292 4805 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/manila-operator-controller-manager-7c79b5df47-d2thc"] Dec 03 14:26:51 crc kubenswrapper[4805]: I1203 14:26:51.140208 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/manila-operator-controller-manager-7c79b5df47-d2thc" Dec 03 14:26:51 crc kubenswrapper[4805]: I1203 14:26:51.143679 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"manila-operator-controller-manager-dockercfg-k7tkq" Dec 03 14:26:51 crc kubenswrapper[4805]: I1203 14:26:51.154073 4805 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/mariadb-operator-controller-manager-56bbcc9d85-n6m2d"] Dec 03 14:26:51 crc kubenswrapper[4805]: I1203 14:26:51.155025 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/mariadb-operator-controller-manager-56bbcc9d85-n6m2d" Dec 03 14:26:51 crc kubenswrapper[4805]: I1203 14:26:51.157336 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/manila-operator-controller-manager-7c79b5df47-d2thc"] Dec 03 14:26:51 crc kubenswrapper[4805]: I1203 14:26:51.164781 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/mariadb-operator-controller-manager-56bbcc9d85-n6m2d"] Dec 03 14:26:51 crc kubenswrapper[4805]: I1203 14:26:51.171651 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s9bwv\" (UniqueName: \"kubernetes.io/projected/cbf3813c-fb86-49d2-8fac-25487ff08601-kube-api-access-s9bwv\") pod \"barbican-operator-controller-manager-7d9dfd778-jc4dw\" (UID: \"cbf3813c-fb86-49d2-8fac-25487ff08601\") " pod="openstack-operators/barbican-operator-controller-manager-7d9dfd778-jc4dw" Dec 03 14:26:51 crc kubenswrapper[4805]: I1203 14:26:51.179138 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"mariadb-operator-controller-manager-dockercfg-kzz5z" Dec 03 14:26:51 crc kubenswrapper[4805]: I1203 14:26:51.179779 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pxqph\" (UniqueName: \"kubernetes.io/projected/b29e4120-0cbc-4de5-a4b6-97b97761f49a-kube-api-access-pxqph\") pod \"horizon-operator-controller-manager-68c6d99b8f-kxbr8\" (UID: \"b29e4120-0cbc-4de5-a4b6-97b97761f49a\") " pod="openstack-operators/horizon-operator-controller-manager-68c6d99b8f-kxbr8" Dec 03 14:26:51 crc kubenswrapper[4805]: I1203 14:26:51.180291 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-89ppm\" (UniqueName: \"kubernetes.io/projected/f9b6b508-1cf6-40aa-829e-b9cf264921e5-kube-api-access-89ppm\") pod \"cinder-operator-controller-manager-859b6ccc6-2nzgx\" (UID: \"f9b6b508-1cf6-40aa-829e-b9cf264921e5\") " pod="openstack-operators/cinder-operator-controller-manager-859b6ccc6-2nzgx" Dec 03 14:26:51 crc kubenswrapper[4805]: I1203 14:26:51.180611 4805 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/neutron-operator-controller-manager-5fdfd5b6b5-4mlhk"] Dec 03 14:26:51 crc kubenswrapper[4805]: I1203 14:26:51.184802 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/neutron-operator-controller-manager-5fdfd5b6b5-4mlhk" Dec 03 14:26:51 crc kubenswrapper[4805]: I1203 14:26:51.195629 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4dpg9\" (UniqueName: \"kubernetes.io/projected/326305da-3798-498c-9b47-0aeb71bd2205-kube-api-access-4dpg9\") pod \"designate-operator-controller-manager-78b4bc895b-w56sr\" (UID: \"326305da-3798-498c-9b47-0aeb71bd2205\") " pod="openstack-operators/designate-operator-controller-manager-78b4bc895b-w56sr" Dec 03 14:26:51 crc kubenswrapper[4805]: I1203 14:26:51.195761 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/neutron-operator-controller-manager-5fdfd5b6b5-4mlhk"] Dec 03 14:26:51 crc kubenswrapper[4805]: I1203 14:26:51.196936 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"neutron-operator-controller-manager-dockercfg-mchz4" Dec 03 14:26:51 crc kubenswrapper[4805]: I1203 14:26:51.197853 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-f9wlh\" (UniqueName: \"kubernetes.io/projected/5615ce63-1ec1-4a77-8e44-49e6d114eec6-kube-api-access-f9wlh\") pod \"glance-operator-controller-manager-77987cd8cd-pt9fw\" (UID: \"5615ce63-1ec1-4a77-8e44-49e6d114eec6\") " pod="openstack-operators/glance-operator-controller-manager-77987cd8cd-pt9fw" Dec 03 14:26:51 crc kubenswrapper[4805]: I1203 14:26:51.197911 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-462zf\" (UniqueName: \"kubernetes.io/projected/a748cfe4-2e54-4a7b-a353-e1dfea6c14b3-kube-api-access-462zf\") pod \"heat-operator-controller-manager-5f64f6f8bb-86sbc\" (UID: \"a748cfe4-2e54-4a7b-a353-e1dfea6c14b3\") " pod="openstack-operators/heat-operator-controller-manager-5f64f6f8bb-86sbc" Dec 03 14:26:51 crc kubenswrapper[4805]: I1203 14:26:51.217816 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hmzsn\" (UniqueName: \"kubernetes.io/projected/ffa72db0-588a-4a78-a1cd-8710f5ce4f41-kube-api-access-hmzsn\") pod \"keystone-operator-controller-manager-7765d96ddf-qzzz9\" (UID: \"ffa72db0-588a-4a78-a1cd-8710f5ce4f41\") " pod="openstack-operators/keystone-operator-controller-manager-7765d96ddf-qzzz9" Dec 03 14:26:51 crc kubenswrapper[4805]: I1203 14:26:51.217869 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/f21790d8-af47-4be7-8758-f038bda76908-cert\") pod \"infra-operator-controller-manager-57548d458d-77t9s\" (UID: \"f21790d8-af47-4be7-8758-f038bda76908\") " pod="openstack-operators/infra-operator-controller-manager-57548d458d-77t9s" Dec 03 14:26:51 crc kubenswrapper[4805]: I1203 14:26:51.217896 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fnmcc\" (UniqueName: \"kubernetes.io/projected/1ad4fa11-2f11-4a06-9944-bba8c7445e84-kube-api-access-fnmcc\") pod \"ironic-operator-controller-manager-6c548fd776-mz4pw\" (UID: \"1ad4fa11-2f11-4a06-9944-bba8c7445e84\") " pod="openstack-operators/ironic-operator-controller-manager-6c548fd776-mz4pw" Dec 03 14:26:51 crc kubenswrapper[4805]: I1203 14:26:51.217921 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-scslw\" (UniqueName: \"kubernetes.io/projected/f21790d8-af47-4be7-8758-f038bda76908-kube-api-access-scslw\") pod \"infra-operator-controller-manager-57548d458d-77t9s\" (UID: \"f21790d8-af47-4be7-8758-f038bda76908\") " pod="openstack-operators/infra-operator-controller-manager-57548d458d-77t9s" Dec 03 14:26:51 crc kubenswrapper[4805]: E1203 14:26:51.218639 4805 secret.go:188] Couldn't get secret openstack-operators/infra-operator-webhook-server-cert: secret "infra-operator-webhook-server-cert" not found Dec 03 14:26:51 crc kubenswrapper[4805]: E1203 14:26:51.218694 4805 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/f21790d8-af47-4be7-8758-f038bda76908-cert podName:f21790d8-af47-4be7-8758-f038bda76908 nodeName:}" failed. No retries permitted until 2025-12-03 14:26:51.71866883 +0000 UTC m=+1041.381585753 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/f21790d8-af47-4be7-8758-f038bda76908-cert") pod "infra-operator-controller-manager-57548d458d-77t9s" (UID: "f21790d8-af47-4be7-8758-f038bda76908") : secret "infra-operator-webhook-server-cert" not found Dec 03 14:26:51 crc kubenswrapper[4805]: I1203 14:26:51.247869 4805 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/nova-operator-controller-manager-697bc559fc-5rl85"] Dec 03 14:26:51 crc kubenswrapper[4805]: I1203 14:26:51.248899 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/nova-operator-controller-manager-697bc559fc-5rl85" Dec 03 14:26:51 crc kubenswrapper[4805]: I1203 14:26:51.250726 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/glance-operator-controller-manager-77987cd8cd-pt9fw" Dec 03 14:26:51 crc kubenswrapper[4805]: I1203 14:26:51.257789 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"nova-operator-controller-manager-dockercfg-rp4sm" Dec 03 14:26:51 crc kubenswrapper[4805]: I1203 14:26:51.279403 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-scslw\" (UniqueName: \"kubernetes.io/projected/f21790d8-af47-4be7-8758-f038bda76908-kube-api-access-scslw\") pod \"infra-operator-controller-manager-57548d458d-77t9s\" (UID: \"f21790d8-af47-4be7-8758-f038bda76908\") " pod="openstack-operators/infra-operator-controller-manager-57548d458d-77t9s" Dec 03 14:26:51 crc kubenswrapper[4805]: I1203 14:26:51.281443 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/nova-operator-controller-manager-697bc559fc-5rl85"] Dec 03 14:26:51 crc kubenswrapper[4805]: I1203 14:26:51.311353 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/heat-operator-controller-manager-5f64f6f8bb-86sbc" Dec 03 14:26:51 crc kubenswrapper[4805]: I1203 14:26:51.319910 4805 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/octavia-operator-controller-manager-998648c74-8mbdb"] Dec 03 14:26:51 crc kubenswrapper[4805]: I1203 14:26:51.322948 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6s4x4\" (UniqueName: \"kubernetes.io/projected/58aa635e-157c-4ae2-ad0c-568c144638d1-kube-api-access-6s4x4\") pod \"neutron-operator-controller-manager-5fdfd5b6b5-4mlhk\" (UID: \"58aa635e-157c-4ae2-ad0c-568c144638d1\") " pod="openstack-operators/neutron-operator-controller-manager-5fdfd5b6b5-4mlhk" Dec 03 14:26:51 crc kubenswrapper[4805]: I1203 14:26:51.322983 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hmzsn\" (UniqueName: \"kubernetes.io/projected/ffa72db0-588a-4a78-a1cd-8710f5ce4f41-kube-api-access-hmzsn\") pod \"keystone-operator-controller-manager-7765d96ddf-qzzz9\" (UID: \"ffa72db0-588a-4a78-a1cd-8710f5ce4f41\") " pod="openstack-operators/keystone-operator-controller-manager-7765d96ddf-qzzz9" Dec 03 14:26:51 crc kubenswrapper[4805]: I1203 14:26:51.323002 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jjz2n\" (UniqueName: \"kubernetes.io/projected/05506111-ddb3-4ab9-93a0-ad2bedff67a4-kube-api-access-jjz2n\") pod \"nova-operator-controller-manager-697bc559fc-5rl85\" (UID: \"05506111-ddb3-4ab9-93a0-ad2bedff67a4\") " pod="openstack-operators/nova-operator-controller-manager-697bc559fc-5rl85" Dec 03 14:26:51 crc kubenswrapper[4805]: I1203 14:26:51.323029 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-v5zjt\" (UniqueName: \"kubernetes.io/projected/2d72009f-b9ef-4c9d-b224-5b133a3cd93a-kube-api-access-v5zjt\") pod \"mariadb-operator-controller-manager-56bbcc9d85-n6m2d\" (UID: \"2d72009f-b9ef-4c9d-b224-5b133a3cd93a\") " pod="openstack-operators/mariadb-operator-controller-manager-56bbcc9d85-n6m2d" Dec 03 14:26:51 crc kubenswrapper[4805]: I1203 14:26:51.323047 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fnmcc\" (UniqueName: \"kubernetes.io/projected/1ad4fa11-2f11-4a06-9944-bba8c7445e84-kube-api-access-fnmcc\") pod \"ironic-operator-controller-manager-6c548fd776-mz4pw\" (UID: \"1ad4fa11-2f11-4a06-9944-bba8c7445e84\") " pod="openstack-operators/ironic-operator-controller-manager-6c548fd776-mz4pw" Dec 03 14:26:51 crc kubenswrapper[4805]: I1203 14:26:51.323072 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sfkq4\" (UniqueName: \"kubernetes.io/projected/967f26e7-3027-4e23-9f02-d111d714a396-kube-api-access-sfkq4\") pod \"manila-operator-controller-manager-7c79b5df47-d2thc\" (UID: \"967f26e7-3027-4e23-9f02-d111d714a396\") " pod="openstack-operators/manila-operator-controller-manager-7c79b5df47-d2thc" Dec 03 14:26:51 crc kubenswrapper[4805]: I1203 14:26:51.323130 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-z9t5w\" (UniqueName: \"kubernetes.io/projected/0f040ce3-0a83-426b-8b9f-b63183d16b4a-kube-api-access-z9t5w\") pod \"octavia-operator-controller-manager-998648c74-8mbdb\" (UID: \"0f040ce3-0a83-426b-8b9f-b63183d16b4a\") " pod="openstack-operators/octavia-operator-controller-manager-998648c74-8mbdb" Dec 03 14:26:51 crc kubenswrapper[4805]: I1203 14:26:51.332935 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/octavia-operator-controller-manager-998648c74-8mbdb" Dec 03 14:26:51 crc kubenswrapper[4805]: I1203 14:26:51.340517 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/octavia-operator-controller-manager-998648c74-8mbdb"] Dec 03 14:26:51 crc kubenswrapper[4805]: I1203 14:26:51.345164 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/horizon-operator-controller-manager-68c6d99b8f-kxbr8" Dec 03 14:26:51 crc kubenswrapper[4805]: I1203 14:26:51.346316 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"octavia-operator-controller-manager-dockercfg-bktzv" Dec 03 14:26:51 crc kubenswrapper[4805]: I1203 14:26:51.355115 4805 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-baremetal-operator-controller-manager-64bc77cfd4x9mdd"] Dec 03 14:26:51 crc kubenswrapper[4805]: I1203 14:26:51.356388 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-baremetal-operator-controller-manager-64bc77cfd4x9mdd" Dec 03 14:26:51 crc kubenswrapper[4805]: I1203 14:26:51.360557 4805 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/placement-operator-controller-manager-78f8948974-b8l4l"] Dec 03 14:26:51 crc kubenswrapper[4805]: I1203 14:26:51.360929 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-baremetal-operator-webhook-server-cert" Dec 03 14:26:51 crc kubenswrapper[4805]: I1203 14:26:51.361423 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/placement-operator-controller-manager-78f8948974-b8l4l" Dec 03 14:26:51 crc kubenswrapper[4805]: I1203 14:26:51.376442 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-baremetal-operator-controller-manager-64bc77cfd4x9mdd"] Dec 03 14:26:51 crc kubenswrapper[4805]: I1203 14:26:51.376670 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-baremetal-operator-controller-manager-dockercfg-s8st6" Dec 03 14:26:51 crc kubenswrapper[4805]: I1203 14:26:51.376812 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"placement-operator-controller-manager-dockercfg-qvtlt" Dec 03 14:26:51 crc kubenswrapper[4805]: I1203 14:26:51.377550 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fnmcc\" (UniqueName: \"kubernetes.io/projected/1ad4fa11-2f11-4a06-9944-bba8c7445e84-kube-api-access-fnmcc\") pod \"ironic-operator-controller-manager-6c548fd776-mz4pw\" (UID: \"1ad4fa11-2f11-4a06-9944-bba8c7445e84\") " pod="openstack-operators/ironic-operator-controller-manager-6c548fd776-mz4pw" Dec 03 14:26:51 crc kubenswrapper[4805]: I1203 14:26:51.392831 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hmzsn\" (UniqueName: \"kubernetes.io/projected/ffa72db0-588a-4a78-a1cd-8710f5ce4f41-kube-api-access-hmzsn\") pod \"keystone-operator-controller-manager-7765d96ddf-qzzz9\" (UID: \"ffa72db0-588a-4a78-a1cd-8710f5ce4f41\") " pod="openstack-operators/keystone-operator-controller-manager-7765d96ddf-qzzz9" Dec 03 14:26:51 crc kubenswrapper[4805]: I1203 14:26:51.395097 4805 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/ovn-operator-controller-manager-b6456fdb6-djzcm"] Dec 03 14:26:51 crc kubenswrapper[4805]: I1203 14:26:51.397056 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/ovn-operator-controller-manager-b6456fdb6-djzcm" Dec 03 14:26:51 crc kubenswrapper[4805]: I1203 14:26:51.411547 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"ovn-operator-controller-manager-dockercfg-gv566" Dec 03 14:26:51 crc kubenswrapper[4805]: I1203 14:26:51.412438 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/placement-operator-controller-manager-78f8948974-b8l4l"] Dec 03 14:26:51 crc kubenswrapper[4805]: I1203 14:26:51.412548 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/ironic-operator-controller-manager-6c548fd776-mz4pw" Dec 03 14:26:51 crc kubenswrapper[4805]: I1203 14:26:51.419407 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/barbican-operator-controller-manager-7d9dfd778-jc4dw" Dec 03 14:26:51 crc kubenswrapper[4805]: I1203 14:26:51.426259 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sfkq4\" (UniqueName: \"kubernetes.io/projected/967f26e7-3027-4e23-9f02-d111d714a396-kube-api-access-sfkq4\") pod \"manila-operator-controller-manager-7c79b5df47-d2thc\" (UID: \"967f26e7-3027-4e23-9f02-d111d714a396\") " pod="openstack-operators/manila-operator-controller-manager-7c79b5df47-d2thc" Dec 03 14:26:51 crc kubenswrapper[4805]: I1203 14:26:51.428429 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-z9t5w\" (UniqueName: \"kubernetes.io/projected/0f040ce3-0a83-426b-8b9f-b63183d16b4a-kube-api-access-z9t5w\") pod \"octavia-operator-controller-manager-998648c74-8mbdb\" (UID: \"0f040ce3-0a83-426b-8b9f-b63183d16b4a\") " pod="openstack-operators/octavia-operator-controller-manager-998648c74-8mbdb" Dec 03 14:26:51 crc kubenswrapper[4805]: I1203 14:26:51.428556 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/ovn-operator-controller-manager-b6456fdb6-djzcm"] Dec 03 14:26:51 crc kubenswrapper[4805]: I1203 14:26:51.428670 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6s4x4\" (UniqueName: \"kubernetes.io/projected/58aa635e-157c-4ae2-ad0c-568c144638d1-kube-api-access-6s4x4\") pod \"neutron-operator-controller-manager-5fdfd5b6b5-4mlhk\" (UID: \"58aa635e-157c-4ae2-ad0c-568c144638d1\") " pod="openstack-operators/neutron-operator-controller-manager-5fdfd5b6b5-4mlhk" Dec 03 14:26:51 crc kubenswrapper[4805]: I1203 14:26:51.428923 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jjz2n\" (UniqueName: \"kubernetes.io/projected/05506111-ddb3-4ab9-93a0-ad2bedff67a4-kube-api-access-jjz2n\") pod \"nova-operator-controller-manager-697bc559fc-5rl85\" (UID: \"05506111-ddb3-4ab9-93a0-ad2bedff67a4\") " pod="openstack-operators/nova-operator-controller-manager-697bc559fc-5rl85" Dec 03 14:26:51 crc kubenswrapper[4805]: I1203 14:26:51.429032 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-v5zjt\" (UniqueName: \"kubernetes.io/projected/2d72009f-b9ef-4c9d-b224-5b133a3cd93a-kube-api-access-v5zjt\") pod \"mariadb-operator-controller-manager-56bbcc9d85-n6m2d\" (UID: \"2d72009f-b9ef-4c9d-b224-5b133a3cd93a\") " pod="openstack-operators/mariadb-operator-controller-manager-56bbcc9d85-n6m2d" Dec 03 14:26:51 crc kubenswrapper[4805]: I1203 14:26:51.460936 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/cinder-operator-controller-manager-859b6ccc6-2nzgx" Dec 03 14:26:51 crc kubenswrapper[4805]: I1203 14:26:51.463496 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/keystone-operator-controller-manager-7765d96ddf-qzzz9" Dec 03 14:26:51 crc kubenswrapper[4805]: I1203 14:26:51.473479 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sfkq4\" (UniqueName: \"kubernetes.io/projected/967f26e7-3027-4e23-9f02-d111d714a396-kube-api-access-sfkq4\") pod \"manila-operator-controller-manager-7c79b5df47-d2thc\" (UID: \"967f26e7-3027-4e23-9f02-d111d714a396\") " pod="openstack-operators/manila-operator-controller-manager-7c79b5df47-d2thc" Dec 03 14:26:51 crc kubenswrapper[4805]: I1203 14:26:51.474379 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/designate-operator-controller-manager-78b4bc895b-w56sr" Dec 03 14:26:51 crc kubenswrapper[4805]: I1203 14:26:51.474897 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-z9t5w\" (UniqueName: \"kubernetes.io/projected/0f040ce3-0a83-426b-8b9f-b63183d16b4a-kube-api-access-z9t5w\") pod \"octavia-operator-controller-manager-998648c74-8mbdb\" (UID: \"0f040ce3-0a83-426b-8b9f-b63183d16b4a\") " pod="openstack-operators/octavia-operator-controller-manager-998648c74-8mbdb" Dec 03 14:26:51 crc kubenswrapper[4805]: I1203 14:26:51.479721 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jjz2n\" (UniqueName: \"kubernetes.io/projected/05506111-ddb3-4ab9-93a0-ad2bedff67a4-kube-api-access-jjz2n\") pod \"nova-operator-controller-manager-697bc559fc-5rl85\" (UID: \"05506111-ddb3-4ab9-93a0-ad2bedff67a4\") " pod="openstack-operators/nova-operator-controller-manager-697bc559fc-5rl85" Dec 03 14:26:51 crc kubenswrapper[4805]: I1203 14:26:51.485693 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-v5zjt\" (UniqueName: \"kubernetes.io/projected/2d72009f-b9ef-4c9d-b224-5b133a3cd93a-kube-api-access-v5zjt\") pod \"mariadb-operator-controller-manager-56bbcc9d85-n6m2d\" (UID: \"2d72009f-b9ef-4c9d-b224-5b133a3cd93a\") " pod="openstack-operators/mariadb-operator-controller-manager-56bbcc9d85-n6m2d" Dec 03 14:26:51 crc kubenswrapper[4805]: I1203 14:26:51.491718 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6s4x4\" (UniqueName: \"kubernetes.io/projected/58aa635e-157c-4ae2-ad0c-568c144638d1-kube-api-access-6s4x4\") pod \"neutron-operator-controller-manager-5fdfd5b6b5-4mlhk\" (UID: \"58aa635e-157c-4ae2-ad0c-568c144638d1\") " pod="openstack-operators/neutron-operator-controller-manager-5fdfd5b6b5-4mlhk" Dec 03 14:26:51 crc kubenswrapper[4805]: I1203 14:26:51.498424 4805 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/swift-operator-controller-manager-5f8c65bbfc-qrvpw"] Dec 03 14:26:51 crc kubenswrapper[4805]: I1203 14:26:51.510898 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/octavia-operator-controller-manager-998648c74-8mbdb" Dec 03 14:26:51 crc kubenswrapper[4805]: I1203 14:26:51.518099 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/swift-operator-controller-manager-5f8c65bbfc-qrvpw" Dec 03 14:26:51 crc kubenswrapper[4805]: I1203 14:26:51.519800 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"swift-operator-controller-manager-dockercfg-gnxz9" Dec 03 14:26:51 crc kubenswrapper[4805]: I1203 14:26:51.550502 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/b9d0830e-0f5d-4639-ab7d-92dcf07a29bc-cert\") pod \"openstack-baremetal-operator-controller-manager-64bc77cfd4x9mdd\" (UID: \"b9d0830e-0f5d-4639-ab7d-92dcf07a29bc\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-64bc77cfd4x9mdd" Dec 03 14:26:51 crc kubenswrapper[4805]: I1203 14:26:51.550553 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-r68wv\" (UniqueName: \"kubernetes.io/projected/f853eae6-829b-42f9-9198-3d7dcabd405f-kube-api-access-r68wv\") pod \"placement-operator-controller-manager-78f8948974-b8l4l\" (UID: \"f853eae6-829b-42f9-9198-3d7dcabd405f\") " pod="openstack-operators/placement-operator-controller-manager-78f8948974-b8l4l" Dec 03 14:26:51 crc kubenswrapper[4805]: I1203 14:26:51.550684 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5g6hd\" (UniqueName: \"kubernetes.io/projected/b9d0830e-0f5d-4639-ab7d-92dcf07a29bc-kube-api-access-5g6hd\") pod \"openstack-baremetal-operator-controller-manager-64bc77cfd4x9mdd\" (UID: \"b9d0830e-0f5d-4639-ab7d-92dcf07a29bc\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-64bc77cfd4x9mdd" Dec 03 14:26:51 crc kubenswrapper[4805]: I1203 14:26:51.550717 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xvkjw\" (UniqueName: \"kubernetes.io/projected/42b37265-f5bd-4ac4-9b96-d758fbe1b31e-kube-api-access-xvkjw\") pod \"ovn-operator-controller-manager-b6456fdb6-djzcm\" (UID: \"42b37265-f5bd-4ac4-9b96-d758fbe1b31e\") " pod="openstack-operators/ovn-operator-controller-manager-b6456fdb6-djzcm" Dec 03 14:26:51 crc kubenswrapper[4805]: I1203 14:26:51.550757 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6mpcm\" (UniqueName: \"kubernetes.io/projected/b51f4884-a980-4643-9deb-8668aead3876-kube-api-access-6mpcm\") pod \"swift-operator-controller-manager-5f8c65bbfc-qrvpw\" (UID: \"b51f4884-a980-4643-9deb-8668aead3876\") " pod="openstack-operators/swift-operator-controller-manager-5f8c65bbfc-qrvpw" Dec 03 14:26:51 crc kubenswrapper[4805]: I1203 14:26:51.585307 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/swift-operator-controller-manager-5f8c65bbfc-qrvpw"] Dec 03 14:26:51 crc kubenswrapper[4805]: I1203 14:26:51.595918 4805 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/telemetry-operator-controller-manager-cc9d8f87b-qfmmx"] Dec 03 14:26:51 crc kubenswrapper[4805]: I1203 14:26:51.596979 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/telemetry-operator-controller-manager-cc9d8f87b-qfmmx" Dec 03 14:26:51 crc kubenswrapper[4805]: I1203 14:26:51.607093 4805 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/test-operator-controller-manager-5854674fcc-d6pvv"] Dec 03 14:26:51 crc kubenswrapper[4805]: I1203 14:26:51.608492 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/test-operator-controller-manager-5854674fcc-d6pvv" Dec 03 14:26:51 crc kubenswrapper[4805]: I1203 14:26:51.612156 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"telemetry-operator-controller-manager-dockercfg-5hsgn" Dec 03 14:26:51 crc kubenswrapper[4805]: I1203 14:26:51.613026 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/telemetry-operator-controller-manager-cc9d8f87b-qfmmx"] Dec 03 14:26:51 crc kubenswrapper[4805]: I1203 14:26:51.617768 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"test-operator-controller-manager-dockercfg-9gttj" Dec 03 14:26:51 crc kubenswrapper[4805]: I1203 14:26:51.644807 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/test-operator-controller-manager-5854674fcc-d6pvv"] Dec 03 14:26:51 crc kubenswrapper[4805]: I1203 14:26:51.651877 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-r68wv\" (UniqueName: \"kubernetes.io/projected/f853eae6-829b-42f9-9198-3d7dcabd405f-kube-api-access-r68wv\") pod \"placement-operator-controller-manager-78f8948974-b8l4l\" (UID: \"f853eae6-829b-42f9-9198-3d7dcabd405f\") " pod="openstack-operators/placement-operator-controller-manager-78f8948974-b8l4l" Dec 03 14:26:51 crc kubenswrapper[4805]: I1203 14:26:51.651976 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5g6hd\" (UniqueName: \"kubernetes.io/projected/b9d0830e-0f5d-4639-ab7d-92dcf07a29bc-kube-api-access-5g6hd\") pod \"openstack-baremetal-operator-controller-manager-64bc77cfd4x9mdd\" (UID: \"b9d0830e-0f5d-4639-ab7d-92dcf07a29bc\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-64bc77cfd4x9mdd" Dec 03 14:26:51 crc kubenswrapper[4805]: I1203 14:26:51.652000 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xvkjw\" (UniqueName: \"kubernetes.io/projected/42b37265-f5bd-4ac4-9b96-d758fbe1b31e-kube-api-access-xvkjw\") pod \"ovn-operator-controller-manager-b6456fdb6-djzcm\" (UID: \"42b37265-f5bd-4ac4-9b96-d758fbe1b31e\") " pod="openstack-operators/ovn-operator-controller-manager-b6456fdb6-djzcm" Dec 03 14:26:51 crc kubenswrapper[4805]: I1203 14:26:51.652030 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6mpcm\" (UniqueName: \"kubernetes.io/projected/b51f4884-a980-4643-9deb-8668aead3876-kube-api-access-6mpcm\") pod \"swift-operator-controller-manager-5f8c65bbfc-qrvpw\" (UID: \"b51f4884-a980-4643-9deb-8668aead3876\") " pod="openstack-operators/swift-operator-controller-manager-5f8c65bbfc-qrvpw" Dec 03 14:26:51 crc kubenswrapper[4805]: I1203 14:26:51.652065 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nxwjv\" (UniqueName: \"kubernetes.io/projected/8cbfcd07-e136-4bfd-b3c6-6615c911d440-kube-api-access-nxwjv\") pod \"telemetry-operator-controller-manager-cc9d8f87b-qfmmx\" (UID: \"8cbfcd07-e136-4bfd-b3c6-6615c911d440\") " pod="openstack-operators/telemetry-operator-controller-manager-cc9d8f87b-qfmmx" Dec 03 14:26:51 crc kubenswrapper[4805]: I1203 14:26:51.652131 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jslvv\" (UniqueName: \"kubernetes.io/projected/ac4ba0f5-635d-4a72-a133-9b4aaf88f097-kube-api-access-jslvv\") pod \"test-operator-controller-manager-5854674fcc-d6pvv\" (UID: \"ac4ba0f5-635d-4a72-a133-9b4aaf88f097\") " pod="openstack-operators/test-operator-controller-manager-5854674fcc-d6pvv" Dec 03 14:26:51 crc kubenswrapper[4805]: I1203 14:26:51.652174 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/b9d0830e-0f5d-4639-ab7d-92dcf07a29bc-cert\") pod \"openstack-baremetal-operator-controller-manager-64bc77cfd4x9mdd\" (UID: \"b9d0830e-0f5d-4639-ab7d-92dcf07a29bc\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-64bc77cfd4x9mdd" Dec 03 14:26:51 crc kubenswrapper[4805]: E1203 14:26:51.652294 4805 secret.go:188] Couldn't get secret openstack-operators/openstack-baremetal-operator-webhook-server-cert: secret "openstack-baremetal-operator-webhook-server-cert" not found Dec 03 14:26:51 crc kubenswrapper[4805]: E1203 14:26:51.652348 4805 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/b9d0830e-0f5d-4639-ab7d-92dcf07a29bc-cert podName:b9d0830e-0f5d-4639-ab7d-92dcf07a29bc nodeName:}" failed. No retries permitted until 2025-12-03 14:26:52.152331651 +0000 UTC m=+1041.815248574 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/b9d0830e-0f5d-4639-ab7d-92dcf07a29bc-cert") pod "openstack-baremetal-operator-controller-manager-64bc77cfd4x9mdd" (UID: "b9d0830e-0f5d-4639-ab7d-92dcf07a29bc") : secret "openstack-baremetal-operator-webhook-server-cert" not found Dec 03 14:26:51 crc kubenswrapper[4805]: I1203 14:26:51.661296 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/manila-operator-controller-manager-7c79b5df47-d2thc" Dec 03 14:26:51 crc kubenswrapper[4805]: I1203 14:26:51.669180 4805 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/watcher-operator-controller-manager-769dc69bc-fl6mm"] Dec 03 14:26:51 crc kubenswrapper[4805]: I1203 14:26:51.670254 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/watcher-operator-controller-manager-769dc69bc-fl6mm" Dec 03 14:26:51 crc kubenswrapper[4805]: I1203 14:26:51.679812 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"watcher-operator-controller-manager-dockercfg-7chr8" Dec 03 14:26:51 crc kubenswrapper[4805]: I1203 14:26:51.683105 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/mariadb-operator-controller-manager-56bbcc9d85-n6m2d" Dec 03 14:26:51 crc kubenswrapper[4805]: I1203 14:26:51.685386 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/watcher-operator-controller-manager-769dc69bc-fl6mm"] Dec 03 14:26:51 crc kubenswrapper[4805]: I1203 14:26:51.689886 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-r68wv\" (UniqueName: \"kubernetes.io/projected/f853eae6-829b-42f9-9198-3d7dcabd405f-kube-api-access-r68wv\") pod \"placement-operator-controller-manager-78f8948974-b8l4l\" (UID: \"f853eae6-829b-42f9-9198-3d7dcabd405f\") " pod="openstack-operators/placement-operator-controller-manager-78f8948974-b8l4l" Dec 03 14:26:51 crc kubenswrapper[4805]: I1203 14:26:51.697604 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xvkjw\" (UniqueName: \"kubernetes.io/projected/42b37265-f5bd-4ac4-9b96-d758fbe1b31e-kube-api-access-xvkjw\") pod \"ovn-operator-controller-manager-b6456fdb6-djzcm\" (UID: \"42b37265-f5bd-4ac4-9b96-d758fbe1b31e\") " pod="openstack-operators/ovn-operator-controller-manager-b6456fdb6-djzcm" Dec 03 14:26:51 crc kubenswrapper[4805]: I1203 14:26:51.698238 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5g6hd\" (UniqueName: \"kubernetes.io/projected/b9d0830e-0f5d-4639-ab7d-92dcf07a29bc-kube-api-access-5g6hd\") pod \"openstack-baremetal-operator-controller-manager-64bc77cfd4x9mdd\" (UID: \"b9d0830e-0f5d-4639-ab7d-92dcf07a29bc\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-64bc77cfd4x9mdd" Dec 03 14:26:51 crc kubenswrapper[4805]: I1203 14:26:51.715242 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/neutron-operator-controller-manager-5fdfd5b6b5-4mlhk" Dec 03 14:26:51 crc kubenswrapper[4805]: I1203 14:26:51.724336 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6mpcm\" (UniqueName: \"kubernetes.io/projected/b51f4884-a980-4643-9deb-8668aead3876-kube-api-access-6mpcm\") pod \"swift-operator-controller-manager-5f8c65bbfc-qrvpw\" (UID: \"b51f4884-a980-4643-9deb-8668aead3876\") " pod="openstack-operators/swift-operator-controller-manager-5f8c65bbfc-qrvpw" Dec 03 14:26:51 crc kubenswrapper[4805]: I1203 14:26:51.732896 4805 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-operator-controller-manager-cc68d789f-pkvr8"] Dec 03 14:26:51 crc kubenswrapper[4805]: I1203 14:26:51.736369 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-controller-manager-cc68d789f-pkvr8" Dec 03 14:26:51 crc kubenswrapper[4805]: I1203 14:26:51.738107 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"metrics-server-cert" Dec 03 14:26:51 crc kubenswrapper[4805]: I1203 14:26:51.738319 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-operator-controller-manager-dockercfg-m9rsn" Dec 03 14:26:51 crc kubenswrapper[4805]: I1203 14:26:51.742341 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"webhook-server-cert" Dec 03 14:26:51 crc kubenswrapper[4805]: I1203 14:26:51.742771 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-controller-manager-cc68d789f-pkvr8"] Dec 03 14:26:51 crc kubenswrapper[4805]: I1203 14:26:51.753648 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jslvv\" (UniqueName: \"kubernetes.io/projected/ac4ba0f5-635d-4a72-a133-9b4aaf88f097-kube-api-access-jslvv\") pod \"test-operator-controller-manager-5854674fcc-d6pvv\" (UID: \"ac4ba0f5-635d-4a72-a133-9b4aaf88f097\") " pod="openstack-operators/test-operator-controller-manager-5854674fcc-d6pvv" Dec 03 14:26:51 crc kubenswrapper[4805]: I1203 14:26:51.753736 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/f21790d8-af47-4be7-8758-f038bda76908-cert\") pod \"infra-operator-controller-manager-57548d458d-77t9s\" (UID: \"f21790d8-af47-4be7-8758-f038bda76908\") " pod="openstack-operators/infra-operator-controller-manager-57548d458d-77t9s" Dec 03 14:26:51 crc kubenswrapper[4805]: I1203 14:26:51.753766 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/41c4618c-1f15-4ceb-825b-546e828e399a-webhook-certs\") pod \"openstack-operator-controller-manager-cc68d789f-pkvr8\" (UID: \"41c4618c-1f15-4ceb-825b-546e828e399a\") " pod="openstack-operators/openstack-operator-controller-manager-cc68d789f-pkvr8" Dec 03 14:26:51 crc kubenswrapper[4805]: I1203 14:26:51.753811 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nxwjv\" (UniqueName: \"kubernetes.io/projected/8cbfcd07-e136-4bfd-b3c6-6615c911d440-kube-api-access-nxwjv\") pod \"telemetry-operator-controller-manager-cc9d8f87b-qfmmx\" (UID: \"8cbfcd07-e136-4bfd-b3c6-6615c911d440\") " pod="openstack-operators/telemetry-operator-controller-manager-cc9d8f87b-qfmmx" Dec 03 14:26:51 crc kubenswrapper[4805]: I1203 14:26:51.753833 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-h74hm\" (UniqueName: \"kubernetes.io/projected/30776d04-200b-4c10-b48d-f51a926e469b-kube-api-access-h74hm\") pod \"watcher-operator-controller-manager-769dc69bc-fl6mm\" (UID: \"30776d04-200b-4c10-b48d-f51a926e469b\") " pod="openstack-operators/watcher-operator-controller-manager-769dc69bc-fl6mm" Dec 03 14:26:51 crc kubenswrapper[4805]: E1203 14:26:51.754090 4805 secret.go:188] Couldn't get secret openstack-operators/infra-operator-webhook-server-cert: secret "infra-operator-webhook-server-cert" not found Dec 03 14:26:51 crc kubenswrapper[4805]: E1203 14:26:51.754137 4805 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/f21790d8-af47-4be7-8758-f038bda76908-cert podName:f21790d8-af47-4be7-8758-f038bda76908 nodeName:}" failed. No retries permitted until 2025-12-03 14:26:52.754122861 +0000 UTC m=+1042.417039784 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/f21790d8-af47-4be7-8758-f038bda76908-cert") pod "infra-operator-controller-manager-57548d458d-77t9s" (UID: "f21790d8-af47-4be7-8758-f038bda76908") : secret "infra-operator-webhook-server-cert" not found Dec 03 14:26:51 crc kubenswrapper[4805]: I1203 14:26:51.762525 4805 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-5q6j2"] Dec 03 14:26:51 crc kubenswrapper[4805]: I1203 14:26:51.763533 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-5q6j2" Dec 03 14:26:51 crc kubenswrapper[4805]: I1203 14:26:51.770155 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"rabbitmq-cluster-operator-controller-manager-dockercfg-v57cx" Dec 03 14:26:51 crc kubenswrapper[4805]: I1203 14:26:51.777754 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/nova-operator-controller-manager-697bc559fc-5rl85" Dec 03 14:26:51 crc kubenswrapper[4805]: I1203 14:26:51.778241 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/41c4618c-1f15-4ceb-825b-546e828e399a-metrics-certs\") pod \"openstack-operator-controller-manager-cc68d789f-pkvr8\" (UID: \"41c4618c-1f15-4ceb-825b-546e828e399a\") " pod="openstack-operators/openstack-operator-controller-manager-cc68d789f-pkvr8" Dec 03 14:26:51 crc kubenswrapper[4805]: I1203 14:26:51.778267 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tsfgm\" (UniqueName: \"kubernetes.io/projected/41c4618c-1f15-4ceb-825b-546e828e399a-kube-api-access-tsfgm\") pod \"openstack-operator-controller-manager-cc68d789f-pkvr8\" (UID: \"41c4618c-1f15-4ceb-825b-546e828e399a\") " pod="openstack-operators/openstack-operator-controller-manager-cc68d789f-pkvr8" Dec 03 14:26:51 crc kubenswrapper[4805]: I1203 14:26:51.781125 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jslvv\" (UniqueName: \"kubernetes.io/projected/ac4ba0f5-635d-4a72-a133-9b4aaf88f097-kube-api-access-jslvv\") pod \"test-operator-controller-manager-5854674fcc-d6pvv\" (UID: \"ac4ba0f5-635d-4a72-a133-9b4aaf88f097\") " pod="openstack-operators/test-operator-controller-manager-5854674fcc-d6pvv" Dec 03 14:26:51 crc kubenswrapper[4805]: I1203 14:26:51.803914 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-5q6j2"] Dec 03 14:26:51 crc kubenswrapper[4805]: I1203 14:26:51.805991 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nxwjv\" (UniqueName: \"kubernetes.io/projected/8cbfcd07-e136-4bfd-b3c6-6615c911d440-kube-api-access-nxwjv\") pod \"telemetry-operator-controller-manager-cc9d8f87b-qfmmx\" (UID: \"8cbfcd07-e136-4bfd-b3c6-6615c911d440\") " pod="openstack-operators/telemetry-operator-controller-manager-cc9d8f87b-qfmmx" Dec 03 14:26:51 crc kubenswrapper[4805]: I1203 14:26:51.854642 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/placement-operator-controller-manager-78f8948974-b8l4l" Dec 03 14:26:51 crc kubenswrapper[4805]: I1203 14:26:51.879224 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/ovn-operator-controller-manager-b6456fdb6-djzcm" Dec 03 14:26:51 crc kubenswrapper[4805]: I1203 14:26:51.887667 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kqs6q\" (UniqueName: \"kubernetes.io/projected/0aa860af-81d3-44ba-9020-0ffef21cefb2-kube-api-access-kqs6q\") pod \"rabbitmq-cluster-operator-manager-668c99d594-5q6j2\" (UID: \"0aa860af-81d3-44ba-9020-0ffef21cefb2\") " pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-5q6j2" Dec 03 14:26:51 crc kubenswrapper[4805]: I1203 14:26:51.887769 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/41c4618c-1f15-4ceb-825b-546e828e399a-webhook-certs\") pod \"openstack-operator-controller-manager-cc68d789f-pkvr8\" (UID: \"41c4618c-1f15-4ceb-825b-546e828e399a\") " pod="openstack-operators/openstack-operator-controller-manager-cc68d789f-pkvr8" Dec 03 14:26:51 crc kubenswrapper[4805]: I1203 14:26:51.887812 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-h74hm\" (UniqueName: \"kubernetes.io/projected/30776d04-200b-4c10-b48d-f51a926e469b-kube-api-access-h74hm\") pod \"watcher-operator-controller-manager-769dc69bc-fl6mm\" (UID: \"30776d04-200b-4c10-b48d-f51a926e469b\") " pod="openstack-operators/watcher-operator-controller-manager-769dc69bc-fl6mm" Dec 03 14:26:51 crc kubenswrapper[4805]: I1203 14:26:51.887851 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/41c4618c-1f15-4ceb-825b-546e828e399a-metrics-certs\") pod \"openstack-operator-controller-manager-cc68d789f-pkvr8\" (UID: \"41c4618c-1f15-4ceb-825b-546e828e399a\") " pod="openstack-operators/openstack-operator-controller-manager-cc68d789f-pkvr8" Dec 03 14:26:51 crc kubenswrapper[4805]: I1203 14:26:51.887870 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tsfgm\" (UniqueName: \"kubernetes.io/projected/41c4618c-1f15-4ceb-825b-546e828e399a-kube-api-access-tsfgm\") pod \"openstack-operator-controller-manager-cc68d789f-pkvr8\" (UID: \"41c4618c-1f15-4ceb-825b-546e828e399a\") " pod="openstack-operators/openstack-operator-controller-manager-cc68d789f-pkvr8" Dec 03 14:26:51 crc kubenswrapper[4805]: E1203 14:26:51.888277 4805 secret.go:188] Couldn't get secret openstack-operators/webhook-server-cert: secret "webhook-server-cert" not found Dec 03 14:26:51 crc kubenswrapper[4805]: E1203 14:26:51.888303 4805 secret.go:188] Couldn't get secret openstack-operators/metrics-server-cert: secret "metrics-server-cert" not found Dec 03 14:26:51 crc kubenswrapper[4805]: E1203 14:26:51.888349 4805 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/41c4618c-1f15-4ceb-825b-546e828e399a-webhook-certs podName:41c4618c-1f15-4ceb-825b-546e828e399a nodeName:}" failed. No retries permitted until 2025-12-03 14:26:52.388327271 +0000 UTC m=+1042.051244194 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "webhook-certs" (UniqueName: "kubernetes.io/secret/41c4618c-1f15-4ceb-825b-546e828e399a-webhook-certs") pod "openstack-operator-controller-manager-cc68d789f-pkvr8" (UID: "41c4618c-1f15-4ceb-825b-546e828e399a") : secret "webhook-server-cert" not found Dec 03 14:26:51 crc kubenswrapper[4805]: E1203 14:26:51.888375 4805 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/41c4618c-1f15-4ceb-825b-546e828e399a-metrics-certs podName:41c4618c-1f15-4ceb-825b-546e828e399a nodeName:}" failed. No retries permitted until 2025-12-03 14:26:52.388369442 +0000 UTC m=+1042.051286365 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/41c4618c-1f15-4ceb-825b-546e828e399a-metrics-certs") pod "openstack-operator-controller-manager-cc68d789f-pkvr8" (UID: "41c4618c-1f15-4ceb-825b-546e828e399a") : secret "metrics-server-cert" not found Dec 03 14:26:51 crc kubenswrapper[4805]: I1203 14:26:51.907772 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tsfgm\" (UniqueName: \"kubernetes.io/projected/41c4618c-1f15-4ceb-825b-546e828e399a-kube-api-access-tsfgm\") pod \"openstack-operator-controller-manager-cc68d789f-pkvr8\" (UID: \"41c4618c-1f15-4ceb-825b-546e828e399a\") " pod="openstack-operators/openstack-operator-controller-manager-cc68d789f-pkvr8" Dec 03 14:26:51 crc kubenswrapper[4805]: I1203 14:26:51.917486 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-h74hm\" (UniqueName: \"kubernetes.io/projected/30776d04-200b-4c10-b48d-f51a926e469b-kube-api-access-h74hm\") pod \"watcher-operator-controller-manager-769dc69bc-fl6mm\" (UID: \"30776d04-200b-4c10-b48d-f51a926e469b\") " pod="openstack-operators/watcher-operator-controller-manager-769dc69bc-fl6mm" Dec 03 14:26:51 crc kubenswrapper[4805]: I1203 14:26:51.935246 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/swift-operator-controller-manager-5f8c65bbfc-qrvpw" Dec 03 14:26:51 crc kubenswrapper[4805]: I1203 14:26:51.964165 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/telemetry-operator-controller-manager-cc9d8f87b-qfmmx" Dec 03 14:26:51 crc kubenswrapper[4805]: I1203 14:26:51.993976 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kqs6q\" (UniqueName: \"kubernetes.io/projected/0aa860af-81d3-44ba-9020-0ffef21cefb2-kube-api-access-kqs6q\") pod \"rabbitmq-cluster-operator-manager-668c99d594-5q6j2\" (UID: \"0aa860af-81d3-44ba-9020-0ffef21cefb2\") " pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-5q6j2" Dec 03 14:26:51 crc kubenswrapper[4805]: I1203 14:26:51.998306 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/test-operator-controller-manager-5854674fcc-d6pvv" Dec 03 14:26:52 crc kubenswrapper[4805]: I1203 14:26:52.010600 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kqs6q\" (UniqueName: \"kubernetes.io/projected/0aa860af-81d3-44ba-9020-0ffef21cefb2-kube-api-access-kqs6q\") pod \"rabbitmq-cluster-operator-manager-668c99d594-5q6j2\" (UID: \"0aa860af-81d3-44ba-9020-0ffef21cefb2\") " pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-5q6j2" Dec 03 14:26:52 crc kubenswrapper[4805]: I1203 14:26:52.028078 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/watcher-operator-controller-manager-769dc69bc-fl6mm" Dec 03 14:26:52 crc kubenswrapper[4805]: I1203 14:26:52.036148 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/glance-operator-controller-manager-77987cd8cd-pt9fw"] Dec 03 14:26:52 crc kubenswrapper[4805]: I1203 14:26:52.048443 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/heat-operator-controller-manager-5f64f6f8bb-86sbc"] Dec 03 14:26:52 crc kubenswrapper[4805]: I1203 14:26:52.107695 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-5q6j2" Dec 03 14:26:52 crc kubenswrapper[4805]: I1203 14:26:52.196944 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/b9d0830e-0f5d-4639-ab7d-92dcf07a29bc-cert\") pod \"openstack-baremetal-operator-controller-manager-64bc77cfd4x9mdd\" (UID: \"b9d0830e-0f5d-4639-ab7d-92dcf07a29bc\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-64bc77cfd4x9mdd" Dec 03 14:26:52 crc kubenswrapper[4805]: E1203 14:26:52.197115 4805 secret.go:188] Couldn't get secret openstack-operators/openstack-baremetal-operator-webhook-server-cert: secret "openstack-baremetal-operator-webhook-server-cert" not found Dec 03 14:26:52 crc kubenswrapper[4805]: E1203 14:26:52.197164 4805 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/b9d0830e-0f5d-4639-ab7d-92dcf07a29bc-cert podName:b9d0830e-0f5d-4639-ab7d-92dcf07a29bc nodeName:}" failed. No retries permitted until 2025-12-03 14:26:53.197150116 +0000 UTC m=+1042.860067039 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/b9d0830e-0f5d-4639-ab7d-92dcf07a29bc-cert") pod "openstack-baremetal-operator-controller-manager-64bc77cfd4x9mdd" (UID: "b9d0830e-0f5d-4639-ab7d-92dcf07a29bc") : secret "openstack-baremetal-operator-webhook-server-cert" not found Dec 03 14:26:52 crc kubenswrapper[4805]: I1203 14:26:52.246789 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/horizon-operator-controller-manager-68c6d99b8f-kxbr8"] Dec 03 14:26:52 crc kubenswrapper[4805]: I1203 14:26:52.389097 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/glance-operator-controller-manager-77987cd8cd-pt9fw" event={"ID":"5615ce63-1ec1-4a77-8e44-49e6d114eec6","Type":"ContainerStarted","Data":"9165f63874b663914dfa12f3488016dde4cef894508c6c64826a7f8550758e5f"} Dec 03 14:26:52 crc kubenswrapper[4805]: I1203 14:26:52.394823 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/heat-operator-controller-manager-5f64f6f8bb-86sbc" event={"ID":"a748cfe4-2e54-4a7b-a353-e1dfea6c14b3","Type":"ContainerStarted","Data":"66a88b96e05dfe4a7544307d5e68bcaba114668a9b39fab4de3b2deb546b37f8"} Dec 03 14:26:52 crc kubenswrapper[4805]: I1203 14:26:52.395887 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/horizon-operator-controller-manager-68c6d99b8f-kxbr8" event={"ID":"b29e4120-0cbc-4de5-a4b6-97b97761f49a","Type":"ContainerStarted","Data":"4dcd751886f35ae6baf789758f9926116e97cb2e214f167cc11086a3210ba39f"} Dec 03 14:26:52 crc kubenswrapper[4805]: I1203 14:26:52.404590 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/41c4618c-1f15-4ceb-825b-546e828e399a-metrics-certs\") pod \"openstack-operator-controller-manager-cc68d789f-pkvr8\" (UID: \"41c4618c-1f15-4ceb-825b-546e828e399a\") " pod="openstack-operators/openstack-operator-controller-manager-cc68d789f-pkvr8" Dec 03 14:26:52 crc kubenswrapper[4805]: I1203 14:26:52.404706 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/41c4618c-1f15-4ceb-825b-546e828e399a-webhook-certs\") pod \"openstack-operator-controller-manager-cc68d789f-pkvr8\" (UID: \"41c4618c-1f15-4ceb-825b-546e828e399a\") " pod="openstack-operators/openstack-operator-controller-manager-cc68d789f-pkvr8" Dec 03 14:26:52 crc kubenswrapper[4805]: E1203 14:26:52.404754 4805 secret.go:188] Couldn't get secret openstack-operators/metrics-server-cert: secret "metrics-server-cert" not found Dec 03 14:26:52 crc kubenswrapper[4805]: E1203 14:26:52.404814 4805 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/41c4618c-1f15-4ceb-825b-546e828e399a-metrics-certs podName:41c4618c-1f15-4ceb-825b-546e828e399a nodeName:}" failed. No retries permitted until 2025-12-03 14:26:53.404795867 +0000 UTC m=+1043.067712790 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/41c4618c-1f15-4ceb-825b-546e828e399a-metrics-certs") pod "openstack-operator-controller-manager-cc68d789f-pkvr8" (UID: "41c4618c-1f15-4ceb-825b-546e828e399a") : secret "metrics-server-cert" not found Dec 03 14:26:52 crc kubenswrapper[4805]: E1203 14:26:52.404861 4805 secret.go:188] Couldn't get secret openstack-operators/webhook-server-cert: secret "webhook-server-cert" not found Dec 03 14:26:52 crc kubenswrapper[4805]: E1203 14:26:52.404909 4805 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/41c4618c-1f15-4ceb-825b-546e828e399a-webhook-certs podName:41c4618c-1f15-4ceb-825b-546e828e399a nodeName:}" failed. No retries permitted until 2025-12-03 14:26:53.40489368 +0000 UTC m=+1043.067810603 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "webhook-certs" (UniqueName: "kubernetes.io/secret/41c4618c-1f15-4ceb-825b-546e828e399a-webhook-certs") pod "openstack-operator-controller-manager-cc68d789f-pkvr8" (UID: "41c4618c-1f15-4ceb-825b-546e828e399a") : secret "webhook-server-cert" not found Dec 03 14:26:52 crc kubenswrapper[4805]: I1203 14:26:52.581946 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/cinder-operator-controller-manager-859b6ccc6-2nzgx"] Dec 03 14:26:52 crc kubenswrapper[4805]: I1203 14:26:52.605500 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/ironic-operator-controller-manager-6c548fd776-mz4pw"] Dec 03 14:26:52 crc kubenswrapper[4805]: I1203 14:26:52.633556 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/designate-operator-controller-manager-78b4bc895b-w56sr"] Dec 03 14:26:52 crc kubenswrapper[4805]: I1203 14:26:52.825107 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/f21790d8-af47-4be7-8758-f038bda76908-cert\") pod \"infra-operator-controller-manager-57548d458d-77t9s\" (UID: \"f21790d8-af47-4be7-8758-f038bda76908\") " pod="openstack-operators/infra-operator-controller-manager-57548d458d-77t9s" Dec 03 14:26:52 crc kubenswrapper[4805]: E1203 14:26:52.825331 4805 secret.go:188] Couldn't get secret openstack-operators/infra-operator-webhook-server-cert: secret "infra-operator-webhook-server-cert" not found Dec 03 14:26:52 crc kubenswrapper[4805]: E1203 14:26:52.825392 4805 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/f21790d8-af47-4be7-8758-f038bda76908-cert podName:f21790d8-af47-4be7-8758-f038bda76908 nodeName:}" failed. No retries permitted until 2025-12-03 14:26:54.825372173 +0000 UTC m=+1044.488289096 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/f21790d8-af47-4be7-8758-f038bda76908-cert") pod "infra-operator-controller-manager-57548d458d-77t9s" (UID: "f21790d8-af47-4be7-8758-f038bda76908") : secret "infra-operator-webhook-server-cert" not found Dec 03 14:26:52 crc kubenswrapper[4805]: I1203 14:26:52.853707 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/placement-operator-controller-manager-78f8948974-b8l4l"] Dec 03 14:26:52 crc kubenswrapper[4805]: I1203 14:26:52.861995 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/telemetry-operator-controller-manager-cc9d8f87b-qfmmx"] Dec 03 14:26:52 crc kubenswrapper[4805]: W1203 14:26:52.871987 4805 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf853eae6_829b_42f9_9198_3d7dcabd405f.slice/crio-dae12cee47aa9324389905e193babd5f18c5f90c999256d8e74bdcecd8b62a97 WatchSource:0}: Error finding container dae12cee47aa9324389905e193babd5f18c5f90c999256d8e74bdcecd8b62a97: Status 404 returned error can't find the container with id dae12cee47aa9324389905e193babd5f18c5f90c999256d8e74bdcecd8b62a97 Dec 03 14:26:52 crc kubenswrapper[4805]: I1203 14:26:52.901926 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/manila-operator-controller-manager-7c79b5df47-d2thc"] Dec 03 14:26:52 crc kubenswrapper[4805]: I1203 14:26:52.907050 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/swift-operator-controller-manager-5f8c65bbfc-qrvpw"] Dec 03 14:26:52 crc kubenswrapper[4805]: W1203 14:26:52.943909 4805 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod967f26e7_3027_4e23_9f02_d111d714a396.slice/crio-34e746f91ab5bcee4aa29e1cea388d4fe5cc5c2043b4f725f180e6b4f0be80b1 WatchSource:0}: Error finding container 34e746f91ab5bcee4aa29e1cea388d4fe5cc5c2043b4f725f180e6b4f0be80b1: Status 404 returned error can't find the container with id 34e746f91ab5bcee4aa29e1cea388d4fe5cc5c2043b4f725f180e6b4f0be80b1 Dec 03 14:26:52 crc kubenswrapper[4805]: W1203 14:26:52.949951 4805 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podb51f4884_a980_4643_9deb_8668aead3876.slice/crio-3e7f1c0ba413a56d65c8e818a598b47e7e7183b7f1e20519b72820cacbcb9ad1 WatchSource:0}: Error finding container 3e7f1c0ba413a56d65c8e818a598b47e7e7183b7f1e20519b72820cacbcb9ad1: Status 404 returned error can't find the container with id 3e7f1c0ba413a56d65c8e818a598b47e7e7183b7f1e20519b72820cacbcb9ad1 Dec 03 14:26:52 crc kubenswrapper[4805]: I1203 14:26:52.951664 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/barbican-operator-controller-manager-7d9dfd778-jc4dw"] Dec 03 14:26:52 crc kubenswrapper[4805]: I1203 14:26:52.957990 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/octavia-operator-controller-manager-998648c74-8mbdb"] Dec 03 14:26:52 crc kubenswrapper[4805]: I1203 14:26:52.963022 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/neutron-operator-controller-manager-5fdfd5b6b5-4mlhk"] Dec 03 14:26:52 crc kubenswrapper[4805]: I1203 14:26:52.992201 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-5q6j2"] Dec 03 14:26:53 crc kubenswrapper[4805]: I1203 14:26:53.001487 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/keystone-operator-controller-manager-7765d96ddf-qzzz9"] Dec 03 14:26:53 crc kubenswrapper[4805]: I1203 14:26:53.005069 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/watcher-operator-controller-manager-769dc69bc-fl6mm"] Dec 03 14:26:53 crc kubenswrapper[4805]: I1203 14:26:53.009759 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/nova-operator-controller-manager-697bc559fc-5rl85"] Dec 03 14:26:53 crc kubenswrapper[4805]: I1203 14:26:53.016066 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/ovn-operator-controller-manager-b6456fdb6-djzcm"] Dec 03 14:26:53 crc kubenswrapper[4805]: I1203 14:26:53.020145 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/test-operator-controller-manager-5854674fcc-d6pvv"] Dec 03 14:26:53 crc kubenswrapper[4805]: W1203 14:26:53.034830 4805 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod42b37265_f5bd_4ac4_9b96_d758fbe1b31e.slice/crio-610d8d993f612cfa89371e3081e20e70c47a0b258bd2a90a9a89a18fde91d173 WatchSource:0}: Error finding container 610d8d993f612cfa89371e3081e20e70c47a0b258bd2a90a9a89a18fde91d173: Status 404 returned error can't find the container with id 610d8d993f612cfa89371e3081e20e70c47a0b258bd2a90a9a89a18fde91d173 Dec 03 14:26:53 crc kubenswrapper[4805]: I1203 14:26:53.039388 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/mariadb-operator-controller-manager-56bbcc9d85-n6m2d"] Dec 03 14:26:53 crc kubenswrapper[4805]: W1203 14:26:53.039572 4805 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod05506111_ddb3_4ab9_93a0_ad2bedff67a4.slice/crio-a150a66e9943bd946086a9783b4a9437b9a47ccc505f01d1580c1f1154c9aef4 WatchSource:0}: Error finding container a150a66e9943bd946086a9783b4a9437b9a47ccc505f01d1580c1f1154c9aef4: Status 404 returned error can't find the container with id a150a66e9943bd946086a9783b4a9437b9a47ccc505f01d1580c1f1154c9aef4 Dec 03 14:26:53 crc kubenswrapper[4805]: E1203 14:26:53.042877 4805 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/nova-operator@sha256:779f0cee6024d0fb8f259b036fe790e62aa5a3b0431ea9bf15a6e7d02e2e5670,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-jjz2n,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod nova-operator-controller-manager-697bc559fc-5rl85_openstack-operators(05506111-ddb3-4ab9-93a0-ad2bedff67a4): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Dec 03 14:26:53 crc kubenswrapper[4805]: E1203 14:26:53.044081 4805 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/ovn-operator@sha256:635a4aef9d6f0b799e8ec91333dbb312160c001d05b3c63f614c124e0b67cb59,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-xvkjw,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod ovn-operator-controller-manager-b6456fdb6-djzcm_openstack-operators(42b37265-f5bd-4ac4-9b96-d758fbe1b31e): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Dec 03 14:26:53 crc kubenswrapper[4805]: E1203 14:26:53.044597 4805 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:kube-rbac-proxy,Image:quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0,Command:[],Args:[--secure-listen-address=0.0.0.0:8443 --upstream=http://127.0.0.1:8080/ --logtostderr=true --v=0],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:https,HostPort:0,ContainerPort:8443,Protocol:TCP,HostIP:,},},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{134217728 0} {} BinarySI},},Requests:ResourceList{cpu: {{5 -3} {} 5m DecimalSI},memory: {{67108864 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-jjz2n,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod nova-operator-controller-manager-697bc559fc-5rl85_openstack-operators(05506111-ddb3-4ab9-93a0-ad2bedff67a4): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Dec 03 14:26:53 crc kubenswrapper[4805]: E1203 14:26:53.045997 4805 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ErrImagePull: \"pull QPS exceeded\"]" pod="openstack-operators/nova-operator-controller-manager-697bc559fc-5rl85" podUID="05506111-ddb3-4ab9-93a0-ad2bedff67a4" Dec 03 14:26:53 crc kubenswrapper[4805]: E1203 14:26:53.046904 4805 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:kube-rbac-proxy,Image:quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0,Command:[],Args:[--secure-listen-address=0.0.0.0:8443 --upstream=http://127.0.0.1:8080/ --logtostderr=true --v=0],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:https,HostPort:0,ContainerPort:8443,Protocol:TCP,HostIP:,},},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{134217728 0} {} BinarySI},},Requests:ResourceList{cpu: {{5 -3} {} 5m DecimalSI},memory: {{67108864 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-xvkjw,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod ovn-operator-controller-manager-b6456fdb6-djzcm_openstack-operators(42b37265-f5bd-4ac4-9b96-d758fbe1b31e): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Dec 03 14:26:53 crc kubenswrapper[4805]: E1203 14:26:53.048142 4805 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ErrImagePull: \"pull QPS exceeded\"]" pod="openstack-operators/ovn-operator-controller-manager-b6456fdb6-djzcm" podUID="42b37265-f5bd-4ac4-9b96-d758fbe1b31e" Dec 03 14:26:53 crc kubenswrapper[4805]: W1203 14:26:53.052105 4805 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podac4ba0f5_635d_4a72_a133_9b4aaf88f097.slice/crio-2a5de6b4b6c34f70bca540c4e08a98d16f40495a22abc3c7f2fb13d7bfc5f691 WatchSource:0}: Error finding container 2a5de6b4b6c34f70bca540c4e08a98d16f40495a22abc3c7f2fb13d7bfc5f691: Status 404 returned error can't find the container with id 2a5de6b4b6c34f70bca540c4e08a98d16f40495a22abc3c7f2fb13d7bfc5f691 Dec 03 14:26:53 crc kubenswrapper[4805]: E1203 14:26:53.052003 4805 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/keystone-operator@sha256:72ad6517987f674af0d0ae092cbb874aeae909c8b8b60188099c311762ebc8f7,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-hmzsn,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod keystone-operator-controller-manager-7765d96ddf-qzzz9_openstack-operators(ffa72db0-588a-4a78-a1cd-8710f5ce4f41): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Dec 03 14:26:53 crc kubenswrapper[4805]: E1203 14:26:53.063418 4805 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:kube-rbac-proxy,Image:quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0,Command:[],Args:[--secure-listen-address=0.0.0.0:8443 --upstream=http://127.0.0.1:8080/ --logtostderr=true --v=0],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:https,HostPort:0,ContainerPort:8443,Protocol:TCP,HostIP:,},},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{134217728 0} {} BinarySI},},Requests:ResourceList{cpu: {{5 -3} {} 5m DecimalSI},memory: {{67108864 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-hmzsn,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod keystone-operator-controller-manager-7765d96ddf-qzzz9_openstack-operators(ffa72db0-588a-4a78-a1cd-8710f5ce4f41): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Dec 03 14:26:53 crc kubenswrapper[4805]: E1203 14:26:53.064538 4805 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/test-operator@sha256:101b3e007d8c9f2e183262d7712f986ad51256448099069bc14f1ea5f997ab94,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-jslvv,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod test-operator-controller-manager-5854674fcc-d6pvv_openstack-operators(ac4ba0f5-635d-4a72-a133-9b4aaf88f097): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Dec 03 14:26:53 crc kubenswrapper[4805]: E1203 14:26:53.064621 4805 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ErrImagePull: \"pull QPS exceeded\"]" pod="openstack-operators/keystone-operator-controller-manager-7765d96ddf-qzzz9" podUID="ffa72db0-588a-4a78-a1cd-8710f5ce4f41" Dec 03 14:26:53 crc kubenswrapper[4805]: E1203 14:26:53.066151 4805 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/mariadb-operator@sha256:600ca007e493d3af0fcc2ebac92e8da5efd2afe812b62d7d3d4dd0115bdf05d7,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-v5zjt,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod mariadb-operator-controller-manager-56bbcc9d85-n6m2d_openstack-operators(2d72009f-b9ef-4c9d-b224-5b133a3cd93a): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Dec 03 14:26:53 crc kubenswrapper[4805]: E1203 14:26:53.075644 4805 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:kube-rbac-proxy,Image:quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0,Command:[],Args:[--secure-listen-address=0.0.0.0:8443 --upstream=http://127.0.0.1:8080/ --logtostderr=true --v=0],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:https,HostPort:0,ContainerPort:8443,Protocol:TCP,HostIP:,},},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{134217728 0} {} BinarySI},},Requests:ResourceList{cpu: {{5 -3} {} 5m DecimalSI},memory: {{67108864 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-jslvv,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod test-operator-controller-manager-5854674fcc-d6pvv_openstack-operators(ac4ba0f5-635d-4a72-a133-9b4aaf88f097): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Dec 03 14:26:53 crc kubenswrapper[4805]: E1203 14:26:53.079134 4805 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:kube-rbac-proxy,Image:quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0,Command:[],Args:[--secure-listen-address=0.0.0.0:8443 --upstream=http://127.0.0.1:8080/ --logtostderr=true --v=0],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:https,HostPort:0,ContainerPort:8443,Protocol:TCP,HostIP:,},},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{134217728 0} {} BinarySI},},Requests:ResourceList{cpu: {{5 -3} {} 5m DecimalSI},memory: {{67108864 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-v5zjt,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod mariadb-operator-controller-manager-56bbcc9d85-n6m2d_openstack-operators(2d72009f-b9ef-4c9d-b224-5b133a3cd93a): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Dec 03 14:26:53 crc kubenswrapper[4805]: E1203 14:26:53.079164 4805 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ErrImagePull: \"pull QPS exceeded\"]" pod="openstack-operators/test-operator-controller-manager-5854674fcc-d6pvv" podUID="ac4ba0f5-635d-4a72-a133-9b4aaf88f097" Dec 03 14:26:53 crc kubenswrapper[4805]: E1203 14:26:53.080301 4805 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ErrImagePull: \"pull QPS exceeded\"]" pod="openstack-operators/mariadb-operator-controller-manager-56bbcc9d85-n6m2d" podUID="2d72009f-b9ef-4c9d-b224-5b133a3cd93a" Dec 03 14:26:53 crc kubenswrapper[4805]: W1203 14:26:53.082958 4805 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod0f040ce3_0a83_426b_8b9f_b63183d16b4a.slice/crio-ced14a3b49f178c86835b109b85512e87bceb8c7934fa3b48e1cd0150a422202 WatchSource:0}: Error finding container ced14a3b49f178c86835b109b85512e87bceb8c7934fa3b48e1cd0150a422202: Status 404 returned error can't find the container with id ced14a3b49f178c86835b109b85512e87bceb8c7934fa3b48e1cd0150a422202 Dec 03 14:26:53 crc kubenswrapper[4805]: I1203 14:26:53.235271 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/b9d0830e-0f5d-4639-ab7d-92dcf07a29bc-cert\") pod \"openstack-baremetal-operator-controller-manager-64bc77cfd4x9mdd\" (UID: \"b9d0830e-0f5d-4639-ab7d-92dcf07a29bc\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-64bc77cfd4x9mdd" Dec 03 14:26:53 crc kubenswrapper[4805]: E1203 14:26:53.235649 4805 secret.go:188] Couldn't get secret openstack-operators/openstack-baremetal-operator-webhook-server-cert: secret "openstack-baremetal-operator-webhook-server-cert" not found Dec 03 14:26:53 crc kubenswrapper[4805]: E1203 14:26:53.235741 4805 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/b9d0830e-0f5d-4639-ab7d-92dcf07a29bc-cert podName:b9d0830e-0f5d-4639-ab7d-92dcf07a29bc nodeName:}" failed. No retries permitted until 2025-12-03 14:26:55.235718221 +0000 UTC m=+1044.898635214 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/b9d0830e-0f5d-4639-ab7d-92dcf07a29bc-cert") pod "openstack-baremetal-operator-controller-manager-64bc77cfd4x9mdd" (UID: "b9d0830e-0f5d-4639-ab7d-92dcf07a29bc") : secret "openstack-baremetal-operator-webhook-server-cert" not found Dec 03 14:26:53 crc kubenswrapper[4805]: I1203 14:26:53.407951 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-5q6j2" event={"ID":"0aa860af-81d3-44ba-9020-0ffef21cefb2","Type":"ContainerStarted","Data":"8b8ccac8373828f85e46a906608732adf1b794eebed6805153c516c9ad8a84e1"} Dec 03 14:26:53 crc kubenswrapper[4805]: I1203 14:26:53.410217 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/watcher-operator-controller-manager-769dc69bc-fl6mm" event={"ID":"30776d04-200b-4c10-b48d-f51a926e469b","Type":"ContainerStarted","Data":"34b4a8e46ac845e3c4a84e41c585abb3eb633a2c32db124cf8324676f8e22c88"} Dec 03 14:26:53 crc kubenswrapper[4805]: I1203 14:26:53.411083 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/test-operator-controller-manager-5854674fcc-d6pvv" event={"ID":"ac4ba0f5-635d-4a72-a133-9b4aaf88f097","Type":"ContainerStarted","Data":"2a5de6b4b6c34f70bca540c4e08a98d16f40495a22abc3c7f2fb13d7bfc5f691"} Dec 03 14:26:53 crc kubenswrapper[4805]: E1203 14:26:53.413462 4805 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/test-operator@sha256:101b3e007d8c9f2e183262d7712f986ad51256448099069bc14f1ea5f997ab94\\\"\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0\\\"\"]" pod="openstack-operators/test-operator-controller-manager-5854674fcc-d6pvv" podUID="ac4ba0f5-635d-4a72-a133-9b4aaf88f097" Dec 03 14:26:53 crc kubenswrapper[4805]: I1203 14:26:53.414300 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/manila-operator-controller-manager-7c79b5df47-d2thc" event={"ID":"967f26e7-3027-4e23-9f02-d111d714a396","Type":"ContainerStarted","Data":"34e746f91ab5bcee4aa29e1cea388d4fe5cc5c2043b4f725f180e6b4f0be80b1"} Dec 03 14:26:53 crc kubenswrapper[4805]: I1203 14:26:53.415336 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/cinder-operator-controller-manager-859b6ccc6-2nzgx" event={"ID":"f9b6b508-1cf6-40aa-829e-b9cf264921e5","Type":"ContainerStarted","Data":"dcc7491d664bde1737aee87b399679a68bde2c46e3ca187769af4b84f863663d"} Dec 03 14:26:53 crc kubenswrapper[4805]: I1203 14:26:53.417114 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/placement-operator-controller-manager-78f8948974-b8l4l" event={"ID":"f853eae6-829b-42f9-9198-3d7dcabd405f","Type":"ContainerStarted","Data":"dae12cee47aa9324389905e193babd5f18c5f90c999256d8e74bdcecd8b62a97"} Dec 03 14:26:53 crc kubenswrapper[4805]: I1203 14:26:53.418545 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/telemetry-operator-controller-manager-cc9d8f87b-qfmmx" event={"ID":"8cbfcd07-e136-4bfd-b3c6-6615c911d440","Type":"ContainerStarted","Data":"fb45e9b68a5246b8b886906e00111323fda4bd01c31c158178c19f52612f5f95"} Dec 03 14:26:53 crc kubenswrapper[4805]: I1203 14:26:53.419716 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/designate-operator-controller-manager-78b4bc895b-w56sr" event={"ID":"326305da-3798-498c-9b47-0aeb71bd2205","Type":"ContainerStarted","Data":"98e012a82f5d5b1cc2a1e79c04f628aebf92abf285a92334863c612d09b1207d"} Dec 03 14:26:53 crc kubenswrapper[4805]: I1203 14:26:53.421760 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/keystone-operator-controller-manager-7765d96ddf-qzzz9" event={"ID":"ffa72db0-588a-4a78-a1cd-8710f5ce4f41","Type":"ContainerStarted","Data":"03c62552aeae8fbe8669eeea38fa2508d11b19dab46e1660dccb0ed3dd268b79"} Dec 03 14:26:53 crc kubenswrapper[4805]: I1203 14:26:53.424348 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ovn-operator-controller-manager-b6456fdb6-djzcm" event={"ID":"42b37265-f5bd-4ac4-9b96-d758fbe1b31e","Type":"ContainerStarted","Data":"610d8d993f612cfa89371e3081e20e70c47a0b258bd2a90a9a89a18fde91d173"} Dec 03 14:26:53 crc kubenswrapper[4805]: E1203 14:26:53.426081 4805 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/keystone-operator@sha256:72ad6517987f674af0d0ae092cbb874aeae909c8b8b60188099c311762ebc8f7\\\"\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0\\\"\"]" pod="openstack-operators/keystone-operator-controller-manager-7765d96ddf-qzzz9" podUID="ffa72db0-588a-4a78-a1cd-8710f5ce4f41" Dec 03 14:26:53 crc kubenswrapper[4805]: I1203 14:26:53.437089 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/mariadb-operator-controller-manager-56bbcc9d85-n6m2d" event={"ID":"2d72009f-b9ef-4c9d-b224-5b133a3cd93a","Type":"ContainerStarted","Data":"d18e7bca0e5bd03ba67d57393ef3f168525adee3f687183cb319c2ab070c1977"} Dec 03 14:26:53 crc kubenswrapper[4805]: I1203 14:26:53.437779 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/41c4618c-1f15-4ceb-825b-546e828e399a-metrics-certs\") pod \"openstack-operator-controller-manager-cc68d789f-pkvr8\" (UID: \"41c4618c-1f15-4ceb-825b-546e828e399a\") " pod="openstack-operators/openstack-operator-controller-manager-cc68d789f-pkvr8" Dec 03 14:26:53 crc kubenswrapper[4805]: I1203 14:26:53.437924 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/41c4618c-1f15-4ceb-825b-546e828e399a-webhook-certs\") pod \"openstack-operator-controller-manager-cc68d789f-pkvr8\" (UID: \"41c4618c-1f15-4ceb-825b-546e828e399a\") " pod="openstack-operators/openstack-operator-controller-manager-cc68d789f-pkvr8" Dec 03 14:26:53 crc kubenswrapper[4805]: E1203 14:26:53.438065 4805 secret.go:188] Couldn't get secret openstack-operators/webhook-server-cert: secret "webhook-server-cert" not found Dec 03 14:26:53 crc kubenswrapper[4805]: E1203 14:26:53.438110 4805 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/41c4618c-1f15-4ceb-825b-546e828e399a-webhook-certs podName:41c4618c-1f15-4ceb-825b-546e828e399a nodeName:}" failed. No retries permitted until 2025-12-03 14:26:55.438094809 +0000 UTC m=+1045.101011732 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "webhook-certs" (UniqueName: "kubernetes.io/secret/41c4618c-1f15-4ceb-825b-546e828e399a-webhook-certs") pod "openstack-operator-controller-manager-cc68d789f-pkvr8" (UID: "41c4618c-1f15-4ceb-825b-546e828e399a") : secret "webhook-server-cert" not found Dec 03 14:26:53 crc kubenswrapper[4805]: E1203 14:26:53.438157 4805 secret.go:188] Couldn't get secret openstack-operators/metrics-server-cert: secret "metrics-server-cert" not found Dec 03 14:26:53 crc kubenswrapper[4805]: E1203 14:26:53.438191 4805 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/41c4618c-1f15-4ceb-825b-546e828e399a-metrics-certs podName:41c4618c-1f15-4ceb-825b-546e828e399a nodeName:}" failed. No retries permitted until 2025-12-03 14:26:55.438182321 +0000 UTC m=+1045.101099244 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/41c4618c-1f15-4ceb-825b-546e828e399a-metrics-certs") pod "openstack-operator-controller-manager-cc68d789f-pkvr8" (UID: "41c4618c-1f15-4ceb-825b-546e828e399a") : secret "metrics-server-cert" not found Dec 03 14:26:53 crc kubenswrapper[4805]: E1203 14:26:53.441262 4805 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/mariadb-operator@sha256:600ca007e493d3af0fcc2ebac92e8da5efd2afe812b62d7d3d4dd0115bdf05d7\\\"\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0\\\"\"]" pod="openstack-operators/mariadb-operator-controller-manager-56bbcc9d85-n6m2d" podUID="2d72009f-b9ef-4c9d-b224-5b133a3cd93a" Dec 03 14:26:53 crc kubenswrapper[4805]: E1203 14:26:53.442645 4805 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/ovn-operator@sha256:635a4aef9d6f0b799e8ec91333dbb312160c001d05b3c63f614c124e0b67cb59\\\"\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0\\\"\"]" pod="openstack-operators/ovn-operator-controller-manager-b6456fdb6-djzcm" podUID="42b37265-f5bd-4ac4-9b96-d758fbe1b31e" Dec 03 14:26:53 crc kubenswrapper[4805]: I1203 14:26:53.443293 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/neutron-operator-controller-manager-5fdfd5b6b5-4mlhk" event={"ID":"58aa635e-157c-4ae2-ad0c-568c144638d1","Type":"ContainerStarted","Data":"4522ae58d646363d6ec2aa86dd0536d08af317b3060f95236ad1e42206bd380d"} Dec 03 14:26:53 crc kubenswrapper[4805]: I1203 14:26:53.445919 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/barbican-operator-controller-manager-7d9dfd778-jc4dw" event={"ID":"cbf3813c-fb86-49d2-8fac-25487ff08601","Type":"ContainerStarted","Data":"1b4b609f11a0f1370997f2b7be971053a09f4450a466a282b1b2c7cd341bf0b3"} Dec 03 14:26:53 crc kubenswrapper[4805]: I1203 14:26:53.448179 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/swift-operator-controller-manager-5f8c65bbfc-qrvpw" event={"ID":"b51f4884-a980-4643-9deb-8668aead3876","Type":"ContainerStarted","Data":"3e7f1c0ba413a56d65c8e818a598b47e7e7183b7f1e20519b72820cacbcb9ad1"} Dec 03 14:26:53 crc kubenswrapper[4805]: I1203 14:26:53.449989 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/nova-operator-controller-manager-697bc559fc-5rl85" event={"ID":"05506111-ddb3-4ab9-93a0-ad2bedff67a4","Type":"ContainerStarted","Data":"a150a66e9943bd946086a9783b4a9437b9a47ccc505f01d1580c1f1154c9aef4"} Dec 03 14:26:53 crc kubenswrapper[4805]: I1203 14:26:53.454392 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ironic-operator-controller-manager-6c548fd776-mz4pw" event={"ID":"1ad4fa11-2f11-4a06-9944-bba8c7445e84","Type":"ContainerStarted","Data":"4678aa56fb358b08052bf83567eb2f1e145709b9015efbf10b2a30c42b136b30"} Dec 03 14:26:53 crc kubenswrapper[4805]: E1203 14:26:53.454555 4805 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/nova-operator@sha256:779f0cee6024d0fb8f259b036fe790e62aa5a3b0431ea9bf15a6e7d02e2e5670\\\"\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0\\\"\"]" pod="openstack-operators/nova-operator-controller-manager-697bc559fc-5rl85" podUID="05506111-ddb3-4ab9-93a0-ad2bedff67a4" Dec 03 14:26:53 crc kubenswrapper[4805]: I1203 14:26:53.457301 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/octavia-operator-controller-manager-998648c74-8mbdb" event={"ID":"0f040ce3-0a83-426b-8b9f-b63183d16b4a","Type":"ContainerStarted","Data":"ced14a3b49f178c86835b109b85512e87bceb8c7934fa3b48e1cd0150a422202"} Dec 03 14:26:54 crc kubenswrapper[4805]: E1203 14:26:54.489035 4805 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/nova-operator@sha256:779f0cee6024d0fb8f259b036fe790e62aa5a3b0431ea9bf15a6e7d02e2e5670\\\"\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0\\\"\"]" pod="openstack-operators/nova-operator-controller-manager-697bc559fc-5rl85" podUID="05506111-ddb3-4ab9-93a0-ad2bedff67a4" Dec 03 14:26:54 crc kubenswrapper[4805]: E1203 14:26:54.489195 4805 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/mariadb-operator@sha256:600ca007e493d3af0fcc2ebac92e8da5efd2afe812b62d7d3d4dd0115bdf05d7\\\"\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0\\\"\"]" pod="openstack-operators/mariadb-operator-controller-manager-56bbcc9d85-n6m2d" podUID="2d72009f-b9ef-4c9d-b224-5b133a3cd93a" Dec 03 14:26:54 crc kubenswrapper[4805]: E1203 14:26:54.489252 4805 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/test-operator@sha256:101b3e007d8c9f2e183262d7712f986ad51256448099069bc14f1ea5f997ab94\\\"\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0\\\"\"]" pod="openstack-operators/test-operator-controller-manager-5854674fcc-d6pvv" podUID="ac4ba0f5-635d-4a72-a133-9b4aaf88f097" Dec 03 14:26:54 crc kubenswrapper[4805]: E1203 14:26:54.490437 4805 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/ovn-operator@sha256:635a4aef9d6f0b799e8ec91333dbb312160c001d05b3c63f614c124e0b67cb59\\\"\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0\\\"\"]" pod="openstack-operators/ovn-operator-controller-manager-b6456fdb6-djzcm" podUID="42b37265-f5bd-4ac4-9b96-d758fbe1b31e" Dec 03 14:26:54 crc kubenswrapper[4805]: E1203 14:26:54.490913 4805 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/keystone-operator@sha256:72ad6517987f674af0d0ae092cbb874aeae909c8b8b60188099c311762ebc8f7\\\"\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0\\\"\"]" pod="openstack-operators/keystone-operator-controller-manager-7765d96ddf-qzzz9" podUID="ffa72db0-588a-4a78-a1cd-8710f5ce4f41" Dec 03 14:26:54 crc kubenswrapper[4805]: I1203 14:26:54.862977 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/f21790d8-af47-4be7-8758-f038bda76908-cert\") pod \"infra-operator-controller-manager-57548d458d-77t9s\" (UID: \"f21790d8-af47-4be7-8758-f038bda76908\") " pod="openstack-operators/infra-operator-controller-manager-57548d458d-77t9s" Dec 03 14:26:54 crc kubenswrapper[4805]: E1203 14:26:54.864114 4805 secret.go:188] Couldn't get secret openstack-operators/infra-operator-webhook-server-cert: secret "infra-operator-webhook-server-cert" not found Dec 03 14:26:54 crc kubenswrapper[4805]: E1203 14:26:54.864181 4805 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/f21790d8-af47-4be7-8758-f038bda76908-cert podName:f21790d8-af47-4be7-8758-f038bda76908 nodeName:}" failed. No retries permitted until 2025-12-03 14:26:58.864160762 +0000 UTC m=+1048.527077735 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/f21790d8-af47-4be7-8758-f038bda76908-cert") pod "infra-operator-controller-manager-57548d458d-77t9s" (UID: "f21790d8-af47-4be7-8758-f038bda76908") : secret "infra-operator-webhook-server-cert" not found Dec 03 14:26:55 crc kubenswrapper[4805]: I1203 14:26:55.269531 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/b9d0830e-0f5d-4639-ab7d-92dcf07a29bc-cert\") pod \"openstack-baremetal-operator-controller-manager-64bc77cfd4x9mdd\" (UID: \"b9d0830e-0f5d-4639-ab7d-92dcf07a29bc\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-64bc77cfd4x9mdd" Dec 03 14:26:55 crc kubenswrapper[4805]: E1203 14:26:55.269923 4805 secret.go:188] Couldn't get secret openstack-operators/openstack-baremetal-operator-webhook-server-cert: secret "openstack-baremetal-operator-webhook-server-cert" not found Dec 03 14:26:55 crc kubenswrapper[4805]: E1203 14:26:55.269975 4805 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/b9d0830e-0f5d-4639-ab7d-92dcf07a29bc-cert podName:b9d0830e-0f5d-4639-ab7d-92dcf07a29bc nodeName:}" failed. No retries permitted until 2025-12-03 14:26:59.269961058 +0000 UTC m=+1048.932877981 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/b9d0830e-0f5d-4639-ab7d-92dcf07a29bc-cert") pod "openstack-baremetal-operator-controller-manager-64bc77cfd4x9mdd" (UID: "b9d0830e-0f5d-4639-ab7d-92dcf07a29bc") : secret "openstack-baremetal-operator-webhook-server-cert" not found Dec 03 14:26:55 crc kubenswrapper[4805]: I1203 14:26:55.474498 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/41c4618c-1f15-4ceb-825b-546e828e399a-metrics-certs\") pod \"openstack-operator-controller-manager-cc68d789f-pkvr8\" (UID: \"41c4618c-1f15-4ceb-825b-546e828e399a\") " pod="openstack-operators/openstack-operator-controller-manager-cc68d789f-pkvr8" Dec 03 14:26:55 crc kubenswrapper[4805]: I1203 14:26:55.474639 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/41c4618c-1f15-4ceb-825b-546e828e399a-webhook-certs\") pod \"openstack-operator-controller-manager-cc68d789f-pkvr8\" (UID: \"41c4618c-1f15-4ceb-825b-546e828e399a\") " pod="openstack-operators/openstack-operator-controller-manager-cc68d789f-pkvr8" Dec 03 14:26:55 crc kubenswrapper[4805]: E1203 14:26:55.474787 4805 secret.go:188] Couldn't get secret openstack-operators/webhook-server-cert: secret "webhook-server-cert" not found Dec 03 14:26:55 crc kubenswrapper[4805]: E1203 14:26:55.474862 4805 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/41c4618c-1f15-4ceb-825b-546e828e399a-webhook-certs podName:41c4618c-1f15-4ceb-825b-546e828e399a nodeName:}" failed. No retries permitted until 2025-12-03 14:26:59.474828263 +0000 UTC m=+1049.137745186 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "webhook-certs" (UniqueName: "kubernetes.io/secret/41c4618c-1f15-4ceb-825b-546e828e399a-webhook-certs") pod "openstack-operator-controller-manager-cc68d789f-pkvr8" (UID: "41c4618c-1f15-4ceb-825b-546e828e399a") : secret "webhook-server-cert" not found Dec 03 14:26:55 crc kubenswrapper[4805]: E1203 14:26:55.474913 4805 secret.go:188] Couldn't get secret openstack-operators/metrics-server-cert: secret "metrics-server-cert" not found Dec 03 14:26:55 crc kubenswrapper[4805]: E1203 14:26:55.474982 4805 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/41c4618c-1f15-4ceb-825b-546e828e399a-metrics-certs podName:41c4618c-1f15-4ceb-825b-546e828e399a nodeName:}" failed. No retries permitted until 2025-12-03 14:26:59.474966397 +0000 UTC m=+1049.137883320 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/41c4618c-1f15-4ceb-825b-546e828e399a-metrics-certs") pod "openstack-operator-controller-manager-cc68d789f-pkvr8" (UID: "41c4618c-1f15-4ceb-825b-546e828e399a") : secret "metrics-server-cert" not found Dec 03 14:26:58 crc kubenswrapper[4805]: I1203 14:26:58.930408 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/f21790d8-af47-4be7-8758-f038bda76908-cert\") pod \"infra-operator-controller-manager-57548d458d-77t9s\" (UID: \"f21790d8-af47-4be7-8758-f038bda76908\") " pod="openstack-operators/infra-operator-controller-manager-57548d458d-77t9s" Dec 03 14:26:58 crc kubenswrapper[4805]: E1203 14:26:58.930600 4805 secret.go:188] Couldn't get secret openstack-operators/infra-operator-webhook-server-cert: secret "infra-operator-webhook-server-cert" not found Dec 03 14:26:58 crc kubenswrapper[4805]: E1203 14:26:58.931883 4805 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/f21790d8-af47-4be7-8758-f038bda76908-cert podName:f21790d8-af47-4be7-8758-f038bda76908 nodeName:}" failed. No retries permitted until 2025-12-03 14:27:06.931828204 +0000 UTC m=+1056.594745177 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/f21790d8-af47-4be7-8758-f038bda76908-cert") pod "infra-operator-controller-manager-57548d458d-77t9s" (UID: "f21790d8-af47-4be7-8758-f038bda76908") : secret "infra-operator-webhook-server-cert" not found Dec 03 14:26:59 crc kubenswrapper[4805]: I1203 14:26:59.337070 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/b9d0830e-0f5d-4639-ab7d-92dcf07a29bc-cert\") pod \"openstack-baremetal-operator-controller-manager-64bc77cfd4x9mdd\" (UID: \"b9d0830e-0f5d-4639-ab7d-92dcf07a29bc\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-64bc77cfd4x9mdd" Dec 03 14:26:59 crc kubenswrapper[4805]: E1203 14:26:59.337296 4805 secret.go:188] Couldn't get secret openstack-operators/openstack-baremetal-operator-webhook-server-cert: secret "openstack-baremetal-operator-webhook-server-cert" not found Dec 03 14:26:59 crc kubenswrapper[4805]: E1203 14:26:59.337382 4805 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/b9d0830e-0f5d-4639-ab7d-92dcf07a29bc-cert podName:b9d0830e-0f5d-4639-ab7d-92dcf07a29bc nodeName:}" failed. No retries permitted until 2025-12-03 14:27:07.337359132 +0000 UTC m=+1057.000276135 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/b9d0830e-0f5d-4639-ab7d-92dcf07a29bc-cert") pod "openstack-baremetal-operator-controller-manager-64bc77cfd4x9mdd" (UID: "b9d0830e-0f5d-4639-ab7d-92dcf07a29bc") : secret "openstack-baremetal-operator-webhook-server-cert" not found Dec 03 14:26:59 crc kubenswrapper[4805]: I1203 14:26:59.540531 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/41c4618c-1f15-4ceb-825b-546e828e399a-webhook-certs\") pod \"openstack-operator-controller-manager-cc68d789f-pkvr8\" (UID: \"41c4618c-1f15-4ceb-825b-546e828e399a\") " pod="openstack-operators/openstack-operator-controller-manager-cc68d789f-pkvr8" Dec 03 14:26:59 crc kubenswrapper[4805]: I1203 14:26:59.540619 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/41c4618c-1f15-4ceb-825b-546e828e399a-metrics-certs\") pod \"openstack-operator-controller-manager-cc68d789f-pkvr8\" (UID: \"41c4618c-1f15-4ceb-825b-546e828e399a\") " pod="openstack-operators/openstack-operator-controller-manager-cc68d789f-pkvr8" Dec 03 14:26:59 crc kubenswrapper[4805]: E1203 14:26:59.540712 4805 secret.go:188] Couldn't get secret openstack-operators/webhook-server-cert: secret "webhook-server-cert" not found Dec 03 14:26:59 crc kubenswrapper[4805]: E1203 14:26:59.540738 4805 secret.go:188] Couldn't get secret openstack-operators/metrics-server-cert: secret "metrics-server-cert" not found Dec 03 14:26:59 crc kubenswrapper[4805]: E1203 14:26:59.540771 4805 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/41c4618c-1f15-4ceb-825b-546e828e399a-webhook-certs podName:41c4618c-1f15-4ceb-825b-546e828e399a nodeName:}" failed. No retries permitted until 2025-12-03 14:27:07.540753387 +0000 UTC m=+1057.203670310 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "webhook-certs" (UniqueName: "kubernetes.io/secret/41c4618c-1f15-4ceb-825b-546e828e399a-webhook-certs") pod "openstack-operator-controller-manager-cc68d789f-pkvr8" (UID: "41c4618c-1f15-4ceb-825b-546e828e399a") : secret "webhook-server-cert" not found Dec 03 14:26:59 crc kubenswrapper[4805]: E1203 14:26:59.540788 4805 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/41c4618c-1f15-4ceb-825b-546e828e399a-metrics-certs podName:41c4618c-1f15-4ceb-825b-546e828e399a nodeName:}" failed. No retries permitted until 2025-12-03 14:27:07.540781258 +0000 UTC m=+1057.203698181 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/41c4618c-1f15-4ceb-825b-546e828e399a-metrics-certs") pod "openstack-operator-controller-manager-cc68d789f-pkvr8" (UID: "41c4618c-1f15-4ceb-825b-546e828e399a") : secret "metrics-server-cert" not found Dec 03 14:27:05 crc kubenswrapper[4805]: E1203 14:27:05.835848 4805 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/rabbitmq-cluster-operator@sha256:893e66303c1b0bc1d00a299a3f0380bad55c8dc813c8a1c6a4aab379f5aa12a2" Dec 03 14:27:05 crc kubenswrapper[4805]: E1203 14:27:05.836593 4805 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:operator,Image:quay.io/openstack-k8s-operators/rabbitmq-cluster-operator@sha256:893e66303c1b0bc1d00a299a3f0380bad55c8dc813c8a1c6a4aab379f5aa12a2,Command:[/manager],Args:[],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:metrics,HostPort:0,ContainerPort:9782,Protocol:TCP,HostIP:,},},Env:[]EnvVar{EnvVar{Name:OPERATOR_NAMESPACE,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:metadata.namespace,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{200 -3} {} 200m DecimalSI},memory: {{524288000 0} {} 500Mi BinarySI},},Requests:ResourceList{cpu: {{5 -3} {} 5m DecimalSI},memory: {{67108864 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-kqs6q,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000660000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod rabbitmq-cluster-operator-manager-668c99d594-5q6j2_openstack-operators(0aa860af-81d3-44ba-9020-0ffef21cefb2): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 03 14:27:05 crc kubenswrapper[4805]: E1203 14:27:05.837929 4805 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"operator\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-5q6j2" podUID="0aa860af-81d3-44ba-9020-0ffef21cefb2" Dec 03 14:27:06 crc kubenswrapper[4805]: E1203 14:27:06.015965 4805 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="38.102.83.154:5001/openstack-k8s-operators/telemetry-operator:e39d5f8652fd394c3fcc2c0989e45436c83851fe" Dec 03 14:27:06 crc kubenswrapper[4805]: E1203 14:27:06.016014 4805 kuberuntime_image.go:55] "Failed to pull image" err="rpc error: code = Canceled desc = copying config: context canceled" image="38.102.83.154:5001/openstack-k8s-operators/telemetry-operator:e39d5f8652fd394c3fcc2c0989e45436c83851fe" Dec 03 14:27:06 crc kubenswrapper[4805]: E1203 14:27:06.016150 4805 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:38.102.83.154:5001/openstack-k8s-operators/telemetry-operator:e39d5f8652fd394c3fcc2c0989e45436c83851fe,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-nxwjv,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod telemetry-operator-controller-manager-cc9d8f87b-qfmmx_openstack-operators(8cbfcd07-e136-4bfd-b3c6-6615c911d440): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 03 14:27:06 crc kubenswrapper[4805]: I1203 14:27:06.299785 4805 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 03 14:27:06 crc kubenswrapper[4805]: I1203 14:27:06.607459 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/glance-operator-controller-manager-77987cd8cd-pt9fw" event={"ID":"5615ce63-1ec1-4a77-8e44-49e6d114eec6","Type":"ContainerStarted","Data":"5fd1ab1869cfae5a8c8d92b13d2d15c1707ffc0bd3d41f01ebac4c8f32bd912b"} Dec 03 14:27:06 crc kubenswrapper[4805]: I1203 14:27:06.615353 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/placement-operator-controller-manager-78f8948974-b8l4l" event={"ID":"f853eae6-829b-42f9-9198-3d7dcabd405f","Type":"ContainerStarted","Data":"dd82191e9d7c3120d38ab57bfa90c4f35e6e52c266f1d0bc80b3e553f01be330"} Dec 03 14:27:06 crc kubenswrapper[4805]: I1203 14:27:06.622889 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/manila-operator-controller-manager-7c79b5df47-d2thc" event={"ID":"967f26e7-3027-4e23-9f02-d111d714a396","Type":"ContainerStarted","Data":"e1042245bb0420f63a0d9a5929ca19144792e9a7c83f1872f1f760082fed01f3"} Dec 03 14:27:06 crc kubenswrapper[4805]: I1203 14:27:06.631526 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/designate-operator-controller-manager-78b4bc895b-w56sr" event={"ID":"326305da-3798-498c-9b47-0aeb71bd2205","Type":"ContainerStarted","Data":"b9ca4caca12af578d5f7677c89de0991a83f04cb7c2fae11bb1376bf17f367e1"} Dec 03 14:27:06 crc kubenswrapper[4805]: I1203 14:27:06.635310 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/watcher-operator-controller-manager-769dc69bc-fl6mm" event={"ID":"30776d04-200b-4c10-b48d-f51a926e469b","Type":"ContainerStarted","Data":"7dd83e1483f0ea7253178184596b21f8be320ab4afde5b4ad19c1a80e521a71c"} Dec 03 14:27:06 crc kubenswrapper[4805]: I1203 14:27:06.638232 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/horizon-operator-controller-manager-68c6d99b8f-kxbr8" event={"ID":"b29e4120-0cbc-4de5-a4b6-97b97761f49a","Type":"ContainerStarted","Data":"ef16bd7c62b29ee5eccd06da98b17ad480504361217920fd65a1ce66137d57d6"} Dec 03 14:27:06 crc kubenswrapper[4805]: I1203 14:27:06.646722 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/neutron-operator-controller-manager-5fdfd5b6b5-4mlhk" event={"ID":"58aa635e-157c-4ae2-ad0c-568c144638d1","Type":"ContainerStarted","Data":"2ca770cb3035ef017c23ad308cd4499177e1289e830d9ec64861fc5d363f98c9"} Dec 03 14:27:06 crc kubenswrapper[4805]: I1203 14:27:06.662400 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/octavia-operator-controller-manager-998648c74-8mbdb" event={"ID":"0f040ce3-0a83-426b-8b9f-b63183d16b4a","Type":"ContainerStarted","Data":"cccb854df57e09df5f255f3be6aa5a97eb91d0d584e714822f740487eab7d76c"} Dec 03 14:27:06 crc kubenswrapper[4805]: E1203 14:27:06.664539 4805 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"operator\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/rabbitmq-cluster-operator@sha256:893e66303c1b0bc1d00a299a3f0380bad55c8dc813c8a1c6a4aab379f5aa12a2\\\"\"" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-5q6j2" podUID="0aa860af-81d3-44ba-9020-0ffef21cefb2" Dec 03 14:27:06 crc kubenswrapper[4805]: I1203 14:27:06.961208 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/f21790d8-af47-4be7-8758-f038bda76908-cert\") pod \"infra-operator-controller-manager-57548d458d-77t9s\" (UID: \"f21790d8-af47-4be7-8758-f038bda76908\") " pod="openstack-operators/infra-operator-controller-manager-57548d458d-77t9s" Dec 03 14:27:06 crc kubenswrapper[4805]: I1203 14:27:06.971579 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/f21790d8-af47-4be7-8758-f038bda76908-cert\") pod \"infra-operator-controller-manager-57548d458d-77t9s\" (UID: \"f21790d8-af47-4be7-8758-f038bda76908\") " pod="openstack-operators/infra-operator-controller-manager-57548d458d-77t9s" Dec 03 14:27:06 crc kubenswrapper[4805]: I1203 14:27:06.978590 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/infra-operator-controller-manager-57548d458d-77t9s" Dec 03 14:27:07 crc kubenswrapper[4805]: I1203 14:27:07.366245 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/b9d0830e-0f5d-4639-ab7d-92dcf07a29bc-cert\") pod \"openstack-baremetal-operator-controller-manager-64bc77cfd4x9mdd\" (UID: \"b9d0830e-0f5d-4639-ab7d-92dcf07a29bc\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-64bc77cfd4x9mdd" Dec 03 14:27:07 crc kubenswrapper[4805]: I1203 14:27:07.378235 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/b9d0830e-0f5d-4639-ab7d-92dcf07a29bc-cert\") pod \"openstack-baremetal-operator-controller-manager-64bc77cfd4x9mdd\" (UID: \"b9d0830e-0f5d-4639-ab7d-92dcf07a29bc\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-64bc77cfd4x9mdd" Dec 03 14:27:07 crc kubenswrapper[4805]: I1203 14:27:07.402277 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/infra-operator-controller-manager-57548d458d-77t9s"] Dec 03 14:27:07 crc kubenswrapper[4805]: I1203 14:27:07.416826 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-baremetal-operator-controller-manager-64bc77cfd4x9mdd" Dec 03 14:27:07 crc kubenswrapper[4805]: I1203 14:27:07.570545 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/41c4618c-1f15-4ceb-825b-546e828e399a-webhook-certs\") pod \"openstack-operator-controller-manager-cc68d789f-pkvr8\" (UID: \"41c4618c-1f15-4ceb-825b-546e828e399a\") " pod="openstack-operators/openstack-operator-controller-manager-cc68d789f-pkvr8" Dec 03 14:27:07 crc kubenswrapper[4805]: I1203 14:27:07.570617 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/41c4618c-1f15-4ceb-825b-546e828e399a-metrics-certs\") pod \"openstack-operator-controller-manager-cc68d789f-pkvr8\" (UID: \"41c4618c-1f15-4ceb-825b-546e828e399a\") " pod="openstack-operators/openstack-operator-controller-manager-cc68d789f-pkvr8" Dec 03 14:27:07 crc kubenswrapper[4805]: I1203 14:27:07.578907 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/41c4618c-1f15-4ceb-825b-546e828e399a-webhook-certs\") pod \"openstack-operator-controller-manager-cc68d789f-pkvr8\" (UID: \"41c4618c-1f15-4ceb-825b-546e828e399a\") " pod="openstack-operators/openstack-operator-controller-manager-cc68d789f-pkvr8" Dec 03 14:27:07 crc kubenswrapper[4805]: I1203 14:27:07.599048 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/41c4618c-1f15-4ceb-825b-546e828e399a-metrics-certs\") pod \"openstack-operator-controller-manager-cc68d789f-pkvr8\" (UID: \"41c4618c-1f15-4ceb-825b-546e828e399a\") " pod="openstack-operators/openstack-operator-controller-manager-cc68d789f-pkvr8" Dec 03 14:27:07 crc kubenswrapper[4805]: I1203 14:27:07.665291 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-controller-manager-cc68d789f-pkvr8" Dec 03 14:27:07 crc kubenswrapper[4805]: I1203 14:27:07.753385 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/swift-operator-controller-manager-5f8c65bbfc-qrvpw" event={"ID":"b51f4884-a980-4643-9deb-8668aead3876","Type":"ContainerStarted","Data":"772be6e57cc6cad5d816001f26aef2cb1f1b9469842b4cd5f9276432a1d98dbd"} Dec 03 14:27:07 crc kubenswrapper[4805]: I1203 14:27:07.820246 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ironic-operator-controller-manager-6c548fd776-mz4pw" event={"ID":"1ad4fa11-2f11-4a06-9944-bba8c7445e84","Type":"ContainerStarted","Data":"7b6c6c82723311e04ca6c842add3623a68c53c4d47452acd552bb24589f3bc76"} Dec 03 14:27:07 crc kubenswrapper[4805]: I1203 14:27:07.839148 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/barbican-operator-controller-manager-7d9dfd778-jc4dw" event={"ID":"cbf3813c-fb86-49d2-8fac-25487ff08601","Type":"ContainerStarted","Data":"cffbe4983bd08ae5f89a34d50f26104e8151111c599bdf770324ced1c10dbca7"} Dec 03 14:27:07 crc kubenswrapper[4805]: I1203 14:27:07.844175 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/cinder-operator-controller-manager-859b6ccc6-2nzgx" event={"ID":"f9b6b508-1cf6-40aa-829e-b9cf264921e5","Type":"ContainerStarted","Data":"d70851c1a781ad22f8dc1bd87c491dcd97b16e46647f76df2aba8fd8b5cf6edd"} Dec 03 14:27:07 crc kubenswrapper[4805]: I1203 14:27:07.848854 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/infra-operator-controller-manager-57548d458d-77t9s" event={"ID":"f21790d8-af47-4be7-8758-f038bda76908","Type":"ContainerStarted","Data":"9b3ee5bc15c39aba01714232d5c443cc17fb6b78fe64ef273d8b7a8483bdb0e6"} Dec 03 14:27:07 crc kubenswrapper[4805]: I1203 14:27:07.853105 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/heat-operator-controller-manager-5f64f6f8bb-86sbc" event={"ID":"a748cfe4-2e54-4a7b-a353-e1dfea6c14b3","Type":"ContainerStarted","Data":"694aec4f0a3ebc4cb68f943a186986826264339358fda8ad0b54c04bcc57d5b1"} Dec 03 14:27:07 crc kubenswrapper[4805]: I1203 14:27:07.979368 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-baremetal-operator-controller-manager-64bc77cfd4x9mdd"] Dec 03 14:27:08 crc kubenswrapper[4805]: I1203 14:27:08.321433 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-controller-manager-cc68d789f-pkvr8"] Dec 03 14:27:08 crc kubenswrapper[4805]: W1203 14:27:08.524991 4805 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod41c4618c_1f15_4ceb_825b_546e828e399a.slice/crio-0c3484cec543d37272968c8e94ad5b6c12121cea6ea453f4066be5be31577551 WatchSource:0}: Error finding container 0c3484cec543d37272968c8e94ad5b6c12121cea6ea453f4066be5be31577551: Status 404 returned error can't find the container with id 0c3484cec543d37272968c8e94ad5b6c12121cea6ea453f4066be5be31577551 Dec 03 14:27:08 crc kubenswrapper[4805]: W1203 14:27:08.529008 4805 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podb9d0830e_0f5d_4639_ab7d_92dcf07a29bc.slice/crio-e949cc209a645b6bdab5f2d6bb48282d1193f994431e19d3c1ed284ccdfa7718 WatchSource:0}: Error finding container e949cc209a645b6bdab5f2d6bb48282d1193f994431e19d3c1ed284ccdfa7718: Status 404 returned error can't find the container with id e949cc209a645b6bdab5f2d6bb48282d1193f994431e19d3c1ed284ccdfa7718 Dec 03 14:27:08 crc kubenswrapper[4805]: I1203 14:27:08.859410 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-manager-cc68d789f-pkvr8" event={"ID":"41c4618c-1f15-4ceb-825b-546e828e399a","Type":"ContainerStarted","Data":"0c3484cec543d37272968c8e94ad5b6c12121cea6ea453f4066be5be31577551"} Dec 03 14:27:08 crc kubenswrapper[4805]: I1203 14:27:08.860856 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-baremetal-operator-controller-manager-64bc77cfd4x9mdd" event={"ID":"b9d0830e-0f5d-4639-ab7d-92dcf07a29bc","Type":"ContainerStarted","Data":"e949cc209a645b6bdab5f2d6bb48282d1193f994431e19d3c1ed284ccdfa7718"} Dec 03 14:27:13 crc kubenswrapper[4805]: I1203 14:27:13.918001 4805 patch_prober.go:28] interesting pod/machine-config-daemon-gskh4 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 03 14:27:13 crc kubenswrapper[4805]: I1203 14:27:13.919011 4805 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-gskh4" podUID="6098937f-e3f6-45e8-a647-4994a79cd711" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 03 14:27:13 crc kubenswrapper[4805]: I1203 14:27:13.919144 4805 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-gskh4" Dec 03 14:27:13 crc kubenswrapper[4805]: I1203 14:27:13.921282 4805 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"5079b541103d89d1ef42b4f7093c19ac19ef0baacd14c63b2eb7fb16b384fba4"} pod="openshift-machine-config-operator/machine-config-daemon-gskh4" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 03 14:27:13 crc kubenswrapper[4805]: I1203 14:27:13.921514 4805 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-gskh4" podUID="6098937f-e3f6-45e8-a647-4994a79cd711" containerName="machine-config-daemon" containerID="cri-o://5079b541103d89d1ef42b4f7093c19ac19ef0baacd14c63b2eb7fb16b384fba4" gracePeriod=600 Dec 03 14:27:14 crc kubenswrapper[4805]: I1203 14:27:14.909972 4805 generic.go:334] "Generic (PLEG): container finished" podID="6098937f-e3f6-45e8-a647-4994a79cd711" containerID="5079b541103d89d1ef42b4f7093c19ac19ef0baacd14c63b2eb7fb16b384fba4" exitCode=0 Dec 03 14:27:14 crc kubenswrapper[4805]: I1203 14:27:14.910023 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-gskh4" event={"ID":"6098937f-e3f6-45e8-a647-4994a79cd711","Type":"ContainerDied","Data":"5079b541103d89d1ef42b4f7093c19ac19ef0baacd14c63b2eb7fb16b384fba4"} Dec 03 14:27:14 crc kubenswrapper[4805]: I1203 14:27:14.910064 4805 scope.go:117] "RemoveContainer" containerID="7e81e89ac6c4e0378da77d246d467c4c9c45debada7dcf3d097d3c2112a7ea0f" Dec 03 14:27:30 crc kubenswrapper[4805]: E1203 14:27:30.393392 4805 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/nova-operator@sha256:779f0cee6024d0fb8f259b036fe790e62aa5a3b0431ea9bf15a6e7d02e2e5670" Dec 03 14:27:30 crc kubenswrapper[4805]: E1203 14:27:30.394178 4805 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/nova-operator@sha256:779f0cee6024d0fb8f259b036fe790e62aa5a3b0431ea9bf15a6e7d02e2e5670,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-jjz2n,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod nova-operator-controller-manager-697bc559fc-5rl85_openstack-operators(05506111-ddb3-4ab9-93a0-ad2bedff67a4): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 03 14:27:30 crc kubenswrapper[4805]: E1203 14:27:30.787891 4805 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/infra-operator@sha256:09a6d0613ee2d3c1c809fc36c22678458ac271e0da87c970aec0a5339f5423f7" Dec 03 14:27:30 crc kubenswrapper[4805]: E1203 14:27:30.788075 4805 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/infra-operator@sha256:09a6d0613ee2d3c1c809fc36c22678458ac271e0da87c970aec0a5339f5423f7,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:true,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{600 -3} {} 600m DecimalSI},memory: {{2147483648 0} {} 2Gi BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{536870912 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:cert,ReadOnly:true,MountPath:/tmp/k8s-webhook-server/serving-certs,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-scslw,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod infra-operator-controller-manager-57548d458d-77t9s_openstack-operators(f21790d8-af47-4be7-8758-f038bda76908): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 03 14:27:31 crc kubenswrapper[4805]: E1203 14:27:31.250981 4805 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/mariadb-operator@sha256:600ca007e493d3af0fcc2ebac92e8da5efd2afe812b62d7d3d4dd0115bdf05d7" Dec 03 14:27:31 crc kubenswrapper[4805]: E1203 14:27:31.251159 4805 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/mariadb-operator@sha256:600ca007e493d3af0fcc2ebac92e8da5efd2afe812b62d7d3d4dd0115bdf05d7,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-v5zjt,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod mariadb-operator-controller-manager-56bbcc9d85-n6m2d_openstack-operators(2d72009f-b9ef-4c9d-b224-5b133a3cd93a): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 03 14:27:31 crc kubenswrapper[4805]: E1203 14:27:31.713997 4805 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/openstack-baremetal-operator@sha256:14cfad6ea2e7f7ecc4cb2aafceb9c61514b3d04b66668832d1e4ac3b19f1ab81" Dec 03 14:27:31 crc kubenswrapper[4805]: E1203 14:27:31.715228 4805 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/openstack-baremetal-operator@sha256:14cfad6ea2e7f7ecc4cb2aafceb9c61514b3d04b66668832d1e4ac3b19f1ab81,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:true,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_AGENT_IMAGE_URL_DEFAULT,Value:quay.io/openstack-k8s-operators/openstack-baremetal-operator-agent:latest,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_ANSIBLEEE_IMAGE_URL_DEFAULT,Value:quay.io/openstack-k8s-operators/openstack-ansibleee-runner:latest,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_AODH_API_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-aodh-api:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_AODH_EVALUATOR_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-aodh-evaluator:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_AODH_LISTENER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-aodh-listener:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_AODH_NOTIFIER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-aodh-notifier:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_APACHE_IMAGE_URL_DEFAULT,Value:registry.redhat.io/ubi9/httpd-24:latest,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_BARBICAN_API_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-barbican-api:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_BARBICAN_KEYSTONE_LISTENER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-barbican-keystone-listener:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_BARBICAN_WORKER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-barbican-worker:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_CEILOMETER_CENTRAL_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-ceilometer-central:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_CEILOMETER_COMPUTE_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-ceilometer-compute:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_CEILOMETER_IPMI_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-ceilometer-ipmi:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_CEILOMETER_MYSQLD_EXPORTER_IMAGE_URL_DEFAULT,Value:quay.io/prometheus/mysqld-exporter:v0.15.1,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_CEILOMETER_NOTIFICATION_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-ceilometer-notification:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_CEILOMETER_SGCORE_IMAGE_URL_DEFAULT,Value:quay.io/openstack-k8s-operators/sg-core:latest,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_CINDER_API_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-cinder-api:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_CINDER_BACKUP_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-cinder-backup:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_CINDER_SCHEDULER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-cinder-scheduler:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_CINDER_VOLUME_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-cinder-volume:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_CLOUDKITTY_API_IMAGE_URL_DEFAULT,Value:quay.rdoproject.org/podified-master-centos10/openstack-cloudkitty-api:current,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_CLOUDKITTY_PROC_IMAGE_URL_DEFAULT,Value:quay.rdoproject.org/podified-master-centos10/openstack-cloudkitty-processor:current,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_DESIGNATE_API_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-designate-api:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_DESIGNATE_BACKENDBIND9_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-designate-backend-bind9:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_DESIGNATE_CENTRAL_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-designate-central:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_DESIGNATE_MDNS_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-designate-mdns:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_DESIGNATE_PRODUCER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-designate-producer:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_DESIGNATE_UNBOUND_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-unbound:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_DESIGNATE_WORKER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-designate-worker:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_EDPM_FRR_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-frr:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_EDPM_ISCSID_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-iscsid:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_EDPM_KEPLER_IMAGE_URL_DEFAULT,Value:quay.io/sustainable_computing_io/kepler:release-0.7.12,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_EDPM_LOGROTATE_CROND_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-cron:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_EDPM_MULTIPATHD_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-multipathd:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_EDPM_NEUTRON_DHCP_AGENT_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-neutron-dhcp-agent:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_EDPM_NEUTRON_METADATA_AGENT_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-neutron-metadata-agent-ovn:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_EDPM_NEUTRON_OVN_AGENT_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-neutron-ovn-agent:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_EDPM_NEUTRON_SRIOV_AGENT_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-neutron-sriov-agent:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_EDPM_NODE_EXPORTER_IMAGE_URL_DEFAULT,Value:quay.io/prometheus/node-exporter:v1.5.0,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_EDPM_OVN_BGP_AGENT_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-ovn-bgp-agent:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_EDPM_PODMAN_EXPORTER_IMAGE_URL_DEFAULT,Value:quay.io/navidys/prometheus-podman-exporter:v1.10.1,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_GLANCE_API_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-glance-api:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_HEAT_API_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-heat-api:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_HEAT_CFNAPI_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-heat-api-cfn:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_HEAT_ENGINE_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-heat-engine:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_HORIZON_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-horizon:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_INFRA_MEMCACHED_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-memcached:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_INFRA_REDIS_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-redis:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_IRONIC_API_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-ironic-api:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_IRONIC_CONDUCTOR_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-ironic-conductor:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_IRONIC_INSPECTOR_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-ironic-inspector:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_IRONIC_NEUTRON_AGENT_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-ironic-neutron-agent:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_IRONIC_PXE_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-ironic-pxe:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_IRONIC_PYTHON_AGENT_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/ironic-python-agent:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_KEYSTONE_API_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-keystone:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_KSM_IMAGE_URL_DEFAULT,Value:registry.k8s.io/kube-state-metrics/kube-state-metrics:v2.15.0,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_MANILA_API_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-manila-api:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_MANILA_SCHEDULER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-manila-scheduler:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_MANILA_SHARE_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-manila-share:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_MARIADB_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-mariadb:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_NET_UTILS_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-netutils:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_NEUTRON_API_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_NOVA_API_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-nova-api:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_NOVA_COMPUTE_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-nova-compute:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_NOVA_CONDUCTOR_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-nova-conductor:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_NOVA_NOVNC_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-nova-novncproxy:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_NOVA_SCHEDULER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-nova-scheduler:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_OCTAVIA_API_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-octavia-api:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_OCTAVIA_HEALTHMANAGER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-octavia-health-manager:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_OCTAVIA_HOUSEKEEPING_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-octavia-housekeeping:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_OCTAVIA_RSYSLOG_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-rsyslog:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_OCTAVIA_WORKER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-octavia-worker:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_OPENSTACK_CLIENT_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-openstackclient:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_OPENSTACK_MUST_GATHER_DEFAULT,Value:quay.io/openstack-k8s-operators/openstack-must-gather:latest,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_OPENSTACK_NETWORK_EXPORTER_IMAGE_URL_DEFAULT,Value:quay.io/openstack-k8s-operators/openstack-network-exporter:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_OS_CONTAINER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/edpm-hardened-uefi:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_OVN_CONTROLLER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-ovn-controller:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_OVN_CONTROLLER_OVS_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-ovn-base:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_OVN_NB_DBCLUSTER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-ovn-nb-db-server:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_OVN_NORTHD_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-ovn-northd:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_OVN_SB_DBCLUSTER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-ovn-sb-db-server:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_PLACEMENT_API_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-placement-api:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_RABBITMQ_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-rabbitmq:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_SWIFT_ACCOUNT_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-swift-account:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_SWIFT_CONTAINER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-swift-container:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_SWIFT_OBJECT_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-swift-object:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_SWIFT_PROXY_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-swift-proxy-server:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_TEST_TEMPEST_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-tempest-all:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_WATCHER_API_IMAGE_URL_DEFAULT,Value:quay.io/podified-master-centos9/openstack-watcher-api:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_WATCHER_APPLIER_IMAGE_URL_DEFAULT,Value:quay.io/podified-master-centos9/openstack-watcher-applier:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_WATCHER_DECISION_ENGINE_IMAGE_URL_DEFAULT,Value:quay.io/podified-master-centos9/openstack-watcher-decision-engine:current-podified,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:cert,ReadOnly:true,MountPath:/tmp/k8s-webhook-server/serving-certs,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-5g6hd,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000660000,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod openstack-baremetal-operator-controller-manager-64bc77cfd4x9mdd_openstack-operators(b9d0830e-0f5d-4639-ab7d-92dcf07a29bc): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 03 14:27:32 crc kubenswrapper[4805]: E1203 14:27:32.105945 4805 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/ovn-operator@sha256:635a4aef9d6f0b799e8ec91333dbb312160c001d05b3c63f614c124e0b67cb59" Dec 03 14:27:32 crc kubenswrapper[4805]: E1203 14:27:32.106188 4805 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/ovn-operator@sha256:635a4aef9d6f0b799e8ec91333dbb312160c001d05b3c63f614c124e0b67cb59,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-xvkjw,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod ovn-operator-controller-manager-b6456fdb6-djzcm_openstack-operators(42b37265-f5bd-4ac4-9b96-d758fbe1b31e): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 03 14:27:32 crc kubenswrapper[4805]: E1203 14:27:32.152319 4805 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying layer: context canceled" image="quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0" Dec 03 14:27:32 crc kubenswrapper[4805]: E1203 14:27:32.152540 4805 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:kube-rbac-proxy,Image:quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0,Command:[],Args:[--secure-listen-address=0.0.0.0:8443 --upstream=http://127.0.0.1:8080/ --logtostderr=true --v=0],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:https,HostPort:0,ContainerPort:8443,Protocol:TCP,HostIP:,},},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{134217728 0} {} BinarySI},},Requests:ResourceList{cpu: {{5 -3} {} 5m DecimalSI},memory: {{67108864 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-89ppm,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod cinder-operator-controller-manager-859b6ccc6-2nzgx_openstack-operators(f9b6b508-1cf6-40aa-829e-b9cf264921e5): ErrImagePull: rpc error: code = Canceled desc = copying layer: context canceled" logger="UnhandledError" Dec 03 14:27:32 crc kubenswrapper[4805]: E1203 14:27:32.154366 4805 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-rbac-proxy\" with ErrImagePull: \"rpc error: code = Canceled desc = copying layer: context canceled\"" pod="openstack-operators/cinder-operator-controller-manager-859b6ccc6-2nzgx" podUID="f9b6b508-1cf6-40aa-829e-b9cf264921e5" Dec 03 14:27:32 crc kubenswrapper[4805]: E1203 14:27:32.247040 4805 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = reading blob sha256:46ba3f23f1d3fb1440deeb279716e4377e79e61736ec2227270349b9618a0fdd: Get \"https://quay.io/v2/openstack-k8s-operators/kube-rbac-proxy/blobs/sha256:46ba3f23f1d3fb1440deeb279716e4377e79e61736ec2227270349b9618a0fdd\": context canceled" image="quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0" Dec 03 14:27:32 crc kubenswrapper[4805]: E1203 14:27:32.247184 4805 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:kube-rbac-proxy,Image:quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0,Command:[],Args:[--secure-listen-address=0.0.0.0:8443 --upstream=http://127.0.0.1:8080/ --logtostderr=true --v=0],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:https,HostPort:0,ContainerPort:8443,Protocol:TCP,HostIP:,},},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{134217728 0} {} BinarySI},},Requests:ResourceList{cpu: {{5 -3} {} 5m DecimalSI},memory: {{67108864 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-s9bwv,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod barbican-operator-controller-manager-7d9dfd778-jc4dw_openstack-operators(cbf3813c-fb86-49d2-8fac-25487ff08601): ErrImagePull: rpc error: code = Canceled desc = reading blob sha256:46ba3f23f1d3fb1440deeb279716e4377e79e61736ec2227270349b9618a0fdd: Get \"https://quay.io/v2/openstack-k8s-operators/kube-rbac-proxy/blobs/sha256:46ba3f23f1d3fb1440deeb279716e4377e79e61736ec2227270349b9618a0fdd\": context canceled" logger="UnhandledError" Dec 03 14:27:32 crc kubenswrapper[4805]: E1203 14:27:32.248505 4805 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-rbac-proxy\" with ErrImagePull: \"rpc error: code = Canceled desc = reading blob sha256:46ba3f23f1d3fb1440deeb279716e4377e79e61736ec2227270349b9618a0fdd: Get \\\"https://quay.io/v2/openstack-k8s-operators/kube-rbac-proxy/blobs/sha256:46ba3f23f1d3fb1440deeb279716e4377e79e61736ec2227270349b9618a0fdd\\\": context canceled\"" pod="openstack-operators/barbican-operator-controller-manager-7d9dfd778-jc4dw" podUID="cbf3813c-fb86-49d2-8fac-25487ff08601" Dec 03 14:27:32 crc kubenswrapper[4805]: E1203 14:27:32.263080 4805 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = reading blob sha256:46ba3f23f1d3fb1440deeb279716e4377e79e61736ec2227270349b9618a0fdd: Get \"https://quay.io/v2/openstack-k8s-operators/kube-rbac-proxy/blobs/sha256:46ba3f23f1d3fb1440deeb279716e4377e79e61736ec2227270349b9618a0fdd\": context canceled" image="quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0" Dec 03 14:27:32 crc kubenswrapper[4805]: E1203 14:27:32.263101 4805 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = reading blob sha256:46ba3f23f1d3fb1440deeb279716e4377e79e61736ec2227270349b9618a0fdd: Get \"https://quay.io/v2/openstack-k8s-operators/kube-rbac-proxy/blobs/sha256:46ba3f23f1d3fb1440deeb279716e4377e79e61736ec2227270349b9618a0fdd\": context canceled" image="quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0" Dec 03 14:27:32 crc kubenswrapper[4805]: E1203 14:27:32.263309 4805 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:kube-rbac-proxy,Image:quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0,Command:[],Args:[--secure-listen-address=0.0.0.0:8443 --upstream=http://127.0.0.1:8080/ --logtostderr=true --v=0],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:https,HostPort:0,ContainerPort:8443,Protocol:TCP,HostIP:,},},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{134217728 0} {} BinarySI},},Requests:ResourceList{cpu: {{5 -3} {} 5m DecimalSI},memory: {{67108864 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-462zf,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod heat-operator-controller-manager-5f64f6f8bb-86sbc_openstack-operators(a748cfe4-2e54-4a7b-a353-e1dfea6c14b3): ErrImagePull: rpc error: code = Canceled desc = reading blob sha256:46ba3f23f1d3fb1440deeb279716e4377e79e61736ec2227270349b9618a0fdd: Get \"https://quay.io/v2/openstack-k8s-operators/kube-rbac-proxy/blobs/sha256:46ba3f23f1d3fb1440deeb279716e4377e79e61736ec2227270349b9618a0fdd\": context canceled" logger="UnhandledError" Dec 03 14:27:32 crc kubenswrapper[4805]: E1203 14:27:32.263303 4805 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:kube-rbac-proxy,Image:quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0,Command:[],Args:[--secure-listen-address=0.0.0.0:8443 --upstream=http://127.0.0.1:8080/ --logtostderr=true --v=0],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:https,HostPort:0,ContainerPort:8443,Protocol:TCP,HostIP:,},},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{134217728 0} {} BinarySI},},Requests:ResourceList{cpu: {{5 -3} {} 5m DecimalSI},memory: {{67108864 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-sfkq4,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod manila-operator-controller-manager-7c79b5df47-d2thc_openstack-operators(967f26e7-3027-4e23-9f02-d111d714a396): ErrImagePull: rpc error: code = Canceled desc = reading blob sha256:46ba3f23f1d3fb1440deeb279716e4377e79e61736ec2227270349b9618a0fdd: Get \"https://quay.io/v2/openstack-k8s-operators/kube-rbac-proxy/blobs/sha256:46ba3f23f1d3fb1440deeb279716e4377e79e61736ec2227270349b9618a0fdd\": context canceled" logger="UnhandledError" Dec 03 14:27:32 crc kubenswrapper[4805]: E1203 14:27:32.264480 4805 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-rbac-proxy\" with ErrImagePull: \"rpc error: code = Canceled desc = reading blob sha256:46ba3f23f1d3fb1440deeb279716e4377e79e61736ec2227270349b9618a0fdd: Get \\\"https://quay.io/v2/openstack-k8s-operators/kube-rbac-proxy/blobs/sha256:46ba3f23f1d3fb1440deeb279716e4377e79e61736ec2227270349b9618a0fdd\\\": context canceled\"" pod="openstack-operators/manila-operator-controller-manager-7c79b5df47-d2thc" podUID="967f26e7-3027-4e23-9f02-d111d714a396" Dec 03 14:27:32 crc kubenswrapper[4805]: E1203 14:27:32.264551 4805 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-rbac-proxy\" with ErrImagePull: \"rpc error: code = Canceled desc = reading blob sha256:46ba3f23f1d3fb1440deeb279716e4377e79e61736ec2227270349b9618a0fdd: Get \\\"https://quay.io/v2/openstack-k8s-operators/kube-rbac-proxy/blobs/sha256:46ba3f23f1d3fb1440deeb279716e4377e79e61736ec2227270349b9618a0fdd\\\": context canceled\"" pod="openstack-operators/heat-operator-controller-manager-5f64f6f8bb-86sbc" podUID="a748cfe4-2e54-4a7b-a353-e1dfea6c14b3" Dec 03 14:27:32 crc kubenswrapper[4805]: E1203 14:27:32.267536 4805 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = reading blob sha256:46ba3f23f1d3fb1440deeb279716e4377e79e61736ec2227270349b9618a0fdd: Get \"https://quay.io/v2/openstack-k8s-operators/kube-rbac-proxy/blobs/sha256:46ba3f23f1d3fb1440deeb279716e4377e79e61736ec2227270349b9618a0fdd\": context canceled" image="quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0" Dec 03 14:27:32 crc kubenswrapper[4805]: E1203 14:27:32.267701 4805 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:kube-rbac-proxy,Image:quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0,Command:[],Args:[--secure-listen-address=0.0.0.0:8443 --upstream=http://127.0.0.1:8080/ --logtostderr=true --v=0],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:https,HostPort:0,ContainerPort:8443,Protocol:TCP,HostIP:,},},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{134217728 0} {} BinarySI},},Requests:ResourceList{cpu: {{5 -3} {} 5m DecimalSI},memory: {{67108864 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-4dpg9,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod designate-operator-controller-manager-78b4bc895b-w56sr_openstack-operators(326305da-3798-498c-9b47-0aeb71bd2205): ErrImagePull: rpc error: code = Canceled desc = reading blob sha256:46ba3f23f1d3fb1440deeb279716e4377e79e61736ec2227270349b9618a0fdd: Get \"https://quay.io/v2/openstack-k8s-operators/kube-rbac-proxy/blobs/sha256:46ba3f23f1d3fb1440deeb279716e4377e79e61736ec2227270349b9618a0fdd\": context canceled" logger="UnhandledError" Dec 03 14:27:32 crc kubenswrapper[4805]: E1203 14:27:32.269246 4805 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-rbac-proxy\" with ErrImagePull: \"rpc error: code = Canceled desc = reading blob sha256:46ba3f23f1d3fb1440deeb279716e4377e79e61736ec2227270349b9618a0fdd: Get \\\"https://quay.io/v2/openstack-k8s-operators/kube-rbac-proxy/blobs/sha256:46ba3f23f1d3fb1440deeb279716e4377e79e61736ec2227270349b9618a0fdd\\\": context canceled\"" pod="openstack-operators/designate-operator-controller-manager-78b4bc895b-w56sr" podUID="326305da-3798-498c-9b47-0aeb71bd2205" Dec 03 14:27:32 crc kubenswrapper[4805]: E1203 14:27:32.324304 4805 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = reading blob sha256:46ba3f23f1d3fb1440deeb279716e4377e79e61736ec2227270349b9618a0fdd: Get \"https://quay.io/v2/openstack-k8s-operators/kube-rbac-proxy/blobs/sha256:46ba3f23f1d3fb1440deeb279716e4377e79e61736ec2227270349b9618a0fdd\": context canceled" image="quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0" Dec 03 14:27:32 crc kubenswrapper[4805]: E1203 14:27:32.324506 4805 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying layer: context canceled" image="quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0" Dec 03 14:27:32 crc kubenswrapper[4805]: E1203 14:27:32.324556 4805 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:kube-rbac-proxy,Image:quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0,Command:[],Args:[--secure-listen-address=0.0.0.0:8443 --upstream=http://127.0.0.1:8080/ --logtostderr=true --v=0],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:https,HostPort:0,ContainerPort:8443,Protocol:TCP,HostIP:,},},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{134217728 0} {} BinarySI},},Requests:ResourceList{cpu: {{5 -3} {} 5m DecimalSI},memory: {{67108864 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-pxqph,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod horizon-operator-controller-manager-68c6d99b8f-kxbr8_openstack-operators(b29e4120-0cbc-4de5-a4b6-97b97761f49a): ErrImagePull: rpc error: code = Canceled desc = reading blob sha256:46ba3f23f1d3fb1440deeb279716e4377e79e61736ec2227270349b9618a0fdd: Get \"https://quay.io/v2/openstack-k8s-operators/kube-rbac-proxy/blobs/sha256:46ba3f23f1d3fb1440deeb279716e4377e79e61736ec2227270349b9618a0fdd\": context canceled" logger="UnhandledError" Dec 03 14:27:32 crc kubenswrapper[4805]: E1203 14:27:32.324686 4805 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying layer: context canceled" image="quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0" Dec 03 14:27:32 crc kubenswrapper[4805]: E1203 14:27:32.324926 4805 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:kube-rbac-proxy,Image:quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0,Command:[],Args:[--secure-listen-address=0.0.0.0:8443 --upstream=http://127.0.0.1:8080/ --logtostderr=true --v=0],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:https,HostPort:0,ContainerPort:8443,Protocol:TCP,HostIP:,},},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{134217728 0} {} BinarySI},},Requests:ResourceList{cpu: {{5 -3} {} 5m DecimalSI},memory: {{67108864 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-r68wv,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod placement-operator-controller-manager-78f8948974-b8l4l_openstack-operators(f853eae6-829b-42f9-9198-3d7dcabd405f): ErrImagePull: rpc error: code = Canceled desc = copying layer: context canceled" logger="UnhandledError" Dec 03 14:27:32 crc kubenswrapper[4805]: E1203 14:27:32.325054 4805 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:kube-rbac-proxy,Image:quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0,Command:[],Args:[--secure-listen-address=0.0.0.0:8443 --upstream=http://127.0.0.1:8080/ --logtostderr=true --v=0],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:https,HostPort:0,ContainerPort:8443,Protocol:TCP,HostIP:,},},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{134217728 0} {} BinarySI},},Requests:ResourceList{cpu: {{5 -3} {} 5m DecimalSI},memory: {{67108864 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-f9wlh,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod glance-operator-controller-manager-77987cd8cd-pt9fw_openstack-operators(5615ce63-1ec1-4a77-8e44-49e6d114eec6): ErrImagePull: rpc error: code = Canceled desc = copying layer: context canceled" logger="UnhandledError" Dec 03 14:27:32 crc kubenswrapper[4805]: E1203 14:27:32.326089 4805 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-rbac-proxy\" with ErrImagePull: \"rpc error: code = Canceled desc = copying layer: context canceled\"" pod="openstack-operators/placement-operator-controller-manager-78f8948974-b8l4l" podUID="f853eae6-829b-42f9-9198-3d7dcabd405f" Dec 03 14:27:32 crc kubenswrapper[4805]: E1203 14:27:32.326156 4805 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-rbac-proxy\" with ErrImagePull: \"rpc error: code = Canceled desc = reading blob sha256:46ba3f23f1d3fb1440deeb279716e4377e79e61736ec2227270349b9618a0fdd: Get \\\"https://quay.io/v2/openstack-k8s-operators/kube-rbac-proxy/blobs/sha256:46ba3f23f1d3fb1440deeb279716e4377e79e61736ec2227270349b9618a0fdd\\\": context canceled\"" pod="openstack-operators/horizon-operator-controller-manager-68c6d99b8f-kxbr8" podUID="b29e4120-0cbc-4de5-a4b6-97b97761f49a" Dec 03 14:27:32 crc kubenswrapper[4805]: E1203 14:27:32.326174 4805 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-rbac-proxy\" with ErrImagePull: \"rpc error: code = Canceled desc = copying layer: context canceled\"" pod="openstack-operators/glance-operator-controller-manager-77987cd8cd-pt9fw" podUID="5615ce63-1ec1-4a77-8e44-49e6d114eec6" Dec 03 14:27:32 crc kubenswrapper[4805]: E1203 14:27:32.340645 4805 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = reading blob sha256:46ba3f23f1d3fb1440deeb279716e4377e79e61736ec2227270349b9618a0fdd: Get \"https://quay.io/v2/openstack-k8s-operators/kube-rbac-proxy/blobs/sha256:46ba3f23f1d3fb1440deeb279716e4377e79e61736ec2227270349b9618a0fdd\": context canceled" image="quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0" Dec 03 14:27:32 crc kubenswrapper[4805]: E1203 14:27:32.341047 4805 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:kube-rbac-proxy,Image:quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0,Command:[],Args:[--secure-listen-address=0.0.0.0:8443 --upstream=http://127.0.0.1:8080/ --logtostderr=true --v=0],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:https,HostPort:0,ContainerPort:8443,Protocol:TCP,HostIP:,},},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{134217728 0} {} BinarySI},},Requests:ResourceList{cpu: {{5 -3} {} 5m DecimalSI},memory: {{67108864 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-h74hm,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod watcher-operator-controller-manager-769dc69bc-fl6mm_openstack-operators(30776d04-200b-4c10-b48d-f51a926e469b): ErrImagePull: rpc error: code = Canceled desc = reading blob sha256:46ba3f23f1d3fb1440deeb279716e4377e79e61736ec2227270349b9618a0fdd: Get \"https://quay.io/v2/openstack-k8s-operators/kube-rbac-proxy/blobs/sha256:46ba3f23f1d3fb1440deeb279716e4377e79e61736ec2227270349b9618a0fdd\": context canceled" logger="UnhandledError" Dec 03 14:27:32 crc kubenswrapper[4805]: E1203 14:27:32.342477 4805 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-rbac-proxy\" with ErrImagePull: \"rpc error: code = Canceled desc = reading blob sha256:46ba3f23f1d3fb1440deeb279716e4377e79e61736ec2227270349b9618a0fdd: Get \\\"https://quay.io/v2/openstack-k8s-operators/kube-rbac-proxy/blobs/sha256:46ba3f23f1d3fb1440deeb279716e4377e79e61736ec2227270349b9618a0fdd\\\": context canceled\"" pod="openstack-operators/watcher-operator-controller-manager-769dc69bc-fl6mm" podUID="30776d04-200b-4c10-b48d-f51a926e469b" Dec 03 14:27:32 crc kubenswrapper[4805]: E1203 14:27:32.693895 4805 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying layer: context canceled" image="quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0" Dec 03 14:27:32 crc kubenswrapper[4805]: E1203 14:27:32.694334 4805 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:kube-rbac-proxy,Image:quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0,Command:[],Args:[--secure-listen-address=0.0.0.0:8443 --upstream=http://127.0.0.1:8080/ --logtostderr=true --v=0],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:https,HostPort:0,ContainerPort:8443,Protocol:TCP,HostIP:,},},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{134217728 0} {} BinarySI},},Requests:ResourceList{cpu: {{5 -3} {} 5m DecimalSI},memory: {{67108864 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-6mpcm,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000660000,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod swift-operator-controller-manager-5f8c65bbfc-qrvpw_openstack-operators(b51f4884-a980-4643-9deb-8668aead3876): ErrImagePull: rpc error: code = Canceled desc = copying layer: context canceled" logger="UnhandledError" Dec 03 14:27:32 crc kubenswrapper[4805]: E1203 14:27:32.695306 4805 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying layer: context canceled" image="quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0" Dec 03 14:27:32 crc kubenswrapper[4805]: E1203 14:27:32.695455 4805 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-rbac-proxy\" with ErrImagePull: \"rpc error: code = Canceled desc = copying layer: context canceled\"" pod="openstack-operators/swift-operator-controller-manager-5f8c65bbfc-qrvpw" podUID="b51f4884-a980-4643-9deb-8668aead3876" Dec 03 14:27:32 crc kubenswrapper[4805]: E1203 14:27:32.695464 4805 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:kube-rbac-proxy,Image:quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0,Command:[],Args:[--secure-listen-address=0.0.0.0:8443 --upstream=http://127.0.0.1:8080/ --logtostderr=true --v=0],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:https,HostPort:0,ContainerPort:8443,Protocol:TCP,HostIP:,},},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{134217728 0} {} BinarySI},},Requests:ResourceList{cpu: {{5 -3} {} 5m DecimalSI},memory: {{67108864 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-fnmcc,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod ironic-operator-controller-manager-6c548fd776-mz4pw_openstack-operators(1ad4fa11-2f11-4a06-9944-bba8c7445e84): ErrImagePull: rpc error: code = Canceled desc = copying layer: context canceled" logger="UnhandledError" Dec 03 14:27:32 crc kubenswrapper[4805]: E1203 14:27:32.695926 4805 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying layer: context canceled" image="quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0" Dec 03 14:27:32 crc kubenswrapper[4805]: E1203 14:27:32.696026 4805 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:kube-rbac-proxy,Image:quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0,Command:[],Args:[--secure-listen-address=0.0.0.0:8443 --upstream=http://127.0.0.1:8080/ --logtostderr=true --v=0],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:https,HostPort:0,ContainerPort:8443,Protocol:TCP,HostIP:,},},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{134217728 0} {} BinarySI},},Requests:ResourceList{cpu: {{5 -3} {} 5m DecimalSI},memory: {{67108864 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-z9t5w,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod octavia-operator-controller-manager-998648c74-8mbdb_openstack-operators(0f040ce3-0a83-426b-8b9f-b63183d16b4a): ErrImagePull: rpc error: code = Canceled desc = copying layer: context canceled" logger="UnhandledError" Dec 03 14:27:32 crc kubenswrapper[4805]: E1203 14:27:32.696567 4805 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-rbac-proxy\" with ErrImagePull: \"rpc error: code = Canceled desc = copying layer: context canceled\"" pod="openstack-operators/ironic-operator-controller-manager-6c548fd776-mz4pw" podUID="1ad4fa11-2f11-4a06-9944-bba8c7445e84" Dec 03 14:27:32 crc kubenswrapper[4805]: E1203 14:27:32.697230 4805 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-rbac-proxy\" with ErrImagePull: \"rpc error: code = Canceled desc = copying layer: context canceled\"" pod="openstack-operators/octavia-operator-controller-manager-998648c74-8mbdb" podUID="0f040ce3-0a83-426b-8b9f-b63183d16b4a" Dec 03 14:27:32 crc kubenswrapper[4805]: E1203 14:27:32.719555 4805 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0" Dec 03 14:27:32 crc kubenswrapper[4805]: E1203 14:27:32.719715 4805 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:kube-rbac-proxy,Image:quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0,Command:[],Args:[--secure-listen-address=0.0.0.0:8443 --upstream=http://127.0.0.1:8080/ --logtostderr=true --v=0],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:https,HostPort:0,ContainerPort:8443,Protocol:TCP,HostIP:,},},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{134217728 0} {} BinarySI},},Requests:ResourceList{cpu: {{5 -3} {} 5m DecimalSI},memory: {{67108864 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-nxwjv,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod telemetry-operator-controller-manager-cc9d8f87b-qfmmx_openstack-operators(8cbfcd07-e136-4bfd-b3c6-6615c911d440): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 03 14:27:32 crc kubenswrapper[4805]: E1203 14:27:32.721155 4805 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"]" pod="openstack-operators/telemetry-operator-controller-manager-cc9d8f87b-qfmmx" podUID="8cbfcd07-e136-4bfd-b3c6-6615c911d440" Dec 03 14:27:32 crc kubenswrapper[4805]: E1203 14:27:32.725822 4805 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0" Dec 03 14:27:32 crc kubenswrapper[4805]: E1203 14:27:32.726015 4805 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:kube-rbac-proxy,Image:quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0,Command:[],Args:[--secure-listen-address=0.0.0.0:8443 --upstream=http://127.0.0.1:8080/ --logtostderr=true --v=0],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:https,HostPort:0,ContainerPort:8443,Protocol:TCP,HostIP:,},},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{134217728 0} {} BinarySI},},Requests:ResourceList{cpu: {{5 -3} {} 5m DecimalSI},memory: {{67108864 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-6s4x4,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod neutron-operator-controller-manager-5fdfd5b6b5-4mlhk_openstack-operators(58aa635e-157c-4ae2-ad0c-568c144638d1): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 03 14:27:32 crc kubenswrapper[4805]: E1203 14:27:32.727188 4805 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-rbac-proxy\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/neutron-operator-controller-manager-5fdfd5b6b5-4mlhk" podUID="58aa635e-157c-4ae2-ad0c-568c144638d1" Dec 03 14:27:33 crc kubenswrapper[4805]: I1203 14:27:33.039829 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-manager-cc68d789f-pkvr8" event={"ID":"41c4618c-1f15-4ceb-825b-546e828e399a","Type":"ContainerStarted","Data":"89f4c508c4cbf47c468be90e0854d7d4d3bcd2cb3700eae10c4dd03b926cf4e8"} Dec 03 14:27:33 crc kubenswrapper[4805]: I1203 14:27:33.040104 4805 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/openstack-operator-controller-manager-cc68d789f-pkvr8" Dec 03 14:27:33 crc kubenswrapper[4805]: I1203 14:27:33.043035 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/keystone-operator-controller-manager-7765d96ddf-qzzz9" event={"ID":"ffa72db0-588a-4a78-a1cd-8710f5ce4f41","Type":"ContainerStarted","Data":"f8c5921e4906e475994f919c7dbeef8e88d368d1ed148ad84e401e1a851a8820"} Dec 03 14:27:33 crc kubenswrapper[4805]: I1203 14:27:33.054602 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/test-operator-controller-manager-5854674fcc-d6pvv" event={"ID":"ac4ba0f5-635d-4a72-a133-9b4aaf88f097","Type":"ContainerStarted","Data":"6d8d17b3d09d1dc2d04758bf26d3e9061f991fa99f7d96bcc17bee8d27bb2357"} Dec 03 14:27:33 crc kubenswrapper[4805]: I1203 14:27:33.057122 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-gskh4" event={"ID":"6098937f-e3f6-45e8-a647-4994a79cd711","Type":"ContainerStarted","Data":"b84d108663e877514e58fb0f23591399246f6b24b642394275299320cb0e52d9"} Dec 03 14:27:33 crc kubenswrapper[4805]: I1203 14:27:33.067064 4805 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-operator-controller-manager-cc68d789f-pkvr8" podStartSLOduration=42.067048027 podStartE2EDuration="42.067048027s" podCreationTimestamp="2025-12-03 14:26:51 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 14:27:33.066163713 +0000 UTC m=+1082.729080636" watchObservedRunningTime="2025-12-03 14:27:33.067048027 +0000 UTC m=+1082.729964950" Dec 03 14:27:33 crc kubenswrapper[4805]: E1203 14:27:33.249613 4805 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/mariadb-operator-controller-manager-56bbcc9d85-n6m2d" podUID="2d72009f-b9ef-4c9d-b224-5b133a3cd93a" Dec 03 14:27:33 crc kubenswrapper[4805]: E1203 14:27:33.249885 4805 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/ovn-operator-controller-manager-b6456fdb6-djzcm" podUID="42b37265-f5bd-4ac4-9b96-d758fbe1b31e" Dec 03 14:27:33 crc kubenswrapper[4805]: E1203 14:27:33.259240 4805 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/openstack-baremetal-operator-controller-manager-64bc77cfd4x9mdd" podUID="b9d0830e-0f5d-4639-ab7d-92dcf07a29bc" Dec 03 14:27:33 crc kubenswrapper[4805]: E1203 14:27:33.276566 4805 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/nova-operator-controller-manager-697bc559fc-5rl85" podUID="05506111-ddb3-4ab9-93a0-ad2bedff67a4" Dec 03 14:27:33 crc kubenswrapper[4805]: E1203 14:27:33.300083 4805 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/infra-operator-controller-manager-57548d458d-77t9s" podUID="f21790d8-af47-4be7-8758-f038bda76908" Dec 03 14:27:34 crc kubenswrapper[4805]: I1203 14:27:34.064628 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-5q6j2" event={"ID":"0aa860af-81d3-44ba-9020-0ffef21cefb2","Type":"ContainerStarted","Data":"be774ba31d7aa015a77e419c9e995db7a8690c07025aacb9dad7657c5742e990"} Dec 03 14:27:34 crc kubenswrapper[4805]: I1203 14:27:34.066313 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/test-operator-controller-manager-5854674fcc-d6pvv" event={"ID":"ac4ba0f5-635d-4a72-a133-9b4aaf88f097","Type":"ContainerStarted","Data":"322b3c36c5622e8b4d347c68dd7c5eb07fa11875dd44d494e5bf0944b85ea855"} Dec 03 14:27:34 crc kubenswrapper[4805]: I1203 14:27:34.066446 4805 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/test-operator-controller-manager-5854674fcc-d6pvv" Dec 03 14:27:34 crc kubenswrapper[4805]: I1203 14:27:34.068059 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/designate-operator-controller-manager-78b4bc895b-w56sr" event={"ID":"326305da-3798-498c-9b47-0aeb71bd2205","Type":"ContainerStarted","Data":"1eb1eba4645935d0068c14c3362348538d514c331bb528d9373583484def91df"} Dec 03 14:27:34 crc kubenswrapper[4805]: I1203 14:27:34.068757 4805 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/designate-operator-controller-manager-78b4bc895b-w56sr" Dec 03 14:27:34 crc kubenswrapper[4805]: I1203 14:27:34.070783 4805 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/designate-operator-controller-manager-78b4bc895b-w56sr" Dec 03 14:27:34 crc kubenswrapper[4805]: I1203 14:27:34.071577 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/octavia-operator-controller-manager-998648c74-8mbdb" event={"ID":"0f040ce3-0a83-426b-8b9f-b63183d16b4a","Type":"ContainerStarted","Data":"0fb2f49c19468c2ab1595a619fcb669285102ae7e30d4a4e36a90375c99f3d1f"} Dec 03 14:27:34 crc kubenswrapper[4805]: I1203 14:27:34.072262 4805 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/octavia-operator-controller-manager-998648c74-8mbdb" Dec 03 14:27:34 crc kubenswrapper[4805]: I1203 14:27:34.073618 4805 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/octavia-operator-controller-manager-998648c74-8mbdb" Dec 03 14:27:34 crc kubenswrapper[4805]: I1203 14:27:34.073849 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/barbican-operator-controller-manager-7d9dfd778-jc4dw" event={"ID":"cbf3813c-fb86-49d2-8fac-25487ff08601","Type":"ContainerStarted","Data":"d3752821cbd89ddbef4b57016cba130199be2230e9b1a0694c6281faca8f5aa4"} Dec 03 14:27:34 crc kubenswrapper[4805]: I1203 14:27:34.074400 4805 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/barbican-operator-controller-manager-7d9dfd778-jc4dw" Dec 03 14:27:34 crc kubenswrapper[4805]: I1203 14:27:34.075546 4805 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/barbican-operator-controller-manager-7d9dfd778-jc4dw" Dec 03 14:27:34 crc kubenswrapper[4805]: I1203 14:27:34.076287 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/cinder-operator-controller-manager-859b6ccc6-2nzgx" event={"ID":"f9b6b508-1cf6-40aa-829e-b9cf264921e5","Type":"ContainerStarted","Data":"3b2c71bf60fe79e2324cf2022d10781fc34f626471768decff8dd2c9a7105ae2"} Dec 03 14:27:34 crc kubenswrapper[4805]: I1203 14:27:34.076576 4805 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/cinder-operator-controller-manager-859b6ccc6-2nzgx" Dec 03 14:27:34 crc kubenswrapper[4805]: I1203 14:27:34.077720 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/mariadb-operator-controller-manager-56bbcc9d85-n6m2d" event={"ID":"2d72009f-b9ef-4c9d-b224-5b133a3cd93a","Type":"ContainerStarted","Data":"bb140f72ae7482c01f6ddbda38fd68b44c1087ac88defd8fb52246427d6205a4"} Dec 03 14:27:34 crc kubenswrapper[4805]: I1203 14:27:34.078518 4805 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/cinder-operator-controller-manager-859b6ccc6-2nzgx" Dec 03 14:27:34 crc kubenswrapper[4805]: E1203 14:27:34.080074 4805 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/mariadb-operator@sha256:600ca007e493d3af0fcc2ebac92e8da5efd2afe812b62d7d3d4dd0115bdf05d7\\\"\"" pod="openstack-operators/mariadb-operator-controller-manager-56bbcc9d85-n6m2d" podUID="2d72009f-b9ef-4c9d-b224-5b133a3cd93a" Dec 03 14:27:34 crc kubenswrapper[4805]: I1203 14:27:34.081850 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/swift-operator-controller-manager-5f8c65bbfc-qrvpw" event={"ID":"b51f4884-a980-4643-9deb-8668aead3876","Type":"ContainerStarted","Data":"4c83be661a2111138959dc300677bb80f304e2c410e67a06a0e8f1a35887871a"} Dec 03 14:27:34 crc kubenswrapper[4805]: I1203 14:27:34.082652 4805 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/swift-operator-controller-manager-5f8c65bbfc-qrvpw" Dec 03 14:27:34 crc kubenswrapper[4805]: I1203 14:27:34.088809 4805 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-5q6j2" podStartSLOduration=3.414493157 podStartE2EDuration="43.088796316s" podCreationTimestamp="2025-12-03 14:26:51 +0000 UTC" firstStartedPulling="2025-12-03 14:26:53.021005188 +0000 UTC m=+1042.683922131" lastFinishedPulling="2025-12-03 14:27:32.695308337 +0000 UTC m=+1082.358225290" observedRunningTime="2025-12-03 14:27:34.084565842 +0000 UTC m=+1083.747482765" watchObservedRunningTime="2025-12-03 14:27:34.088796316 +0000 UTC m=+1083.751713239" Dec 03 14:27:34 crc kubenswrapper[4805]: I1203 14:27:34.094065 4805 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/swift-operator-controller-manager-5f8c65bbfc-qrvpw" Dec 03 14:27:34 crc kubenswrapper[4805]: I1203 14:27:34.099023 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ironic-operator-controller-manager-6c548fd776-mz4pw" event={"ID":"1ad4fa11-2f11-4a06-9944-bba8c7445e84","Type":"ContainerStarted","Data":"d894f4fc23ec95a4cb73d9d88f1b1eb35041c4c7f91e8a197876cb63f09ffb0f"} Dec 03 14:27:34 crc kubenswrapper[4805]: I1203 14:27:34.099330 4805 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/ironic-operator-controller-manager-6c548fd776-mz4pw" Dec 03 14:27:34 crc kubenswrapper[4805]: I1203 14:27:34.101485 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/keystone-operator-controller-manager-7765d96ddf-qzzz9" event={"ID":"ffa72db0-588a-4a78-a1cd-8710f5ce4f41","Type":"ContainerStarted","Data":"0a8c4b0810a2a06355557228e69bf5976c32dd9f46078425d22705d568bb871d"} Dec 03 14:27:34 crc kubenswrapper[4805]: I1203 14:27:34.102145 4805 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/keystone-operator-controller-manager-7765d96ddf-qzzz9" Dec 03 14:27:34 crc kubenswrapper[4805]: I1203 14:27:34.102595 4805 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/ironic-operator-controller-manager-6c548fd776-mz4pw" Dec 03 14:27:34 crc kubenswrapper[4805]: I1203 14:27:34.103664 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-baremetal-operator-controller-manager-64bc77cfd4x9mdd" event={"ID":"b9d0830e-0f5d-4639-ab7d-92dcf07a29bc","Type":"ContainerStarted","Data":"46757d969c8a673ead2b4d79f39bb2478d744b1c13391db2b2cc7d7769061eff"} Dec 03 14:27:34 crc kubenswrapper[4805]: E1203 14:27:34.106436 4805 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/openstack-baremetal-operator@sha256:14cfad6ea2e7f7ecc4cb2aafceb9c61514b3d04b66668832d1e4ac3b19f1ab81\\\"\"" pod="openstack-operators/openstack-baremetal-operator-controller-manager-64bc77cfd4x9mdd" podUID="b9d0830e-0f5d-4639-ab7d-92dcf07a29bc" Dec 03 14:27:34 crc kubenswrapper[4805]: I1203 14:27:34.115647 4805 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/test-operator-controller-manager-5854674fcc-d6pvv" podStartSLOduration=4.005909136 podStartE2EDuration="43.115633334s" podCreationTimestamp="2025-12-03 14:26:51 +0000 UTC" firstStartedPulling="2025-12-03 14:26:53.064406595 +0000 UTC m=+1042.727323518" lastFinishedPulling="2025-12-03 14:27:32.174130793 +0000 UTC m=+1081.837047716" observedRunningTime="2025-12-03 14:27:34.114799012 +0000 UTC m=+1083.777715935" watchObservedRunningTime="2025-12-03 14:27:34.115633334 +0000 UTC m=+1083.778550257" Dec 03 14:27:34 crc kubenswrapper[4805]: I1203 14:27:34.116788 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/glance-operator-controller-manager-77987cd8cd-pt9fw" event={"ID":"5615ce63-1ec1-4a77-8e44-49e6d114eec6","Type":"ContainerStarted","Data":"d06cdc6852d690ac14ab39a3c2ca3c6d47ee3a7a0e58a51e088d802ba683cbe4"} Dec 03 14:27:34 crc kubenswrapper[4805]: I1203 14:27:34.118055 4805 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/glance-operator-controller-manager-77987cd8cd-pt9fw" Dec 03 14:27:34 crc kubenswrapper[4805]: I1203 14:27:34.122046 4805 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/glance-operator-controller-manager-77987cd8cd-pt9fw" Dec 03 14:27:34 crc kubenswrapper[4805]: I1203 14:27:34.123992 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/placement-operator-controller-manager-78f8948974-b8l4l" event={"ID":"f853eae6-829b-42f9-9198-3d7dcabd405f","Type":"ContainerStarted","Data":"7286c5c7960f753f546eb2c379610387635f0c9dbe9b4d4599e1bda037acf884"} Dec 03 14:27:34 crc kubenswrapper[4805]: I1203 14:27:34.125545 4805 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/placement-operator-controller-manager-78f8948974-b8l4l" Dec 03 14:27:34 crc kubenswrapper[4805]: I1203 14:27:34.132072 4805 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/placement-operator-controller-manager-78f8948974-b8l4l" Dec 03 14:27:34 crc kubenswrapper[4805]: I1203 14:27:34.136176 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/watcher-operator-controller-manager-769dc69bc-fl6mm" event={"ID":"30776d04-200b-4c10-b48d-f51a926e469b","Type":"ContainerStarted","Data":"3bde2cedf12f44eac732c9c018e474a2063d8ad93d3efef9d21c2ed27b0052a5"} Dec 03 14:27:34 crc kubenswrapper[4805]: I1203 14:27:34.136969 4805 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/watcher-operator-controller-manager-769dc69bc-fl6mm" Dec 03 14:27:34 crc kubenswrapper[4805]: I1203 14:27:34.139052 4805 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/watcher-operator-controller-manager-769dc69bc-fl6mm" Dec 03 14:27:34 crc kubenswrapper[4805]: I1203 14:27:34.148449 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/horizon-operator-controller-manager-68c6d99b8f-kxbr8" event={"ID":"b29e4120-0cbc-4de5-a4b6-97b97761f49a","Type":"ContainerStarted","Data":"4053f9031af09530db9cf53aa9e3916c76f04247f84022b886633c136cc872d3"} Dec 03 14:27:34 crc kubenswrapper[4805]: I1203 14:27:34.153857 4805 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/horizon-operator-controller-manager-68c6d99b8f-kxbr8" Dec 03 14:27:34 crc kubenswrapper[4805]: I1203 14:27:34.159423 4805 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/horizon-operator-controller-manager-68c6d99b8f-kxbr8" Dec 03 14:27:34 crc kubenswrapper[4805]: I1203 14:27:34.159583 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ovn-operator-controller-manager-b6456fdb6-djzcm" event={"ID":"42b37265-f5bd-4ac4-9b96-d758fbe1b31e","Type":"ContainerStarted","Data":"915bcabc161ae5f426835f49d170daa4adcbddc8ef10cc006fd639c06dd7285d"} Dec 03 14:27:34 crc kubenswrapper[4805]: E1203 14:27:34.161987 4805 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/ovn-operator@sha256:635a4aef9d6f0b799e8ec91333dbb312160c001d05b3c63f614c124e0b67cb59\\\"\"" pod="openstack-operators/ovn-operator-controller-manager-b6456fdb6-djzcm" podUID="42b37265-f5bd-4ac4-9b96-d758fbe1b31e" Dec 03 14:27:34 crc kubenswrapper[4805]: I1203 14:27:34.170106 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/neutron-operator-controller-manager-5fdfd5b6b5-4mlhk" event={"ID":"58aa635e-157c-4ae2-ad0c-568c144638d1","Type":"ContainerStarted","Data":"023dc8c9fd15f3b8a70db6f6eab372dc8fed9c22cda381578cbe6c5a8e083e1a"} Dec 03 14:27:34 crc kubenswrapper[4805]: I1203 14:27:34.170853 4805 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/neutron-operator-controller-manager-5fdfd5b6b5-4mlhk" Dec 03 14:27:34 crc kubenswrapper[4805]: I1203 14:27:34.172120 4805 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/neutron-operator-controller-manager-5fdfd5b6b5-4mlhk" Dec 03 14:27:34 crc kubenswrapper[4805]: I1203 14:27:34.183823 4805 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/barbican-operator-controller-manager-7d9dfd778-jc4dw" podStartSLOduration=31.160620988 podStartE2EDuration="44.183811113s" podCreationTimestamp="2025-12-03 14:26:50 +0000 UTC" firstStartedPulling="2025-12-03 14:26:52.983851341 +0000 UTC m=+1042.646768264" lastFinishedPulling="2025-12-03 14:27:06.007041456 +0000 UTC m=+1055.669958389" observedRunningTime="2025-12-03 14:27:34.146392178 +0000 UTC m=+1083.809309101" watchObservedRunningTime="2025-12-03 14:27:34.183811113 +0000 UTC m=+1083.846728036" Dec 03 14:27:34 crc kubenswrapper[4805]: I1203 14:27:34.186558 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/infra-operator-controller-manager-57548d458d-77t9s" event={"ID":"f21790d8-af47-4be7-8758-f038bda76908","Type":"ContainerStarted","Data":"42a8189f6b3744631e33ffdde4c11eaf0ded5d1c14c8a8143a08ef90b2b7c7a8"} Dec 03 14:27:34 crc kubenswrapper[4805]: E1203 14:27:34.189202 4805 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/infra-operator@sha256:09a6d0613ee2d3c1c809fc36c22678458ac271e0da87c970aec0a5339f5423f7\\\"\"" pod="openstack-operators/infra-operator-controller-manager-57548d458d-77t9s" podUID="f21790d8-af47-4be7-8758-f038bda76908" Dec 03 14:27:34 crc kubenswrapper[4805]: I1203 14:27:34.192023 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/heat-operator-controller-manager-5f64f6f8bb-86sbc" event={"ID":"a748cfe4-2e54-4a7b-a353-e1dfea6c14b3","Type":"ContainerStarted","Data":"a0c7be0377c9602825113101127efa49e473f9f7f8382084f1338c15af3e8964"} Dec 03 14:27:34 crc kubenswrapper[4805]: I1203 14:27:34.192942 4805 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/heat-operator-controller-manager-5f64f6f8bb-86sbc" Dec 03 14:27:34 crc kubenswrapper[4805]: I1203 14:27:34.201371 4805 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/heat-operator-controller-manager-5f64f6f8bb-86sbc" Dec 03 14:27:34 crc kubenswrapper[4805]: I1203 14:27:34.226282 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/telemetry-operator-controller-manager-cc9d8f87b-qfmmx" event={"ID":"8cbfcd07-e136-4bfd-b3c6-6615c911d440","Type":"ContainerStarted","Data":"f663a6b573d1c3914b2188e924cf8eb2f69435fc731e25752bd2495b248b4f6c"} Dec 03 14:27:34 crc kubenswrapper[4805]: I1203 14:27:34.226737 4805 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/octavia-operator-controller-manager-998648c74-8mbdb" podStartSLOduration=30.311943202 podStartE2EDuration="43.226720797s" podCreationTimestamp="2025-12-03 14:26:51 +0000 UTC" firstStartedPulling="2025-12-03 14:26:53.099307162 +0000 UTC m=+1042.762224085" lastFinishedPulling="2025-12-03 14:27:06.014084767 +0000 UTC m=+1055.677001680" observedRunningTime="2025-12-03 14:27:34.226399418 +0000 UTC m=+1083.889316331" watchObservedRunningTime="2025-12-03 14:27:34.226720797 +0000 UTC m=+1083.889637720" Dec 03 14:27:34 crc kubenswrapper[4805]: I1203 14:27:34.229297 4805 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/designate-operator-controller-manager-78b4bc895b-w56sr" podStartSLOduration=30.842632485 podStartE2EDuration="44.229285846s" podCreationTimestamp="2025-12-03 14:26:50 +0000 UTC" firstStartedPulling="2025-12-03 14:26:52.60879989 +0000 UTC m=+1042.271716813" lastFinishedPulling="2025-12-03 14:27:05.995453211 +0000 UTC m=+1055.658370174" observedRunningTime="2025-12-03 14:27:34.18516013 +0000 UTC m=+1083.848077053" watchObservedRunningTime="2025-12-03 14:27:34.229285846 +0000 UTC m=+1083.892202759" Dec 03 14:27:34 crc kubenswrapper[4805]: I1203 14:27:34.232383 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/manila-operator-controller-manager-7c79b5df47-d2thc" event={"ID":"967f26e7-3027-4e23-9f02-d111d714a396","Type":"ContainerStarted","Data":"9f98a06d260673001a3f931525b34266e2db887d3842401b3832972755e20ee1"} Dec 03 14:27:34 crc kubenswrapper[4805]: I1203 14:27:34.232873 4805 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/manila-operator-controller-manager-7c79b5df47-d2thc" Dec 03 14:27:34 crc kubenswrapper[4805]: I1203 14:27:34.244237 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/nova-operator-controller-manager-697bc559fc-5rl85" event={"ID":"05506111-ddb3-4ab9-93a0-ad2bedff67a4","Type":"ContainerStarted","Data":"f2e0d4533b1cc883833a8d160b0c39bfeed16e1e9fb2f3839be654520bc873f9"} Dec 03 14:27:34 crc kubenswrapper[4805]: I1203 14:27:34.245153 4805 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/manila-operator-controller-manager-7c79b5df47-d2thc" Dec 03 14:27:34 crc kubenswrapper[4805]: E1203 14:27:34.245606 4805 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/nova-operator@sha256:779f0cee6024d0fb8f259b036fe790e62aa5a3b0431ea9bf15a6e7d02e2e5670\\\"\"" pod="openstack-operators/nova-operator-controller-manager-697bc559fc-5rl85" podUID="05506111-ddb3-4ab9-93a0-ad2bedff67a4" Dec 03 14:27:34 crc kubenswrapper[4805]: I1203 14:27:34.294798 4805 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/cinder-operator-controller-manager-859b6ccc6-2nzgx" podStartSLOduration=30.859110392 podStartE2EDuration="44.294782373s" podCreationTimestamp="2025-12-03 14:26:50 +0000 UTC" firstStartedPulling="2025-12-03 14:26:52.577252584 +0000 UTC m=+1042.240169527" lastFinishedPulling="2025-12-03 14:27:06.012924585 +0000 UTC m=+1055.675841508" observedRunningTime="2025-12-03 14:27:34.293898829 +0000 UTC m=+1083.956815752" watchObservedRunningTime="2025-12-03 14:27:34.294782373 +0000 UTC m=+1083.957699296" Dec 03 14:27:34 crc kubenswrapper[4805]: I1203 14:27:34.334172 4805 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/swift-operator-controller-manager-5f8c65bbfc-qrvpw" podStartSLOduration=30.219460005 podStartE2EDuration="43.334158211s" podCreationTimestamp="2025-12-03 14:26:51 +0000 UTC" firstStartedPulling="2025-12-03 14:26:52.961427243 +0000 UTC m=+1042.624344166" lastFinishedPulling="2025-12-03 14:27:06.076125449 +0000 UTC m=+1055.739042372" observedRunningTime="2025-12-03 14:27:34.332451904 +0000 UTC m=+1083.995368847" watchObservedRunningTime="2025-12-03 14:27:34.334158211 +0000 UTC m=+1083.997075134" Dec 03 14:27:34 crc kubenswrapper[4805]: I1203 14:27:34.370621 4805 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/heat-operator-controller-manager-5f64f6f8bb-86sbc" podStartSLOduration=30.459491035 podStartE2EDuration="44.370602579s" podCreationTimestamp="2025-12-03 14:26:50 +0000 UTC" firstStartedPulling="2025-12-03 14:26:52.099093337 +0000 UTC m=+1041.762010260" lastFinishedPulling="2025-12-03 14:27:06.010204881 +0000 UTC m=+1055.673121804" observedRunningTime="2025-12-03 14:27:34.363902107 +0000 UTC m=+1084.026819040" watchObservedRunningTime="2025-12-03 14:27:34.370602579 +0000 UTC m=+1084.033519492" Dec 03 14:27:34 crc kubenswrapper[4805]: I1203 14:27:34.425122 4805 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/placement-operator-controller-manager-78f8948974-b8l4l" podStartSLOduration=30.271348661 podStartE2EDuration="43.425104137s" podCreationTimestamp="2025-12-03 14:26:51 +0000 UTC" firstStartedPulling="2025-12-03 14:26:52.892275577 +0000 UTC m=+1042.555192500" lastFinishedPulling="2025-12-03 14:27:06.046031043 +0000 UTC m=+1055.708947976" observedRunningTime="2025-12-03 14:27:34.392649117 +0000 UTC m=+1084.055566040" watchObservedRunningTime="2025-12-03 14:27:34.425104137 +0000 UTC m=+1084.088021060" Dec 03 14:27:34 crc kubenswrapper[4805]: I1203 14:27:34.450725 4805 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/glance-operator-controller-manager-77987cd8cd-pt9fw" podStartSLOduration=30.585159462 podStartE2EDuration="44.450706721s" podCreationTimestamp="2025-12-03 14:26:50 +0000 UTC" firstStartedPulling="2025-12-03 14:26:52.129360358 +0000 UTC m=+1041.792277281" lastFinishedPulling="2025-12-03 14:27:05.994907617 +0000 UTC m=+1055.657824540" observedRunningTime="2025-12-03 14:27:34.449002975 +0000 UTC m=+1084.111919898" watchObservedRunningTime="2025-12-03 14:27:34.450706721 +0000 UTC m=+1084.113623644" Dec 03 14:27:34 crc kubenswrapper[4805]: I1203 14:27:34.549545 4805 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/horizon-operator-controller-manager-68c6d99b8f-kxbr8" podStartSLOduration=30.851505965 podStartE2EDuration="44.549528821s" podCreationTimestamp="2025-12-03 14:26:50 +0000 UTC" firstStartedPulling="2025-12-03 14:26:52.310503 +0000 UTC m=+1041.973419923" lastFinishedPulling="2025-12-03 14:27:06.008525856 +0000 UTC m=+1055.671442779" observedRunningTime="2025-12-03 14:27:34.53991075 +0000 UTC m=+1084.202827683" watchObservedRunningTime="2025-12-03 14:27:34.549528821 +0000 UTC m=+1084.212445744" Dec 03 14:27:34 crc kubenswrapper[4805]: I1203 14:27:34.572342 4805 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/ironic-operator-controller-manager-6c548fd776-mz4pw" podStartSLOduration=31.171934445 podStartE2EDuration="44.572317139s" podCreationTimestamp="2025-12-03 14:26:50 +0000 UTC" firstStartedPulling="2025-12-03 14:26:52.608483431 +0000 UTC m=+1042.271400354" lastFinishedPulling="2025-12-03 14:27:06.008866115 +0000 UTC m=+1055.671783048" observedRunningTime="2025-12-03 14:27:34.564014344 +0000 UTC m=+1084.226931267" watchObservedRunningTime="2025-12-03 14:27:34.572317139 +0000 UTC m=+1084.235234062" Dec 03 14:27:34 crc kubenswrapper[4805]: I1203 14:27:34.630892 4805 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/keystone-operator-controller-manager-7765d96ddf-qzzz9" podStartSLOduration=5.42308661 podStartE2EDuration="44.630876117s" podCreationTimestamp="2025-12-03 14:26:50 +0000 UTC" firstStartedPulling="2025-12-03 14:26:53.050597671 +0000 UTC m=+1042.713514594" lastFinishedPulling="2025-12-03 14:27:32.258387168 +0000 UTC m=+1081.921304101" observedRunningTime="2025-12-03 14:27:34.59998831 +0000 UTC m=+1084.262905233" watchObservedRunningTime="2025-12-03 14:27:34.630876117 +0000 UTC m=+1084.293793040" Dec 03 14:27:34 crc kubenswrapper[4805]: I1203 14:27:34.654247 4805 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/manila-operator-controller-manager-7c79b5df47-d2thc" podStartSLOduration=31.587077794 podStartE2EDuration="44.654231521s" podCreationTimestamp="2025-12-03 14:26:50 +0000 UTC" firstStartedPulling="2025-12-03 14:26:52.947502275 +0000 UTC m=+1042.610419188" lastFinishedPulling="2025-12-03 14:27:06.014655982 +0000 UTC m=+1055.677572915" observedRunningTime="2025-12-03 14:27:34.631595017 +0000 UTC m=+1084.294511940" watchObservedRunningTime="2025-12-03 14:27:34.654231521 +0000 UTC m=+1084.317148434" Dec 03 14:27:34 crc kubenswrapper[4805]: I1203 14:27:34.735433 4805 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/watcher-operator-controller-manager-769dc69bc-fl6mm" podStartSLOduration=30.755324326 podStartE2EDuration="43.735413912s" podCreationTimestamp="2025-12-03 14:26:51 +0000 UTC" firstStartedPulling="2025-12-03 14:26:53.034976087 +0000 UTC m=+1042.697893010" lastFinishedPulling="2025-12-03 14:27:06.015065673 +0000 UTC m=+1055.677982596" observedRunningTime="2025-12-03 14:27:34.681287154 +0000 UTC m=+1084.344204087" watchObservedRunningTime="2025-12-03 14:27:34.735413912 +0000 UTC m=+1084.398330835" Dec 03 14:27:34 crc kubenswrapper[4805]: I1203 14:27:34.766068 4805 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/neutron-operator-controller-manager-5fdfd5b6b5-4mlhk" podStartSLOduration=30.787719524 podStartE2EDuration="43.766052043s" podCreationTimestamp="2025-12-03 14:26:51 +0000 UTC" firstStartedPulling="2025-12-03 14:26:53.019381514 +0000 UTC m=+1042.682298437" lastFinishedPulling="2025-12-03 14:27:05.997713993 +0000 UTC m=+1055.660630956" observedRunningTime="2025-12-03 14:27:34.764889782 +0000 UTC m=+1084.427806715" watchObservedRunningTime="2025-12-03 14:27:34.766052043 +0000 UTC m=+1084.428968966" Dec 03 14:27:35 crc kubenswrapper[4805]: I1203 14:27:35.254577 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/telemetry-operator-controller-manager-cc9d8f87b-qfmmx" event={"ID":"8cbfcd07-e136-4bfd-b3c6-6615c911d440","Type":"ContainerStarted","Data":"3e7896ec9ea345c6c0e1e4ba84145fe96bd80e209c80913ddd5238d6686c4c63"} Dec 03 14:27:35 crc kubenswrapper[4805]: I1203 14:27:35.259162 4805 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/telemetry-operator-controller-manager-cc9d8f87b-qfmmx" Dec 03 14:27:35 crc kubenswrapper[4805]: E1203 14:27:35.260366 4805 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/infra-operator@sha256:09a6d0613ee2d3c1c809fc36c22678458ac271e0da87c970aec0a5339f5423f7\\\"\"" pod="openstack-operators/infra-operator-controller-manager-57548d458d-77t9s" podUID="f21790d8-af47-4be7-8758-f038bda76908" Dec 03 14:27:35 crc kubenswrapper[4805]: E1203 14:27:35.260955 4805 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/openstack-baremetal-operator@sha256:14cfad6ea2e7f7ecc4cb2aafceb9c61514b3d04b66668832d1e4ac3b19f1ab81\\\"\"" pod="openstack-operators/openstack-baremetal-operator-controller-manager-64bc77cfd4x9mdd" podUID="b9d0830e-0f5d-4639-ab7d-92dcf07a29bc" Dec 03 14:27:35 crc kubenswrapper[4805]: I1203 14:27:35.278981 4805 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/telemetry-operator-controller-manager-cc9d8f87b-qfmmx" podStartSLOduration=3.772585821 podStartE2EDuration="44.278956113s" podCreationTimestamp="2025-12-03 14:26:51 +0000 UTC" firstStartedPulling="2025-12-03 14:26:52.892252067 +0000 UTC m=+1042.555168990" lastFinishedPulling="2025-12-03 14:27:33.398622359 +0000 UTC m=+1083.061539282" observedRunningTime="2025-12-03 14:27:35.2718649 +0000 UTC m=+1084.934781823" watchObservedRunningTime="2025-12-03 14:27:35.278956113 +0000 UTC m=+1084.941873036" Dec 03 14:27:37 crc kubenswrapper[4805]: I1203 14:27:37.673120 4805 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/openstack-operator-controller-manager-cc68d789f-pkvr8" Dec 03 14:27:41 crc kubenswrapper[4805]: I1203 14:27:41.468333 4805 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/keystone-operator-controller-manager-7765d96ddf-qzzz9" Dec 03 14:27:41 crc kubenswrapper[4805]: I1203 14:27:41.967508 4805 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/telemetry-operator-controller-manager-cc9d8f87b-qfmmx" Dec 03 14:27:42 crc kubenswrapper[4805]: I1203 14:27:42.003195 4805 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/test-operator-controller-manager-5854674fcc-d6pvv" Dec 03 14:27:47 crc kubenswrapper[4805]: E1203 14:27:47.697261 4805 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/mariadb-operator@sha256:600ca007e493d3af0fcc2ebac92e8da5efd2afe812b62d7d3d4dd0115bdf05d7\\\"\"" pod="openstack-operators/mariadb-operator-controller-manager-56bbcc9d85-n6m2d" podUID="2d72009f-b9ef-4c9d-b224-5b133a3cd93a" Dec 03 14:27:47 crc kubenswrapper[4805]: E1203 14:27:47.697318 4805 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/ovn-operator@sha256:635a4aef9d6f0b799e8ec91333dbb312160c001d05b3c63f614c124e0b67cb59\\\"\"" pod="openstack-operators/ovn-operator-controller-manager-b6456fdb6-djzcm" podUID="42b37265-f5bd-4ac4-9b96-d758fbe1b31e" Dec 03 14:27:47 crc kubenswrapper[4805]: E1203 14:27:47.697394 4805 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/nova-operator@sha256:779f0cee6024d0fb8f259b036fe790e62aa5a3b0431ea9bf15a6e7d02e2e5670\\\"\"" pod="openstack-operators/nova-operator-controller-manager-697bc559fc-5rl85" podUID="05506111-ddb3-4ab9-93a0-ad2bedff67a4" Dec 03 14:27:49 crc kubenswrapper[4805]: I1203 14:27:49.356906 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/infra-operator-controller-manager-57548d458d-77t9s" event={"ID":"f21790d8-af47-4be7-8758-f038bda76908","Type":"ContainerStarted","Data":"63872f687948317497108adf326af9c7a72c8bd3d3e4794040569ee8512bf575"} Dec 03 14:27:49 crc kubenswrapper[4805]: I1203 14:27:49.357410 4805 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/infra-operator-controller-manager-57548d458d-77t9s" Dec 03 14:27:49 crc kubenswrapper[4805]: I1203 14:27:49.377425 4805 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/infra-operator-controller-manager-57548d458d-77t9s" podStartSLOduration=18.672929883 podStartE2EDuration="59.377407928s" podCreationTimestamp="2025-12-03 14:26:50 +0000 UTC" firstStartedPulling="2025-12-03 14:27:07.470985337 +0000 UTC m=+1057.133902260" lastFinishedPulling="2025-12-03 14:27:48.175463392 +0000 UTC m=+1097.838380305" observedRunningTime="2025-12-03 14:27:49.371047025 +0000 UTC m=+1099.033963948" watchObservedRunningTime="2025-12-03 14:27:49.377407928 +0000 UTC m=+1099.040324851" Dec 03 14:27:51 crc kubenswrapper[4805]: I1203 14:27:51.372533 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-baremetal-operator-controller-manager-64bc77cfd4x9mdd" event={"ID":"b9d0830e-0f5d-4639-ab7d-92dcf07a29bc","Type":"ContainerStarted","Data":"9f40cb4440f55adc44458d65c2770316d29c652e41dde6ad2dc7ff10edd65d4c"} Dec 03 14:27:51 crc kubenswrapper[4805]: I1203 14:27:51.373368 4805 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/openstack-baremetal-operator-controller-manager-64bc77cfd4x9mdd" Dec 03 14:27:51 crc kubenswrapper[4805]: I1203 14:27:51.403205 4805 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-baremetal-operator-controller-manager-64bc77cfd4x9mdd" podStartSLOduration=17.758624736 podStartE2EDuration="1m0.403185245s" podCreationTimestamp="2025-12-03 14:26:51 +0000 UTC" firstStartedPulling="2025-12-03 14:27:08.536031499 +0000 UTC m=+1058.198948422" lastFinishedPulling="2025-12-03 14:27:51.180591988 +0000 UTC m=+1100.843508931" observedRunningTime="2025-12-03 14:27:51.397636744 +0000 UTC m=+1101.060553667" watchObservedRunningTime="2025-12-03 14:27:51.403185245 +0000 UTC m=+1101.066102168" Dec 03 14:27:56 crc kubenswrapper[4805]: I1203 14:27:56.987989 4805 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/infra-operator-controller-manager-57548d458d-77t9s" Dec 03 14:27:57 crc kubenswrapper[4805]: I1203 14:27:57.423311 4805 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/openstack-baremetal-operator-controller-manager-64bc77cfd4x9mdd" Dec 03 14:28:00 crc kubenswrapper[4805]: I1203 14:28:00.451834 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/mariadb-operator-controller-manager-56bbcc9d85-n6m2d" event={"ID":"2d72009f-b9ef-4c9d-b224-5b133a3cd93a","Type":"ContainerStarted","Data":"eb94f9313ab20110ea2634d3329ff94aa57d251f695dc8f4dd3ab510c82bc55b"} Dec 03 14:28:00 crc kubenswrapper[4805]: I1203 14:28:00.452661 4805 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/mariadb-operator-controller-manager-56bbcc9d85-n6m2d" Dec 03 14:28:00 crc kubenswrapper[4805]: I1203 14:28:00.455031 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/nova-operator-controller-manager-697bc559fc-5rl85" event={"ID":"05506111-ddb3-4ab9-93a0-ad2bedff67a4","Type":"ContainerStarted","Data":"00c1411335f31f10311ee36d1303bb5db5038a9730dbfbf26a077be4e696ee25"} Dec 03 14:28:00 crc kubenswrapper[4805]: I1203 14:28:00.455211 4805 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/nova-operator-controller-manager-697bc559fc-5rl85" Dec 03 14:28:00 crc kubenswrapper[4805]: I1203 14:28:00.471209 4805 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/mariadb-operator-controller-manager-56bbcc9d85-n6m2d" podStartSLOduration=2.387922389 podStartE2EDuration="1m9.471195591s" podCreationTimestamp="2025-12-03 14:26:51 +0000 UTC" firstStartedPulling="2025-12-03 14:26:53.066007109 +0000 UTC m=+1042.728924032" lastFinishedPulling="2025-12-03 14:28:00.149280311 +0000 UTC m=+1109.812197234" observedRunningTime="2025-12-03 14:28:00.47078727 +0000 UTC m=+1110.133704193" watchObservedRunningTime="2025-12-03 14:28:00.471195591 +0000 UTC m=+1110.134112514" Dec 03 14:28:00 crc kubenswrapper[4805]: I1203 14:28:00.493140 4805 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/nova-operator-controller-manager-697bc559fc-5rl85" podStartSLOduration=2.385827833 podStartE2EDuration="1m9.493125416s" podCreationTimestamp="2025-12-03 14:26:51 +0000 UTC" firstStartedPulling="2025-12-03 14:26:53.042701127 +0000 UTC m=+1042.705618050" lastFinishedPulling="2025-12-03 14:28:00.14999871 +0000 UTC m=+1109.812915633" observedRunningTime="2025-12-03 14:28:00.488934942 +0000 UTC m=+1110.151851855" watchObservedRunningTime="2025-12-03 14:28:00.493125416 +0000 UTC m=+1110.156042339" Dec 03 14:28:02 crc kubenswrapper[4805]: I1203 14:28:02.474254 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ovn-operator-controller-manager-b6456fdb6-djzcm" event={"ID":"42b37265-f5bd-4ac4-9b96-d758fbe1b31e","Type":"ContainerStarted","Data":"56859e8adca82e0745a0b76b10359c6dd66955f2380f3a9bda01703de031fdbb"} Dec 03 14:28:02 crc kubenswrapper[4805]: I1203 14:28:02.474808 4805 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/ovn-operator-controller-manager-b6456fdb6-djzcm" Dec 03 14:28:11 crc kubenswrapper[4805]: I1203 14:28:11.687411 4805 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/mariadb-operator-controller-manager-56bbcc9d85-n6m2d" Dec 03 14:28:11 crc kubenswrapper[4805]: I1203 14:28:11.710575 4805 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/ovn-operator-controller-manager-b6456fdb6-djzcm" podStartSLOduration=11.659787814 podStartE2EDuration="1m20.710551492s" podCreationTimestamp="2025-12-03 14:26:51 +0000 UTC" firstStartedPulling="2025-12-03 14:26:53.043976831 +0000 UTC m=+1042.706893754" lastFinishedPulling="2025-12-03 14:28:02.094740509 +0000 UTC m=+1111.757657432" observedRunningTime="2025-12-03 14:28:02.505042587 +0000 UTC m=+1112.167959510" watchObservedRunningTime="2025-12-03 14:28:11.710551492 +0000 UTC m=+1121.373468435" Dec 03 14:28:11 crc kubenswrapper[4805]: I1203 14:28:11.781641 4805 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/nova-operator-controller-manager-697bc559fc-5rl85" Dec 03 14:28:11 crc kubenswrapper[4805]: I1203 14:28:11.882600 4805 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/ovn-operator-controller-manager-b6456fdb6-djzcm" Dec 03 14:28:27 crc kubenswrapper[4805]: I1203 14:28:27.363784 4805 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-675f4bcbfc-5hg5w"] Dec 03 14:28:27 crc kubenswrapper[4805]: I1203 14:28:27.372319 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-675f4bcbfc-5hg5w" Dec 03 14:28:27 crc kubenswrapper[4805]: I1203 14:28:27.373325 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-675f4bcbfc-5hg5w"] Dec 03 14:28:27 crc kubenswrapper[4805]: I1203 14:28:27.374636 4805 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"dns" Dec 03 14:28:27 crc kubenswrapper[4805]: I1203 14:28:27.375855 4805 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openshift-service-ca.crt" Dec 03 14:28:27 crc kubenswrapper[4805]: I1203 14:28:27.377923 4805 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"kube-root-ca.crt" Dec 03 14:28:27 crc kubenswrapper[4805]: I1203 14:28:27.377989 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dnsmasq-dns-dockercfg-f6lcc" Dec 03 14:28:27 crc kubenswrapper[4805]: I1203 14:28:27.415440 4805 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-78dd6ddcc-ql7t4"] Dec 03 14:28:27 crc kubenswrapper[4805]: I1203 14:28:27.416605 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-78dd6ddcc-ql7t4" Dec 03 14:28:27 crc kubenswrapper[4805]: I1203 14:28:27.419596 4805 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"dns-svc" Dec 03 14:28:27 crc kubenswrapper[4805]: I1203 14:28:27.423388 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-78dd6ddcc-ql7t4"] Dec 03 14:28:27 crc kubenswrapper[4805]: I1203 14:28:27.455297 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7jj28\" (UniqueName: \"kubernetes.io/projected/388dedb5-6949-4ccb-9289-520b6f9ba4ce-kube-api-access-7jj28\") pod \"dnsmasq-dns-675f4bcbfc-5hg5w\" (UID: \"388dedb5-6949-4ccb-9289-520b6f9ba4ce\") " pod="openstack/dnsmasq-dns-675f4bcbfc-5hg5w" Dec 03 14:28:27 crc kubenswrapper[4805]: I1203 14:28:27.455455 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rfqvq\" (UniqueName: \"kubernetes.io/projected/6c4701fa-549e-427c-b680-2f3096aafd00-kube-api-access-rfqvq\") pod \"dnsmasq-dns-78dd6ddcc-ql7t4\" (UID: \"6c4701fa-549e-427c-b680-2f3096aafd00\") " pod="openstack/dnsmasq-dns-78dd6ddcc-ql7t4" Dec 03 14:28:27 crc kubenswrapper[4805]: I1203 14:28:27.455557 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6c4701fa-549e-427c-b680-2f3096aafd00-config\") pod \"dnsmasq-dns-78dd6ddcc-ql7t4\" (UID: \"6c4701fa-549e-427c-b680-2f3096aafd00\") " pod="openstack/dnsmasq-dns-78dd6ddcc-ql7t4" Dec 03 14:28:27 crc kubenswrapper[4805]: I1203 14:28:27.455651 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/388dedb5-6949-4ccb-9289-520b6f9ba4ce-config\") pod \"dnsmasq-dns-675f4bcbfc-5hg5w\" (UID: \"388dedb5-6949-4ccb-9289-520b6f9ba4ce\") " pod="openstack/dnsmasq-dns-675f4bcbfc-5hg5w" Dec 03 14:28:27 crc kubenswrapper[4805]: I1203 14:28:27.455734 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/6c4701fa-549e-427c-b680-2f3096aafd00-dns-svc\") pod \"dnsmasq-dns-78dd6ddcc-ql7t4\" (UID: \"6c4701fa-549e-427c-b680-2f3096aafd00\") " pod="openstack/dnsmasq-dns-78dd6ddcc-ql7t4" Dec 03 14:28:27 crc kubenswrapper[4805]: I1203 14:28:27.557051 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7jj28\" (UniqueName: \"kubernetes.io/projected/388dedb5-6949-4ccb-9289-520b6f9ba4ce-kube-api-access-7jj28\") pod \"dnsmasq-dns-675f4bcbfc-5hg5w\" (UID: \"388dedb5-6949-4ccb-9289-520b6f9ba4ce\") " pod="openstack/dnsmasq-dns-675f4bcbfc-5hg5w" Dec 03 14:28:27 crc kubenswrapper[4805]: I1203 14:28:27.557201 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rfqvq\" (UniqueName: \"kubernetes.io/projected/6c4701fa-549e-427c-b680-2f3096aafd00-kube-api-access-rfqvq\") pod \"dnsmasq-dns-78dd6ddcc-ql7t4\" (UID: \"6c4701fa-549e-427c-b680-2f3096aafd00\") " pod="openstack/dnsmasq-dns-78dd6ddcc-ql7t4" Dec 03 14:28:27 crc kubenswrapper[4805]: I1203 14:28:27.557278 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6c4701fa-549e-427c-b680-2f3096aafd00-config\") pod \"dnsmasq-dns-78dd6ddcc-ql7t4\" (UID: \"6c4701fa-549e-427c-b680-2f3096aafd00\") " pod="openstack/dnsmasq-dns-78dd6ddcc-ql7t4" Dec 03 14:28:27 crc kubenswrapper[4805]: I1203 14:28:27.557321 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/388dedb5-6949-4ccb-9289-520b6f9ba4ce-config\") pod \"dnsmasq-dns-675f4bcbfc-5hg5w\" (UID: \"388dedb5-6949-4ccb-9289-520b6f9ba4ce\") " pod="openstack/dnsmasq-dns-675f4bcbfc-5hg5w" Dec 03 14:28:27 crc kubenswrapper[4805]: I1203 14:28:27.557357 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/6c4701fa-549e-427c-b680-2f3096aafd00-dns-svc\") pod \"dnsmasq-dns-78dd6ddcc-ql7t4\" (UID: \"6c4701fa-549e-427c-b680-2f3096aafd00\") " pod="openstack/dnsmasq-dns-78dd6ddcc-ql7t4" Dec 03 14:28:27 crc kubenswrapper[4805]: I1203 14:28:27.558193 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/388dedb5-6949-4ccb-9289-520b6f9ba4ce-config\") pod \"dnsmasq-dns-675f4bcbfc-5hg5w\" (UID: \"388dedb5-6949-4ccb-9289-520b6f9ba4ce\") " pod="openstack/dnsmasq-dns-675f4bcbfc-5hg5w" Dec 03 14:28:27 crc kubenswrapper[4805]: I1203 14:28:27.558208 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/6c4701fa-549e-427c-b680-2f3096aafd00-dns-svc\") pod \"dnsmasq-dns-78dd6ddcc-ql7t4\" (UID: \"6c4701fa-549e-427c-b680-2f3096aafd00\") " pod="openstack/dnsmasq-dns-78dd6ddcc-ql7t4" Dec 03 14:28:27 crc kubenswrapper[4805]: I1203 14:28:27.558351 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6c4701fa-549e-427c-b680-2f3096aafd00-config\") pod \"dnsmasq-dns-78dd6ddcc-ql7t4\" (UID: \"6c4701fa-549e-427c-b680-2f3096aafd00\") " pod="openstack/dnsmasq-dns-78dd6ddcc-ql7t4" Dec 03 14:28:27 crc kubenswrapper[4805]: I1203 14:28:27.578637 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7jj28\" (UniqueName: \"kubernetes.io/projected/388dedb5-6949-4ccb-9289-520b6f9ba4ce-kube-api-access-7jj28\") pod \"dnsmasq-dns-675f4bcbfc-5hg5w\" (UID: \"388dedb5-6949-4ccb-9289-520b6f9ba4ce\") " pod="openstack/dnsmasq-dns-675f4bcbfc-5hg5w" Dec 03 14:28:27 crc kubenswrapper[4805]: I1203 14:28:27.579244 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rfqvq\" (UniqueName: \"kubernetes.io/projected/6c4701fa-549e-427c-b680-2f3096aafd00-kube-api-access-rfqvq\") pod \"dnsmasq-dns-78dd6ddcc-ql7t4\" (UID: \"6c4701fa-549e-427c-b680-2f3096aafd00\") " pod="openstack/dnsmasq-dns-78dd6ddcc-ql7t4" Dec 03 14:28:27 crc kubenswrapper[4805]: I1203 14:28:27.689288 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-675f4bcbfc-5hg5w" Dec 03 14:28:27 crc kubenswrapper[4805]: I1203 14:28:27.734491 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-78dd6ddcc-ql7t4" Dec 03 14:28:28 crc kubenswrapper[4805]: I1203 14:28:28.153457 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-675f4bcbfc-5hg5w"] Dec 03 14:28:28 crc kubenswrapper[4805]: W1203 14:28:28.160047 4805 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod388dedb5_6949_4ccb_9289_520b6f9ba4ce.slice/crio-50c8fc4de0bd4a7db91845f4194f401d0ce5dd9a58c29d9d96249806e7583a12 WatchSource:0}: Error finding container 50c8fc4de0bd4a7db91845f4194f401d0ce5dd9a58c29d9d96249806e7583a12: Status 404 returned error can't find the container with id 50c8fc4de0bd4a7db91845f4194f401d0ce5dd9a58c29d9d96249806e7583a12 Dec 03 14:28:28 crc kubenswrapper[4805]: I1203 14:28:28.228061 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-78dd6ddcc-ql7t4"] Dec 03 14:28:28 crc kubenswrapper[4805]: W1203 14:28:28.230109 4805 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod6c4701fa_549e_427c_b680_2f3096aafd00.slice/crio-b646b0a0968d43dfbffc43d927635219d05dfa54928609163da0a26dc543acef WatchSource:0}: Error finding container b646b0a0968d43dfbffc43d927635219d05dfa54928609163da0a26dc543acef: Status 404 returned error can't find the container with id b646b0a0968d43dfbffc43d927635219d05dfa54928609163da0a26dc543acef Dec 03 14:28:28 crc kubenswrapper[4805]: I1203 14:28:28.669077 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-78dd6ddcc-ql7t4" event={"ID":"6c4701fa-549e-427c-b680-2f3096aafd00","Type":"ContainerStarted","Data":"b646b0a0968d43dfbffc43d927635219d05dfa54928609163da0a26dc543acef"} Dec 03 14:28:28 crc kubenswrapper[4805]: I1203 14:28:28.671356 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-675f4bcbfc-5hg5w" event={"ID":"388dedb5-6949-4ccb-9289-520b6f9ba4ce","Type":"ContainerStarted","Data":"50c8fc4de0bd4a7db91845f4194f401d0ce5dd9a58c29d9d96249806e7583a12"} Dec 03 14:28:30 crc kubenswrapper[4805]: I1203 14:28:30.511963 4805 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-675f4bcbfc-5hg5w"] Dec 03 14:28:30 crc kubenswrapper[4805]: I1203 14:28:30.562180 4805 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-666b6646f7-x75rp"] Dec 03 14:28:30 crc kubenswrapper[4805]: I1203 14:28:30.565206 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-666b6646f7-x75rp" Dec 03 14:28:30 crc kubenswrapper[4805]: I1203 14:28:30.576371 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-666b6646f7-x75rp"] Dec 03 14:28:30 crc kubenswrapper[4805]: I1203 14:28:30.613605 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-djw29\" (UniqueName: \"kubernetes.io/projected/eba109b2-1336-48fc-a849-f2d039eae8e1-kube-api-access-djw29\") pod \"dnsmasq-dns-666b6646f7-x75rp\" (UID: \"eba109b2-1336-48fc-a849-f2d039eae8e1\") " pod="openstack/dnsmasq-dns-666b6646f7-x75rp" Dec 03 14:28:30 crc kubenswrapper[4805]: I1203 14:28:30.613663 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/eba109b2-1336-48fc-a849-f2d039eae8e1-dns-svc\") pod \"dnsmasq-dns-666b6646f7-x75rp\" (UID: \"eba109b2-1336-48fc-a849-f2d039eae8e1\") " pod="openstack/dnsmasq-dns-666b6646f7-x75rp" Dec 03 14:28:30 crc kubenswrapper[4805]: I1203 14:28:30.613703 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/eba109b2-1336-48fc-a849-f2d039eae8e1-config\") pod \"dnsmasq-dns-666b6646f7-x75rp\" (UID: \"eba109b2-1336-48fc-a849-f2d039eae8e1\") " pod="openstack/dnsmasq-dns-666b6646f7-x75rp" Dec 03 14:28:30 crc kubenswrapper[4805]: I1203 14:28:30.715062 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-djw29\" (UniqueName: \"kubernetes.io/projected/eba109b2-1336-48fc-a849-f2d039eae8e1-kube-api-access-djw29\") pod \"dnsmasq-dns-666b6646f7-x75rp\" (UID: \"eba109b2-1336-48fc-a849-f2d039eae8e1\") " pod="openstack/dnsmasq-dns-666b6646f7-x75rp" Dec 03 14:28:30 crc kubenswrapper[4805]: I1203 14:28:30.715415 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/eba109b2-1336-48fc-a849-f2d039eae8e1-dns-svc\") pod \"dnsmasq-dns-666b6646f7-x75rp\" (UID: \"eba109b2-1336-48fc-a849-f2d039eae8e1\") " pod="openstack/dnsmasq-dns-666b6646f7-x75rp" Dec 03 14:28:30 crc kubenswrapper[4805]: I1203 14:28:30.715544 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/eba109b2-1336-48fc-a849-f2d039eae8e1-config\") pod \"dnsmasq-dns-666b6646f7-x75rp\" (UID: \"eba109b2-1336-48fc-a849-f2d039eae8e1\") " pod="openstack/dnsmasq-dns-666b6646f7-x75rp" Dec 03 14:28:30 crc kubenswrapper[4805]: I1203 14:28:30.716319 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/eba109b2-1336-48fc-a849-f2d039eae8e1-dns-svc\") pod \"dnsmasq-dns-666b6646f7-x75rp\" (UID: \"eba109b2-1336-48fc-a849-f2d039eae8e1\") " pod="openstack/dnsmasq-dns-666b6646f7-x75rp" Dec 03 14:28:30 crc kubenswrapper[4805]: I1203 14:28:30.717188 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/eba109b2-1336-48fc-a849-f2d039eae8e1-config\") pod \"dnsmasq-dns-666b6646f7-x75rp\" (UID: \"eba109b2-1336-48fc-a849-f2d039eae8e1\") " pod="openstack/dnsmasq-dns-666b6646f7-x75rp" Dec 03 14:28:30 crc kubenswrapper[4805]: I1203 14:28:30.762569 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-djw29\" (UniqueName: \"kubernetes.io/projected/eba109b2-1336-48fc-a849-f2d039eae8e1-kube-api-access-djw29\") pod \"dnsmasq-dns-666b6646f7-x75rp\" (UID: \"eba109b2-1336-48fc-a849-f2d039eae8e1\") " pod="openstack/dnsmasq-dns-666b6646f7-x75rp" Dec 03 14:28:30 crc kubenswrapper[4805]: I1203 14:28:30.895887 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-666b6646f7-x75rp" Dec 03 14:28:30 crc kubenswrapper[4805]: I1203 14:28:30.896912 4805 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-78dd6ddcc-ql7t4"] Dec 03 14:28:30 crc kubenswrapper[4805]: I1203 14:28:30.921214 4805 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-57d769cc4f-fthjg"] Dec 03 14:28:30 crc kubenswrapper[4805]: I1203 14:28:30.922604 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-57d769cc4f-fthjg" Dec 03 14:28:30 crc kubenswrapper[4805]: I1203 14:28:30.932901 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-57d769cc4f-fthjg"] Dec 03 14:28:31 crc kubenswrapper[4805]: I1203 14:28:31.021934 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8jl2l\" (UniqueName: \"kubernetes.io/projected/2af0398e-b188-49ec-ae9e-391cb1122753-kube-api-access-8jl2l\") pod \"dnsmasq-dns-57d769cc4f-fthjg\" (UID: \"2af0398e-b188-49ec-ae9e-391cb1122753\") " pod="openstack/dnsmasq-dns-57d769cc4f-fthjg" Dec 03 14:28:31 crc kubenswrapper[4805]: I1203 14:28:31.022146 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2af0398e-b188-49ec-ae9e-391cb1122753-config\") pod \"dnsmasq-dns-57d769cc4f-fthjg\" (UID: \"2af0398e-b188-49ec-ae9e-391cb1122753\") " pod="openstack/dnsmasq-dns-57d769cc4f-fthjg" Dec 03 14:28:31 crc kubenswrapper[4805]: I1203 14:28:31.022257 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/2af0398e-b188-49ec-ae9e-391cb1122753-dns-svc\") pod \"dnsmasq-dns-57d769cc4f-fthjg\" (UID: \"2af0398e-b188-49ec-ae9e-391cb1122753\") " pod="openstack/dnsmasq-dns-57d769cc4f-fthjg" Dec 03 14:28:31 crc kubenswrapper[4805]: I1203 14:28:31.123376 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/2af0398e-b188-49ec-ae9e-391cb1122753-dns-svc\") pod \"dnsmasq-dns-57d769cc4f-fthjg\" (UID: \"2af0398e-b188-49ec-ae9e-391cb1122753\") " pod="openstack/dnsmasq-dns-57d769cc4f-fthjg" Dec 03 14:28:31 crc kubenswrapper[4805]: I1203 14:28:31.123483 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8jl2l\" (UniqueName: \"kubernetes.io/projected/2af0398e-b188-49ec-ae9e-391cb1122753-kube-api-access-8jl2l\") pod \"dnsmasq-dns-57d769cc4f-fthjg\" (UID: \"2af0398e-b188-49ec-ae9e-391cb1122753\") " pod="openstack/dnsmasq-dns-57d769cc4f-fthjg" Dec 03 14:28:31 crc kubenswrapper[4805]: I1203 14:28:31.123532 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2af0398e-b188-49ec-ae9e-391cb1122753-config\") pod \"dnsmasq-dns-57d769cc4f-fthjg\" (UID: \"2af0398e-b188-49ec-ae9e-391cb1122753\") " pod="openstack/dnsmasq-dns-57d769cc4f-fthjg" Dec 03 14:28:31 crc kubenswrapper[4805]: I1203 14:28:31.124511 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2af0398e-b188-49ec-ae9e-391cb1122753-config\") pod \"dnsmasq-dns-57d769cc4f-fthjg\" (UID: \"2af0398e-b188-49ec-ae9e-391cb1122753\") " pod="openstack/dnsmasq-dns-57d769cc4f-fthjg" Dec 03 14:28:31 crc kubenswrapper[4805]: I1203 14:28:31.124729 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/2af0398e-b188-49ec-ae9e-391cb1122753-dns-svc\") pod \"dnsmasq-dns-57d769cc4f-fthjg\" (UID: \"2af0398e-b188-49ec-ae9e-391cb1122753\") " pod="openstack/dnsmasq-dns-57d769cc4f-fthjg" Dec 03 14:28:31 crc kubenswrapper[4805]: I1203 14:28:31.141669 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8jl2l\" (UniqueName: \"kubernetes.io/projected/2af0398e-b188-49ec-ae9e-391cb1122753-kube-api-access-8jl2l\") pod \"dnsmasq-dns-57d769cc4f-fthjg\" (UID: \"2af0398e-b188-49ec-ae9e-391cb1122753\") " pod="openstack/dnsmasq-dns-57d769cc4f-fthjg" Dec 03 14:28:31 crc kubenswrapper[4805]: I1203 14:28:31.266905 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-57d769cc4f-fthjg" Dec 03 14:28:31 crc kubenswrapper[4805]: I1203 14:28:31.722960 4805 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/rabbitmq-server-0"] Dec 03 14:28:31 crc kubenswrapper[4805]: I1203 14:28:31.726294 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Dec 03 14:28:31 crc kubenswrapper[4805]: I1203 14:28:31.729615 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-default-user" Dec 03 14:28:31 crc kubenswrapper[4805]: I1203 14:28:31.729795 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-rabbitmq-svc" Dec 03 14:28:31 crc kubenswrapper[4805]: I1203 14:28:31.729946 4805 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-config-data" Dec 03 14:28:31 crc kubenswrapper[4805]: I1203 14:28:31.730263 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-server-dockercfg-7db6l" Dec 03 14:28:31 crc kubenswrapper[4805]: I1203 14:28:31.730415 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-erlang-cookie" Dec 03 14:28:31 crc kubenswrapper[4805]: I1203 14:28:31.730652 4805 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-server-conf" Dec 03 14:28:31 crc kubenswrapper[4805]: I1203 14:28:31.730766 4805 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-plugins-conf" Dec 03 14:28:31 crc kubenswrapper[4805]: I1203 14:28:31.742787 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-server-0"] Dec 03 14:28:31 crc kubenswrapper[4805]: I1203 14:28:31.831715 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tbx6c\" (UniqueName: \"kubernetes.io/projected/1b34fa5c-2867-4dbe-9285-80c327699864-kube-api-access-tbx6c\") pod \"rabbitmq-server-0\" (UID: \"1b34fa5c-2867-4dbe-9285-80c327699864\") " pod="openstack/rabbitmq-server-0" Dec 03 14:28:31 crc kubenswrapper[4805]: I1203 14:28:31.831768 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/1b34fa5c-2867-4dbe-9285-80c327699864-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"1b34fa5c-2867-4dbe-9285-80c327699864\") " pod="openstack/rabbitmq-server-0" Dec 03 14:28:31 crc kubenswrapper[4805]: I1203 14:28:31.831795 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/1b34fa5c-2867-4dbe-9285-80c327699864-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"1b34fa5c-2867-4dbe-9285-80c327699864\") " pod="openstack/rabbitmq-server-0" Dec 03 14:28:31 crc kubenswrapper[4805]: I1203 14:28:31.831856 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/1b34fa5c-2867-4dbe-9285-80c327699864-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"1b34fa5c-2867-4dbe-9285-80c327699864\") " pod="openstack/rabbitmq-server-0" Dec 03 14:28:31 crc kubenswrapper[4805]: I1203 14:28:31.831884 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/1b34fa5c-2867-4dbe-9285-80c327699864-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"1b34fa5c-2867-4dbe-9285-80c327699864\") " pod="openstack/rabbitmq-server-0" Dec 03 14:28:31 crc kubenswrapper[4805]: I1203 14:28:31.831910 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/1b34fa5c-2867-4dbe-9285-80c327699864-config-data\") pod \"rabbitmq-server-0\" (UID: \"1b34fa5c-2867-4dbe-9285-80c327699864\") " pod="openstack/rabbitmq-server-0" Dec 03 14:28:31 crc kubenswrapper[4805]: I1203 14:28:31.831947 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/1b34fa5c-2867-4dbe-9285-80c327699864-server-conf\") pod \"rabbitmq-server-0\" (UID: \"1b34fa5c-2867-4dbe-9285-80c327699864\") " pod="openstack/rabbitmq-server-0" Dec 03 14:28:31 crc kubenswrapper[4805]: I1203 14:28:31.831974 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/1b34fa5c-2867-4dbe-9285-80c327699864-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"1b34fa5c-2867-4dbe-9285-80c327699864\") " pod="openstack/rabbitmq-server-0" Dec 03 14:28:31 crc kubenswrapper[4805]: I1203 14:28:31.832012 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/1b34fa5c-2867-4dbe-9285-80c327699864-pod-info\") pod \"rabbitmq-server-0\" (UID: \"1b34fa5c-2867-4dbe-9285-80c327699864\") " pod="openstack/rabbitmq-server-0" Dec 03 14:28:31 crc kubenswrapper[4805]: I1203 14:28:31.832042 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/1b34fa5c-2867-4dbe-9285-80c327699864-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"1b34fa5c-2867-4dbe-9285-80c327699864\") " pod="openstack/rabbitmq-server-0" Dec 03 14:28:31 crc kubenswrapper[4805]: I1203 14:28:31.832070 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"rabbitmq-server-0\" (UID: \"1b34fa5c-2867-4dbe-9285-80c327699864\") " pod="openstack/rabbitmq-server-0" Dec 03 14:28:31 crc kubenswrapper[4805]: I1203 14:28:31.940679 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/1b34fa5c-2867-4dbe-9285-80c327699864-server-conf\") pod \"rabbitmq-server-0\" (UID: \"1b34fa5c-2867-4dbe-9285-80c327699864\") " pod="openstack/rabbitmq-server-0" Dec 03 14:28:31 crc kubenswrapper[4805]: I1203 14:28:31.940736 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/1b34fa5c-2867-4dbe-9285-80c327699864-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"1b34fa5c-2867-4dbe-9285-80c327699864\") " pod="openstack/rabbitmq-server-0" Dec 03 14:28:31 crc kubenswrapper[4805]: I1203 14:28:31.940774 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/1b34fa5c-2867-4dbe-9285-80c327699864-pod-info\") pod \"rabbitmq-server-0\" (UID: \"1b34fa5c-2867-4dbe-9285-80c327699864\") " pod="openstack/rabbitmq-server-0" Dec 03 14:28:31 crc kubenswrapper[4805]: I1203 14:28:31.940802 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/1b34fa5c-2867-4dbe-9285-80c327699864-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"1b34fa5c-2867-4dbe-9285-80c327699864\") " pod="openstack/rabbitmq-server-0" Dec 03 14:28:31 crc kubenswrapper[4805]: I1203 14:28:31.940830 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"rabbitmq-server-0\" (UID: \"1b34fa5c-2867-4dbe-9285-80c327699864\") " pod="openstack/rabbitmq-server-0" Dec 03 14:28:31 crc kubenswrapper[4805]: I1203 14:28:31.940887 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/1b34fa5c-2867-4dbe-9285-80c327699864-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"1b34fa5c-2867-4dbe-9285-80c327699864\") " pod="openstack/rabbitmq-server-0" Dec 03 14:28:31 crc kubenswrapper[4805]: I1203 14:28:31.940907 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tbx6c\" (UniqueName: \"kubernetes.io/projected/1b34fa5c-2867-4dbe-9285-80c327699864-kube-api-access-tbx6c\") pod \"rabbitmq-server-0\" (UID: \"1b34fa5c-2867-4dbe-9285-80c327699864\") " pod="openstack/rabbitmq-server-0" Dec 03 14:28:31 crc kubenswrapper[4805]: I1203 14:28:31.940934 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/1b34fa5c-2867-4dbe-9285-80c327699864-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"1b34fa5c-2867-4dbe-9285-80c327699864\") " pod="openstack/rabbitmq-server-0" Dec 03 14:28:31 crc kubenswrapper[4805]: I1203 14:28:31.940975 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/1b34fa5c-2867-4dbe-9285-80c327699864-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"1b34fa5c-2867-4dbe-9285-80c327699864\") " pod="openstack/rabbitmq-server-0" Dec 03 14:28:31 crc kubenswrapper[4805]: I1203 14:28:31.940999 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/1b34fa5c-2867-4dbe-9285-80c327699864-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"1b34fa5c-2867-4dbe-9285-80c327699864\") " pod="openstack/rabbitmq-server-0" Dec 03 14:28:31 crc kubenswrapper[4805]: I1203 14:28:31.941023 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/1b34fa5c-2867-4dbe-9285-80c327699864-config-data\") pod \"rabbitmq-server-0\" (UID: \"1b34fa5c-2867-4dbe-9285-80c327699864\") " pod="openstack/rabbitmq-server-0" Dec 03 14:28:31 crc kubenswrapper[4805]: I1203 14:28:31.942276 4805 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"rabbitmq-server-0\" (UID: \"1b34fa5c-2867-4dbe-9285-80c327699864\") device mount path \"/mnt/openstack/pv03\"" pod="openstack/rabbitmq-server-0" Dec 03 14:28:31 crc kubenswrapper[4805]: I1203 14:28:31.942751 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/1b34fa5c-2867-4dbe-9285-80c327699864-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"1b34fa5c-2867-4dbe-9285-80c327699864\") " pod="openstack/rabbitmq-server-0" Dec 03 14:28:31 crc kubenswrapper[4805]: I1203 14:28:31.942801 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/1b34fa5c-2867-4dbe-9285-80c327699864-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"1b34fa5c-2867-4dbe-9285-80c327699864\") " pod="openstack/rabbitmq-server-0" Dec 03 14:28:31 crc kubenswrapper[4805]: I1203 14:28:31.942891 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/1b34fa5c-2867-4dbe-9285-80c327699864-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"1b34fa5c-2867-4dbe-9285-80c327699864\") " pod="openstack/rabbitmq-server-0" Dec 03 14:28:31 crc kubenswrapper[4805]: I1203 14:28:31.943277 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/1b34fa5c-2867-4dbe-9285-80c327699864-config-data\") pod \"rabbitmq-server-0\" (UID: \"1b34fa5c-2867-4dbe-9285-80c327699864\") " pod="openstack/rabbitmq-server-0" Dec 03 14:28:31 crc kubenswrapper[4805]: I1203 14:28:31.943804 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/1b34fa5c-2867-4dbe-9285-80c327699864-server-conf\") pod \"rabbitmq-server-0\" (UID: \"1b34fa5c-2867-4dbe-9285-80c327699864\") " pod="openstack/rabbitmq-server-0" Dec 03 14:28:31 crc kubenswrapper[4805]: I1203 14:28:31.949371 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/1b34fa5c-2867-4dbe-9285-80c327699864-pod-info\") pod \"rabbitmq-server-0\" (UID: \"1b34fa5c-2867-4dbe-9285-80c327699864\") " pod="openstack/rabbitmq-server-0" Dec 03 14:28:31 crc kubenswrapper[4805]: I1203 14:28:31.951309 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/1b34fa5c-2867-4dbe-9285-80c327699864-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"1b34fa5c-2867-4dbe-9285-80c327699864\") " pod="openstack/rabbitmq-server-0" Dec 03 14:28:31 crc kubenswrapper[4805]: I1203 14:28:31.951322 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/1b34fa5c-2867-4dbe-9285-80c327699864-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"1b34fa5c-2867-4dbe-9285-80c327699864\") " pod="openstack/rabbitmq-server-0" Dec 03 14:28:31 crc kubenswrapper[4805]: I1203 14:28:31.953179 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/1b34fa5c-2867-4dbe-9285-80c327699864-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"1b34fa5c-2867-4dbe-9285-80c327699864\") " pod="openstack/rabbitmq-server-0" Dec 03 14:28:31 crc kubenswrapper[4805]: I1203 14:28:31.962095 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tbx6c\" (UniqueName: \"kubernetes.io/projected/1b34fa5c-2867-4dbe-9285-80c327699864-kube-api-access-tbx6c\") pod \"rabbitmq-server-0\" (UID: \"1b34fa5c-2867-4dbe-9285-80c327699864\") " pod="openstack/rabbitmq-server-0" Dec 03 14:28:31 crc kubenswrapper[4805]: I1203 14:28:31.974648 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"rabbitmq-server-0\" (UID: \"1b34fa5c-2867-4dbe-9285-80c327699864\") " pod="openstack/rabbitmq-server-0" Dec 03 14:28:32 crc kubenswrapper[4805]: I1203 14:28:32.032574 4805 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Dec 03 14:28:32 crc kubenswrapper[4805]: I1203 14:28:32.034225 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Dec 03 14:28:32 crc kubenswrapper[4805]: I1203 14:28:32.035557 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-default-user" Dec 03 14:28:32 crc kubenswrapper[4805]: I1203 14:28:32.036187 4805 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-server-conf" Dec 03 14:28:32 crc kubenswrapper[4805]: I1203 14:28:32.036212 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-erlang-cookie" Dec 03 14:28:32 crc kubenswrapper[4805]: I1203 14:28:32.036324 4805 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-plugins-conf" Dec 03 14:28:32 crc kubenswrapper[4805]: I1203 14:28:32.036375 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-server-dockercfg-vpb8f" Dec 03 14:28:32 crc kubenswrapper[4805]: I1203 14:28:32.036422 4805 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-config-data" Dec 03 14:28:32 crc kubenswrapper[4805]: I1203 14:28:32.038748 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-rabbitmq-cell1-svc" Dec 03 14:28:32 crc kubenswrapper[4805]: I1203 14:28:32.048242 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Dec 03 14:28:32 crc kubenswrapper[4805]: I1203 14:28:32.067732 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Dec 03 14:28:32 crc kubenswrapper[4805]: I1203 14:28:32.145157 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/656f691c-1188-47b9-b8dd-b393a230602a-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"656f691c-1188-47b9-b8dd-b393a230602a\") " pod="openstack/rabbitmq-cell1-server-0" Dec 03 14:28:32 crc kubenswrapper[4805]: I1203 14:28:32.145206 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/656f691c-1188-47b9-b8dd-b393a230602a-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"656f691c-1188-47b9-b8dd-b393a230602a\") " pod="openstack/rabbitmq-cell1-server-0" Dec 03 14:28:32 crc kubenswrapper[4805]: I1203 14:28:32.145260 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/656f691c-1188-47b9-b8dd-b393a230602a-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"656f691c-1188-47b9-b8dd-b393a230602a\") " pod="openstack/rabbitmq-cell1-server-0" Dec 03 14:28:32 crc kubenswrapper[4805]: I1203 14:28:32.145292 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/656f691c-1188-47b9-b8dd-b393a230602a-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"656f691c-1188-47b9-b8dd-b393a230602a\") " pod="openstack/rabbitmq-cell1-server-0" Dec 03 14:28:32 crc kubenswrapper[4805]: I1203 14:28:32.145318 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/656f691c-1188-47b9-b8dd-b393a230602a-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"656f691c-1188-47b9-b8dd-b393a230602a\") " pod="openstack/rabbitmq-cell1-server-0" Dec 03 14:28:32 crc kubenswrapper[4805]: I1203 14:28:32.145334 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/656f691c-1188-47b9-b8dd-b393a230602a-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"656f691c-1188-47b9-b8dd-b393a230602a\") " pod="openstack/rabbitmq-cell1-server-0" Dec 03 14:28:32 crc kubenswrapper[4805]: I1203 14:28:32.145366 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9qnqg\" (UniqueName: \"kubernetes.io/projected/656f691c-1188-47b9-b8dd-b393a230602a-kube-api-access-9qnqg\") pod \"rabbitmq-cell1-server-0\" (UID: \"656f691c-1188-47b9-b8dd-b393a230602a\") " pod="openstack/rabbitmq-cell1-server-0" Dec 03 14:28:32 crc kubenswrapper[4805]: I1203 14:28:32.145382 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/656f691c-1188-47b9-b8dd-b393a230602a-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"656f691c-1188-47b9-b8dd-b393a230602a\") " pod="openstack/rabbitmq-cell1-server-0" Dec 03 14:28:32 crc kubenswrapper[4805]: I1203 14:28:32.145472 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"656f691c-1188-47b9-b8dd-b393a230602a\") " pod="openstack/rabbitmq-cell1-server-0" Dec 03 14:28:32 crc kubenswrapper[4805]: I1203 14:28:32.145629 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/656f691c-1188-47b9-b8dd-b393a230602a-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"656f691c-1188-47b9-b8dd-b393a230602a\") " pod="openstack/rabbitmq-cell1-server-0" Dec 03 14:28:32 crc kubenswrapper[4805]: I1203 14:28:32.145663 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/656f691c-1188-47b9-b8dd-b393a230602a-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"656f691c-1188-47b9-b8dd-b393a230602a\") " pod="openstack/rabbitmq-cell1-server-0" Dec 03 14:28:32 crc kubenswrapper[4805]: I1203 14:28:32.246939 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/656f691c-1188-47b9-b8dd-b393a230602a-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"656f691c-1188-47b9-b8dd-b393a230602a\") " pod="openstack/rabbitmq-cell1-server-0" Dec 03 14:28:32 crc kubenswrapper[4805]: I1203 14:28:32.247007 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/656f691c-1188-47b9-b8dd-b393a230602a-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"656f691c-1188-47b9-b8dd-b393a230602a\") " pod="openstack/rabbitmq-cell1-server-0" Dec 03 14:28:32 crc kubenswrapper[4805]: I1203 14:28:32.247033 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/656f691c-1188-47b9-b8dd-b393a230602a-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"656f691c-1188-47b9-b8dd-b393a230602a\") " pod="openstack/rabbitmq-cell1-server-0" Dec 03 14:28:32 crc kubenswrapper[4805]: I1203 14:28:32.247054 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/656f691c-1188-47b9-b8dd-b393a230602a-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"656f691c-1188-47b9-b8dd-b393a230602a\") " pod="openstack/rabbitmq-cell1-server-0" Dec 03 14:28:32 crc kubenswrapper[4805]: I1203 14:28:32.247096 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/656f691c-1188-47b9-b8dd-b393a230602a-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"656f691c-1188-47b9-b8dd-b393a230602a\") " pod="openstack/rabbitmq-cell1-server-0" Dec 03 14:28:32 crc kubenswrapper[4805]: I1203 14:28:32.247116 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9qnqg\" (UniqueName: \"kubernetes.io/projected/656f691c-1188-47b9-b8dd-b393a230602a-kube-api-access-9qnqg\") pod \"rabbitmq-cell1-server-0\" (UID: \"656f691c-1188-47b9-b8dd-b393a230602a\") " pod="openstack/rabbitmq-cell1-server-0" Dec 03 14:28:32 crc kubenswrapper[4805]: I1203 14:28:32.247143 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"656f691c-1188-47b9-b8dd-b393a230602a\") " pod="openstack/rabbitmq-cell1-server-0" Dec 03 14:28:32 crc kubenswrapper[4805]: I1203 14:28:32.247180 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/656f691c-1188-47b9-b8dd-b393a230602a-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"656f691c-1188-47b9-b8dd-b393a230602a\") " pod="openstack/rabbitmq-cell1-server-0" Dec 03 14:28:32 crc kubenswrapper[4805]: I1203 14:28:32.247201 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/656f691c-1188-47b9-b8dd-b393a230602a-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"656f691c-1188-47b9-b8dd-b393a230602a\") " pod="openstack/rabbitmq-cell1-server-0" Dec 03 14:28:32 crc kubenswrapper[4805]: I1203 14:28:32.247226 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/656f691c-1188-47b9-b8dd-b393a230602a-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"656f691c-1188-47b9-b8dd-b393a230602a\") " pod="openstack/rabbitmq-cell1-server-0" Dec 03 14:28:32 crc kubenswrapper[4805]: I1203 14:28:32.247253 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/656f691c-1188-47b9-b8dd-b393a230602a-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"656f691c-1188-47b9-b8dd-b393a230602a\") " pod="openstack/rabbitmq-cell1-server-0" Dec 03 14:28:32 crc kubenswrapper[4805]: I1203 14:28:32.249259 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/656f691c-1188-47b9-b8dd-b393a230602a-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"656f691c-1188-47b9-b8dd-b393a230602a\") " pod="openstack/rabbitmq-cell1-server-0" Dec 03 14:28:32 crc kubenswrapper[4805]: I1203 14:28:32.249948 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/656f691c-1188-47b9-b8dd-b393a230602a-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"656f691c-1188-47b9-b8dd-b393a230602a\") " pod="openstack/rabbitmq-cell1-server-0" Dec 03 14:28:32 crc kubenswrapper[4805]: I1203 14:28:32.249973 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/656f691c-1188-47b9-b8dd-b393a230602a-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"656f691c-1188-47b9-b8dd-b393a230602a\") " pod="openstack/rabbitmq-cell1-server-0" Dec 03 14:28:32 crc kubenswrapper[4805]: I1203 14:28:32.250546 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/656f691c-1188-47b9-b8dd-b393a230602a-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"656f691c-1188-47b9-b8dd-b393a230602a\") " pod="openstack/rabbitmq-cell1-server-0" Dec 03 14:28:32 crc kubenswrapper[4805]: I1203 14:28:32.251462 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/656f691c-1188-47b9-b8dd-b393a230602a-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"656f691c-1188-47b9-b8dd-b393a230602a\") " pod="openstack/rabbitmq-cell1-server-0" Dec 03 14:28:32 crc kubenswrapper[4805]: I1203 14:28:32.251755 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/656f691c-1188-47b9-b8dd-b393a230602a-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"656f691c-1188-47b9-b8dd-b393a230602a\") " pod="openstack/rabbitmq-cell1-server-0" Dec 03 14:28:32 crc kubenswrapper[4805]: I1203 14:28:32.252243 4805 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"656f691c-1188-47b9-b8dd-b393a230602a\") device mount path \"/mnt/openstack/pv01\"" pod="openstack/rabbitmq-cell1-server-0" Dec 03 14:28:32 crc kubenswrapper[4805]: I1203 14:28:32.252445 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/656f691c-1188-47b9-b8dd-b393a230602a-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"656f691c-1188-47b9-b8dd-b393a230602a\") " pod="openstack/rabbitmq-cell1-server-0" Dec 03 14:28:32 crc kubenswrapper[4805]: I1203 14:28:32.254342 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/656f691c-1188-47b9-b8dd-b393a230602a-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"656f691c-1188-47b9-b8dd-b393a230602a\") " pod="openstack/rabbitmq-cell1-server-0" Dec 03 14:28:32 crc kubenswrapper[4805]: I1203 14:28:32.259209 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/656f691c-1188-47b9-b8dd-b393a230602a-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"656f691c-1188-47b9-b8dd-b393a230602a\") " pod="openstack/rabbitmq-cell1-server-0" Dec 03 14:28:32 crc kubenswrapper[4805]: I1203 14:28:32.273265 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9qnqg\" (UniqueName: \"kubernetes.io/projected/656f691c-1188-47b9-b8dd-b393a230602a-kube-api-access-9qnqg\") pod \"rabbitmq-cell1-server-0\" (UID: \"656f691c-1188-47b9-b8dd-b393a230602a\") " pod="openstack/rabbitmq-cell1-server-0" Dec 03 14:28:32 crc kubenswrapper[4805]: I1203 14:28:32.292478 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"656f691c-1188-47b9-b8dd-b393a230602a\") " pod="openstack/rabbitmq-cell1-server-0" Dec 03 14:28:32 crc kubenswrapper[4805]: I1203 14:28:32.363418 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Dec 03 14:28:33 crc kubenswrapper[4805]: I1203 14:28:33.348489 4805 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/openstack-galera-0"] Dec 03 14:28:33 crc kubenswrapper[4805]: I1203 14:28:33.350606 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-galera-0" Dec 03 14:28:33 crc kubenswrapper[4805]: I1203 14:28:33.358708 4805 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-config-data" Dec 03 14:28:33 crc kubenswrapper[4805]: I1203 14:28:33.358986 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"galera-openstack-dockercfg-9q9gg" Dec 03 14:28:33 crc kubenswrapper[4805]: I1203 14:28:33.363125 4805 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-scripts" Dec 03 14:28:33 crc kubenswrapper[4805]: I1203 14:28:33.363425 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-galera-openstack-svc" Dec 03 14:28:33 crc kubenswrapper[4805]: I1203 14:28:33.406479 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"combined-ca-bundle" Dec 03 14:28:33 crc kubenswrapper[4805]: I1203 14:28:33.447792 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstack-galera-0"] Dec 03 14:28:33 crc kubenswrapper[4805]: I1203 14:28:33.472191 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"openstack-galera-0\" (UID: \"5b4bb902-dba6-4ec6-97bf-532d996e918d\") " pod="openstack/openstack-galera-0" Dec 03 14:28:33 crc kubenswrapper[4805]: I1203 14:28:33.472280 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/5b4bb902-dba6-4ec6-97bf-532d996e918d-kolla-config\") pod \"openstack-galera-0\" (UID: \"5b4bb902-dba6-4ec6-97bf-532d996e918d\") " pod="openstack/openstack-galera-0" Dec 03 14:28:33 crc kubenswrapper[4805]: I1203 14:28:33.472315 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5b4bb902-dba6-4ec6-97bf-532d996e918d-combined-ca-bundle\") pod \"openstack-galera-0\" (UID: \"5b4bb902-dba6-4ec6-97bf-532d996e918d\") " pod="openstack/openstack-galera-0" Dec 03 14:28:33 crc kubenswrapper[4805]: I1203 14:28:33.472346 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/5b4bb902-dba6-4ec6-97bf-532d996e918d-config-data-default\") pod \"openstack-galera-0\" (UID: \"5b4bb902-dba6-4ec6-97bf-532d996e918d\") " pod="openstack/openstack-galera-0" Dec 03 14:28:33 crc kubenswrapper[4805]: I1203 14:28:33.472384 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/5b4bb902-dba6-4ec6-97bf-532d996e918d-operator-scripts\") pod \"openstack-galera-0\" (UID: \"5b4bb902-dba6-4ec6-97bf-532d996e918d\") " pod="openstack/openstack-galera-0" Dec 03 14:28:33 crc kubenswrapper[4805]: I1203 14:28:33.472441 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/5b4bb902-dba6-4ec6-97bf-532d996e918d-galera-tls-certs\") pod \"openstack-galera-0\" (UID: \"5b4bb902-dba6-4ec6-97bf-532d996e918d\") " pod="openstack/openstack-galera-0" Dec 03 14:28:33 crc kubenswrapper[4805]: I1203 14:28:33.472465 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8flv5\" (UniqueName: \"kubernetes.io/projected/5b4bb902-dba6-4ec6-97bf-532d996e918d-kube-api-access-8flv5\") pod \"openstack-galera-0\" (UID: \"5b4bb902-dba6-4ec6-97bf-532d996e918d\") " pod="openstack/openstack-galera-0" Dec 03 14:28:33 crc kubenswrapper[4805]: I1203 14:28:33.472492 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/5b4bb902-dba6-4ec6-97bf-532d996e918d-config-data-generated\") pod \"openstack-galera-0\" (UID: \"5b4bb902-dba6-4ec6-97bf-532d996e918d\") " pod="openstack/openstack-galera-0" Dec 03 14:28:33 crc kubenswrapper[4805]: I1203 14:28:33.574241 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/5b4bb902-dba6-4ec6-97bf-532d996e918d-galera-tls-certs\") pod \"openstack-galera-0\" (UID: \"5b4bb902-dba6-4ec6-97bf-532d996e918d\") " pod="openstack/openstack-galera-0" Dec 03 14:28:33 crc kubenswrapper[4805]: I1203 14:28:33.574296 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8flv5\" (UniqueName: \"kubernetes.io/projected/5b4bb902-dba6-4ec6-97bf-532d996e918d-kube-api-access-8flv5\") pod \"openstack-galera-0\" (UID: \"5b4bb902-dba6-4ec6-97bf-532d996e918d\") " pod="openstack/openstack-galera-0" Dec 03 14:28:33 crc kubenswrapper[4805]: I1203 14:28:33.574329 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/5b4bb902-dba6-4ec6-97bf-532d996e918d-config-data-generated\") pod \"openstack-galera-0\" (UID: \"5b4bb902-dba6-4ec6-97bf-532d996e918d\") " pod="openstack/openstack-galera-0" Dec 03 14:28:33 crc kubenswrapper[4805]: I1203 14:28:33.574363 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"openstack-galera-0\" (UID: \"5b4bb902-dba6-4ec6-97bf-532d996e918d\") " pod="openstack/openstack-galera-0" Dec 03 14:28:33 crc kubenswrapper[4805]: I1203 14:28:33.574404 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/5b4bb902-dba6-4ec6-97bf-532d996e918d-kolla-config\") pod \"openstack-galera-0\" (UID: \"5b4bb902-dba6-4ec6-97bf-532d996e918d\") " pod="openstack/openstack-galera-0" Dec 03 14:28:33 crc kubenswrapper[4805]: I1203 14:28:33.574433 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5b4bb902-dba6-4ec6-97bf-532d996e918d-combined-ca-bundle\") pod \"openstack-galera-0\" (UID: \"5b4bb902-dba6-4ec6-97bf-532d996e918d\") " pod="openstack/openstack-galera-0" Dec 03 14:28:33 crc kubenswrapper[4805]: I1203 14:28:33.574461 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/5b4bb902-dba6-4ec6-97bf-532d996e918d-config-data-default\") pod \"openstack-galera-0\" (UID: \"5b4bb902-dba6-4ec6-97bf-532d996e918d\") " pod="openstack/openstack-galera-0" Dec 03 14:28:33 crc kubenswrapper[4805]: I1203 14:28:33.574507 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/5b4bb902-dba6-4ec6-97bf-532d996e918d-operator-scripts\") pod \"openstack-galera-0\" (UID: \"5b4bb902-dba6-4ec6-97bf-532d996e918d\") " pod="openstack/openstack-galera-0" Dec 03 14:28:33 crc kubenswrapper[4805]: I1203 14:28:33.575419 4805 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"openstack-galera-0\" (UID: \"5b4bb902-dba6-4ec6-97bf-532d996e918d\") device mount path \"/mnt/openstack/pv05\"" pod="openstack/openstack-galera-0" Dec 03 14:28:33 crc kubenswrapper[4805]: I1203 14:28:33.575909 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/5b4bb902-dba6-4ec6-97bf-532d996e918d-config-data-generated\") pod \"openstack-galera-0\" (UID: \"5b4bb902-dba6-4ec6-97bf-532d996e918d\") " pod="openstack/openstack-galera-0" Dec 03 14:28:33 crc kubenswrapper[4805]: I1203 14:28:33.576094 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/5b4bb902-dba6-4ec6-97bf-532d996e918d-kolla-config\") pod \"openstack-galera-0\" (UID: \"5b4bb902-dba6-4ec6-97bf-532d996e918d\") " pod="openstack/openstack-galera-0" Dec 03 14:28:33 crc kubenswrapper[4805]: I1203 14:28:33.576477 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/5b4bb902-dba6-4ec6-97bf-532d996e918d-operator-scripts\") pod \"openstack-galera-0\" (UID: \"5b4bb902-dba6-4ec6-97bf-532d996e918d\") " pod="openstack/openstack-galera-0" Dec 03 14:28:33 crc kubenswrapper[4805]: I1203 14:28:33.576488 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/5b4bb902-dba6-4ec6-97bf-532d996e918d-config-data-default\") pod \"openstack-galera-0\" (UID: \"5b4bb902-dba6-4ec6-97bf-532d996e918d\") " pod="openstack/openstack-galera-0" Dec 03 14:28:33 crc kubenswrapper[4805]: I1203 14:28:33.580515 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/5b4bb902-dba6-4ec6-97bf-532d996e918d-galera-tls-certs\") pod \"openstack-galera-0\" (UID: \"5b4bb902-dba6-4ec6-97bf-532d996e918d\") " pod="openstack/openstack-galera-0" Dec 03 14:28:33 crc kubenswrapper[4805]: I1203 14:28:33.582210 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5b4bb902-dba6-4ec6-97bf-532d996e918d-combined-ca-bundle\") pod \"openstack-galera-0\" (UID: \"5b4bb902-dba6-4ec6-97bf-532d996e918d\") " pod="openstack/openstack-galera-0" Dec 03 14:28:33 crc kubenswrapper[4805]: I1203 14:28:33.607907 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"openstack-galera-0\" (UID: \"5b4bb902-dba6-4ec6-97bf-532d996e918d\") " pod="openstack/openstack-galera-0" Dec 03 14:28:33 crc kubenswrapper[4805]: I1203 14:28:33.611170 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8flv5\" (UniqueName: \"kubernetes.io/projected/5b4bb902-dba6-4ec6-97bf-532d996e918d-kube-api-access-8flv5\") pod \"openstack-galera-0\" (UID: \"5b4bb902-dba6-4ec6-97bf-532d996e918d\") " pod="openstack/openstack-galera-0" Dec 03 14:28:33 crc kubenswrapper[4805]: I1203 14:28:33.689382 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-galera-0" Dec 03 14:28:34 crc kubenswrapper[4805]: I1203 14:28:34.626459 4805 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/openstack-cell1-galera-0"] Dec 03 14:28:34 crc kubenswrapper[4805]: I1203 14:28:34.627607 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-cell1-galera-0" Dec 03 14:28:34 crc kubenswrapper[4805]: I1203 14:28:34.630096 4805 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-cell1-scripts" Dec 03 14:28:34 crc kubenswrapper[4805]: I1203 14:28:34.630576 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"galera-openstack-cell1-dockercfg-ldzp2" Dec 03 14:28:34 crc kubenswrapper[4805]: I1203 14:28:34.630603 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-galera-openstack-cell1-svc" Dec 03 14:28:34 crc kubenswrapper[4805]: I1203 14:28:34.631258 4805 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-cell1-config-data" Dec 03 14:28:34 crc kubenswrapper[4805]: I1203 14:28:34.640520 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstack-cell1-galera-0"] Dec 03 14:28:34 crc kubenswrapper[4805]: I1203 14:28:34.792801 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/c13cbc51-432a-419a-9469-37fd31a82dd9-operator-scripts\") pod \"openstack-cell1-galera-0\" (UID: \"c13cbc51-432a-419a-9469-37fd31a82dd9\") " pod="openstack/openstack-cell1-galera-0" Dec 03 14:28:34 crc kubenswrapper[4805]: I1203 14:28:34.792868 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/c13cbc51-432a-419a-9469-37fd31a82dd9-config-data-default\") pod \"openstack-cell1-galera-0\" (UID: \"c13cbc51-432a-419a-9469-37fd31a82dd9\") " pod="openstack/openstack-cell1-galera-0" Dec 03 14:28:34 crc kubenswrapper[4805]: I1203 14:28:34.792898 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/c13cbc51-432a-419a-9469-37fd31a82dd9-galera-tls-certs\") pod \"openstack-cell1-galera-0\" (UID: \"c13cbc51-432a-419a-9469-37fd31a82dd9\") " pod="openstack/openstack-cell1-galera-0" Dec 03 14:28:34 crc kubenswrapper[4805]: I1203 14:28:34.792919 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9c6dz\" (UniqueName: \"kubernetes.io/projected/c13cbc51-432a-419a-9469-37fd31a82dd9-kube-api-access-9c6dz\") pod \"openstack-cell1-galera-0\" (UID: \"c13cbc51-432a-419a-9469-37fd31a82dd9\") " pod="openstack/openstack-cell1-galera-0" Dec 03 14:28:34 crc kubenswrapper[4805]: I1203 14:28:34.792986 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"openstack-cell1-galera-0\" (UID: \"c13cbc51-432a-419a-9469-37fd31a82dd9\") " pod="openstack/openstack-cell1-galera-0" Dec 03 14:28:34 crc kubenswrapper[4805]: I1203 14:28:34.793021 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c13cbc51-432a-419a-9469-37fd31a82dd9-combined-ca-bundle\") pod \"openstack-cell1-galera-0\" (UID: \"c13cbc51-432a-419a-9469-37fd31a82dd9\") " pod="openstack/openstack-cell1-galera-0" Dec 03 14:28:34 crc kubenswrapper[4805]: I1203 14:28:34.793048 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/c13cbc51-432a-419a-9469-37fd31a82dd9-config-data-generated\") pod \"openstack-cell1-galera-0\" (UID: \"c13cbc51-432a-419a-9469-37fd31a82dd9\") " pod="openstack/openstack-cell1-galera-0" Dec 03 14:28:34 crc kubenswrapper[4805]: I1203 14:28:34.793073 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/c13cbc51-432a-419a-9469-37fd31a82dd9-kolla-config\") pod \"openstack-cell1-galera-0\" (UID: \"c13cbc51-432a-419a-9469-37fd31a82dd9\") " pod="openstack/openstack-cell1-galera-0" Dec 03 14:28:34 crc kubenswrapper[4805]: I1203 14:28:34.894629 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/c13cbc51-432a-419a-9469-37fd31a82dd9-operator-scripts\") pod \"openstack-cell1-galera-0\" (UID: \"c13cbc51-432a-419a-9469-37fd31a82dd9\") " pod="openstack/openstack-cell1-galera-0" Dec 03 14:28:34 crc kubenswrapper[4805]: I1203 14:28:34.894711 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/c13cbc51-432a-419a-9469-37fd31a82dd9-config-data-default\") pod \"openstack-cell1-galera-0\" (UID: \"c13cbc51-432a-419a-9469-37fd31a82dd9\") " pod="openstack/openstack-cell1-galera-0" Dec 03 14:28:34 crc kubenswrapper[4805]: I1203 14:28:34.894739 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/c13cbc51-432a-419a-9469-37fd31a82dd9-galera-tls-certs\") pod \"openstack-cell1-galera-0\" (UID: \"c13cbc51-432a-419a-9469-37fd31a82dd9\") " pod="openstack/openstack-cell1-galera-0" Dec 03 14:28:34 crc kubenswrapper[4805]: I1203 14:28:34.894782 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9c6dz\" (UniqueName: \"kubernetes.io/projected/c13cbc51-432a-419a-9469-37fd31a82dd9-kube-api-access-9c6dz\") pod \"openstack-cell1-galera-0\" (UID: \"c13cbc51-432a-419a-9469-37fd31a82dd9\") " pod="openstack/openstack-cell1-galera-0" Dec 03 14:28:34 crc kubenswrapper[4805]: I1203 14:28:34.894826 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"openstack-cell1-galera-0\" (UID: \"c13cbc51-432a-419a-9469-37fd31a82dd9\") " pod="openstack/openstack-cell1-galera-0" Dec 03 14:28:34 crc kubenswrapper[4805]: I1203 14:28:34.894953 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c13cbc51-432a-419a-9469-37fd31a82dd9-combined-ca-bundle\") pod \"openstack-cell1-galera-0\" (UID: \"c13cbc51-432a-419a-9469-37fd31a82dd9\") " pod="openstack/openstack-cell1-galera-0" Dec 03 14:28:34 crc kubenswrapper[4805]: I1203 14:28:34.895010 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/c13cbc51-432a-419a-9469-37fd31a82dd9-config-data-generated\") pod \"openstack-cell1-galera-0\" (UID: \"c13cbc51-432a-419a-9469-37fd31a82dd9\") " pod="openstack/openstack-cell1-galera-0" Dec 03 14:28:34 crc kubenswrapper[4805]: I1203 14:28:34.895041 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/c13cbc51-432a-419a-9469-37fd31a82dd9-kolla-config\") pod \"openstack-cell1-galera-0\" (UID: \"c13cbc51-432a-419a-9469-37fd31a82dd9\") " pod="openstack/openstack-cell1-galera-0" Dec 03 14:28:34 crc kubenswrapper[4805]: I1203 14:28:34.895491 4805 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"openstack-cell1-galera-0\" (UID: \"c13cbc51-432a-419a-9469-37fd31a82dd9\") device mount path \"/mnt/openstack/pv02\"" pod="openstack/openstack-cell1-galera-0" Dec 03 14:28:34 crc kubenswrapper[4805]: I1203 14:28:34.896032 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/c13cbc51-432a-419a-9469-37fd31a82dd9-kolla-config\") pod \"openstack-cell1-galera-0\" (UID: \"c13cbc51-432a-419a-9469-37fd31a82dd9\") " pod="openstack/openstack-cell1-galera-0" Dec 03 14:28:34 crc kubenswrapper[4805]: I1203 14:28:34.896436 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/c13cbc51-432a-419a-9469-37fd31a82dd9-operator-scripts\") pod \"openstack-cell1-galera-0\" (UID: \"c13cbc51-432a-419a-9469-37fd31a82dd9\") " pod="openstack/openstack-cell1-galera-0" Dec 03 14:28:34 crc kubenswrapper[4805]: I1203 14:28:34.896662 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/c13cbc51-432a-419a-9469-37fd31a82dd9-config-data-generated\") pod \"openstack-cell1-galera-0\" (UID: \"c13cbc51-432a-419a-9469-37fd31a82dd9\") " pod="openstack/openstack-cell1-galera-0" Dec 03 14:28:34 crc kubenswrapper[4805]: I1203 14:28:34.899490 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/c13cbc51-432a-419a-9469-37fd31a82dd9-galera-tls-certs\") pod \"openstack-cell1-galera-0\" (UID: \"c13cbc51-432a-419a-9469-37fd31a82dd9\") " pod="openstack/openstack-cell1-galera-0" Dec 03 14:28:34 crc kubenswrapper[4805]: I1203 14:28:34.918197 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/c13cbc51-432a-419a-9469-37fd31a82dd9-config-data-default\") pod \"openstack-cell1-galera-0\" (UID: \"c13cbc51-432a-419a-9469-37fd31a82dd9\") " pod="openstack/openstack-cell1-galera-0" Dec 03 14:28:34 crc kubenswrapper[4805]: I1203 14:28:34.919014 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c13cbc51-432a-419a-9469-37fd31a82dd9-combined-ca-bundle\") pod \"openstack-cell1-galera-0\" (UID: \"c13cbc51-432a-419a-9469-37fd31a82dd9\") " pod="openstack/openstack-cell1-galera-0" Dec 03 14:28:34 crc kubenswrapper[4805]: I1203 14:28:34.921749 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9c6dz\" (UniqueName: \"kubernetes.io/projected/c13cbc51-432a-419a-9469-37fd31a82dd9-kube-api-access-9c6dz\") pod \"openstack-cell1-galera-0\" (UID: \"c13cbc51-432a-419a-9469-37fd31a82dd9\") " pod="openstack/openstack-cell1-galera-0" Dec 03 14:28:34 crc kubenswrapper[4805]: I1203 14:28:34.966283 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"openstack-cell1-galera-0\" (UID: \"c13cbc51-432a-419a-9469-37fd31a82dd9\") " pod="openstack/openstack-cell1-galera-0" Dec 03 14:28:35 crc kubenswrapper[4805]: I1203 14:28:35.040812 4805 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/memcached-0"] Dec 03 14:28:35 crc kubenswrapper[4805]: I1203 14:28:35.044828 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/memcached-0" Dec 03 14:28:35 crc kubenswrapper[4805]: I1203 14:28:35.047225 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-memcached-svc" Dec 03 14:28:35 crc kubenswrapper[4805]: I1203 14:28:35.047354 4805 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"memcached-config-data" Dec 03 14:28:35 crc kubenswrapper[4805]: I1203 14:28:35.049755 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"memcached-memcached-dockercfg-ptlcs" Dec 03 14:28:35 crc kubenswrapper[4805]: I1203 14:28:35.055051 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/memcached-0"] Dec 03 14:28:35 crc kubenswrapper[4805]: I1203 14:28:35.200057 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/3c84e84f-43fa-4163-81b2-489ec6a0d834-kolla-config\") pod \"memcached-0\" (UID: \"3c84e84f-43fa-4163-81b2-489ec6a0d834\") " pod="openstack/memcached-0" Dec 03 14:28:35 crc kubenswrapper[4805]: I1203 14:28:35.200448 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hnkns\" (UniqueName: \"kubernetes.io/projected/3c84e84f-43fa-4163-81b2-489ec6a0d834-kube-api-access-hnkns\") pod \"memcached-0\" (UID: \"3c84e84f-43fa-4163-81b2-489ec6a0d834\") " pod="openstack/memcached-0" Dec 03 14:28:35 crc kubenswrapper[4805]: I1203 14:28:35.200511 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"memcached-tls-certs\" (UniqueName: \"kubernetes.io/secret/3c84e84f-43fa-4163-81b2-489ec6a0d834-memcached-tls-certs\") pod \"memcached-0\" (UID: \"3c84e84f-43fa-4163-81b2-489ec6a0d834\") " pod="openstack/memcached-0" Dec 03 14:28:35 crc kubenswrapper[4805]: I1203 14:28:35.200552 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/3c84e84f-43fa-4163-81b2-489ec6a0d834-config-data\") pod \"memcached-0\" (UID: \"3c84e84f-43fa-4163-81b2-489ec6a0d834\") " pod="openstack/memcached-0" Dec 03 14:28:35 crc kubenswrapper[4805]: I1203 14:28:35.200630 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3c84e84f-43fa-4163-81b2-489ec6a0d834-combined-ca-bundle\") pod \"memcached-0\" (UID: \"3c84e84f-43fa-4163-81b2-489ec6a0d834\") " pod="openstack/memcached-0" Dec 03 14:28:35 crc kubenswrapper[4805]: I1203 14:28:35.245037 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-cell1-galera-0" Dec 03 14:28:35 crc kubenswrapper[4805]: I1203 14:28:35.302904 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/3c84e84f-43fa-4163-81b2-489ec6a0d834-kolla-config\") pod \"memcached-0\" (UID: \"3c84e84f-43fa-4163-81b2-489ec6a0d834\") " pod="openstack/memcached-0" Dec 03 14:28:35 crc kubenswrapper[4805]: I1203 14:28:35.302957 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hnkns\" (UniqueName: \"kubernetes.io/projected/3c84e84f-43fa-4163-81b2-489ec6a0d834-kube-api-access-hnkns\") pod \"memcached-0\" (UID: \"3c84e84f-43fa-4163-81b2-489ec6a0d834\") " pod="openstack/memcached-0" Dec 03 14:28:35 crc kubenswrapper[4805]: I1203 14:28:35.303012 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"memcached-tls-certs\" (UniqueName: \"kubernetes.io/secret/3c84e84f-43fa-4163-81b2-489ec6a0d834-memcached-tls-certs\") pod \"memcached-0\" (UID: \"3c84e84f-43fa-4163-81b2-489ec6a0d834\") " pod="openstack/memcached-0" Dec 03 14:28:35 crc kubenswrapper[4805]: I1203 14:28:35.303045 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/3c84e84f-43fa-4163-81b2-489ec6a0d834-config-data\") pod \"memcached-0\" (UID: \"3c84e84f-43fa-4163-81b2-489ec6a0d834\") " pod="openstack/memcached-0" Dec 03 14:28:35 crc kubenswrapper[4805]: I1203 14:28:35.303088 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3c84e84f-43fa-4163-81b2-489ec6a0d834-combined-ca-bundle\") pod \"memcached-0\" (UID: \"3c84e84f-43fa-4163-81b2-489ec6a0d834\") " pod="openstack/memcached-0" Dec 03 14:28:35 crc kubenswrapper[4805]: I1203 14:28:35.303930 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/3c84e84f-43fa-4163-81b2-489ec6a0d834-kolla-config\") pod \"memcached-0\" (UID: \"3c84e84f-43fa-4163-81b2-489ec6a0d834\") " pod="openstack/memcached-0" Dec 03 14:28:35 crc kubenswrapper[4805]: I1203 14:28:35.305096 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/3c84e84f-43fa-4163-81b2-489ec6a0d834-config-data\") pod \"memcached-0\" (UID: \"3c84e84f-43fa-4163-81b2-489ec6a0d834\") " pod="openstack/memcached-0" Dec 03 14:28:35 crc kubenswrapper[4805]: I1203 14:28:35.309959 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"memcached-tls-certs\" (UniqueName: \"kubernetes.io/secret/3c84e84f-43fa-4163-81b2-489ec6a0d834-memcached-tls-certs\") pod \"memcached-0\" (UID: \"3c84e84f-43fa-4163-81b2-489ec6a0d834\") " pod="openstack/memcached-0" Dec 03 14:28:35 crc kubenswrapper[4805]: I1203 14:28:35.331629 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3c84e84f-43fa-4163-81b2-489ec6a0d834-combined-ca-bundle\") pod \"memcached-0\" (UID: \"3c84e84f-43fa-4163-81b2-489ec6a0d834\") " pod="openstack/memcached-0" Dec 03 14:28:35 crc kubenswrapper[4805]: I1203 14:28:35.336269 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hnkns\" (UniqueName: \"kubernetes.io/projected/3c84e84f-43fa-4163-81b2-489ec6a0d834-kube-api-access-hnkns\") pod \"memcached-0\" (UID: \"3c84e84f-43fa-4163-81b2-489ec6a0d834\") " pod="openstack/memcached-0" Dec 03 14:28:35 crc kubenswrapper[4805]: I1203 14:28:35.364472 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/memcached-0" Dec 03 14:28:37 crc kubenswrapper[4805]: I1203 14:28:37.207878 4805 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/kube-state-metrics-0"] Dec 03 14:28:37 crc kubenswrapper[4805]: I1203 14:28:37.209125 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Dec 03 14:28:37 crc kubenswrapper[4805]: I1203 14:28:37.210962 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"telemetry-ceilometer-dockercfg-wpxxk" Dec 03 14:28:37 crc kubenswrapper[4805]: I1203 14:28:37.215491 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/kube-state-metrics-0"] Dec 03 14:28:37 crc kubenswrapper[4805]: I1203 14:28:37.328864 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mmj8n\" (UniqueName: \"kubernetes.io/projected/978876d6-1a61-47c1-a0af-eaec298dbe47-kube-api-access-mmj8n\") pod \"kube-state-metrics-0\" (UID: \"978876d6-1a61-47c1-a0af-eaec298dbe47\") " pod="openstack/kube-state-metrics-0" Dec 03 14:28:37 crc kubenswrapper[4805]: I1203 14:28:37.430686 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mmj8n\" (UniqueName: \"kubernetes.io/projected/978876d6-1a61-47c1-a0af-eaec298dbe47-kube-api-access-mmj8n\") pod \"kube-state-metrics-0\" (UID: \"978876d6-1a61-47c1-a0af-eaec298dbe47\") " pod="openstack/kube-state-metrics-0" Dec 03 14:28:37 crc kubenswrapper[4805]: I1203 14:28:37.460110 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mmj8n\" (UniqueName: \"kubernetes.io/projected/978876d6-1a61-47c1-a0af-eaec298dbe47-kube-api-access-mmj8n\") pod \"kube-state-metrics-0\" (UID: \"978876d6-1a61-47c1-a0af-eaec298dbe47\") " pod="openstack/kube-state-metrics-0" Dec 03 14:28:37 crc kubenswrapper[4805]: I1203 14:28:37.530145 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Dec 03 14:28:40 crc kubenswrapper[4805]: I1203 14:28:40.162226 4805 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-zhnwt"] Dec 03 14:28:40 crc kubenswrapper[4805]: I1203 14:28:40.163563 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-zhnwt" Dec 03 14:28:40 crc kubenswrapper[4805]: I1203 14:28:40.166994 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ovncontroller-ovncontroller-dockercfg-c9hg9" Dec 03 14:28:40 crc kubenswrapper[4805]: I1203 14:28:40.167229 4805 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovncontroller-scripts" Dec 03 14:28:40 crc kubenswrapper[4805]: I1203 14:28:40.167407 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovncontroller-ovndbs" Dec 03 14:28:40 crc kubenswrapper[4805]: I1203 14:28:40.179494 4805 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-ovs-dqwfk"] Dec 03 14:28:40 crc kubenswrapper[4805]: I1203 14:28:40.181617 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-ovs-dqwfk" Dec 03 14:28:40 crc kubenswrapper[4805]: I1203 14:28:40.185456 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-zhnwt"] Dec 03 14:28:40 crc kubenswrapper[4805]: I1203 14:28:40.214012 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-ovs-dqwfk"] Dec 03 14:28:40 crc kubenswrapper[4805]: I1203 14:28:40.275124 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/139abefe-12d4-4d8f-bdc6-43e71c818a65-var-run-ovn\") pod \"ovn-controller-zhnwt\" (UID: \"139abefe-12d4-4d8f-bdc6-43e71c818a65\") " pod="openstack/ovn-controller-zhnwt" Dec 03 14:28:40 crc kubenswrapper[4805]: I1203 14:28:40.275199 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-ovs\" (UniqueName: \"kubernetes.io/host-path/4335461c-d070-4c88-9d40-980109c57b4c-etc-ovs\") pod \"ovn-controller-ovs-dqwfk\" (UID: \"4335461c-d070-4c88-9d40-980109c57b4c\") " pod="openstack/ovn-controller-ovs-dqwfk" Dec 03 14:28:40 crc kubenswrapper[4805]: I1203 14:28:40.275231 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/4335461c-d070-4c88-9d40-980109c57b4c-var-run\") pod \"ovn-controller-ovs-dqwfk\" (UID: \"4335461c-d070-4c88-9d40-980109c57b4c\") " pod="openstack/ovn-controller-ovs-dqwfk" Dec 03 14:28:40 crc kubenswrapper[4805]: I1203 14:28:40.275267 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-controller-tls-certs\" (UniqueName: \"kubernetes.io/secret/139abefe-12d4-4d8f-bdc6-43e71c818a65-ovn-controller-tls-certs\") pod \"ovn-controller-zhnwt\" (UID: \"139abefe-12d4-4d8f-bdc6-43e71c818a65\") " pod="openstack/ovn-controller-zhnwt" Dec 03 14:28:40 crc kubenswrapper[4805]: I1203 14:28:40.275399 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/4335461c-d070-4c88-9d40-980109c57b4c-var-log\") pod \"ovn-controller-ovs-dqwfk\" (UID: \"4335461c-d070-4c88-9d40-980109c57b4c\") " pod="openstack/ovn-controller-ovs-dqwfk" Dec 03 14:28:40 crc kubenswrapper[4805]: I1203 14:28:40.275462 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib\" (UniqueName: \"kubernetes.io/host-path/4335461c-d070-4c88-9d40-980109c57b4c-var-lib\") pod \"ovn-controller-ovs-dqwfk\" (UID: \"4335461c-d070-4c88-9d40-980109c57b4c\") " pod="openstack/ovn-controller-ovs-dqwfk" Dec 03 14:28:40 crc kubenswrapper[4805]: I1203 14:28:40.275508 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/139abefe-12d4-4d8f-bdc6-43e71c818a65-var-run\") pod \"ovn-controller-zhnwt\" (UID: \"139abefe-12d4-4d8f-bdc6-43e71c818a65\") " pod="openstack/ovn-controller-zhnwt" Dec 03 14:28:40 crc kubenswrapper[4805]: I1203 14:28:40.275532 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/4335461c-d070-4c88-9d40-980109c57b4c-scripts\") pod \"ovn-controller-ovs-dqwfk\" (UID: \"4335461c-d070-4c88-9d40-980109c57b4c\") " pod="openstack/ovn-controller-ovs-dqwfk" Dec 03 14:28:40 crc kubenswrapper[4805]: I1203 14:28:40.275618 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/139abefe-12d4-4d8f-bdc6-43e71c818a65-scripts\") pod \"ovn-controller-zhnwt\" (UID: \"139abefe-12d4-4d8f-bdc6-43e71c818a65\") " pod="openstack/ovn-controller-zhnwt" Dec 03 14:28:40 crc kubenswrapper[4805]: I1203 14:28:40.275651 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8jlbn\" (UniqueName: \"kubernetes.io/projected/4335461c-d070-4c88-9d40-980109c57b4c-kube-api-access-8jlbn\") pod \"ovn-controller-ovs-dqwfk\" (UID: \"4335461c-d070-4c88-9d40-980109c57b4c\") " pod="openstack/ovn-controller-ovs-dqwfk" Dec 03 14:28:40 crc kubenswrapper[4805]: I1203 14:28:40.275702 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/139abefe-12d4-4d8f-bdc6-43e71c818a65-var-log-ovn\") pod \"ovn-controller-zhnwt\" (UID: \"139abefe-12d4-4d8f-bdc6-43e71c818a65\") " pod="openstack/ovn-controller-zhnwt" Dec 03 14:28:40 crc kubenswrapper[4805]: I1203 14:28:40.275720 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mmdm9\" (UniqueName: \"kubernetes.io/projected/139abefe-12d4-4d8f-bdc6-43e71c818a65-kube-api-access-mmdm9\") pod \"ovn-controller-zhnwt\" (UID: \"139abefe-12d4-4d8f-bdc6-43e71c818a65\") " pod="openstack/ovn-controller-zhnwt" Dec 03 14:28:40 crc kubenswrapper[4805]: I1203 14:28:40.275761 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/139abefe-12d4-4d8f-bdc6-43e71c818a65-combined-ca-bundle\") pod \"ovn-controller-zhnwt\" (UID: \"139abefe-12d4-4d8f-bdc6-43e71c818a65\") " pod="openstack/ovn-controller-zhnwt" Dec 03 14:28:40 crc kubenswrapper[4805]: I1203 14:28:40.376867 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/139abefe-12d4-4d8f-bdc6-43e71c818a65-var-log-ovn\") pod \"ovn-controller-zhnwt\" (UID: \"139abefe-12d4-4d8f-bdc6-43e71c818a65\") " pod="openstack/ovn-controller-zhnwt" Dec 03 14:28:40 crc kubenswrapper[4805]: I1203 14:28:40.376910 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mmdm9\" (UniqueName: \"kubernetes.io/projected/139abefe-12d4-4d8f-bdc6-43e71c818a65-kube-api-access-mmdm9\") pod \"ovn-controller-zhnwt\" (UID: \"139abefe-12d4-4d8f-bdc6-43e71c818a65\") " pod="openstack/ovn-controller-zhnwt" Dec 03 14:28:40 crc kubenswrapper[4805]: I1203 14:28:40.376944 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/139abefe-12d4-4d8f-bdc6-43e71c818a65-combined-ca-bundle\") pod \"ovn-controller-zhnwt\" (UID: \"139abefe-12d4-4d8f-bdc6-43e71c818a65\") " pod="openstack/ovn-controller-zhnwt" Dec 03 14:28:40 crc kubenswrapper[4805]: I1203 14:28:40.376969 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/139abefe-12d4-4d8f-bdc6-43e71c818a65-var-run-ovn\") pod \"ovn-controller-zhnwt\" (UID: \"139abefe-12d4-4d8f-bdc6-43e71c818a65\") " pod="openstack/ovn-controller-zhnwt" Dec 03 14:28:40 crc kubenswrapper[4805]: I1203 14:28:40.376995 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-ovs\" (UniqueName: \"kubernetes.io/host-path/4335461c-d070-4c88-9d40-980109c57b4c-etc-ovs\") pod \"ovn-controller-ovs-dqwfk\" (UID: \"4335461c-d070-4c88-9d40-980109c57b4c\") " pod="openstack/ovn-controller-ovs-dqwfk" Dec 03 14:28:40 crc kubenswrapper[4805]: I1203 14:28:40.377023 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/4335461c-d070-4c88-9d40-980109c57b4c-var-run\") pod \"ovn-controller-ovs-dqwfk\" (UID: \"4335461c-d070-4c88-9d40-980109c57b4c\") " pod="openstack/ovn-controller-ovs-dqwfk" Dec 03 14:28:40 crc kubenswrapper[4805]: I1203 14:28:40.377062 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-controller-tls-certs\" (UniqueName: \"kubernetes.io/secret/139abefe-12d4-4d8f-bdc6-43e71c818a65-ovn-controller-tls-certs\") pod \"ovn-controller-zhnwt\" (UID: \"139abefe-12d4-4d8f-bdc6-43e71c818a65\") " pod="openstack/ovn-controller-zhnwt" Dec 03 14:28:40 crc kubenswrapper[4805]: I1203 14:28:40.377094 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/4335461c-d070-4c88-9d40-980109c57b4c-var-log\") pod \"ovn-controller-ovs-dqwfk\" (UID: \"4335461c-d070-4c88-9d40-980109c57b4c\") " pod="openstack/ovn-controller-ovs-dqwfk" Dec 03 14:28:40 crc kubenswrapper[4805]: I1203 14:28:40.377121 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib\" (UniqueName: \"kubernetes.io/host-path/4335461c-d070-4c88-9d40-980109c57b4c-var-lib\") pod \"ovn-controller-ovs-dqwfk\" (UID: \"4335461c-d070-4c88-9d40-980109c57b4c\") " pod="openstack/ovn-controller-ovs-dqwfk" Dec 03 14:28:40 crc kubenswrapper[4805]: I1203 14:28:40.377154 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/139abefe-12d4-4d8f-bdc6-43e71c818a65-var-run\") pod \"ovn-controller-zhnwt\" (UID: \"139abefe-12d4-4d8f-bdc6-43e71c818a65\") " pod="openstack/ovn-controller-zhnwt" Dec 03 14:28:40 crc kubenswrapper[4805]: I1203 14:28:40.377174 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/4335461c-d070-4c88-9d40-980109c57b4c-scripts\") pod \"ovn-controller-ovs-dqwfk\" (UID: \"4335461c-d070-4c88-9d40-980109c57b4c\") " pod="openstack/ovn-controller-ovs-dqwfk" Dec 03 14:28:40 crc kubenswrapper[4805]: I1203 14:28:40.377217 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/139abefe-12d4-4d8f-bdc6-43e71c818a65-scripts\") pod \"ovn-controller-zhnwt\" (UID: \"139abefe-12d4-4d8f-bdc6-43e71c818a65\") " pod="openstack/ovn-controller-zhnwt" Dec 03 14:28:40 crc kubenswrapper[4805]: I1203 14:28:40.377236 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8jlbn\" (UniqueName: \"kubernetes.io/projected/4335461c-d070-4c88-9d40-980109c57b4c-kube-api-access-8jlbn\") pod \"ovn-controller-ovs-dqwfk\" (UID: \"4335461c-d070-4c88-9d40-980109c57b4c\") " pod="openstack/ovn-controller-ovs-dqwfk" Dec 03 14:28:40 crc kubenswrapper[4805]: I1203 14:28:40.377727 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib\" (UniqueName: \"kubernetes.io/host-path/4335461c-d070-4c88-9d40-980109c57b4c-var-lib\") pod \"ovn-controller-ovs-dqwfk\" (UID: \"4335461c-d070-4c88-9d40-980109c57b4c\") " pod="openstack/ovn-controller-ovs-dqwfk" Dec 03 14:28:40 crc kubenswrapper[4805]: I1203 14:28:40.377796 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/4335461c-d070-4c88-9d40-980109c57b4c-var-run\") pod \"ovn-controller-ovs-dqwfk\" (UID: \"4335461c-d070-4c88-9d40-980109c57b4c\") " pod="openstack/ovn-controller-ovs-dqwfk" Dec 03 14:28:40 crc kubenswrapper[4805]: I1203 14:28:40.377828 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/139abefe-12d4-4d8f-bdc6-43e71c818a65-var-run\") pod \"ovn-controller-zhnwt\" (UID: \"139abefe-12d4-4d8f-bdc6-43e71c818a65\") " pod="openstack/ovn-controller-zhnwt" Dec 03 14:28:40 crc kubenswrapper[4805]: I1203 14:28:40.377868 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/139abefe-12d4-4d8f-bdc6-43e71c818a65-var-run-ovn\") pod \"ovn-controller-zhnwt\" (UID: \"139abefe-12d4-4d8f-bdc6-43e71c818a65\") " pod="openstack/ovn-controller-zhnwt" Dec 03 14:28:40 crc kubenswrapper[4805]: I1203 14:28:40.377943 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/4335461c-d070-4c88-9d40-980109c57b4c-var-log\") pod \"ovn-controller-ovs-dqwfk\" (UID: \"4335461c-d070-4c88-9d40-980109c57b4c\") " pod="openstack/ovn-controller-ovs-dqwfk" Dec 03 14:28:40 crc kubenswrapper[4805]: I1203 14:28:40.378033 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-ovs\" (UniqueName: \"kubernetes.io/host-path/4335461c-d070-4c88-9d40-980109c57b4c-etc-ovs\") pod \"ovn-controller-ovs-dqwfk\" (UID: \"4335461c-d070-4c88-9d40-980109c57b4c\") " pod="openstack/ovn-controller-ovs-dqwfk" Dec 03 14:28:40 crc kubenswrapper[4805]: I1203 14:28:40.379377 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/139abefe-12d4-4d8f-bdc6-43e71c818a65-var-log-ovn\") pod \"ovn-controller-zhnwt\" (UID: \"139abefe-12d4-4d8f-bdc6-43e71c818a65\") " pod="openstack/ovn-controller-zhnwt" Dec 03 14:28:40 crc kubenswrapper[4805]: I1203 14:28:40.379591 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/4335461c-d070-4c88-9d40-980109c57b4c-scripts\") pod \"ovn-controller-ovs-dqwfk\" (UID: \"4335461c-d070-4c88-9d40-980109c57b4c\") " pod="openstack/ovn-controller-ovs-dqwfk" Dec 03 14:28:40 crc kubenswrapper[4805]: I1203 14:28:40.381771 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/139abefe-12d4-4d8f-bdc6-43e71c818a65-scripts\") pod \"ovn-controller-zhnwt\" (UID: \"139abefe-12d4-4d8f-bdc6-43e71c818a65\") " pod="openstack/ovn-controller-zhnwt" Dec 03 14:28:40 crc kubenswrapper[4805]: I1203 14:28:40.382660 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/139abefe-12d4-4d8f-bdc6-43e71c818a65-combined-ca-bundle\") pod \"ovn-controller-zhnwt\" (UID: \"139abefe-12d4-4d8f-bdc6-43e71c818a65\") " pod="openstack/ovn-controller-zhnwt" Dec 03 14:28:40 crc kubenswrapper[4805]: I1203 14:28:40.384267 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-controller-tls-certs\" (UniqueName: \"kubernetes.io/secret/139abefe-12d4-4d8f-bdc6-43e71c818a65-ovn-controller-tls-certs\") pod \"ovn-controller-zhnwt\" (UID: \"139abefe-12d4-4d8f-bdc6-43e71c818a65\") " pod="openstack/ovn-controller-zhnwt" Dec 03 14:28:40 crc kubenswrapper[4805]: I1203 14:28:40.393123 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mmdm9\" (UniqueName: \"kubernetes.io/projected/139abefe-12d4-4d8f-bdc6-43e71c818a65-kube-api-access-mmdm9\") pod \"ovn-controller-zhnwt\" (UID: \"139abefe-12d4-4d8f-bdc6-43e71c818a65\") " pod="openstack/ovn-controller-zhnwt" Dec 03 14:28:40 crc kubenswrapper[4805]: I1203 14:28:40.398879 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8jlbn\" (UniqueName: \"kubernetes.io/projected/4335461c-d070-4c88-9d40-980109c57b4c-kube-api-access-8jlbn\") pod \"ovn-controller-ovs-dqwfk\" (UID: \"4335461c-d070-4c88-9d40-980109c57b4c\") " pod="openstack/ovn-controller-ovs-dqwfk" Dec 03 14:28:40 crc kubenswrapper[4805]: I1203 14:28:40.487814 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-zhnwt" Dec 03 14:28:40 crc kubenswrapper[4805]: I1203 14:28:40.498818 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-ovs-dqwfk" Dec 03 14:28:41 crc kubenswrapper[4805]: I1203 14:28:41.053132 4805 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovsdbserver-nb-0"] Dec 03 14:28:41 crc kubenswrapper[4805]: I1203 14:28:41.054603 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-nb-0" Dec 03 14:28:41 crc kubenswrapper[4805]: I1203 14:28:41.057832 4805 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovndbcluster-nb-config" Dec 03 14:28:41 crc kubenswrapper[4805]: I1203 14:28:41.058195 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ovncluster-ovndbcluster-nb-dockercfg-65cnm" Dec 03 14:28:41 crc kubenswrapper[4805]: I1203 14:28:41.058359 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovn-metrics" Dec 03 14:28:41 crc kubenswrapper[4805]: I1203 14:28:41.058537 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovndbcluster-nb-ovndbs" Dec 03 14:28:41 crc kubenswrapper[4805]: I1203 14:28:41.059806 4805 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovndbcluster-nb-scripts" Dec 03 14:28:41 crc kubenswrapper[4805]: I1203 14:28:41.071452 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-nb-0"] Dec 03 14:28:41 crc kubenswrapper[4805]: I1203 14:28:41.102012 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"ovsdbserver-nb-0\" (UID: \"00ac9536-1677-4a3a-be82-7c8d8df91656\") " pod="openstack/ovsdbserver-nb-0" Dec 03 14:28:41 crc kubenswrapper[4805]: I1203 14:28:41.102064 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb-tls-certs\" (UniqueName: \"kubernetes.io/secret/00ac9536-1677-4a3a-be82-7c8d8df91656-ovsdbserver-nb-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"00ac9536-1677-4a3a-be82-7c8d8df91656\") " pod="openstack/ovsdbserver-nb-0" Dec 03 14:28:41 crc kubenswrapper[4805]: I1203 14:28:41.102100 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/00ac9536-1677-4a3a-be82-7c8d8df91656-scripts\") pod \"ovsdbserver-nb-0\" (UID: \"00ac9536-1677-4a3a-be82-7c8d8df91656\") " pod="openstack/ovsdbserver-nb-0" Dec 03 14:28:41 crc kubenswrapper[4805]: I1203 14:28:41.102137 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8vlfd\" (UniqueName: \"kubernetes.io/projected/00ac9536-1677-4a3a-be82-7c8d8df91656-kube-api-access-8vlfd\") pod \"ovsdbserver-nb-0\" (UID: \"00ac9536-1677-4a3a-be82-7c8d8df91656\") " pod="openstack/ovsdbserver-nb-0" Dec 03 14:28:41 crc kubenswrapper[4805]: I1203 14:28:41.102158 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/00ac9536-1677-4a3a-be82-7c8d8df91656-metrics-certs-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"00ac9536-1677-4a3a-be82-7c8d8df91656\") " pod="openstack/ovsdbserver-nb-0" Dec 03 14:28:41 crc kubenswrapper[4805]: I1203 14:28:41.102181 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/00ac9536-1677-4a3a-be82-7c8d8df91656-config\") pod \"ovsdbserver-nb-0\" (UID: \"00ac9536-1677-4a3a-be82-7c8d8df91656\") " pod="openstack/ovsdbserver-nb-0" Dec 03 14:28:41 crc kubenswrapper[4805]: I1203 14:28:41.102196 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/00ac9536-1677-4a3a-be82-7c8d8df91656-ovsdb-rundir\") pod \"ovsdbserver-nb-0\" (UID: \"00ac9536-1677-4a3a-be82-7c8d8df91656\") " pod="openstack/ovsdbserver-nb-0" Dec 03 14:28:41 crc kubenswrapper[4805]: I1203 14:28:41.102218 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/00ac9536-1677-4a3a-be82-7c8d8df91656-combined-ca-bundle\") pod \"ovsdbserver-nb-0\" (UID: \"00ac9536-1677-4a3a-be82-7c8d8df91656\") " pod="openstack/ovsdbserver-nb-0" Dec 03 14:28:41 crc kubenswrapper[4805]: I1203 14:28:41.203993 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8vlfd\" (UniqueName: \"kubernetes.io/projected/00ac9536-1677-4a3a-be82-7c8d8df91656-kube-api-access-8vlfd\") pod \"ovsdbserver-nb-0\" (UID: \"00ac9536-1677-4a3a-be82-7c8d8df91656\") " pod="openstack/ovsdbserver-nb-0" Dec 03 14:28:41 crc kubenswrapper[4805]: I1203 14:28:41.204050 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/00ac9536-1677-4a3a-be82-7c8d8df91656-metrics-certs-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"00ac9536-1677-4a3a-be82-7c8d8df91656\") " pod="openstack/ovsdbserver-nb-0" Dec 03 14:28:41 crc kubenswrapper[4805]: I1203 14:28:41.204089 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/00ac9536-1677-4a3a-be82-7c8d8df91656-config\") pod \"ovsdbserver-nb-0\" (UID: \"00ac9536-1677-4a3a-be82-7c8d8df91656\") " pod="openstack/ovsdbserver-nb-0" Dec 03 14:28:41 crc kubenswrapper[4805]: I1203 14:28:41.204112 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/00ac9536-1677-4a3a-be82-7c8d8df91656-ovsdb-rundir\") pod \"ovsdbserver-nb-0\" (UID: \"00ac9536-1677-4a3a-be82-7c8d8df91656\") " pod="openstack/ovsdbserver-nb-0" Dec 03 14:28:41 crc kubenswrapper[4805]: I1203 14:28:41.204146 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/00ac9536-1677-4a3a-be82-7c8d8df91656-combined-ca-bundle\") pod \"ovsdbserver-nb-0\" (UID: \"00ac9536-1677-4a3a-be82-7c8d8df91656\") " pod="openstack/ovsdbserver-nb-0" Dec 03 14:28:41 crc kubenswrapper[4805]: I1203 14:28:41.204297 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"ovsdbserver-nb-0\" (UID: \"00ac9536-1677-4a3a-be82-7c8d8df91656\") " pod="openstack/ovsdbserver-nb-0" Dec 03 14:28:41 crc kubenswrapper[4805]: I1203 14:28:41.204357 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb-tls-certs\" (UniqueName: \"kubernetes.io/secret/00ac9536-1677-4a3a-be82-7c8d8df91656-ovsdbserver-nb-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"00ac9536-1677-4a3a-be82-7c8d8df91656\") " pod="openstack/ovsdbserver-nb-0" Dec 03 14:28:41 crc kubenswrapper[4805]: I1203 14:28:41.204957 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/00ac9536-1677-4a3a-be82-7c8d8df91656-ovsdb-rundir\") pod \"ovsdbserver-nb-0\" (UID: \"00ac9536-1677-4a3a-be82-7c8d8df91656\") " pod="openstack/ovsdbserver-nb-0" Dec 03 14:28:41 crc kubenswrapper[4805]: I1203 14:28:41.207119 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/00ac9536-1677-4a3a-be82-7c8d8df91656-scripts\") pod \"ovsdbserver-nb-0\" (UID: \"00ac9536-1677-4a3a-be82-7c8d8df91656\") " pod="openstack/ovsdbserver-nb-0" Dec 03 14:28:41 crc kubenswrapper[4805]: I1203 14:28:41.207998 4805 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"ovsdbserver-nb-0\" (UID: \"00ac9536-1677-4a3a-be82-7c8d8df91656\") device mount path \"/mnt/openstack/pv04\"" pod="openstack/ovsdbserver-nb-0" Dec 03 14:28:41 crc kubenswrapper[4805]: I1203 14:28:41.208300 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/00ac9536-1677-4a3a-be82-7c8d8df91656-config\") pod \"ovsdbserver-nb-0\" (UID: \"00ac9536-1677-4a3a-be82-7c8d8df91656\") " pod="openstack/ovsdbserver-nb-0" Dec 03 14:28:41 crc kubenswrapper[4805]: I1203 14:28:41.208384 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/00ac9536-1677-4a3a-be82-7c8d8df91656-scripts\") pod \"ovsdbserver-nb-0\" (UID: \"00ac9536-1677-4a3a-be82-7c8d8df91656\") " pod="openstack/ovsdbserver-nb-0" Dec 03 14:28:41 crc kubenswrapper[4805]: I1203 14:28:41.211760 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/00ac9536-1677-4a3a-be82-7c8d8df91656-combined-ca-bundle\") pod \"ovsdbserver-nb-0\" (UID: \"00ac9536-1677-4a3a-be82-7c8d8df91656\") " pod="openstack/ovsdbserver-nb-0" Dec 03 14:28:41 crc kubenswrapper[4805]: I1203 14:28:41.212065 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/00ac9536-1677-4a3a-be82-7c8d8df91656-metrics-certs-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"00ac9536-1677-4a3a-be82-7c8d8df91656\") " pod="openstack/ovsdbserver-nb-0" Dec 03 14:28:41 crc kubenswrapper[4805]: I1203 14:28:41.225404 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb-tls-certs\" (UniqueName: \"kubernetes.io/secret/00ac9536-1677-4a3a-be82-7c8d8df91656-ovsdbserver-nb-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"00ac9536-1677-4a3a-be82-7c8d8df91656\") " pod="openstack/ovsdbserver-nb-0" Dec 03 14:28:41 crc kubenswrapper[4805]: I1203 14:28:41.243909 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8vlfd\" (UniqueName: \"kubernetes.io/projected/00ac9536-1677-4a3a-be82-7c8d8df91656-kube-api-access-8vlfd\") pod \"ovsdbserver-nb-0\" (UID: \"00ac9536-1677-4a3a-be82-7c8d8df91656\") " pod="openstack/ovsdbserver-nb-0" Dec 03 14:28:41 crc kubenswrapper[4805]: I1203 14:28:41.245327 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"ovsdbserver-nb-0\" (UID: \"00ac9536-1677-4a3a-be82-7c8d8df91656\") " pod="openstack/ovsdbserver-nb-0" Dec 03 14:28:41 crc kubenswrapper[4805]: I1203 14:28:41.377927 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-nb-0" Dec 03 14:28:44 crc kubenswrapper[4805]: I1203 14:28:44.634332 4805 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovsdbserver-sb-0"] Dec 03 14:28:44 crc kubenswrapper[4805]: I1203 14:28:44.636976 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-sb-0" Dec 03 14:28:44 crc kubenswrapper[4805]: I1203 14:28:44.639352 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ovncluster-ovndbcluster-sb-dockercfg-chftw" Dec 03 14:28:44 crc kubenswrapper[4805]: I1203 14:28:44.639490 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovndbcluster-sb-ovndbs" Dec 03 14:28:44 crc kubenswrapper[4805]: I1203 14:28:44.641045 4805 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovndbcluster-sb-config" Dec 03 14:28:44 crc kubenswrapper[4805]: I1203 14:28:44.641119 4805 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovndbcluster-sb-scripts" Dec 03 14:28:44 crc kubenswrapper[4805]: I1203 14:28:44.644029 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-server-0"] Dec 03 14:28:44 crc kubenswrapper[4805]: I1203 14:28:44.660011 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-sb-0"] Dec 03 14:28:44 crc kubenswrapper[4805]: I1203 14:28:44.666070 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/17f337d8-be50-4c6c-8364-208b394be6d3-ovsdb-rundir\") pod \"ovsdbserver-sb-0\" (UID: \"17f337d8-be50-4c6c-8364-208b394be6d3\") " pod="openstack/ovsdbserver-sb-0" Dec 03 14:28:44 crc kubenswrapper[4805]: I1203 14:28:44.666152 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/17f337d8-be50-4c6c-8364-208b394be6d3-combined-ca-bundle\") pod \"ovsdbserver-sb-0\" (UID: \"17f337d8-be50-4c6c-8364-208b394be6d3\") " pod="openstack/ovsdbserver-sb-0" Dec 03 14:28:44 crc kubenswrapper[4805]: I1203 14:28:44.666201 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"ovsdbserver-sb-0\" (UID: \"17f337d8-be50-4c6c-8364-208b394be6d3\") " pod="openstack/ovsdbserver-sb-0" Dec 03 14:28:44 crc kubenswrapper[4805]: I1203 14:28:44.666246 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/17f337d8-be50-4c6c-8364-208b394be6d3-config\") pod \"ovsdbserver-sb-0\" (UID: \"17f337d8-be50-4c6c-8364-208b394be6d3\") " pod="openstack/ovsdbserver-sb-0" Dec 03 14:28:44 crc kubenswrapper[4805]: I1203 14:28:44.666384 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-x84dw\" (UniqueName: \"kubernetes.io/projected/17f337d8-be50-4c6c-8364-208b394be6d3-kube-api-access-x84dw\") pod \"ovsdbserver-sb-0\" (UID: \"17f337d8-be50-4c6c-8364-208b394be6d3\") " pod="openstack/ovsdbserver-sb-0" Dec 03 14:28:44 crc kubenswrapper[4805]: I1203 14:28:44.666457 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/17f337d8-be50-4c6c-8364-208b394be6d3-metrics-certs-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"17f337d8-be50-4c6c-8364-208b394be6d3\") " pod="openstack/ovsdbserver-sb-0" Dec 03 14:28:44 crc kubenswrapper[4805]: I1203 14:28:44.666527 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/17f337d8-be50-4c6c-8364-208b394be6d3-scripts\") pod \"ovsdbserver-sb-0\" (UID: \"17f337d8-be50-4c6c-8364-208b394be6d3\") " pod="openstack/ovsdbserver-sb-0" Dec 03 14:28:44 crc kubenswrapper[4805]: I1203 14:28:44.666643 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb-tls-certs\" (UniqueName: \"kubernetes.io/secret/17f337d8-be50-4c6c-8364-208b394be6d3-ovsdbserver-sb-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"17f337d8-be50-4c6c-8364-208b394be6d3\") " pod="openstack/ovsdbserver-sb-0" Dec 03 14:28:44 crc kubenswrapper[4805]: I1203 14:28:44.767330 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/17f337d8-be50-4c6c-8364-208b394be6d3-scripts\") pod \"ovsdbserver-sb-0\" (UID: \"17f337d8-be50-4c6c-8364-208b394be6d3\") " pod="openstack/ovsdbserver-sb-0" Dec 03 14:28:44 crc kubenswrapper[4805]: I1203 14:28:44.767428 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb-tls-certs\" (UniqueName: \"kubernetes.io/secret/17f337d8-be50-4c6c-8364-208b394be6d3-ovsdbserver-sb-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"17f337d8-be50-4c6c-8364-208b394be6d3\") " pod="openstack/ovsdbserver-sb-0" Dec 03 14:28:44 crc kubenswrapper[4805]: I1203 14:28:44.767471 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/17f337d8-be50-4c6c-8364-208b394be6d3-ovsdb-rundir\") pod \"ovsdbserver-sb-0\" (UID: \"17f337d8-be50-4c6c-8364-208b394be6d3\") " pod="openstack/ovsdbserver-sb-0" Dec 03 14:28:44 crc kubenswrapper[4805]: I1203 14:28:44.767517 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/17f337d8-be50-4c6c-8364-208b394be6d3-combined-ca-bundle\") pod \"ovsdbserver-sb-0\" (UID: \"17f337d8-be50-4c6c-8364-208b394be6d3\") " pod="openstack/ovsdbserver-sb-0" Dec 03 14:28:44 crc kubenswrapper[4805]: I1203 14:28:44.767546 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"ovsdbserver-sb-0\" (UID: \"17f337d8-be50-4c6c-8364-208b394be6d3\") " pod="openstack/ovsdbserver-sb-0" Dec 03 14:28:44 crc kubenswrapper[4805]: I1203 14:28:44.767576 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/17f337d8-be50-4c6c-8364-208b394be6d3-config\") pod \"ovsdbserver-sb-0\" (UID: \"17f337d8-be50-4c6c-8364-208b394be6d3\") " pod="openstack/ovsdbserver-sb-0" Dec 03 14:28:44 crc kubenswrapper[4805]: I1203 14:28:44.767633 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-x84dw\" (UniqueName: \"kubernetes.io/projected/17f337d8-be50-4c6c-8364-208b394be6d3-kube-api-access-x84dw\") pod \"ovsdbserver-sb-0\" (UID: \"17f337d8-be50-4c6c-8364-208b394be6d3\") " pod="openstack/ovsdbserver-sb-0" Dec 03 14:28:44 crc kubenswrapper[4805]: I1203 14:28:44.767674 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/17f337d8-be50-4c6c-8364-208b394be6d3-metrics-certs-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"17f337d8-be50-4c6c-8364-208b394be6d3\") " pod="openstack/ovsdbserver-sb-0" Dec 03 14:28:44 crc kubenswrapper[4805]: I1203 14:28:44.767994 4805 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"ovsdbserver-sb-0\" (UID: \"17f337d8-be50-4c6c-8364-208b394be6d3\") device mount path \"/mnt/openstack/pv06\"" pod="openstack/ovsdbserver-sb-0" Dec 03 14:28:44 crc kubenswrapper[4805]: I1203 14:28:44.768122 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/17f337d8-be50-4c6c-8364-208b394be6d3-ovsdb-rundir\") pod \"ovsdbserver-sb-0\" (UID: \"17f337d8-be50-4c6c-8364-208b394be6d3\") " pod="openstack/ovsdbserver-sb-0" Dec 03 14:28:44 crc kubenswrapper[4805]: I1203 14:28:44.768770 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/17f337d8-be50-4c6c-8364-208b394be6d3-config\") pod \"ovsdbserver-sb-0\" (UID: \"17f337d8-be50-4c6c-8364-208b394be6d3\") " pod="openstack/ovsdbserver-sb-0" Dec 03 14:28:44 crc kubenswrapper[4805]: I1203 14:28:44.770543 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/17f337d8-be50-4c6c-8364-208b394be6d3-scripts\") pod \"ovsdbserver-sb-0\" (UID: \"17f337d8-be50-4c6c-8364-208b394be6d3\") " pod="openstack/ovsdbserver-sb-0" Dec 03 14:28:44 crc kubenswrapper[4805]: I1203 14:28:44.775889 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/17f337d8-be50-4c6c-8364-208b394be6d3-combined-ca-bundle\") pod \"ovsdbserver-sb-0\" (UID: \"17f337d8-be50-4c6c-8364-208b394be6d3\") " pod="openstack/ovsdbserver-sb-0" Dec 03 14:28:44 crc kubenswrapper[4805]: I1203 14:28:44.787915 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb-tls-certs\" (UniqueName: \"kubernetes.io/secret/17f337d8-be50-4c6c-8364-208b394be6d3-ovsdbserver-sb-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"17f337d8-be50-4c6c-8364-208b394be6d3\") " pod="openstack/ovsdbserver-sb-0" Dec 03 14:28:44 crc kubenswrapper[4805]: I1203 14:28:44.795647 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/17f337d8-be50-4c6c-8364-208b394be6d3-metrics-certs-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"17f337d8-be50-4c6c-8364-208b394be6d3\") " pod="openstack/ovsdbserver-sb-0" Dec 03 14:28:44 crc kubenswrapper[4805]: I1203 14:28:44.796297 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-x84dw\" (UniqueName: \"kubernetes.io/projected/17f337d8-be50-4c6c-8364-208b394be6d3-kube-api-access-x84dw\") pod \"ovsdbserver-sb-0\" (UID: \"17f337d8-be50-4c6c-8364-208b394be6d3\") " pod="openstack/ovsdbserver-sb-0" Dec 03 14:28:44 crc kubenswrapper[4805]: I1203 14:28:44.812866 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"ovsdbserver-sb-0\" (UID: \"17f337d8-be50-4c6c-8364-208b394be6d3\") " pod="openstack/ovsdbserver-sb-0" Dec 03 14:28:44 crc kubenswrapper[4805]: I1203 14:28:44.972774 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-sb-0" Dec 03 14:28:45 crc kubenswrapper[4805]: E1203 14:28:45.308359 4805 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified" Dec 03 14:28:45 crc kubenswrapper[4805]: E1203 14:28:45.308520 4805 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:init,Image:quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified,Command:[/bin/bash],Args:[-c dnsmasq --interface=* --conf-dir=/etc/dnsmasq.d --hostsdir=/etc/dnsmasq.d/hosts --keep-in-foreground --log-debug --bind-interfaces --listen-address=$(POD_IP) --port 5353 --log-facility=- --no-hosts --domain-needed --no-resolv --bogus-priv --log-queries --test],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:nffh5bdhf4h5f8h79h55h77h58fh56dh7bh6fh578hbch55dh68h56bhd9h65dh57ch658hc9h566h666h688h58h65dh684h5d7h6ch575h5d6h88q,ValueFrom:nil,},EnvVar{Name:POD_IP,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:status.podIP,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config,ReadOnly:true,MountPath:/etc/dnsmasq.d/config.cfg,SubPath:dns,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-7jj28,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000650000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod dnsmasq-dns-675f4bcbfc-5hg5w_openstack(388dedb5-6949-4ccb-9289-520b6f9ba4ce): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 03 14:28:45 crc kubenswrapper[4805]: E1203 14:28:45.309835 4805 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"init\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/dnsmasq-dns-675f4bcbfc-5hg5w" podUID="388dedb5-6949-4ccb-9289-520b6f9ba4ce" Dec 03 14:28:45 crc kubenswrapper[4805]: W1203 14:28:45.319554 4805 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod1b34fa5c_2867_4dbe_9285_80c327699864.slice/crio-ac82af3d82d8e8e05ac0c4a1dd4039bc0b98c7443be5ab903606c15c6c452a3b WatchSource:0}: Error finding container ac82af3d82d8e8e05ac0c4a1dd4039bc0b98c7443be5ab903606c15c6c452a3b: Status 404 returned error can't find the container with id ac82af3d82d8e8e05ac0c4a1dd4039bc0b98c7443be5ab903606c15c6c452a3b Dec 03 14:28:45 crc kubenswrapper[4805]: E1203 14:28:45.332682 4805 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified" Dec 03 14:28:45 crc kubenswrapper[4805]: E1203 14:28:45.332943 4805 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:init,Image:quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified,Command:[/bin/bash],Args:[-c dnsmasq --interface=* --conf-dir=/etc/dnsmasq.d --hostsdir=/etc/dnsmasq.d/hosts --keep-in-foreground --log-debug --bind-interfaces --listen-address=$(POD_IP) --port 5353 --log-facility=- --no-hosts --domain-needed --no-resolv --bogus-priv --log-queries --test],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:ndfhb5h667h568h584h5f9h58dh565h664h587h597h577h64bh5c4h66fh647hbdh68ch5c5h68dh686h5f7h64hd7hc6h55fh57bh98h57fh87h5fh57fq,ValueFrom:nil,},EnvVar{Name:POD_IP,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:status.podIP,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config,ReadOnly:true,MountPath:/etc/dnsmasq.d/config.cfg,SubPath:dns,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:dns-svc,ReadOnly:true,MountPath:/etc/dnsmasq.d/hosts/dns-svc,SubPath:dns-svc,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-rfqvq,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000650000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod dnsmasq-dns-78dd6ddcc-ql7t4_openstack(6c4701fa-549e-427c-b680-2f3096aafd00): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 03 14:28:45 crc kubenswrapper[4805]: E1203 14:28:45.335230 4805 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"init\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/dnsmasq-dns-78dd6ddcc-ql7t4" podUID="6c4701fa-549e-427c-b680-2f3096aafd00" Dec 03 14:28:45 crc kubenswrapper[4805]: I1203 14:28:45.806508 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"1b34fa5c-2867-4dbe-9285-80c327699864","Type":"ContainerStarted","Data":"ac82af3d82d8e8e05ac0c4a1dd4039bc0b98c7443be5ab903606c15c6c452a3b"} Dec 03 14:28:46 crc kubenswrapper[4805]: I1203 14:28:46.448441 4805 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-675f4bcbfc-5hg5w" Dec 03 14:28:46 crc kubenswrapper[4805]: I1203 14:28:46.453653 4805 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-78dd6ddcc-ql7t4" Dec 03 14:28:46 crc kubenswrapper[4805]: I1203 14:28:46.476025 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-zhnwt"] Dec 03 14:28:46 crc kubenswrapper[4805]: I1203 14:28:46.504616 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7jj28\" (UniqueName: \"kubernetes.io/projected/388dedb5-6949-4ccb-9289-520b6f9ba4ce-kube-api-access-7jj28\") pod \"388dedb5-6949-4ccb-9289-520b6f9ba4ce\" (UID: \"388dedb5-6949-4ccb-9289-520b6f9ba4ce\") " Dec 03 14:28:46 crc kubenswrapper[4805]: I1203 14:28:46.504711 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/6c4701fa-549e-427c-b680-2f3096aafd00-dns-svc\") pod \"6c4701fa-549e-427c-b680-2f3096aafd00\" (UID: \"6c4701fa-549e-427c-b680-2f3096aafd00\") " Dec 03 14:28:46 crc kubenswrapper[4805]: I1203 14:28:46.504833 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/388dedb5-6949-4ccb-9289-520b6f9ba4ce-config\") pod \"388dedb5-6949-4ccb-9289-520b6f9ba4ce\" (UID: \"388dedb5-6949-4ccb-9289-520b6f9ba4ce\") " Dec 03 14:28:46 crc kubenswrapper[4805]: I1203 14:28:46.504885 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rfqvq\" (UniqueName: \"kubernetes.io/projected/6c4701fa-549e-427c-b680-2f3096aafd00-kube-api-access-rfqvq\") pod \"6c4701fa-549e-427c-b680-2f3096aafd00\" (UID: \"6c4701fa-549e-427c-b680-2f3096aafd00\") " Dec 03 14:28:46 crc kubenswrapper[4805]: I1203 14:28:46.505328 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6c4701fa-549e-427c-b680-2f3096aafd00-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "6c4701fa-549e-427c-b680-2f3096aafd00" (UID: "6c4701fa-549e-427c-b680-2f3096aafd00"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 14:28:46 crc kubenswrapper[4805]: I1203 14:28:46.505654 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/388dedb5-6949-4ccb-9289-520b6f9ba4ce-config" (OuterVolumeSpecName: "config") pod "388dedb5-6949-4ccb-9289-520b6f9ba4ce" (UID: "388dedb5-6949-4ccb-9289-520b6f9ba4ce"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 14:28:46 crc kubenswrapper[4805]: I1203 14:28:46.505814 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6c4701fa-549e-427c-b680-2f3096aafd00-config\") pod \"6c4701fa-549e-427c-b680-2f3096aafd00\" (UID: \"6c4701fa-549e-427c-b680-2f3096aafd00\") " Dec 03 14:28:46 crc kubenswrapper[4805]: I1203 14:28:46.506629 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6c4701fa-549e-427c-b680-2f3096aafd00-config" (OuterVolumeSpecName: "config") pod "6c4701fa-549e-427c-b680-2f3096aafd00" (UID: "6c4701fa-549e-427c-b680-2f3096aafd00"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 14:28:46 crc kubenswrapper[4805]: I1203 14:28:46.508237 4805 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6c4701fa-549e-427c-b680-2f3096aafd00-config\") on node \"crc\" DevicePath \"\"" Dec 03 14:28:46 crc kubenswrapper[4805]: I1203 14:28:46.508277 4805 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/6c4701fa-549e-427c-b680-2f3096aafd00-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 03 14:28:46 crc kubenswrapper[4805]: I1203 14:28:46.508290 4805 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/388dedb5-6949-4ccb-9289-520b6f9ba4ce-config\") on node \"crc\" DevicePath \"\"" Dec 03 14:28:46 crc kubenswrapper[4805]: I1203 14:28:46.518782 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/388dedb5-6949-4ccb-9289-520b6f9ba4ce-kube-api-access-7jj28" (OuterVolumeSpecName: "kube-api-access-7jj28") pod "388dedb5-6949-4ccb-9289-520b6f9ba4ce" (UID: "388dedb5-6949-4ccb-9289-520b6f9ba4ce"). InnerVolumeSpecName "kube-api-access-7jj28". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 14:28:46 crc kubenswrapper[4805]: I1203 14:28:46.521560 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6c4701fa-549e-427c-b680-2f3096aafd00-kube-api-access-rfqvq" (OuterVolumeSpecName: "kube-api-access-rfqvq") pod "6c4701fa-549e-427c-b680-2f3096aafd00" (UID: "6c4701fa-549e-427c-b680-2f3096aafd00"). InnerVolumeSpecName "kube-api-access-rfqvq". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 14:28:46 crc kubenswrapper[4805]: I1203 14:28:46.552680 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/kube-state-metrics-0"] Dec 03 14:28:46 crc kubenswrapper[4805]: I1203 14:28:46.560926 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/memcached-0"] Dec 03 14:28:46 crc kubenswrapper[4805]: W1203 14:28:46.584202 4805 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod656f691c_1188_47b9_b8dd_b393a230602a.slice/crio-2e965419aa5bcdde632d972d20b4f13ceff9f46a20482645583589d483f55028 WatchSource:0}: Error finding container 2e965419aa5bcdde632d972d20b4f13ceff9f46a20482645583589d483f55028: Status 404 returned error can't find the container with id 2e965419aa5bcdde632d972d20b4f13ceff9f46a20482645583589d483f55028 Dec 03 14:28:46 crc kubenswrapper[4805]: I1203 14:28:46.586004 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Dec 03 14:28:46 crc kubenswrapper[4805]: I1203 14:28:46.606620 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstack-cell1-galera-0"] Dec 03 14:28:46 crc kubenswrapper[4805]: I1203 14:28:46.609331 4805 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rfqvq\" (UniqueName: \"kubernetes.io/projected/6c4701fa-549e-427c-b680-2f3096aafd00-kube-api-access-rfqvq\") on node \"crc\" DevicePath \"\"" Dec 03 14:28:46 crc kubenswrapper[4805]: I1203 14:28:46.609360 4805 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7jj28\" (UniqueName: \"kubernetes.io/projected/388dedb5-6949-4ccb-9289-520b6f9ba4ce-kube-api-access-7jj28\") on node \"crc\" DevicePath \"\"" Dec 03 14:28:46 crc kubenswrapper[4805]: W1203 14:28:46.610224 4805 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod5b4bb902_dba6_4ec6_97bf_532d996e918d.slice/crio-886674b917503eb0255d47bc1780dbac0703c39440f571f80448f74de784602a WatchSource:0}: Error finding container 886674b917503eb0255d47bc1780dbac0703c39440f571f80448f74de784602a: Status 404 returned error can't find the container with id 886674b917503eb0255d47bc1780dbac0703c39440f571f80448f74de784602a Dec 03 14:28:46 crc kubenswrapper[4805]: I1203 14:28:46.616697 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-666b6646f7-x75rp"] Dec 03 14:28:46 crc kubenswrapper[4805]: I1203 14:28:46.624228 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstack-galera-0"] Dec 03 14:28:46 crc kubenswrapper[4805]: W1203 14:28:46.632555 4805 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod00ac9536_1677_4a3a_be82_7c8d8df91656.slice/crio-ddf13354a1069cd94de5c8f2523b92f8ef63be2236a7c51679261cc97178c6c7 WatchSource:0}: Error finding container ddf13354a1069cd94de5c8f2523b92f8ef63be2236a7c51679261cc97178c6c7: Status 404 returned error can't find the container with id ddf13354a1069cd94de5c8f2523b92f8ef63be2236a7c51679261cc97178c6c7 Dec 03 14:28:46 crc kubenswrapper[4805]: I1203 14:28:46.632607 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-57d769cc4f-fthjg"] Dec 03 14:28:46 crc kubenswrapper[4805]: I1203 14:28:46.641859 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-nb-0"] Dec 03 14:28:46 crc kubenswrapper[4805]: I1203 14:28:46.814070 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/memcached-0" event={"ID":"3c84e84f-43fa-4163-81b2-489ec6a0d834","Type":"ContainerStarted","Data":"ac85c469993a5f20d0e842da5cee2ba84f4b407eff4ba7cc1595e9cbf172ca25"} Dec 03 14:28:46 crc kubenswrapper[4805]: I1203 14:28:46.816192 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"5b4bb902-dba6-4ec6-97bf-532d996e918d","Type":"ContainerStarted","Data":"886674b917503eb0255d47bc1780dbac0703c39440f571f80448f74de784602a"} Dec 03 14:28:46 crc kubenswrapper[4805]: I1203 14:28:46.817213 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-zhnwt" event={"ID":"139abefe-12d4-4d8f-bdc6-43e71c818a65","Type":"ContainerStarted","Data":"2f9f338f5d1f77570fc89d8a692d886b309e36bc0db096c96948cb157d0896f3"} Dec 03 14:28:46 crc kubenswrapper[4805]: I1203 14:28:46.818112 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"656f691c-1188-47b9-b8dd-b393a230602a","Type":"ContainerStarted","Data":"2e965419aa5bcdde632d972d20b4f13ceff9f46a20482645583589d483f55028"} Dec 03 14:28:46 crc kubenswrapper[4805]: I1203 14:28:46.819222 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-78dd6ddcc-ql7t4" event={"ID":"6c4701fa-549e-427c-b680-2f3096aafd00","Type":"ContainerDied","Data":"b646b0a0968d43dfbffc43d927635219d05dfa54928609163da0a26dc543acef"} Dec 03 14:28:46 crc kubenswrapper[4805]: I1203 14:28:46.819247 4805 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-78dd6ddcc-ql7t4" Dec 03 14:28:46 crc kubenswrapper[4805]: I1203 14:28:46.820233 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-57d769cc4f-fthjg" event={"ID":"2af0398e-b188-49ec-ae9e-391cb1122753","Type":"ContainerStarted","Data":"f674510e91a8b892788668806a350ee018c948cdc7b18fc4d3d00d606b500b33"} Dec 03 14:28:46 crc kubenswrapper[4805]: I1203 14:28:46.821361 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-666b6646f7-x75rp" event={"ID":"eba109b2-1336-48fc-a849-f2d039eae8e1","Type":"ContainerStarted","Data":"0a92596606eb5fd52e10e78fa863b2fbee542e7cf1c03cecef092378ab26e9f5"} Dec 03 14:28:46 crc kubenswrapper[4805]: I1203 14:28:46.822169 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"978876d6-1a61-47c1-a0af-eaec298dbe47","Type":"ContainerStarted","Data":"4019ce98731b0c3fb180a1b5df437792cc076f25c91ddb833823160217cc22f5"} Dec 03 14:28:46 crc kubenswrapper[4805]: I1203 14:28:46.825111 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-0" event={"ID":"00ac9536-1677-4a3a-be82-7c8d8df91656","Type":"ContainerStarted","Data":"ddf13354a1069cd94de5c8f2523b92f8ef63be2236a7c51679261cc97178c6c7"} Dec 03 14:28:46 crc kubenswrapper[4805]: I1203 14:28:46.826712 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-675f4bcbfc-5hg5w" event={"ID":"388dedb5-6949-4ccb-9289-520b6f9ba4ce","Type":"ContainerDied","Data":"50c8fc4de0bd4a7db91845f4194f401d0ce5dd9a58c29d9d96249806e7583a12"} Dec 03 14:28:46 crc kubenswrapper[4805]: I1203 14:28:46.826733 4805 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-675f4bcbfc-5hg5w" Dec 03 14:28:46 crc kubenswrapper[4805]: I1203 14:28:46.827917 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"c13cbc51-432a-419a-9469-37fd31a82dd9","Type":"ContainerStarted","Data":"0546d3cdafbd4fa4c7a9ccc8d93983c18549894abf5e338912a3fdedac42477d"} Dec 03 14:28:46 crc kubenswrapper[4805]: I1203 14:28:46.858432 4805 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-78dd6ddcc-ql7t4"] Dec 03 14:28:46 crc kubenswrapper[4805]: I1203 14:28:46.863508 4805 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-78dd6ddcc-ql7t4"] Dec 03 14:28:46 crc kubenswrapper[4805]: I1203 14:28:46.893168 4805 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-675f4bcbfc-5hg5w"] Dec 03 14:28:46 crc kubenswrapper[4805]: I1203 14:28:46.900654 4805 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-675f4bcbfc-5hg5w"] Dec 03 14:28:47 crc kubenswrapper[4805]: I1203 14:28:47.231173 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-ovs-dqwfk"] Dec 03 14:28:47 crc kubenswrapper[4805]: I1203 14:28:47.692952 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-sb-0"] Dec 03 14:28:47 crc kubenswrapper[4805]: W1203 14:28:47.696725 4805 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod17f337d8_be50_4c6c_8364_208b394be6d3.slice/crio-f149a489fb767c9192547b5347c75f988ad875a7b06795cd67cb36aed8d8de8b WatchSource:0}: Error finding container f149a489fb767c9192547b5347c75f988ad875a7b06795cd67cb36aed8d8de8b: Status 404 returned error can't find the container with id f149a489fb767c9192547b5347c75f988ad875a7b06795cd67cb36aed8d8de8b Dec 03 14:28:47 crc kubenswrapper[4805]: I1203 14:28:47.843222 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-0" event={"ID":"17f337d8-be50-4c6c-8364-208b394be6d3","Type":"ContainerStarted","Data":"f149a489fb767c9192547b5347c75f988ad875a7b06795cd67cb36aed8d8de8b"} Dec 03 14:28:47 crc kubenswrapper[4805]: I1203 14:28:47.844034 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-dqwfk" event={"ID":"4335461c-d070-4c88-9d40-980109c57b4c","Type":"ContainerStarted","Data":"195f181a7164baa834eb5fc101f4fd960ea9f3fcc62c9dc35e9dbdf6c1c281ce"} Dec 03 14:28:48 crc kubenswrapper[4805]: I1203 14:28:48.711658 4805 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="388dedb5-6949-4ccb-9289-520b6f9ba4ce" path="/var/lib/kubelet/pods/388dedb5-6949-4ccb-9289-520b6f9ba4ce/volumes" Dec 03 14:28:48 crc kubenswrapper[4805]: I1203 14:28:48.712584 4805 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6c4701fa-549e-427c-b680-2f3096aafd00" path="/var/lib/kubelet/pods/6c4701fa-549e-427c-b680-2f3096aafd00/volumes" Dec 03 14:28:49 crc kubenswrapper[4805]: I1203 14:28:49.865033 4805 generic.go:334] "Generic (PLEG): container finished" podID="2af0398e-b188-49ec-ae9e-391cb1122753" containerID="8b70115a80a67bf0ded6b6883dda9ff0036d3720338a2ef80b70fcae73736348" exitCode=0 Dec 03 14:28:49 crc kubenswrapper[4805]: I1203 14:28:49.865122 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-57d769cc4f-fthjg" event={"ID":"2af0398e-b188-49ec-ae9e-391cb1122753","Type":"ContainerDied","Data":"8b70115a80a67bf0ded6b6883dda9ff0036d3720338a2ef80b70fcae73736348"} Dec 03 14:28:49 crc kubenswrapper[4805]: I1203 14:28:49.868316 4805 generic.go:334] "Generic (PLEG): container finished" podID="eba109b2-1336-48fc-a849-f2d039eae8e1" containerID="4e7aa0ff2ac6d5331aa4f6d2c86466a7274ce5897b7c11798e99c07d1bc72687" exitCode=0 Dec 03 14:28:49 crc kubenswrapper[4805]: I1203 14:28:49.868355 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-666b6646f7-x75rp" event={"ID":"eba109b2-1336-48fc-a849-f2d039eae8e1","Type":"ContainerDied","Data":"4e7aa0ff2ac6d5331aa4f6d2c86466a7274ce5897b7c11798e99c07d1bc72687"} Dec 03 14:28:56 crc kubenswrapper[4805]: I1203 14:28:56.941248 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-57d769cc4f-fthjg" event={"ID":"2af0398e-b188-49ec-ae9e-391cb1122753","Type":"ContainerStarted","Data":"b704ee8929e1278fce1a4d68a59d5b34da60bcace6cc40b496d450dce3b2ff98"} Dec 03 14:28:56 crc kubenswrapper[4805]: I1203 14:28:56.941937 4805 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-57d769cc4f-fthjg" Dec 03 14:28:56 crc kubenswrapper[4805]: I1203 14:28:56.944645 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-666b6646f7-x75rp" event={"ID":"eba109b2-1336-48fc-a849-f2d039eae8e1","Type":"ContainerStarted","Data":"af79900fcf86fa035a9726c5c728cacfe56afa771bbdcc93b96783c84f36fdc3"} Dec 03 14:28:56 crc kubenswrapper[4805]: I1203 14:28:56.945292 4805 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-666b6646f7-x75rp" Dec 03 14:28:56 crc kubenswrapper[4805]: I1203 14:28:56.948410 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-dqwfk" event={"ID":"4335461c-d070-4c88-9d40-980109c57b4c","Type":"ContainerStarted","Data":"164b8fc8d2b0c0a6806108ca7d141198c07e2c1124921e58cd8988975ca7658e"} Dec 03 14:28:56 crc kubenswrapper[4805]: I1203 14:28:56.953100 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/memcached-0" event={"ID":"3c84e84f-43fa-4163-81b2-489ec6a0d834","Type":"ContainerStarted","Data":"bd52ff6a34e8b10609f5ec8f537c4fb85d81793fc416c440758ec101987abd2f"} Dec 03 14:28:56 crc kubenswrapper[4805]: I1203 14:28:56.953217 4805 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/memcached-0" Dec 03 14:28:56 crc kubenswrapper[4805]: I1203 14:28:56.967075 4805 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-57d769cc4f-fthjg" podStartSLOduration=24.553180513 podStartE2EDuration="26.967053204s" podCreationTimestamp="2025-12-03 14:28:30 +0000 UTC" firstStartedPulling="2025-12-03 14:28:46.602403585 +0000 UTC m=+1156.265320508" lastFinishedPulling="2025-12-03 14:28:49.016276286 +0000 UTC m=+1158.679193199" observedRunningTime="2025-12-03 14:28:56.962404808 +0000 UTC m=+1166.625321731" watchObservedRunningTime="2025-12-03 14:28:56.967053204 +0000 UTC m=+1166.629970137" Dec 03 14:28:57 crc kubenswrapper[4805]: I1203 14:28:57.012534 4805 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-666b6646f7-x75rp" podStartSLOduration=24.67070753 podStartE2EDuration="27.012519557s" podCreationTimestamp="2025-12-03 14:28:30 +0000 UTC" firstStartedPulling="2025-12-03 14:28:46.606712131 +0000 UTC m=+1156.269629054" lastFinishedPulling="2025-12-03 14:28:48.948524158 +0000 UTC m=+1158.611441081" observedRunningTime="2025-12-03 14:28:57.008964401 +0000 UTC m=+1166.671881334" watchObservedRunningTime="2025-12-03 14:28:57.012519557 +0000 UTC m=+1166.675436470" Dec 03 14:28:57 crc kubenswrapper[4805]: I1203 14:28:57.016452 4805 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/memcached-0" podStartSLOduration=12.832920445 podStartE2EDuration="22.016401042s" podCreationTimestamp="2025-12-03 14:28:35 +0000 UTC" firstStartedPulling="2025-12-03 14:28:46.554380052 +0000 UTC m=+1156.217296975" lastFinishedPulling="2025-12-03 14:28:55.737860649 +0000 UTC m=+1165.400777572" observedRunningTime="2025-12-03 14:28:56.991952739 +0000 UTC m=+1166.654869672" watchObservedRunningTime="2025-12-03 14:28:57.016401042 +0000 UTC m=+1166.679317965" Dec 03 14:28:57 crc kubenswrapper[4805]: I1203 14:28:57.961143 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"1b34fa5c-2867-4dbe-9285-80c327699864","Type":"ContainerStarted","Data":"2e0f381f35c1220d9cac0d2d4c6085458b076fdcc37439c164c7ebd5a322440b"} Dec 03 14:28:57 crc kubenswrapper[4805]: I1203 14:28:57.964188 4805 generic.go:334] "Generic (PLEG): container finished" podID="4335461c-d070-4c88-9d40-980109c57b4c" containerID="164b8fc8d2b0c0a6806108ca7d141198c07e2c1124921e58cd8988975ca7658e" exitCode=0 Dec 03 14:28:57 crc kubenswrapper[4805]: I1203 14:28:57.964255 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-dqwfk" event={"ID":"4335461c-d070-4c88-9d40-980109c57b4c","Type":"ContainerDied","Data":"164b8fc8d2b0c0a6806108ca7d141198c07e2c1124921e58cd8988975ca7658e"} Dec 03 14:28:57 crc kubenswrapper[4805]: I1203 14:28:57.966279 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-zhnwt" event={"ID":"139abefe-12d4-4d8f-bdc6-43e71c818a65","Type":"ContainerStarted","Data":"feff2d36f0a5b15ba87c0c837cdf932e5428d233a791330536239a8304eb3fdf"} Dec 03 14:28:57 crc kubenswrapper[4805]: I1203 14:28:57.966371 4805 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovn-controller-zhnwt" Dec 03 14:28:57 crc kubenswrapper[4805]: I1203 14:28:57.968473 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"c13cbc51-432a-419a-9469-37fd31a82dd9","Type":"ContainerStarted","Data":"860b1712352b1e14def5d9057871da99d65fbdcce05423a307ac73ce697a415e"} Dec 03 14:28:57 crc kubenswrapper[4805]: I1203 14:28:57.972789 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"5b4bb902-dba6-4ec6-97bf-532d996e918d","Type":"ContainerStarted","Data":"837bce806cd5be22704075197396e52bd05cc8c2b364b8b7e09d291129690d6a"} Dec 03 14:28:57 crc kubenswrapper[4805]: I1203 14:28:57.974562 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-0" event={"ID":"17f337d8-be50-4c6c-8364-208b394be6d3","Type":"ContainerStarted","Data":"b2e248bacadf9b617fdfee8ea42dcd83c02089a1a3d7ec535ab217a6ec731325"} Dec 03 14:28:57 crc kubenswrapper[4805]: I1203 14:28:57.976682 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"978876d6-1a61-47c1-a0af-eaec298dbe47","Type":"ContainerStarted","Data":"21beea7642ced06d70a91581568a623fd5e9cda04779e8724086621a25095b2f"} Dec 03 14:28:57 crc kubenswrapper[4805]: I1203 14:28:57.977252 4805 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/kube-state-metrics-0" Dec 03 14:28:57 crc kubenswrapper[4805]: I1203 14:28:57.979816 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-0" event={"ID":"00ac9536-1677-4a3a-be82-7c8d8df91656","Type":"ContainerStarted","Data":"1b1394043144625410473eae14968d096baba244b4c75d2f01be0dcddd6c484e"} Dec 03 14:28:58 crc kubenswrapper[4805]: I1203 14:28:58.012855 4805 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-controller-zhnwt" podStartSLOduration=8.437729847 podStartE2EDuration="18.012810004s" podCreationTimestamp="2025-12-03 14:28:40 +0000 UTC" firstStartedPulling="2025-12-03 14:28:46.518139609 +0000 UTC m=+1156.181056532" lastFinishedPulling="2025-12-03 14:28:56.093219756 +0000 UTC m=+1165.756136689" observedRunningTime="2025-12-03 14:28:58.012550637 +0000 UTC m=+1167.675467560" watchObservedRunningTime="2025-12-03 14:28:58.012810004 +0000 UTC m=+1167.675726927" Dec 03 14:28:58 crc kubenswrapper[4805]: I1203 14:28:58.054400 4805 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/kube-state-metrics-0" podStartSLOduration=11.047049402 podStartE2EDuration="21.054380902s" podCreationTimestamp="2025-12-03 14:28:37 +0000 UTC" firstStartedPulling="2025-12-03 14:28:46.554499675 +0000 UTC m=+1156.217416598" lastFinishedPulling="2025-12-03 14:28:56.561831175 +0000 UTC m=+1166.224748098" observedRunningTime="2025-12-03 14:28:58.043505497 +0000 UTC m=+1167.706422420" watchObservedRunningTime="2025-12-03 14:28:58.054380902 +0000 UTC m=+1167.717297825" Dec 03 14:28:58 crc kubenswrapper[4805]: I1203 14:28:58.996693 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"656f691c-1188-47b9-b8dd-b393a230602a","Type":"ContainerStarted","Data":"0d085016bd6e7629b80b349c936428485e82d12ba8205b13e1f636404a6741cc"} Dec 03 14:28:59 crc kubenswrapper[4805]: I1203 14:28:59.004262 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-dqwfk" event={"ID":"4335461c-d070-4c88-9d40-980109c57b4c","Type":"ContainerStarted","Data":"ad2b799c6f12c75bd18819fae9c93ab93fd494263c290c02de24ac025f418623"} Dec 03 14:28:59 crc kubenswrapper[4805]: I1203 14:28:59.004317 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-dqwfk" event={"ID":"4335461c-d070-4c88-9d40-980109c57b4c","Type":"ContainerStarted","Data":"0527d3a20e6733671ea996f1b3daa5e4da4dcd6f254c85c93d60302441c0f713"} Dec 03 14:28:59 crc kubenswrapper[4805]: I1203 14:28:59.004562 4805 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovn-controller-ovs-dqwfk" Dec 03 14:28:59 crc kubenswrapper[4805]: I1203 14:28:59.004719 4805 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovn-controller-ovs-dqwfk" Dec 03 14:28:59 crc kubenswrapper[4805]: I1203 14:28:59.061759 4805 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-controller-ovs-dqwfk" podStartSLOduration=10.557452052 podStartE2EDuration="19.06173908s" podCreationTimestamp="2025-12-03 14:28:40 +0000 UTC" firstStartedPulling="2025-12-03 14:28:47.23463642 +0000 UTC m=+1156.897553343" lastFinishedPulling="2025-12-03 14:28:55.738923438 +0000 UTC m=+1165.401840371" observedRunningTime="2025-12-03 14:28:59.054069052 +0000 UTC m=+1168.716985995" watchObservedRunningTime="2025-12-03 14:28:59.06173908 +0000 UTC m=+1168.724656023" Dec 03 14:29:01 crc kubenswrapper[4805]: I1203 14:29:01.272161 4805 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-57d769cc4f-fthjg" Dec 03 14:29:01 crc kubenswrapper[4805]: I1203 14:29:01.334120 4805 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-666b6646f7-x75rp"] Dec 03 14:29:01 crc kubenswrapper[4805]: I1203 14:29:01.334403 4805 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-666b6646f7-x75rp" podUID="eba109b2-1336-48fc-a849-f2d039eae8e1" containerName="dnsmasq-dns" containerID="cri-o://af79900fcf86fa035a9726c5c728cacfe56afa771bbdcc93b96783c84f36fdc3" gracePeriod=10 Dec 03 14:29:01 crc kubenswrapper[4805]: I1203 14:29:01.339861 4805 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-666b6646f7-x75rp" Dec 03 14:29:01 crc kubenswrapper[4805]: I1203 14:29:01.740295 4805 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-666b6646f7-x75rp" Dec 03 14:29:01 crc kubenswrapper[4805]: I1203 14:29:01.853609 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/eba109b2-1336-48fc-a849-f2d039eae8e1-config\") pod \"eba109b2-1336-48fc-a849-f2d039eae8e1\" (UID: \"eba109b2-1336-48fc-a849-f2d039eae8e1\") " Dec 03 14:29:01 crc kubenswrapper[4805]: I1203 14:29:01.854491 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/eba109b2-1336-48fc-a849-f2d039eae8e1-dns-svc\") pod \"eba109b2-1336-48fc-a849-f2d039eae8e1\" (UID: \"eba109b2-1336-48fc-a849-f2d039eae8e1\") " Dec 03 14:29:01 crc kubenswrapper[4805]: I1203 14:29:01.854548 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-djw29\" (UniqueName: \"kubernetes.io/projected/eba109b2-1336-48fc-a849-f2d039eae8e1-kube-api-access-djw29\") pod \"eba109b2-1336-48fc-a849-f2d039eae8e1\" (UID: \"eba109b2-1336-48fc-a849-f2d039eae8e1\") " Dec 03 14:29:01 crc kubenswrapper[4805]: I1203 14:29:01.858796 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/eba109b2-1336-48fc-a849-f2d039eae8e1-kube-api-access-djw29" (OuterVolumeSpecName: "kube-api-access-djw29") pod "eba109b2-1336-48fc-a849-f2d039eae8e1" (UID: "eba109b2-1336-48fc-a849-f2d039eae8e1"). InnerVolumeSpecName "kube-api-access-djw29". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 14:29:01 crc kubenswrapper[4805]: I1203 14:29:01.901068 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/eba109b2-1336-48fc-a849-f2d039eae8e1-config" (OuterVolumeSpecName: "config") pod "eba109b2-1336-48fc-a849-f2d039eae8e1" (UID: "eba109b2-1336-48fc-a849-f2d039eae8e1"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 14:29:01 crc kubenswrapper[4805]: I1203 14:29:01.910229 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/eba109b2-1336-48fc-a849-f2d039eae8e1-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "eba109b2-1336-48fc-a849-f2d039eae8e1" (UID: "eba109b2-1336-48fc-a849-f2d039eae8e1"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 14:29:01 crc kubenswrapper[4805]: I1203 14:29:01.956618 4805 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-djw29\" (UniqueName: \"kubernetes.io/projected/eba109b2-1336-48fc-a849-f2d039eae8e1-kube-api-access-djw29\") on node \"crc\" DevicePath \"\"" Dec 03 14:29:01 crc kubenswrapper[4805]: I1203 14:29:01.956651 4805 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/eba109b2-1336-48fc-a849-f2d039eae8e1-config\") on node \"crc\" DevicePath \"\"" Dec 03 14:29:01 crc kubenswrapper[4805]: I1203 14:29:01.956660 4805 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/eba109b2-1336-48fc-a849-f2d039eae8e1-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 03 14:29:02 crc kubenswrapper[4805]: I1203 14:29:02.031656 4805 generic.go:334] "Generic (PLEG): container finished" podID="eba109b2-1336-48fc-a849-f2d039eae8e1" containerID="af79900fcf86fa035a9726c5c728cacfe56afa771bbdcc93b96783c84f36fdc3" exitCode=0 Dec 03 14:29:02 crc kubenswrapper[4805]: I1203 14:29:02.031717 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-666b6646f7-x75rp" event={"ID":"eba109b2-1336-48fc-a849-f2d039eae8e1","Type":"ContainerDied","Data":"af79900fcf86fa035a9726c5c728cacfe56afa771bbdcc93b96783c84f36fdc3"} Dec 03 14:29:02 crc kubenswrapper[4805]: I1203 14:29:02.031745 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-666b6646f7-x75rp" event={"ID":"eba109b2-1336-48fc-a849-f2d039eae8e1","Type":"ContainerDied","Data":"0a92596606eb5fd52e10e78fa863b2fbee542e7cf1c03cecef092378ab26e9f5"} Dec 03 14:29:02 crc kubenswrapper[4805]: I1203 14:29:02.031772 4805 scope.go:117] "RemoveContainer" containerID="af79900fcf86fa035a9726c5c728cacfe56afa771bbdcc93b96783c84f36fdc3" Dec 03 14:29:02 crc kubenswrapper[4805]: I1203 14:29:02.031914 4805 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-666b6646f7-x75rp" Dec 03 14:29:02 crc kubenswrapper[4805]: I1203 14:29:02.037773 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-0" event={"ID":"17f337d8-be50-4c6c-8364-208b394be6d3","Type":"ContainerStarted","Data":"577c247d02738dd81c684e2e7d24e9471ecd9e705de69bd9db75ee40f123cf4a"} Dec 03 14:29:02 crc kubenswrapper[4805]: I1203 14:29:02.039796 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-0" event={"ID":"00ac9536-1677-4a3a-be82-7c8d8df91656","Type":"ContainerStarted","Data":"5ca64542fb58b2a01f9b0b5fc3bbbdebe86a51dc29decf72249eea1fed682953"} Dec 03 14:29:02 crc kubenswrapper[4805]: I1203 14:29:02.043630 4805 generic.go:334] "Generic (PLEG): container finished" podID="c13cbc51-432a-419a-9469-37fd31a82dd9" containerID="860b1712352b1e14def5d9057871da99d65fbdcce05423a307ac73ce697a415e" exitCode=0 Dec 03 14:29:02 crc kubenswrapper[4805]: I1203 14:29:02.043689 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"c13cbc51-432a-419a-9469-37fd31a82dd9","Type":"ContainerDied","Data":"860b1712352b1e14def5d9057871da99d65fbdcce05423a307ac73ce697a415e"} Dec 03 14:29:02 crc kubenswrapper[4805]: I1203 14:29:02.055979 4805 scope.go:117] "RemoveContainer" containerID="4e7aa0ff2ac6d5331aa4f6d2c86466a7274ce5897b7c11798e99c07d1bc72687" Dec 03 14:29:02 crc kubenswrapper[4805]: I1203 14:29:02.077989 4805 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovsdbserver-sb-0" podStartSLOduration=5.606205192 podStartE2EDuration="19.077969225s" podCreationTimestamp="2025-12-03 14:28:43 +0000 UTC" firstStartedPulling="2025-12-03 14:28:47.702208531 +0000 UTC m=+1157.365125474" lastFinishedPulling="2025-12-03 14:29:01.173972584 +0000 UTC m=+1170.836889507" observedRunningTime="2025-12-03 14:29:02.058228151 +0000 UTC m=+1171.721145074" watchObservedRunningTime="2025-12-03 14:29:02.077969225 +0000 UTC m=+1171.740886148" Dec 03 14:29:02 crc kubenswrapper[4805]: I1203 14:29:02.100920 4805 scope.go:117] "RemoveContainer" containerID="af79900fcf86fa035a9726c5c728cacfe56afa771bbdcc93b96783c84f36fdc3" Dec 03 14:29:02 crc kubenswrapper[4805]: E1203 14:29:02.106671 4805 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"af79900fcf86fa035a9726c5c728cacfe56afa771bbdcc93b96783c84f36fdc3\": container with ID starting with af79900fcf86fa035a9726c5c728cacfe56afa771bbdcc93b96783c84f36fdc3 not found: ID does not exist" containerID="af79900fcf86fa035a9726c5c728cacfe56afa771bbdcc93b96783c84f36fdc3" Dec 03 14:29:02 crc kubenswrapper[4805]: I1203 14:29:02.107614 4805 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"af79900fcf86fa035a9726c5c728cacfe56afa771bbdcc93b96783c84f36fdc3"} err="failed to get container status \"af79900fcf86fa035a9726c5c728cacfe56afa771bbdcc93b96783c84f36fdc3\": rpc error: code = NotFound desc = could not find container \"af79900fcf86fa035a9726c5c728cacfe56afa771bbdcc93b96783c84f36fdc3\": container with ID starting with af79900fcf86fa035a9726c5c728cacfe56afa771bbdcc93b96783c84f36fdc3 not found: ID does not exist" Dec 03 14:29:02 crc kubenswrapper[4805]: I1203 14:29:02.107649 4805 scope.go:117] "RemoveContainer" containerID="4e7aa0ff2ac6d5331aa4f6d2c86466a7274ce5897b7c11798e99c07d1bc72687" Dec 03 14:29:02 crc kubenswrapper[4805]: E1203 14:29:02.109740 4805 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4e7aa0ff2ac6d5331aa4f6d2c86466a7274ce5897b7c11798e99c07d1bc72687\": container with ID starting with 4e7aa0ff2ac6d5331aa4f6d2c86466a7274ce5897b7c11798e99c07d1bc72687 not found: ID does not exist" containerID="4e7aa0ff2ac6d5331aa4f6d2c86466a7274ce5897b7c11798e99c07d1bc72687" Dec 03 14:29:02 crc kubenswrapper[4805]: I1203 14:29:02.109805 4805 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4e7aa0ff2ac6d5331aa4f6d2c86466a7274ce5897b7c11798e99c07d1bc72687"} err="failed to get container status \"4e7aa0ff2ac6d5331aa4f6d2c86466a7274ce5897b7c11798e99c07d1bc72687\": rpc error: code = NotFound desc = could not find container \"4e7aa0ff2ac6d5331aa4f6d2c86466a7274ce5897b7c11798e99c07d1bc72687\": container with ID starting with 4e7aa0ff2ac6d5331aa4f6d2c86466a7274ce5897b7c11798e99c07d1bc72687 not found: ID does not exist" Dec 03 14:29:02 crc kubenswrapper[4805]: I1203 14:29:02.114699 4805 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovsdbserver-nb-0" podStartSLOduration=7.593824844 podStartE2EDuration="22.114679109s" podCreationTimestamp="2025-12-03 14:28:40 +0000 UTC" firstStartedPulling="2025-12-03 14:28:46.634549496 +0000 UTC m=+1156.297466409" lastFinishedPulling="2025-12-03 14:29:01.155403751 +0000 UTC m=+1170.818320674" observedRunningTime="2025-12-03 14:29:02.078320835 +0000 UTC m=+1171.741237758" watchObservedRunningTime="2025-12-03 14:29:02.114679109 +0000 UTC m=+1171.777596032" Dec 03 14:29:02 crc kubenswrapper[4805]: I1203 14:29:02.120100 4805 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-666b6646f7-x75rp"] Dec 03 14:29:02 crc kubenswrapper[4805]: I1203 14:29:02.126097 4805 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-666b6646f7-x75rp"] Dec 03 14:29:02 crc kubenswrapper[4805]: I1203 14:29:02.378257 4805 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/ovsdbserver-nb-0" Dec 03 14:29:02 crc kubenswrapper[4805]: I1203 14:29:02.415106 4805 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/ovsdbserver-nb-0" Dec 03 14:29:02 crc kubenswrapper[4805]: I1203 14:29:02.705162 4805 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="eba109b2-1336-48fc-a849-f2d039eae8e1" path="/var/lib/kubelet/pods/eba109b2-1336-48fc-a849-f2d039eae8e1/volumes" Dec 03 14:29:02 crc kubenswrapper[4805]: I1203 14:29:02.973607 4805 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/ovsdbserver-sb-0" Dec 03 14:29:03 crc kubenswrapper[4805]: I1203 14:29:03.018714 4805 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/ovsdbserver-sb-0" Dec 03 14:29:03 crc kubenswrapper[4805]: I1203 14:29:03.052705 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"c13cbc51-432a-419a-9469-37fd31a82dd9","Type":"ContainerStarted","Data":"3212f8f52fa39abfcf90c7b4e0c1b862de420b51337cafef22dd81564334e00c"} Dec 03 14:29:03 crc kubenswrapper[4805]: I1203 14:29:03.054296 4805 generic.go:334] "Generic (PLEG): container finished" podID="5b4bb902-dba6-4ec6-97bf-532d996e918d" containerID="837bce806cd5be22704075197396e52bd05cc8c2b364b8b7e09d291129690d6a" exitCode=0 Dec 03 14:29:03 crc kubenswrapper[4805]: I1203 14:29:03.054357 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"5b4bb902-dba6-4ec6-97bf-532d996e918d","Type":"ContainerDied","Data":"837bce806cd5be22704075197396e52bd05cc8c2b364b8b7e09d291129690d6a"} Dec 03 14:29:03 crc kubenswrapper[4805]: I1203 14:29:03.056015 4805 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovsdbserver-sb-0" Dec 03 14:29:03 crc kubenswrapper[4805]: I1203 14:29:03.056288 4805 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovsdbserver-nb-0" Dec 03 14:29:03 crc kubenswrapper[4805]: I1203 14:29:03.082231 4805 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/openstack-cell1-galera-0" podStartSLOduration=20.580518342 podStartE2EDuration="30.082210309s" podCreationTimestamp="2025-12-03 14:28:33 +0000 UTC" firstStartedPulling="2025-12-03 14:28:46.590033119 +0000 UTC m=+1156.252950042" lastFinishedPulling="2025-12-03 14:28:56.091725086 +0000 UTC m=+1165.754642009" observedRunningTime="2025-12-03 14:29:03.075926689 +0000 UTC m=+1172.738843612" watchObservedRunningTime="2025-12-03 14:29:03.082210309 +0000 UTC m=+1172.745127242" Dec 03 14:29:03 crc kubenswrapper[4805]: I1203 14:29:03.103609 4805 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovsdbserver-nb-0" Dec 03 14:29:03 crc kubenswrapper[4805]: I1203 14:29:03.113173 4805 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovsdbserver-sb-0" Dec 03 14:29:03 crc kubenswrapper[4805]: I1203 14:29:03.284777 4805 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-7fd796d7df-7l7rr"] Dec 03 14:29:03 crc kubenswrapper[4805]: E1203 14:29:03.285162 4805 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="eba109b2-1336-48fc-a849-f2d039eae8e1" containerName="init" Dec 03 14:29:03 crc kubenswrapper[4805]: I1203 14:29:03.285179 4805 state_mem.go:107] "Deleted CPUSet assignment" podUID="eba109b2-1336-48fc-a849-f2d039eae8e1" containerName="init" Dec 03 14:29:03 crc kubenswrapper[4805]: E1203 14:29:03.285198 4805 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="eba109b2-1336-48fc-a849-f2d039eae8e1" containerName="dnsmasq-dns" Dec 03 14:29:03 crc kubenswrapper[4805]: I1203 14:29:03.285205 4805 state_mem.go:107] "Deleted CPUSet assignment" podUID="eba109b2-1336-48fc-a849-f2d039eae8e1" containerName="dnsmasq-dns" Dec 03 14:29:03 crc kubenswrapper[4805]: I1203 14:29:03.285351 4805 memory_manager.go:354] "RemoveStaleState removing state" podUID="eba109b2-1336-48fc-a849-f2d039eae8e1" containerName="dnsmasq-dns" Dec 03 14:29:03 crc kubenswrapper[4805]: I1203 14:29:03.286181 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7fd796d7df-7l7rr" Dec 03 14:29:03 crc kubenswrapper[4805]: I1203 14:29:03.289469 4805 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovsdbserver-nb" Dec 03 14:29:03 crc kubenswrapper[4805]: I1203 14:29:03.312182 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-7fd796d7df-7l7rr"] Dec 03 14:29:03 crc kubenswrapper[4805]: I1203 14:29:03.383212 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/0e091270-fda5-4856-9dd8-894e830d07e3-dns-svc\") pod \"dnsmasq-dns-7fd796d7df-7l7rr\" (UID: \"0e091270-fda5-4856-9dd8-894e830d07e3\") " pod="openstack/dnsmasq-dns-7fd796d7df-7l7rr" Dec 03 14:29:03 crc kubenswrapper[4805]: I1203 14:29:03.383270 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0e091270-fda5-4856-9dd8-894e830d07e3-config\") pod \"dnsmasq-dns-7fd796d7df-7l7rr\" (UID: \"0e091270-fda5-4856-9dd8-894e830d07e3\") " pod="openstack/dnsmasq-dns-7fd796d7df-7l7rr" Dec 03 14:29:03 crc kubenswrapper[4805]: I1203 14:29:03.383343 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/0e091270-fda5-4856-9dd8-894e830d07e3-ovsdbserver-nb\") pod \"dnsmasq-dns-7fd796d7df-7l7rr\" (UID: \"0e091270-fda5-4856-9dd8-894e830d07e3\") " pod="openstack/dnsmasq-dns-7fd796d7df-7l7rr" Dec 03 14:29:03 crc kubenswrapper[4805]: I1203 14:29:03.383534 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ktm7p\" (UniqueName: \"kubernetes.io/projected/0e091270-fda5-4856-9dd8-894e830d07e3-kube-api-access-ktm7p\") pod \"dnsmasq-dns-7fd796d7df-7l7rr\" (UID: \"0e091270-fda5-4856-9dd8-894e830d07e3\") " pod="openstack/dnsmasq-dns-7fd796d7df-7l7rr" Dec 03 14:29:03 crc kubenswrapper[4805]: I1203 14:29:03.484763 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/0e091270-fda5-4856-9dd8-894e830d07e3-ovsdbserver-nb\") pod \"dnsmasq-dns-7fd796d7df-7l7rr\" (UID: \"0e091270-fda5-4856-9dd8-894e830d07e3\") " pod="openstack/dnsmasq-dns-7fd796d7df-7l7rr" Dec 03 14:29:03 crc kubenswrapper[4805]: I1203 14:29:03.484860 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ktm7p\" (UniqueName: \"kubernetes.io/projected/0e091270-fda5-4856-9dd8-894e830d07e3-kube-api-access-ktm7p\") pod \"dnsmasq-dns-7fd796d7df-7l7rr\" (UID: \"0e091270-fda5-4856-9dd8-894e830d07e3\") " pod="openstack/dnsmasq-dns-7fd796d7df-7l7rr" Dec 03 14:29:03 crc kubenswrapper[4805]: I1203 14:29:03.484929 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/0e091270-fda5-4856-9dd8-894e830d07e3-dns-svc\") pod \"dnsmasq-dns-7fd796d7df-7l7rr\" (UID: \"0e091270-fda5-4856-9dd8-894e830d07e3\") " pod="openstack/dnsmasq-dns-7fd796d7df-7l7rr" Dec 03 14:29:03 crc kubenswrapper[4805]: I1203 14:29:03.484976 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0e091270-fda5-4856-9dd8-894e830d07e3-config\") pod \"dnsmasq-dns-7fd796d7df-7l7rr\" (UID: \"0e091270-fda5-4856-9dd8-894e830d07e3\") " pod="openstack/dnsmasq-dns-7fd796d7df-7l7rr" Dec 03 14:29:03 crc kubenswrapper[4805]: I1203 14:29:03.485570 4805 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-metrics-7kmnp"] Dec 03 14:29:03 crc kubenswrapper[4805]: I1203 14:29:03.486052 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0e091270-fda5-4856-9dd8-894e830d07e3-config\") pod \"dnsmasq-dns-7fd796d7df-7l7rr\" (UID: \"0e091270-fda5-4856-9dd8-894e830d07e3\") " pod="openstack/dnsmasq-dns-7fd796d7df-7l7rr" Dec 03 14:29:03 crc kubenswrapper[4805]: I1203 14:29:03.486586 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/0e091270-fda5-4856-9dd8-894e830d07e3-ovsdbserver-nb\") pod \"dnsmasq-dns-7fd796d7df-7l7rr\" (UID: \"0e091270-fda5-4856-9dd8-894e830d07e3\") " pod="openstack/dnsmasq-dns-7fd796d7df-7l7rr" Dec 03 14:29:03 crc kubenswrapper[4805]: I1203 14:29:03.487119 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/0e091270-fda5-4856-9dd8-894e830d07e3-dns-svc\") pod \"dnsmasq-dns-7fd796d7df-7l7rr\" (UID: \"0e091270-fda5-4856-9dd8-894e830d07e3\") " pod="openstack/dnsmasq-dns-7fd796d7df-7l7rr" Dec 03 14:29:03 crc kubenswrapper[4805]: I1203 14:29:03.487137 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-metrics-7kmnp" Dec 03 14:29:03 crc kubenswrapper[4805]: I1203 14:29:03.494799 4805 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovncontroller-metrics-config" Dec 03 14:29:03 crc kubenswrapper[4805]: I1203 14:29:03.511702 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ktm7p\" (UniqueName: \"kubernetes.io/projected/0e091270-fda5-4856-9dd8-894e830d07e3-kube-api-access-ktm7p\") pod \"dnsmasq-dns-7fd796d7df-7l7rr\" (UID: \"0e091270-fda5-4856-9dd8-894e830d07e3\") " pod="openstack/dnsmasq-dns-7fd796d7df-7l7rr" Dec 03 14:29:03 crc kubenswrapper[4805]: I1203 14:29:03.514539 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-metrics-7kmnp"] Dec 03 14:29:03 crc kubenswrapper[4805]: I1203 14:29:03.563517 4805 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-northd-0"] Dec 03 14:29:03 crc kubenswrapper[4805]: I1203 14:29:03.566042 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-northd-0" Dec 03 14:29:03 crc kubenswrapper[4805]: I1203 14:29:03.571139 4805 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovnnorthd-scripts" Dec 03 14:29:03 crc kubenswrapper[4805]: I1203 14:29:03.571314 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovnnorthd-ovndbs" Dec 03 14:29:03 crc kubenswrapper[4805]: I1203 14:29:03.571473 4805 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovnnorthd-config" Dec 03 14:29:03 crc kubenswrapper[4805]: I1203 14:29:03.571483 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ovnnorthd-ovnnorthd-dockercfg-c24wf" Dec 03 14:29:03 crc kubenswrapper[4805]: I1203 14:29:03.574978 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-northd-0"] Dec 03 14:29:03 crc kubenswrapper[4805]: I1203 14:29:03.586137 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/76002c62-88cb-4f74-816a-3b2be4f8ba4c-config\") pod \"ovn-controller-metrics-7kmnp\" (UID: \"76002c62-88cb-4f74-816a-3b2be4f8ba4c\") " pod="openstack/ovn-controller-metrics-7kmnp" Dec 03 14:29:03 crc kubenswrapper[4805]: I1203 14:29:03.586232 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/host-path/76002c62-88cb-4f74-816a-3b2be4f8ba4c-ovn-rundir\") pod \"ovn-controller-metrics-7kmnp\" (UID: \"76002c62-88cb-4f74-816a-3b2be4f8ba4c\") " pod="openstack/ovn-controller-metrics-7kmnp" Dec 03 14:29:03 crc kubenswrapper[4805]: I1203 14:29:03.586271 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/76002c62-88cb-4f74-816a-3b2be4f8ba4c-combined-ca-bundle\") pod \"ovn-controller-metrics-7kmnp\" (UID: \"76002c62-88cb-4f74-816a-3b2be4f8ba4c\") " pod="openstack/ovn-controller-metrics-7kmnp" Dec 03 14:29:03 crc kubenswrapper[4805]: I1203 14:29:03.586294 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovs-rundir\" (UniqueName: \"kubernetes.io/host-path/76002c62-88cb-4f74-816a-3b2be4f8ba4c-ovs-rundir\") pod \"ovn-controller-metrics-7kmnp\" (UID: \"76002c62-88cb-4f74-816a-3b2be4f8ba4c\") " pod="openstack/ovn-controller-metrics-7kmnp" Dec 03 14:29:03 crc kubenswrapper[4805]: I1203 14:29:03.586326 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-82wwn\" (UniqueName: \"kubernetes.io/projected/76002c62-88cb-4f74-816a-3b2be4f8ba4c-kube-api-access-82wwn\") pod \"ovn-controller-metrics-7kmnp\" (UID: \"76002c62-88cb-4f74-816a-3b2be4f8ba4c\") " pod="openstack/ovn-controller-metrics-7kmnp" Dec 03 14:29:03 crc kubenswrapper[4805]: I1203 14:29:03.586415 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/76002c62-88cb-4f74-816a-3b2be4f8ba4c-metrics-certs-tls-certs\") pod \"ovn-controller-metrics-7kmnp\" (UID: \"76002c62-88cb-4f74-816a-3b2be4f8ba4c\") " pod="openstack/ovn-controller-metrics-7kmnp" Dec 03 14:29:03 crc kubenswrapper[4805]: I1203 14:29:03.601710 4805 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-7fd796d7df-7l7rr"] Dec 03 14:29:03 crc kubenswrapper[4805]: I1203 14:29:03.602384 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7fd796d7df-7l7rr" Dec 03 14:29:03 crc kubenswrapper[4805]: I1203 14:29:03.641448 4805 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-86db49b7ff-x7rps"] Dec 03 14:29:03 crc kubenswrapper[4805]: I1203 14:29:03.645146 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-86db49b7ff-x7rps" Dec 03 14:29:03 crc kubenswrapper[4805]: I1203 14:29:03.650699 4805 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovsdbserver-sb" Dec 03 14:29:03 crc kubenswrapper[4805]: I1203 14:29:03.651935 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-86db49b7ff-x7rps"] Dec 03 14:29:03 crc kubenswrapper[4805]: I1203 14:29:03.687771 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8e6ca40c-f7f2-445e-8e22-71f8d4d4693b-config\") pod \"ovn-northd-0\" (UID: \"8e6ca40c-f7f2-445e-8e22-71f8d4d4693b\") " pod="openstack/ovn-northd-0" Dec 03 14:29:03 crc kubenswrapper[4805]: I1203 14:29:03.688305 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/76002c62-88cb-4f74-816a-3b2be4f8ba4c-combined-ca-bundle\") pod \"ovn-controller-metrics-7kmnp\" (UID: \"76002c62-88cb-4f74-816a-3b2be4f8ba4c\") " pod="openstack/ovn-controller-metrics-7kmnp" Dec 03 14:29:03 crc kubenswrapper[4805]: I1203 14:29:03.688399 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovs-rundir\" (UniqueName: \"kubernetes.io/host-path/76002c62-88cb-4f74-816a-3b2be4f8ba4c-ovs-rundir\") pod \"ovn-controller-metrics-7kmnp\" (UID: \"76002c62-88cb-4f74-816a-3b2be4f8ba4c\") " pod="openstack/ovn-controller-metrics-7kmnp" Dec 03 14:29:03 crc kubenswrapper[4805]: I1203 14:29:03.688498 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-82wwn\" (UniqueName: \"kubernetes.io/projected/76002c62-88cb-4f74-816a-3b2be4f8ba4c-kube-api-access-82wwn\") pod \"ovn-controller-metrics-7kmnp\" (UID: \"76002c62-88cb-4f74-816a-3b2be4f8ba4c\") " pod="openstack/ovn-controller-metrics-7kmnp" Dec 03 14:29:03 crc kubenswrapper[4805]: I1203 14:29:03.688585 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-f8xgx\" (UniqueName: \"kubernetes.io/projected/8e6ca40c-f7f2-445e-8e22-71f8d4d4693b-kube-api-access-f8xgx\") pod \"ovn-northd-0\" (UID: \"8e6ca40c-f7f2-445e-8e22-71f8d4d4693b\") " pod="openstack/ovn-northd-0" Dec 03 14:29:03 crc kubenswrapper[4805]: I1203 14:29:03.688777 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/8e6ca40c-f7f2-445e-8e22-71f8d4d4693b-scripts\") pod \"ovn-northd-0\" (UID: \"8e6ca40c-f7f2-445e-8e22-71f8d4d4693b\") " pod="openstack/ovn-northd-0" Dec 03 14:29:03 crc kubenswrapper[4805]: I1203 14:29:03.688899 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8e6ca40c-f7f2-445e-8e22-71f8d4d4693b-combined-ca-bundle\") pod \"ovn-northd-0\" (UID: \"8e6ca40c-f7f2-445e-8e22-71f8d4d4693b\") " pod="openstack/ovn-northd-0" Dec 03 14:29:03 crc kubenswrapper[4805]: I1203 14:29:03.688985 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/8e6ca40c-f7f2-445e-8e22-71f8d4d4693b-metrics-certs-tls-certs\") pod \"ovn-northd-0\" (UID: \"8e6ca40c-f7f2-445e-8e22-71f8d4d4693b\") " pod="openstack/ovn-northd-0" Dec 03 14:29:03 crc kubenswrapper[4805]: I1203 14:29:03.689086 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-northd-tls-certs\" (UniqueName: \"kubernetes.io/secret/8e6ca40c-f7f2-445e-8e22-71f8d4d4693b-ovn-northd-tls-certs\") pod \"ovn-northd-0\" (UID: \"8e6ca40c-f7f2-445e-8e22-71f8d4d4693b\") " pod="openstack/ovn-northd-0" Dec 03 14:29:03 crc kubenswrapper[4805]: I1203 14:29:03.689174 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/empty-dir/8e6ca40c-f7f2-445e-8e22-71f8d4d4693b-ovn-rundir\") pod \"ovn-northd-0\" (UID: \"8e6ca40c-f7f2-445e-8e22-71f8d4d4693b\") " pod="openstack/ovn-northd-0" Dec 03 14:29:03 crc kubenswrapper[4805]: I1203 14:29:03.689303 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/76002c62-88cb-4f74-816a-3b2be4f8ba4c-metrics-certs-tls-certs\") pod \"ovn-controller-metrics-7kmnp\" (UID: \"76002c62-88cb-4f74-816a-3b2be4f8ba4c\") " pod="openstack/ovn-controller-metrics-7kmnp" Dec 03 14:29:03 crc kubenswrapper[4805]: I1203 14:29:03.689388 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/76002c62-88cb-4f74-816a-3b2be4f8ba4c-config\") pod \"ovn-controller-metrics-7kmnp\" (UID: \"76002c62-88cb-4f74-816a-3b2be4f8ba4c\") " pod="openstack/ovn-controller-metrics-7kmnp" Dec 03 14:29:03 crc kubenswrapper[4805]: I1203 14:29:03.689544 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/host-path/76002c62-88cb-4f74-816a-3b2be4f8ba4c-ovn-rundir\") pod \"ovn-controller-metrics-7kmnp\" (UID: \"76002c62-88cb-4f74-816a-3b2be4f8ba4c\") " pod="openstack/ovn-controller-metrics-7kmnp" Dec 03 14:29:03 crc kubenswrapper[4805]: I1203 14:29:03.691169 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/host-path/76002c62-88cb-4f74-816a-3b2be4f8ba4c-ovn-rundir\") pod \"ovn-controller-metrics-7kmnp\" (UID: \"76002c62-88cb-4f74-816a-3b2be4f8ba4c\") " pod="openstack/ovn-controller-metrics-7kmnp" Dec 03 14:29:03 crc kubenswrapper[4805]: I1203 14:29:03.691320 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovs-rundir\" (UniqueName: \"kubernetes.io/host-path/76002c62-88cb-4f74-816a-3b2be4f8ba4c-ovs-rundir\") pod \"ovn-controller-metrics-7kmnp\" (UID: \"76002c62-88cb-4f74-816a-3b2be4f8ba4c\") " pod="openstack/ovn-controller-metrics-7kmnp" Dec 03 14:29:03 crc kubenswrapper[4805]: I1203 14:29:03.693039 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/76002c62-88cb-4f74-816a-3b2be4f8ba4c-combined-ca-bundle\") pod \"ovn-controller-metrics-7kmnp\" (UID: \"76002c62-88cb-4f74-816a-3b2be4f8ba4c\") " pod="openstack/ovn-controller-metrics-7kmnp" Dec 03 14:29:03 crc kubenswrapper[4805]: I1203 14:29:03.693542 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/76002c62-88cb-4f74-816a-3b2be4f8ba4c-config\") pod \"ovn-controller-metrics-7kmnp\" (UID: \"76002c62-88cb-4f74-816a-3b2be4f8ba4c\") " pod="openstack/ovn-controller-metrics-7kmnp" Dec 03 14:29:03 crc kubenswrapper[4805]: I1203 14:29:03.699743 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/76002c62-88cb-4f74-816a-3b2be4f8ba4c-metrics-certs-tls-certs\") pod \"ovn-controller-metrics-7kmnp\" (UID: \"76002c62-88cb-4f74-816a-3b2be4f8ba4c\") " pod="openstack/ovn-controller-metrics-7kmnp" Dec 03 14:29:03 crc kubenswrapper[4805]: I1203 14:29:03.718434 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-82wwn\" (UniqueName: \"kubernetes.io/projected/76002c62-88cb-4f74-816a-3b2be4f8ba4c-kube-api-access-82wwn\") pod \"ovn-controller-metrics-7kmnp\" (UID: \"76002c62-88cb-4f74-816a-3b2be4f8ba4c\") " pod="openstack/ovn-controller-metrics-7kmnp" Dec 03 14:29:03 crc kubenswrapper[4805]: I1203 14:29:03.791469 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/e7db8f49-10ce-4818-8d9b-963d6e2e49ce-ovsdbserver-sb\") pod \"dnsmasq-dns-86db49b7ff-x7rps\" (UID: \"e7db8f49-10ce-4818-8d9b-963d6e2e49ce\") " pod="openstack/dnsmasq-dns-86db49b7ff-x7rps" Dec 03 14:29:03 crc kubenswrapper[4805]: I1203 14:29:03.791548 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cp8vh\" (UniqueName: \"kubernetes.io/projected/e7db8f49-10ce-4818-8d9b-963d6e2e49ce-kube-api-access-cp8vh\") pod \"dnsmasq-dns-86db49b7ff-x7rps\" (UID: \"e7db8f49-10ce-4818-8d9b-963d6e2e49ce\") " pod="openstack/dnsmasq-dns-86db49b7ff-x7rps" Dec 03 14:29:03 crc kubenswrapper[4805]: I1203 14:29:03.791605 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8e6ca40c-f7f2-445e-8e22-71f8d4d4693b-config\") pod \"ovn-northd-0\" (UID: \"8e6ca40c-f7f2-445e-8e22-71f8d4d4693b\") " pod="openstack/ovn-northd-0" Dec 03 14:29:03 crc kubenswrapper[4805]: I1203 14:29:03.791639 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/e7db8f49-10ce-4818-8d9b-963d6e2e49ce-ovsdbserver-nb\") pod \"dnsmasq-dns-86db49b7ff-x7rps\" (UID: \"e7db8f49-10ce-4818-8d9b-963d6e2e49ce\") " pod="openstack/dnsmasq-dns-86db49b7ff-x7rps" Dec 03 14:29:03 crc kubenswrapper[4805]: I1203 14:29:03.791674 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/e7db8f49-10ce-4818-8d9b-963d6e2e49ce-dns-svc\") pod \"dnsmasq-dns-86db49b7ff-x7rps\" (UID: \"e7db8f49-10ce-4818-8d9b-963d6e2e49ce\") " pod="openstack/dnsmasq-dns-86db49b7ff-x7rps" Dec 03 14:29:03 crc kubenswrapper[4805]: I1203 14:29:03.791703 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-f8xgx\" (UniqueName: \"kubernetes.io/projected/8e6ca40c-f7f2-445e-8e22-71f8d4d4693b-kube-api-access-f8xgx\") pod \"ovn-northd-0\" (UID: \"8e6ca40c-f7f2-445e-8e22-71f8d4d4693b\") " pod="openstack/ovn-northd-0" Dec 03 14:29:03 crc kubenswrapper[4805]: I1203 14:29:03.791735 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/8e6ca40c-f7f2-445e-8e22-71f8d4d4693b-scripts\") pod \"ovn-northd-0\" (UID: \"8e6ca40c-f7f2-445e-8e22-71f8d4d4693b\") " pod="openstack/ovn-northd-0" Dec 03 14:29:03 crc kubenswrapper[4805]: I1203 14:29:03.791779 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8e6ca40c-f7f2-445e-8e22-71f8d4d4693b-combined-ca-bundle\") pod \"ovn-northd-0\" (UID: \"8e6ca40c-f7f2-445e-8e22-71f8d4d4693b\") " pod="openstack/ovn-northd-0" Dec 03 14:29:03 crc kubenswrapper[4805]: I1203 14:29:03.791804 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/8e6ca40c-f7f2-445e-8e22-71f8d4d4693b-metrics-certs-tls-certs\") pod \"ovn-northd-0\" (UID: \"8e6ca40c-f7f2-445e-8e22-71f8d4d4693b\") " pod="openstack/ovn-northd-0" Dec 03 14:29:03 crc kubenswrapper[4805]: I1203 14:29:03.791854 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-northd-tls-certs\" (UniqueName: \"kubernetes.io/secret/8e6ca40c-f7f2-445e-8e22-71f8d4d4693b-ovn-northd-tls-certs\") pod \"ovn-northd-0\" (UID: \"8e6ca40c-f7f2-445e-8e22-71f8d4d4693b\") " pod="openstack/ovn-northd-0" Dec 03 14:29:03 crc kubenswrapper[4805]: I1203 14:29:03.791878 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/empty-dir/8e6ca40c-f7f2-445e-8e22-71f8d4d4693b-ovn-rundir\") pod \"ovn-northd-0\" (UID: \"8e6ca40c-f7f2-445e-8e22-71f8d4d4693b\") " pod="openstack/ovn-northd-0" Dec 03 14:29:03 crc kubenswrapper[4805]: I1203 14:29:03.791947 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e7db8f49-10ce-4818-8d9b-963d6e2e49ce-config\") pod \"dnsmasq-dns-86db49b7ff-x7rps\" (UID: \"e7db8f49-10ce-4818-8d9b-963d6e2e49ce\") " pod="openstack/dnsmasq-dns-86db49b7ff-x7rps" Dec 03 14:29:03 crc kubenswrapper[4805]: I1203 14:29:03.793326 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8e6ca40c-f7f2-445e-8e22-71f8d4d4693b-config\") pod \"ovn-northd-0\" (UID: \"8e6ca40c-f7f2-445e-8e22-71f8d4d4693b\") " pod="openstack/ovn-northd-0" Dec 03 14:29:03 crc kubenswrapper[4805]: I1203 14:29:03.793610 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/empty-dir/8e6ca40c-f7f2-445e-8e22-71f8d4d4693b-ovn-rundir\") pod \"ovn-northd-0\" (UID: \"8e6ca40c-f7f2-445e-8e22-71f8d4d4693b\") " pod="openstack/ovn-northd-0" Dec 03 14:29:03 crc kubenswrapper[4805]: I1203 14:29:03.794025 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/8e6ca40c-f7f2-445e-8e22-71f8d4d4693b-scripts\") pod \"ovn-northd-0\" (UID: \"8e6ca40c-f7f2-445e-8e22-71f8d4d4693b\") " pod="openstack/ovn-northd-0" Dec 03 14:29:03 crc kubenswrapper[4805]: I1203 14:29:03.797698 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/8e6ca40c-f7f2-445e-8e22-71f8d4d4693b-metrics-certs-tls-certs\") pod \"ovn-northd-0\" (UID: \"8e6ca40c-f7f2-445e-8e22-71f8d4d4693b\") " pod="openstack/ovn-northd-0" Dec 03 14:29:03 crc kubenswrapper[4805]: I1203 14:29:03.798983 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8e6ca40c-f7f2-445e-8e22-71f8d4d4693b-combined-ca-bundle\") pod \"ovn-northd-0\" (UID: \"8e6ca40c-f7f2-445e-8e22-71f8d4d4693b\") " pod="openstack/ovn-northd-0" Dec 03 14:29:03 crc kubenswrapper[4805]: I1203 14:29:03.801769 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-northd-tls-certs\" (UniqueName: \"kubernetes.io/secret/8e6ca40c-f7f2-445e-8e22-71f8d4d4693b-ovn-northd-tls-certs\") pod \"ovn-northd-0\" (UID: \"8e6ca40c-f7f2-445e-8e22-71f8d4d4693b\") " pod="openstack/ovn-northd-0" Dec 03 14:29:03 crc kubenswrapper[4805]: I1203 14:29:03.804161 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-metrics-7kmnp" Dec 03 14:29:03 crc kubenswrapper[4805]: I1203 14:29:03.825615 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-f8xgx\" (UniqueName: \"kubernetes.io/projected/8e6ca40c-f7f2-445e-8e22-71f8d4d4693b-kube-api-access-f8xgx\") pod \"ovn-northd-0\" (UID: \"8e6ca40c-f7f2-445e-8e22-71f8d4d4693b\") " pod="openstack/ovn-northd-0" Dec 03 14:29:03 crc kubenswrapper[4805]: I1203 14:29:03.893254 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-northd-0" Dec 03 14:29:03 crc kubenswrapper[4805]: I1203 14:29:03.893671 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/e7db8f49-10ce-4818-8d9b-963d6e2e49ce-dns-svc\") pod \"dnsmasq-dns-86db49b7ff-x7rps\" (UID: \"e7db8f49-10ce-4818-8d9b-963d6e2e49ce\") " pod="openstack/dnsmasq-dns-86db49b7ff-x7rps" Dec 03 14:29:03 crc kubenswrapper[4805]: I1203 14:29:03.893777 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e7db8f49-10ce-4818-8d9b-963d6e2e49ce-config\") pod \"dnsmasq-dns-86db49b7ff-x7rps\" (UID: \"e7db8f49-10ce-4818-8d9b-963d6e2e49ce\") " pod="openstack/dnsmasq-dns-86db49b7ff-x7rps" Dec 03 14:29:03 crc kubenswrapper[4805]: I1203 14:29:03.893808 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/e7db8f49-10ce-4818-8d9b-963d6e2e49ce-ovsdbserver-sb\") pod \"dnsmasq-dns-86db49b7ff-x7rps\" (UID: \"e7db8f49-10ce-4818-8d9b-963d6e2e49ce\") " pod="openstack/dnsmasq-dns-86db49b7ff-x7rps" Dec 03 14:29:03 crc kubenswrapper[4805]: I1203 14:29:03.893846 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cp8vh\" (UniqueName: \"kubernetes.io/projected/e7db8f49-10ce-4818-8d9b-963d6e2e49ce-kube-api-access-cp8vh\") pod \"dnsmasq-dns-86db49b7ff-x7rps\" (UID: \"e7db8f49-10ce-4818-8d9b-963d6e2e49ce\") " pod="openstack/dnsmasq-dns-86db49b7ff-x7rps" Dec 03 14:29:03 crc kubenswrapper[4805]: I1203 14:29:03.893877 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/e7db8f49-10ce-4818-8d9b-963d6e2e49ce-ovsdbserver-nb\") pod \"dnsmasq-dns-86db49b7ff-x7rps\" (UID: \"e7db8f49-10ce-4818-8d9b-963d6e2e49ce\") " pod="openstack/dnsmasq-dns-86db49b7ff-x7rps" Dec 03 14:29:03 crc kubenswrapper[4805]: I1203 14:29:03.894423 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/e7db8f49-10ce-4818-8d9b-963d6e2e49ce-dns-svc\") pod \"dnsmasq-dns-86db49b7ff-x7rps\" (UID: \"e7db8f49-10ce-4818-8d9b-963d6e2e49ce\") " pod="openstack/dnsmasq-dns-86db49b7ff-x7rps" Dec 03 14:29:03 crc kubenswrapper[4805]: I1203 14:29:03.898147 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/e7db8f49-10ce-4818-8d9b-963d6e2e49ce-ovsdbserver-nb\") pod \"dnsmasq-dns-86db49b7ff-x7rps\" (UID: \"e7db8f49-10ce-4818-8d9b-963d6e2e49ce\") " pod="openstack/dnsmasq-dns-86db49b7ff-x7rps" Dec 03 14:29:03 crc kubenswrapper[4805]: I1203 14:29:03.898317 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e7db8f49-10ce-4818-8d9b-963d6e2e49ce-config\") pod \"dnsmasq-dns-86db49b7ff-x7rps\" (UID: \"e7db8f49-10ce-4818-8d9b-963d6e2e49ce\") " pod="openstack/dnsmasq-dns-86db49b7ff-x7rps" Dec 03 14:29:03 crc kubenswrapper[4805]: I1203 14:29:03.898511 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/e7db8f49-10ce-4818-8d9b-963d6e2e49ce-ovsdbserver-sb\") pod \"dnsmasq-dns-86db49b7ff-x7rps\" (UID: \"e7db8f49-10ce-4818-8d9b-963d6e2e49ce\") " pod="openstack/dnsmasq-dns-86db49b7ff-x7rps" Dec 03 14:29:03 crc kubenswrapper[4805]: I1203 14:29:03.917799 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cp8vh\" (UniqueName: \"kubernetes.io/projected/e7db8f49-10ce-4818-8d9b-963d6e2e49ce-kube-api-access-cp8vh\") pod \"dnsmasq-dns-86db49b7ff-x7rps\" (UID: \"e7db8f49-10ce-4818-8d9b-963d6e2e49ce\") " pod="openstack/dnsmasq-dns-86db49b7ff-x7rps" Dec 03 14:29:04 crc kubenswrapper[4805]: I1203 14:29:04.067246 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"5b4bb902-dba6-4ec6-97bf-532d996e918d","Type":"ContainerStarted","Data":"92590cb576baa29db9b9495e8f7cc182b7211e173c3b3ec2ece6e5d05c6e36a5"} Dec 03 14:29:04 crc kubenswrapper[4805]: I1203 14:29:04.082027 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-86db49b7ff-x7rps" Dec 03 14:29:04 crc kubenswrapper[4805]: I1203 14:29:04.098771 4805 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/openstack-galera-0" podStartSLOduration=22.760975155 podStartE2EDuration="32.098743136s" podCreationTimestamp="2025-12-03 14:28:32 +0000 UTC" firstStartedPulling="2025-12-03 14:28:46.612230541 +0000 UTC m=+1156.275147474" lastFinishedPulling="2025-12-03 14:28:55.949998532 +0000 UTC m=+1165.612915455" observedRunningTime="2025-12-03 14:29:04.090595646 +0000 UTC m=+1173.753512569" watchObservedRunningTime="2025-12-03 14:29:04.098743136 +0000 UTC m=+1173.761660059" Dec 03 14:29:04 crc kubenswrapper[4805]: I1203 14:29:04.286388 4805 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-7fd796d7df-7l7rr"] Dec 03 14:29:04 crc kubenswrapper[4805]: I1203 14:29:04.356154 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-metrics-7kmnp"] Dec 03 14:29:04 crc kubenswrapper[4805]: W1203 14:29:04.366864 4805 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod76002c62_88cb_4f74_816a_3b2be4f8ba4c.slice/crio-d3fed19b9b50048643e3e63c91c61e29e6edc53396a6eef2f0d45cd7317e3d82 WatchSource:0}: Error finding container d3fed19b9b50048643e3e63c91c61e29e6edc53396a6eef2f0d45cd7317e3d82: Status 404 returned error can't find the container with id d3fed19b9b50048643e3e63c91c61e29e6edc53396a6eef2f0d45cd7317e3d82 Dec 03 14:29:04 crc kubenswrapper[4805]: I1203 14:29:04.441278 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-northd-0"] Dec 03 14:29:04 crc kubenswrapper[4805]: I1203 14:29:04.552785 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-86db49b7ff-x7rps"] Dec 03 14:29:04 crc kubenswrapper[4805]: W1203 14:29:04.555770 4805 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pode7db8f49_10ce_4818_8d9b_963d6e2e49ce.slice/crio-fea65c6cb163df9486ceb2477217e70087be12c99955148404fc637c49a092ef WatchSource:0}: Error finding container fea65c6cb163df9486ceb2477217e70087be12c99955148404fc637c49a092ef: Status 404 returned error can't find the container with id fea65c6cb163df9486ceb2477217e70087be12c99955148404fc637c49a092ef Dec 03 14:29:05 crc kubenswrapper[4805]: I1203 14:29:05.081052 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-metrics-7kmnp" event={"ID":"76002c62-88cb-4f74-816a-3b2be4f8ba4c","Type":"ContainerStarted","Data":"4808aa7f851c2bc222f006dd5aebec4b842fab57104dc85962c80385ed4a0163"} Dec 03 14:29:05 crc kubenswrapper[4805]: I1203 14:29:05.081494 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-metrics-7kmnp" event={"ID":"76002c62-88cb-4f74-816a-3b2be4f8ba4c","Type":"ContainerStarted","Data":"d3fed19b9b50048643e3e63c91c61e29e6edc53396a6eef2f0d45cd7317e3d82"} Dec 03 14:29:05 crc kubenswrapper[4805]: I1203 14:29:05.082646 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-northd-0" event={"ID":"8e6ca40c-f7f2-445e-8e22-71f8d4d4693b","Type":"ContainerStarted","Data":"71910f8847a028ffb6b5c510542d194f206d24152336835ca149988fa1498ee3"} Dec 03 14:29:05 crc kubenswrapper[4805]: I1203 14:29:05.084736 4805 generic.go:334] "Generic (PLEG): container finished" podID="e7db8f49-10ce-4818-8d9b-963d6e2e49ce" containerID="ef9c08e3a8b1a0b765a1ed21255c4481ef432d9841cf9d7887a4b0e721a54f17" exitCode=0 Dec 03 14:29:05 crc kubenswrapper[4805]: I1203 14:29:05.084811 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-86db49b7ff-x7rps" event={"ID":"e7db8f49-10ce-4818-8d9b-963d6e2e49ce","Type":"ContainerDied","Data":"ef9c08e3a8b1a0b765a1ed21255c4481ef432d9841cf9d7887a4b0e721a54f17"} Dec 03 14:29:05 crc kubenswrapper[4805]: I1203 14:29:05.084855 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-86db49b7ff-x7rps" event={"ID":"e7db8f49-10ce-4818-8d9b-963d6e2e49ce","Type":"ContainerStarted","Data":"fea65c6cb163df9486ceb2477217e70087be12c99955148404fc637c49a092ef"} Dec 03 14:29:05 crc kubenswrapper[4805]: I1203 14:29:05.093514 4805 generic.go:334] "Generic (PLEG): container finished" podID="0e091270-fda5-4856-9dd8-894e830d07e3" containerID="0314cfb0613c10f8a4364a5ab79b1c4427dcdbfba93b738b3a3f4116191b409a" exitCode=0 Dec 03 14:29:05 crc kubenswrapper[4805]: I1203 14:29:05.093606 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7fd796d7df-7l7rr" event={"ID":"0e091270-fda5-4856-9dd8-894e830d07e3","Type":"ContainerDied","Data":"0314cfb0613c10f8a4364a5ab79b1c4427dcdbfba93b738b3a3f4116191b409a"} Dec 03 14:29:05 crc kubenswrapper[4805]: I1203 14:29:05.093670 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7fd796d7df-7l7rr" event={"ID":"0e091270-fda5-4856-9dd8-894e830d07e3","Type":"ContainerStarted","Data":"4798401de9a95112f99cd441fb41793324a35185a63af43b9d5c73dee271d147"} Dec 03 14:29:05 crc kubenswrapper[4805]: I1203 14:29:05.109016 4805 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-controller-metrics-7kmnp" podStartSLOduration=2.108998204 podStartE2EDuration="2.108998204s" podCreationTimestamp="2025-12-03 14:29:03 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 14:29:05.103731671 +0000 UTC m=+1174.766648594" watchObservedRunningTime="2025-12-03 14:29:05.108998204 +0000 UTC m=+1174.771915127" Dec 03 14:29:05 crc kubenswrapper[4805]: I1203 14:29:05.245635 4805 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/openstack-cell1-galera-0" Dec 03 14:29:05 crc kubenswrapper[4805]: I1203 14:29:05.246768 4805 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/openstack-cell1-galera-0" Dec 03 14:29:05 crc kubenswrapper[4805]: I1203 14:29:05.365999 4805 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/memcached-0" Dec 03 14:29:05 crc kubenswrapper[4805]: I1203 14:29:05.933770 4805 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7fd796d7df-7l7rr" Dec 03 14:29:06 crc kubenswrapper[4805]: I1203 14:29:06.038055 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/0e091270-fda5-4856-9dd8-894e830d07e3-dns-svc\") pod \"0e091270-fda5-4856-9dd8-894e830d07e3\" (UID: \"0e091270-fda5-4856-9dd8-894e830d07e3\") " Dec 03 14:29:06 crc kubenswrapper[4805]: I1203 14:29:06.038150 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/0e091270-fda5-4856-9dd8-894e830d07e3-ovsdbserver-nb\") pod \"0e091270-fda5-4856-9dd8-894e830d07e3\" (UID: \"0e091270-fda5-4856-9dd8-894e830d07e3\") " Dec 03 14:29:06 crc kubenswrapper[4805]: I1203 14:29:06.038217 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0e091270-fda5-4856-9dd8-894e830d07e3-config\") pod \"0e091270-fda5-4856-9dd8-894e830d07e3\" (UID: \"0e091270-fda5-4856-9dd8-894e830d07e3\") " Dec 03 14:29:06 crc kubenswrapper[4805]: I1203 14:29:06.038275 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ktm7p\" (UniqueName: \"kubernetes.io/projected/0e091270-fda5-4856-9dd8-894e830d07e3-kube-api-access-ktm7p\") pod \"0e091270-fda5-4856-9dd8-894e830d07e3\" (UID: \"0e091270-fda5-4856-9dd8-894e830d07e3\") " Dec 03 14:29:06 crc kubenswrapper[4805]: I1203 14:29:06.041810 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0e091270-fda5-4856-9dd8-894e830d07e3-kube-api-access-ktm7p" (OuterVolumeSpecName: "kube-api-access-ktm7p") pod "0e091270-fda5-4856-9dd8-894e830d07e3" (UID: "0e091270-fda5-4856-9dd8-894e830d07e3"). InnerVolumeSpecName "kube-api-access-ktm7p". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 14:29:06 crc kubenswrapper[4805]: I1203 14:29:06.064648 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0e091270-fda5-4856-9dd8-894e830d07e3-config" (OuterVolumeSpecName: "config") pod "0e091270-fda5-4856-9dd8-894e830d07e3" (UID: "0e091270-fda5-4856-9dd8-894e830d07e3"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 14:29:06 crc kubenswrapper[4805]: I1203 14:29:06.070005 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0e091270-fda5-4856-9dd8-894e830d07e3-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "0e091270-fda5-4856-9dd8-894e830d07e3" (UID: "0e091270-fda5-4856-9dd8-894e830d07e3"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 14:29:06 crc kubenswrapper[4805]: I1203 14:29:06.073314 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0e091270-fda5-4856-9dd8-894e830d07e3-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "0e091270-fda5-4856-9dd8-894e830d07e3" (UID: "0e091270-fda5-4856-9dd8-894e830d07e3"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 14:29:06 crc kubenswrapper[4805]: I1203 14:29:06.103574 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-northd-0" event={"ID":"8e6ca40c-f7f2-445e-8e22-71f8d4d4693b","Type":"ContainerStarted","Data":"079096b86b2a569934c267b5e2e0518e498987cd0d2bd17ba7953c7a2aa6c587"} Dec 03 14:29:06 crc kubenswrapper[4805]: I1203 14:29:06.107333 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-86db49b7ff-x7rps" event={"ID":"e7db8f49-10ce-4818-8d9b-963d6e2e49ce","Type":"ContainerStarted","Data":"0eec91f5631e0ed7c3bc78c7ce05a3e4effb735201f47c86b2313526b7653276"} Dec 03 14:29:06 crc kubenswrapper[4805]: I1203 14:29:06.107394 4805 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-86db49b7ff-x7rps" Dec 03 14:29:06 crc kubenswrapper[4805]: I1203 14:29:06.110611 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7fd796d7df-7l7rr" event={"ID":"0e091270-fda5-4856-9dd8-894e830d07e3","Type":"ContainerDied","Data":"4798401de9a95112f99cd441fb41793324a35185a63af43b9d5c73dee271d147"} Dec 03 14:29:06 crc kubenswrapper[4805]: I1203 14:29:06.110669 4805 scope.go:117] "RemoveContainer" containerID="0314cfb0613c10f8a4364a5ab79b1c4427dcdbfba93b738b3a3f4116191b409a" Dec 03 14:29:06 crc kubenswrapper[4805]: I1203 14:29:06.110788 4805 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7fd796d7df-7l7rr" Dec 03 14:29:06 crc kubenswrapper[4805]: I1203 14:29:06.125064 4805 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-86db49b7ff-x7rps" podStartSLOduration=3.125051007 podStartE2EDuration="3.125051007s" podCreationTimestamp="2025-12-03 14:29:03 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 14:29:06.12328169 +0000 UTC m=+1175.786198613" watchObservedRunningTime="2025-12-03 14:29:06.125051007 +0000 UTC m=+1175.787967920" Dec 03 14:29:06 crc kubenswrapper[4805]: I1203 14:29:06.139756 4805 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0e091270-fda5-4856-9dd8-894e830d07e3-config\") on node \"crc\" DevicePath \"\"" Dec 03 14:29:06 crc kubenswrapper[4805]: I1203 14:29:06.139795 4805 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ktm7p\" (UniqueName: \"kubernetes.io/projected/0e091270-fda5-4856-9dd8-894e830d07e3-kube-api-access-ktm7p\") on node \"crc\" DevicePath \"\"" Dec 03 14:29:06 crc kubenswrapper[4805]: I1203 14:29:06.139806 4805 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/0e091270-fda5-4856-9dd8-894e830d07e3-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 03 14:29:06 crc kubenswrapper[4805]: I1203 14:29:06.139816 4805 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/0e091270-fda5-4856-9dd8-894e830d07e3-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Dec 03 14:29:06 crc kubenswrapper[4805]: I1203 14:29:06.196453 4805 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-7fd796d7df-7l7rr"] Dec 03 14:29:06 crc kubenswrapper[4805]: I1203 14:29:06.203599 4805 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-7fd796d7df-7l7rr"] Dec 03 14:29:06 crc kubenswrapper[4805]: I1203 14:29:06.708139 4805 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0e091270-fda5-4856-9dd8-894e830d07e3" path="/var/lib/kubelet/pods/0e091270-fda5-4856-9dd8-894e830d07e3/volumes" Dec 03 14:29:07 crc kubenswrapper[4805]: I1203 14:29:07.118622 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-northd-0" event={"ID":"8e6ca40c-f7f2-445e-8e22-71f8d4d4693b","Type":"ContainerStarted","Data":"5c9c173a3f86add4b1380a4a29403a5e521ae6bef4a06ba6c8812e1258523a90"} Dec 03 14:29:07 crc kubenswrapper[4805]: I1203 14:29:07.118719 4805 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovn-northd-0" Dec 03 14:29:07 crc kubenswrapper[4805]: I1203 14:29:07.135651 4805 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-northd-0" podStartSLOduration=2.894745623 podStartE2EDuration="4.135634104s" podCreationTimestamp="2025-12-03 14:29:03 +0000 UTC" firstStartedPulling="2025-12-03 14:29:04.442670416 +0000 UTC m=+1174.105587339" lastFinishedPulling="2025-12-03 14:29:05.683558897 +0000 UTC m=+1175.346475820" observedRunningTime="2025-12-03 14:29:07.13216094 +0000 UTC m=+1176.795077863" watchObservedRunningTime="2025-12-03 14:29:07.135634104 +0000 UTC m=+1176.798551027" Dec 03 14:29:07 crc kubenswrapper[4805]: I1203 14:29:07.547954 4805 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/kube-state-metrics-0" Dec 03 14:29:07 crc kubenswrapper[4805]: I1203 14:29:07.571465 4805 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/openstack-cell1-galera-0" Dec 03 14:29:07 crc kubenswrapper[4805]: I1203 14:29:07.577045 4805 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-86db49b7ff-x7rps"] Dec 03 14:29:07 crc kubenswrapper[4805]: I1203 14:29:07.615744 4805 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-698758b865-l46rr"] Dec 03 14:29:07 crc kubenswrapper[4805]: E1203 14:29:07.616157 4805 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0e091270-fda5-4856-9dd8-894e830d07e3" containerName="init" Dec 03 14:29:07 crc kubenswrapper[4805]: I1203 14:29:07.616173 4805 state_mem.go:107] "Deleted CPUSet assignment" podUID="0e091270-fda5-4856-9dd8-894e830d07e3" containerName="init" Dec 03 14:29:07 crc kubenswrapper[4805]: I1203 14:29:07.616378 4805 memory_manager.go:354] "RemoveStaleState removing state" podUID="0e091270-fda5-4856-9dd8-894e830d07e3" containerName="init" Dec 03 14:29:07 crc kubenswrapper[4805]: I1203 14:29:07.617376 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-698758b865-l46rr" Dec 03 14:29:07 crc kubenswrapper[4805]: I1203 14:29:07.643911 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-698758b865-l46rr"] Dec 03 14:29:07 crc kubenswrapper[4805]: I1203 14:29:07.664342 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/83e323ed-ad2c-4635-a583-4e5b21d8e268-dns-svc\") pod \"dnsmasq-dns-698758b865-l46rr\" (UID: \"83e323ed-ad2c-4635-a583-4e5b21d8e268\") " pod="openstack/dnsmasq-dns-698758b865-l46rr" Dec 03 14:29:07 crc kubenswrapper[4805]: I1203 14:29:07.664666 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/83e323ed-ad2c-4635-a583-4e5b21d8e268-ovsdbserver-sb\") pod \"dnsmasq-dns-698758b865-l46rr\" (UID: \"83e323ed-ad2c-4635-a583-4e5b21d8e268\") " pod="openstack/dnsmasq-dns-698758b865-l46rr" Dec 03 14:29:07 crc kubenswrapper[4805]: I1203 14:29:07.664793 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qj4w8\" (UniqueName: \"kubernetes.io/projected/83e323ed-ad2c-4635-a583-4e5b21d8e268-kube-api-access-qj4w8\") pod \"dnsmasq-dns-698758b865-l46rr\" (UID: \"83e323ed-ad2c-4635-a583-4e5b21d8e268\") " pod="openstack/dnsmasq-dns-698758b865-l46rr" Dec 03 14:29:07 crc kubenswrapper[4805]: I1203 14:29:07.664935 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/83e323ed-ad2c-4635-a583-4e5b21d8e268-ovsdbserver-nb\") pod \"dnsmasq-dns-698758b865-l46rr\" (UID: \"83e323ed-ad2c-4635-a583-4e5b21d8e268\") " pod="openstack/dnsmasq-dns-698758b865-l46rr" Dec 03 14:29:07 crc kubenswrapper[4805]: I1203 14:29:07.665111 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/83e323ed-ad2c-4635-a583-4e5b21d8e268-config\") pod \"dnsmasq-dns-698758b865-l46rr\" (UID: \"83e323ed-ad2c-4635-a583-4e5b21d8e268\") " pod="openstack/dnsmasq-dns-698758b865-l46rr" Dec 03 14:29:07 crc kubenswrapper[4805]: I1203 14:29:07.721520 4805 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/openstack-cell1-galera-0" Dec 03 14:29:07 crc kubenswrapper[4805]: I1203 14:29:07.766865 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/83e323ed-ad2c-4635-a583-4e5b21d8e268-dns-svc\") pod \"dnsmasq-dns-698758b865-l46rr\" (UID: \"83e323ed-ad2c-4635-a583-4e5b21d8e268\") " pod="openstack/dnsmasq-dns-698758b865-l46rr" Dec 03 14:29:07 crc kubenswrapper[4805]: I1203 14:29:07.766923 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/83e323ed-ad2c-4635-a583-4e5b21d8e268-ovsdbserver-sb\") pod \"dnsmasq-dns-698758b865-l46rr\" (UID: \"83e323ed-ad2c-4635-a583-4e5b21d8e268\") " pod="openstack/dnsmasq-dns-698758b865-l46rr" Dec 03 14:29:07 crc kubenswrapper[4805]: I1203 14:29:07.766955 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qj4w8\" (UniqueName: \"kubernetes.io/projected/83e323ed-ad2c-4635-a583-4e5b21d8e268-kube-api-access-qj4w8\") pod \"dnsmasq-dns-698758b865-l46rr\" (UID: \"83e323ed-ad2c-4635-a583-4e5b21d8e268\") " pod="openstack/dnsmasq-dns-698758b865-l46rr" Dec 03 14:29:07 crc kubenswrapper[4805]: I1203 14:29:07.766979 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/83e323ed-ad2c-4635-a583-4e5b21d8e268-ovsdbserver-nb\") pod \"dnsmasq-dns-698758b865-l46rr\" (UID: \"83e323ed-ad2c-4635-a583-4e5b21d8e268\") " pod="openstack/dnsmasq-dns-698758b865-l46rr" Dec 03 14:29:07 crc kubenswrapper[4805]: I1203 14:29:07.767012 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/83e323ed-ad2c-4635-a583-4e5b21d8e268-config\") pod \"dnsmasq-dns-698758b865-l46rr\" (UID: \"83e323ed-ad2c-4635-a583-4e5b21d8e268\") " pod="openstack/dnsmasq-dns-698758b865-l46rr" Dec 03 14:29:07 crc kubenswrapper[4805]: I1203 14:29:07.767963 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/83e323ed-ad2c-4635-a583-4e5b21d8e268-config\") pod \"dnsmasq-dns-698758b865-l46rr\" (UID: \"83e323ed-ad2c-4635-a583-4e5b21d8e268\") " pod="openstack/dnsmasq-dns-698758b865-l46rr" Dec 03 14:29:07 crc kubenswrapper[4805]: I1203 14:29:07.768066 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/83e323ed-ad2c-4635-a583-4e5b21d8e268-ovsdbserver-sb\") pod \"dnsmasq-dns-698758b865-l46rr\" (UID: \"83e323ed-ad2c-4635-a583-4e5b21d8e268\") " pod="openstack/dnsmasq-dns-698758b865-l46rr" Dec 03 14:29:07 crc kubenswrapper[4805]: I1203 14:29:07.768132 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/83e323ed-ad2c-4635-a583-4e5b21d8e268-dns-svc\") pod \"dnsmasq-dns-698758b865-l46rr\" (UID: \"83e323ed-ad2c-4635-a583-4e5b21d8e268\") " pod="openstack/dnsmasq-dns-698758b865-l46rr" Dec 03 14:29:07 crc kubenswrapper[4805]: I1203 14:29:07.768441 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/83e323ed-ad2c-4635-a583-4e5b21d8e268-ovsdbserver-nb\") pod \"dnsmasq-dns-698758b865-l46rr\" (UID: \"83e323ed-ad2c-4635-a583-4e5b21d8e268\") " pod="openstack/dnsmasq-dns-698758b865-l46rr" Dec 03 14:29:07 crc kubenswrapper[4805]: I1203 14:29:07.790977 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qj4w8\" (UniqueName: \"kubernetes.io/projected/83e323ed-ad2c-4635-a583-4e5b21d8e268-kube-api-access-qj4w8\") pod \"dnsmasq-dns-698758b865-l46rr\" (UID: \"83e323ed-ad2c-4635-a583-4e5b21d8e268\") " pod="openstack/dnsmasq-dns-698758b865-l46rr" Dec 03 14:29:07 crc kubenswrapper[4805]: I1203 14:29:07.934503 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-698758b865-l46rr" Dec 03 14:29:08 crc kubenswrapper[4805]: I1203 14:29:08.127938 4805 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-86db49b7ff-x7rps" podUID="e7db8f49-10ce-4818-8d9b-963d6e2e49ce" containerName="dnsmasq-dns" containerID="cri-o://0eec91f5631e0ed7c3bc78c7ce05a3e4effb735201f47c86b2313526b7653276" gracePeriod=10 Dec 03 14:29:08 crc kubenswrapper[4805]: I1203 14:29:08.239400 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-698758b865-l46rr"] Dec 03 14:29:08 crc kubenswrapper[4805]: I1203 14:29:08.456689 4805 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-86db49b7ff-x7rps" Dec 03 14:29:08 crc kubenswrapper[4805]: I1203 14:29:08.598322 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/e7db8f49-10ce-4818-8d9b-963d6e2e49ce-dns-svc\") pod \"e7db8f49-10ce-4818-8d9b-963d6e2e49ce\" (UID: \"e7db8f49-10ce-4818-8d9b-963d6e2e49ce\") " Dec 03 14:29:08 crc kubenswrapper[4805]: I1203 14:29:08.598439 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/e7db8f49-10ce-4818-8d9b-963d6e2e49ce-ovsdbserver-sb\") pod \"e7db8f49-10ce-4818-8d9b-963d6e2e49ce\" (UID: \"e7db8f49-10ce-4818-8d9b-963d6e2e49ce\") " Dec 03 14:29:08 crc kubenswrapper[4805]: I1203 14:29:08.598499 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e7db8f49-10ce-4818-8d9b-963d6e2e49ce-config\") pod \"e7db8f49-10ce-4818-8d9b-963d6e2e49ce\" (UID: \"e7db8f49-10ce-4818-8d9b-963d6e2e49ce\") " Dec 03 14:29:08 crc kubenswrapper[4805]: I1203 14:29:08.598577 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/e7db8f49-10ce-4818-8d9b-963d6e2e49ce-ovsdbserver-nb\") pod \"e7db8f49-10ce-4818-8d9b-963d6e2e49ce\" (UID: \"e7db8f49-10ce-4818-8d9b-963d6e2e49ce\") " Dec 03 14:29:08 crc kubenswrapper[4805]: I1203 14:29:08.598702 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cp8vh\" (UniqueName: \"kubernetes.io/projected/e7db8f49-10ce-4818-8d9b-963d6e2e49ce-kube-api-access-cp8vh\") pod \"e7db8f49-10ce-4818-8d9b-963d6e2e49ce\" (UID: \"e7db8f49-10ce-4818-8d9b-963d6e2e49ce\") " Dec 03 14:29:08 crc kubenswrapper[4805]: I1203 14:29:08.603667 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e7db8f49-10ce-4818-8d9b-963d6e2e49ce-kube-api-access-cp8vh" (OuterVolumeSpecName: "kube-api-access-cp8vh") pod "e7db8f49-10ce-4818-8d9b-963d6e2e49ce" (UID: "e7db8f49-10ce-4818-8d9b-963d6e2e49ce"). InnerVolumeSpecName "kube-api-access-cp8vh". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 14:29:08 crc kubenswrapper[4805]: I1203 14:29:08.645349 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e7db8f49-10ce-4818-8d9b-963d6e2e49ce-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "e7db8f49-10ce-4818-8d9b-963d6e2e49ce" (UID: "e7db8f49-10ce-4818-8d9b-963d6e2e49ce"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 14:29:08 crc kubenswrapper[4805]: I1203 14:29:08.646497 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e7db8f49-10ce-4818-8d9b-963d6e2e49ce-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "e7db8f49-10ce-4818-8d9b-963d6e2e49ce" (UID: "e7db8f49-10ce-4818-8d9b-963d6e2e49ce"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 14:29:08 crc kubenswrapper[4805]: I1203 14:29:08.665519 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e7db8f49-10ce-4818-8d9b-963d6e2e49ce-config" (OuterVolumeSpecName: "config") pod "e7db8f49-10ce-4818-8d9b-963d6e2e49ce" (UID: "e7db8f49-10ce-4818-8d9b-963d6e2e49ce"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 14:29:08 crc kubenswrapper[4805]: I1203 14:29:08.673753 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e7db8f49-10ce-4818-8d9b-963d6e2e49ce-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "e7db8f49-10ce-4818-8d9b-963d6e2e49ce" (UID: "e7db8f49-10ce-4818-8d9b-963d6e2e49ce"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 14:29:08 crc kubenswrapper[4805]: I1203 14:29:08.702296 4805 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/e7db8f49-10ce-4818-8d9b-963d6e2e49ce-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Dec 03 14:29:08 crc kubenswrapper[4805]: I1203 14:29:08.702354 4805 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e7db8f49-10ce-4818-8d9b-963d6e2e49ce-config\") on node \"crc\" DevicePath \"\"" Dec 03 14:29:08 crc kubenswrapper[4805]: I1203 14:29:08.702372 4805 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/e7db8f49-10ce-4818-8d9b-963d6e2e49ce-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Dec 03 14:29:08 crc kubenswrapper[4805]: I1203 14:29:08.702391 4805 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cp8vh\" (UniqueName: \"kubernetes.io/projected/e7db8f49-10ce-4818-8d9b-963d6e2e49ce-kube-api-access-cp8vh\") on node \"crc\" DevicePath \"\"" Dec 03 14:29:08 crc kubenswrapper[4805]: I1203 14:29:08.702407 4805 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/e7db8f49-10ce-4818-8d9b-963d6e2e49ce-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 03 14:29:08 crc kubenswrapper[4805]: I1203 14:29:08.713643 4805 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/swift-storage-0"] Dec 03 14:29:08 crc kubenswrapper[4805]: E1203 14:29:08.714294 4805 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e7db8f49-10ce-4818-8d9b-963d6e2e49ce" containerName="dnsmasq-dns" Dec 03 14:29:08 crc kubenswrapper[4805]: I1203 14:29:08.714391 4805 state_mem.go:107] "Deleted CPUSet assignment" podUID="e7db8f49-10ce-4818-8d9b-963d6e2e49ce" containerName="dnsmasq-dns" Dec 03 14:29:08 crc kubenswrapper[4805]: E1203 14:29:08.714506 4805 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e7db8f49-10ce-4818-8d9b-963d6e2e49ce" containerName="init" Dec 03 14:29:08 crc kubenswrapper[4805]: I1203 14:29:08.714600 4805 state_mem.go:107] "Deleted CPUSet assignment" podUID="e7db8f49-10ce-4818-8d9b-963d6e2e49ce" containerName="init" Dec 03 14:29:08 crc kubenswrapper[4805]: I1203 14:29:08.714950 4805 memory_manager.go:354] "RemoveStaleState removing state" podUID="e7db8f49-10ce-4818-8d9b-963d6e2e49ce" containerName="dnsmasq-dns" Dec 03 14:29:08 crc kubenswrapper[4805]: I1203 14:29:08.720187 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-storage-0" Dec 03 14:29:08 crc kubenswrapper[4805]: I1203 14:29:08.726271 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"swift-conf" Dec 03 14:29:08 crc kubenswrapper[4805]: I1203 14:29:08.726399 4805 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"swift-storage-config-data" Dec 03 14:29:08 crc kubenswrapper[4805]: I1203 14:29:08.734188 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-storage-0"] Dec 03 14:29:08 crc kubenswrapper[4805]: I1203 14:29:08.737177 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"swift-swift-dockercfg-67p62" Dec 03 14:29:08 crc kubenswrapper[4805]: I1203 14:29:08.737420 4805 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"swift-ring-files" Dec 03 14:29:08 crc kubenswrapper[4805]: I1203 14:29:08.804014 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lock\" (UniqueName: \"kubernetes.io/empty-dir/4fbc8d7f-d47f-4d91-915b-504820aac08a-lock\") pod \"swift-storage-0\" (UID: \"4fbc8d7f-d47f-4d91-915b-504820aac08a\") " pod="openstack/swift-storage-0" Dec 03 14:29:08 crc kubenswrapper[4805]: I1203 14:29:08.804124 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"swift-storage-0\" (UID: \"4fbc8d7f-d47f-4d91-915b-504820aac08a\") " pod="openstack/swift-storage-0" Dec 03 14:29:08 crc kubenswrapper[4805]: I1203 14:29:08.804189 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/4fbc8d7f-d47f-4d91-915b-504820aac08a-etc-swift\") pod \"swift-storage-0\" (UID: \"4fbc8d7f-d47f-4d91-915b-504820aac08a\") " pod="openstack/swift-storage-0" Dec 03 14:29:08 crc kubenswrapper[4805]: I1203 14:29:08.804373 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cache\" (UniqueName: \"kubernetes.io/empty-dir/4fbc8d7f-d47f-4d91-915b-504820aac08a-cache\") pod \"swift-storage-0\" (UID: \"4fbc8d7f-d47f-4d91-915b-504820aac08a\") " pod="openstack/swift-storage-0" Dec 03 14:29:08 crc kubenswrapper[4805]: I1203 14:29:08.804458 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-79wdl\" (UniqueName: \"kubernetes.io/projected/4fbc8d7f-d47f-4d91-915b-504820aac08a-kube-api-access-79wdl\") pod \"swift-storage-0\" (UID: \"4fbc8d7f-d47f-4d91-915b-504820aac08a\") " pod="openstack/swift-storage-0" Dec 03 14:29:08 crc kubenswrapper[4805]: I1203 14:29:08.906033 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"swift-storage-0\" (UID: \"4fbc8d7f-d47f-4d91-915b-504820aac08a\") " pod="openstack/swift-storage-0" Dec 03 14:29:08 crc kubenswrapper[4805]: I1203 14:29:08.906106 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/4fbc8d7f-d47f-4d91-915b-504820aac08a-etc-swift\") pod \"swift-storage-0\" (UID: \"4fbc8d7f-d47f-4d91-915b-504820aac08a\") " pod="openstack/swift-storage-0" Dec 03 14:29:08 crc kubenswrapper[4805]: I1203 14:29:08.906152 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cache\" (UniqueName: \"kubernetes.io/empty-dir/4fbc8d7f-d47f-4d91-915b-504820aac08a-cache\") pod \"swift-storage-0\" (UID: \"4fbc8d7f-d47f-4d91-915b-504820aac08a\") " pod="openstack/swift-storage-0" Dec 03 14:29:08 crc kubenswrapper[4805]: I1203 14:29:08.906203 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-79wdl\" (UniqueName: \"kubernetes.io/projected/4fbc8d7f-d47f-4d91-915b-504820aac08a-kube-api-access-79wdl\") pod \"swift-storage-0\" (UID: \"4fbc8d7f-d47f-4d91-915b-504820aac08a\") " pod="openstack/swift-storage-0" Dec 03 14:29:08 crc kubenswrapper[4805]: I1203 14:29:08.906252 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"lock\" (UniqueName: \"kubernetes.io/empty-dir/4fbc8d7f-d47f-4d91-915b-504820aac08a-lock\") pod \"swift-storage-0\" (UID: \"4fbc8d7f-d47f-4d91-915b-504820aac08a\") " pod="openstack/swift-storage-0" Dec 03 14:29:08 crc kubenswrapper[4805]: I1203 14:29:08.906673 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"lock\" (UniqueName: \"kubernetes.io/empty-dir/4fbc8d7f-d47f-4d91-915b-504820aac08a-lock\") pod \"swift-storage-0\" (UID: \"4fbc8d7f-d47f-4d91-915b-504820aac08a\") " pod="openstack/swift-storage-0" Dec 03 14:29:08 crc kubenswrapper[4805]: I1203 14:29:08.906945 4805 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"swift-storage-0\" (UID: \"4fbc8d7f-d47f-4d91-915b-504820aac08a\") device mount path \"/mnt/openstack/pv10\"" pod="openstack/swift-storage-0" Dec 03 14:29:08 crc kubenswrapper[4805]: E1203 14:29:08.907157 4805 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Dec 03 14:29:08 crc kubenswrapper[4805]: E1203 14:29:08.907194 4805 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Dec 03 14:29:08 crc kubenswrapper[4805]: E1203 14:29:08.907254 4805 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/4fbc8d7f-d47f-4d91-915b-504820aac08a-etc-swift podName:4fbc8d7f-d47f-4d91-915b-504820aac08a nodeName:}" failed. No retries permitted until 2025-12-03 14:29:09.407236351 +0000 UTC m=+1179.070153274 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/4fbc8d7f-d47f-4d91-915b-504820aac08a-etc-swift") pod "swift-storage-0" (UID: "4fbc8d7f-d47f-4d91-915b-504820aac08a") : configmap "swift-ring-files" not found Dec 03 14:29:08 crc kubenswrapper[4805]: I1203 14:29:08.907447 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cache\" (UniqueName: \"kubernetes.io/empty-dir/4fbc8d7f-d47f-4d91-915b-504820aac08a-cache\") pod \"swift-storage-0\" (UID: \"4fbc8d7f-d47f-4d91-915b-504820aac08a\") " pod="openstack/swift-storage-0" Dec 03 14:29:08 crc kubenswrapper[4805]: I1203 14:29:08.930100 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"swift-storage-0\" (UID: \"4fbc8d7f-d47f-4d91-915b-504820aac08a\") " pod="openstack/swift-storage-0" Dec 03 14:29:08 crc kubenswrapper[4805]: I1203 14:29:08.938867 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-79wdl\" (UniqueName: \"kubernetes.io/projected/4fbc8d7f-d47f-4d91-915b-504820aac08a-kube-api-access-79wdl\") pod \"swift-storage-0\" (UID: \"4fbc8d7f-d47f-4d91-915b-504820aac08a\") " pod="openstack/swift-storage-0" Dec 03 14:29:09 crc kubenswrapper[4805]: I1203 14:29:09.137274 4805 generic.go:334] "Generic (PLEG): container finished" podID="83e323ed-ad2c-4635-a583-4e5b21d8e268" containerID="4d555253ea8f012425b96f097443fb842d8d499049efb2c8bb321d17e8ee0788" exitCode=0 Dec 03 14:29:09 crc kubenswrapper[4805]: I1203 14:29:09.137940 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-698758b865-l46rr" event={"ID":"83e323ed-ad2c-4635-a583-4e5b21d8e268","Type":"ContainerDied","Data":"4d555253ea8f012425b96f097443fb842d8d499049efb2c8bb321d17e8ee0788"} Dec 03 14:29:09 crc kubenswrapper[4805]: I1203 14:29:09.138455 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-698758b865-l46rr" event={"ID":"83e323ed-ad2c-4635-a583-4e5b21d8e268","Type":"ContainerStarted","Data":"38e9567f9c900cc2c54caf706a009e55dc558787efb3fb6195719b107e2a2247"} Dec 03 14:29:09 crc kubenswrapper[4805]: I1203 14:29:09.140963 4805 generic.go:334] "Generic (PLEG): container finished" podID="e7db8f49-10ce-4818-8d9b-963d6e2e49ce" containerID="0eec91f5631e0ed7c3bc78c7ce05a3e4effb735201f47c86b2313526b7653276" exitCode=0 Dec 03 14:29:09 crc kubenswrapper[4805]: I1203 14:29:09.141004 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-86db49b7ff-x7rps" event={"ID":"e7db8f49-10ce-4818-8d9b-963d6e2e49ce","Type":"ContainerDied","Data":"0eec91f5631e0ed7c3bc78c7ce05a3e4effb735201f47c86b2313526b7653276"} Dec 03 14:29:09 crc kubenswrapper[4805]: I1203 14:29:09.141031 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-86db49b7ff-x7rps" event={"ID":"e7db8f49-10ce-4818-8d9b-963d6e2e49ce","Type":"ContainerDied","Data":"fea65c6cb163df9486ceb2477217e70087be12c99955148404fc637c49a092ef"} Dec 03 14:29:09 crc kubenswrapper[4805]: I1203 14:29:09.141046 4805 scope.go:117] "RemoveContainer" containerID="0eec91f5631e0ed7c3bc78c7ce05a3e4effb735201f47c86b2313526b7653276" Dec 03 14:29:09 crc kubenswrapper[4805]: I1203 14:29:09.141170 4805 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-86db49b7ff-x7rps" Dec 03 14:29:09 crc kubenswrapper[4805]: I1203 14:29:09.198915 4805 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/swift-ring-rebalance-h55gd"] Dec 03 14:29:09 crc kubenswrapper[4805]: I1203 14:29:09.200152 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-h55gd" Dec 03 14:29:09 crc kubenswrapper[4805]: I1203 14:29:09.206111 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"swift-proxy-config-data" Dec 03 14:29:09 crc kubenswrapper[4805]: I1203 14:29:09.206380 4805 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"swift-ring-config-data" Dec 03 14:29:09 crc kubenswrapper[4805]: I1203 14:29:09.206543 4805 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"swift-ring-scripts" Dec 03 14:29:09 crc kubenswrapper[4805]: I1203 14:29:09.258533 4805 scope.go:117] "RemoveContainer" containerID="ef9c08e3a8b1a0b765a1ed21255c4481ef432d9841cf9d7887a4b0e721a54f17" Dec 03 14:29:09 crc kubenswrapper[4805]: I1203 14:29:09.277736 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-ring-rebalance-h55gd"] Dec 03 14:29:09 crc kubenswrapper[4805]: I1203 14:29:09.313917 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/acfd02d5-fb0a-4427-95e0-4c9f956ea137-dispersionconf\") pod \"swift-ring-rebalance-h55gd\" (UID: \"acfd02d5-fb0a-4427-95e0-4c9f956ea137\") " pod="openstack/swift-ring-rebalance-h55gd" Dec 03 14:29:09 crc kubenswrapper[4805]: I1203 14:29:09.313963 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/acfd02d5-fb0a-4427-95e0-4c9f956ea137-ring-data-devices\") pod \"swift-ring-rebalance-h55gd\" (UID: \"acfd02d5-fb0a-4427-95e0-4c9f956ea137\") " pod="openstack/swift-ring-rebalance-h55gd" Dec 03 14:29:09 crc kubenswrapper[4805]: I1203 14:29:09.313988 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lx4q5\" (UniqueName: \"kubernetes.io/projected/acfd02d5-fb0a-4427-95e0-4c9f956ea137-kube-api-access-lx4q5\") pod \"swift-ring-rebalance-h55gd\" (UID: \"acfd02d5-fb0a-4427-95e0-4c9f956ea137\") " pod="openstack/swift-ring-rebalance-h55gd" Dec 03 14:29:09 crc kubenswrapper[4805]: I1203 14:29:09.314036 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/acfd02d5-fb0a-4427-95e0-4c9f956ea137-etc-swift\") pod \"swift-ring-rebalance-h55gd\" (UID: \"acfd02d5-fb0a-4427-95e0-4c9f956ea137\") " pod="openstack/swift-ring-rebalance-h55gd" Dec 03 14:29:09 crc kubenswrapper[4805]: I1203 14:29:09.314067 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/acfd02d5-fb0a-4427-95e0-4c9f956ea137-swiftconf\") pod \"swift-ring-rebalance-h55gd\" (UID: \"acfd02d5-fb0a-4427-95e0-4c9f956ea137\") " pod="openstack/swift-ring-rebalance-h55gd" Dec 03 14:29:09 crc kubenswrapper[4805]: I1203 14:29:09.314117 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/acfd02d5-fb0a-4427-95e0-4c9f956ea137-combined-ca-bundle\") pod \"swift-ring-rebalance-h55gd\" (UID: \"acfd02d5-fb0a-4427-95e0-4c9f956ea137\") " pod="openstack/swift-ring-rebalance-h55gd" Dec 03 14:29:09 crc kubenswrapper[4805]: I1203 14:29:09.314146 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/acfd02d5-fb0a-4427-95e0-4c9f956ea137-scripts\") pod \"swift-ring-rebalance-h55gd\" (UID: \"acfd02d5-fb0a-4427-95e0-4c9f956ea137\") " pod="openstack/swift-ring-rebalance-h55gd" Dec 03 14:29:09 crc kubenswrapper[4805]: I1203 14:29:09.346987 4805 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-86db49b7ff-x7rps"] Dec 03 14:29:09 crc kubenswrapper[4805]: I1203 14:29:09.347150 4805 scope.go:117] "RemoveContainer" containerID="0eec91f5631e0ed7c3bc78c7ce05a3e4effb735201f47c86b2313526b7653276" Dec 03 14:29:09 crc kubenswrapper[4805]: E1203 14:29:09.347584 4805 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0eec91f5631e0ed7c3bc78c7ce05a3e4effb735201f47c86b2313526b7653276\": container with ID starting with 0eec91f5631e0ed7c3bc78c7ce05a3e4effb735201f47c86b2313526b7653276 not found: ID does not exist" containerID="0eec91f5631e0ed7c3bc78c7ce05a3e4effb735201f47c86b2313526b7653276" Dec 03 14:29:09 crc kubenswrapper[4805]: I1203 14:29:09.347615 4805 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0eec91f5631e0ed7c3bc78c7ce05a3e4effb735201f47c86b2313526b7653276"} err="failed to get container status \"0eec91f5631e0ed7c3bc78c7ce05a3e4effb735201f47c86b2313526b7653276\": rpc error: code = NotFound desc = could not find container \"0eec91f5631e0ed7c3bc78c7ce05a3e4effb735201f47c86b2313526b7653276\": container with ID starting with 0eec91f5631e0ed7c3bc78c7ce05a3e4effb735201f47c86b2313526b7653276 not found: ID does not exist" Dec 03 14:29:09 crc kubenswrapper[4805]: I1203 14:29:09.347636 4805 scope.go:117] "RemoveContainer" containerID="ef9c08e3a8b1a0b765a1ed21255c4481ef432d9841cf9d7887a4b0e721a54f17" Dec 03 14:29:09 crc kubenswrapper[4805]: E1203 14:29:09.348059 4805 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ef9c08e3a8b1a0b765a1ed21255c4481ef432d9841cf9d7887a4b0e721a54f17\": container with ID starting with ef9c08e3a8b1a0b765a1ed21255c4481ef432d9841cf9d7887a4b0e721a54f17 not found: ID does not exist" containerID="ef9c08e3a8b1a0b765a1ed21255c4481ef432d9841cf9d7887a4b0e721a54f17" Dec 03 14:29:09 crc kubenswrapper[4805]: I1203 14:29:09.348080 4805 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ef9c08e3a8b1a0b765a1ed21255c4481ef432d9841cf9d7887a4b0e721a54f17"} err="failed to get container status \"ef9c08e3a8b1a0b765a1ed21255c4481ef432d9841cf9d7887a4b0e721a54f17\": rpc error: code = NotFound desc = could not find container \"ef9c08e3a8b1a0b765a1ed21255c4481ef432d9841cf9d7887a4b0e721a54f17\": container with ID starting with ef9c08e3a8b1a0b765a1ed21255c4481ef432d9841cf9d7887a4b0e721a54f17 not found: ID does not exist" Dec 03 14:29:09 crc kubenswrapper[4805]: I1203 14:29:09.359925 4805 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-86db49b7ff-x7rps"] Dec 03 14:29:09 crc kubenswrapper[4805]: I1203 14:29:09.415029 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/acfd02d5-fb0a-4427-95e0-4c9f956ea137-combined-ca-bundle\") pod \"swift-ring-rebalance-h55gd\" (UID: \"acfd02d5-fb0a-4427-95e0-4c9f956ea137\") " pod="openstack/swift-ring-rebalance-h55gd" Dec 03 14:29:09 crc kubenswrapper[4805]: I1203 14:29:09.415090 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/acfd02d5-fb0a-4427-95e0-4c9f956ea137-scripts\") pod \"swift-ring-rebalance-h55gd\" (UID: \"acfd02d5-fb0a-4427-95e0-4c9f956ea137\") " pod="openstack/swift-ring-rebalance-h55gd" Dec 03 14:29:09 crc kubenswrapper[4805]: I1203 14:29:09.415137 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/acfd02d5-fb0a-4427-95e0-4c9f956ea137-dispersionconf\") pod \"swift-ring-rebalance-h55gd\" (UID: \"acfd02d5-fb0a-4427-95e0-4c9f956ea137\") " pod="openstack/swift-ring-rebalance-h55gd" Dec 03 14:29:09 crc kubenswrapper[4805]: I1203 14:29:09.415160 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/acfd02d5-fb0a-4427-95e0-4c9f956ea137-ring-data-devices\") pod \"swift-ring-rebalance-h55gd\" (UID: \"acfd02d5-fb0a-4427-95e0-4c9f956ea137\") " pod="openstack/swift-ring-rebalance-h55gd" Dec 03 14:29:09 crc kubenswrapper[4805]: I1203 14:29:09.415181 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lx4q5\" (UniqueName: \"kubernetes.io/projected/acfd02d5-fb0a-4427-95e0-4c9f956ea137-kube-api-access-lx4q5\") pod \"swift-ring-rebalance-h55gd\" (UID: \"acfd02d5-fb0a-4427-95e0-4c9f956ea137\") " pod="openstack/swift-ring-rebalance-h55gd" Dec 03 14:29:09 crc kubenswrapper[4805]: I1203 14:29:09.415232 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/4fbc8d7f-d47f-4d91-915b-504820aac08a-etc-swift\") pod \"swift-storage-0\" (UID: \"4fbc8d7f-d47f-4d91-915b-504820aac08a\") " pod="openstack/swift-storage-0" Dec 03 14:29:09 crc kubenswrapper[4805]: I1203 14:29:09.415249 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/acfd02d5-fb0a-4427-95e0-4c9f956ea137-etc-swift\") pod \"swift-ring-rebalance-h55gd\" (UID: \"acfd02d5-fb0a-4427-95e0-4c9f956ea137\") " pod="openstack/swift-ring-rebalance-h55gd" Dec 03 14:29:09 crc kubenswrapper[4805]: I1203 14:29:09.415276 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/acfd02d5-fb0a-4427-95e0-4c9f956ea137-swiftconf\") pod \"swift-ring-rebalance-h55gd\" (UID: \"acfd02d5-fb0a-4427-95e0-4c9f956ea137\") " pod="openstack/swift-ring-rebalance-h55gd" Dec 03 14:29:09 crc kubenswrapper[4805]: E1203 14:29:09.415513 4805 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Dec 03 14:29:09 crc kubenswrapper[4805]: E1203 14:29:09.415583 4805 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Dec 03 14:29:09 crc kubenswrapper[4805]: E1203 14:29:09.415673 4805 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/4fbc8d7f-d47f-4d91-915b-504820aac08a-etc-swift podName:4fbc8d7f-d47f-4d91-915b-504820aac08a nodeName:}" failed. No retries permitted until 2025-12-03 14:29:10.415656633 +0000 UTC m=+1180.078573556 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/4fbc8d7f-d47f-4d91-915b-504820aac08a-etc-swift") pod "swift-storage-0" (UID: "4fbc8d7f-d47f-4d91-915b-504820aac08a") : configmap "swift-ring-files" not found Dec 03 14:29:09 crc kubenswrapper[4805]: I1203 14:29:09.416173 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/acfd02d5-fb0a-4427-95e0-4c9f956ea137-etc-swift\") pod \"swift-ring-rebalance-h55gd\" (UID: \"acfd02d5-fb0a-4427-95e0-4c9f956ea137\") " pod="openstack/swift-ring-rebalance-h55gd" Dec 03 14:29:09 crc kubenswrapper[4805]: I1203 14:29:09.416243 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/acfd02d5-fb0a-4427-95e0-4c9f956ea137-scripts\") pod \"swift-ring-rebalance-h55gd\" (UID: \"acfd02d5-fb0a-4427-95e0-4c9f956ea137\") " pod="openstack/swift-ring-rebalance-h55gd" Dec 03 14:29:09 crc kubenswrapper[4805]: I1203 14:29:09.416333 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/acfd02d5-fb0a-4427-95e0-4c9f956ea137-ring-data-devices\") pod \"swift-ring-rebalance-h55gd\" (UID: \"acfd02d5-fb0a-4427-95e0-4c9f956ea137\") " pod="openstack/swift-ring-rebalance-h55gd" Dec 03 14:29:09 crc kubenswrapper[4805]: I1203 14:29:09.419231 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/acfd02d5-fb0a-4427-95e0-4c9f956ea137-swiftconf\") pod \"swift-ring-rebalance-h55gd\" (UID: \"acfd02d5-fb0a-4427-95e0-4c9f956ea137\") " pod="openstack/swift-ring-rebalance-h55gd" Dec 03 14:29:09 crc kubenswrapper[4805]: I1203 14:29:09.419291 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/acfd02d5-fb0a-4427-95e0-4c9f956ea137-combined-ca-bundle\") pod \"swift-ring-rebalance-h55gd\" (UID: \"acfd02d5-fb0a-4427-95e0-4c9f956ea137\") " pod="openstack/swift-ring-rebalance-h55gd" Dec 03 14:29:09 crc kubenswrapper[4805]: I1203 14:29:09.419967 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/acfd02d5-fb0a-4427-95e0-4c9f956ea137-dispersionconf\") pod \"swift-ring-rebalance-h55gd\" (UID: \"acfd02d5-fb0a-4427-95e0-4c9f956ea137\") " pod="openstack/swift-ring-rebalance-h55gd" Dec 03 14:29:09 crc kubenswrapper[4805]: I1203 14:29:09.436497 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lx4q5\" (UniqueName: \"kubernetes.io/projected/acfd02d5-fb0a-4427-95e0-4c9f956ea137-kube-api-access-lx4q5\") pod \"swift-ring-rebalance-h55gd\" (UID: \"acfd02d5-fb0a-4427-95e0-4c9f956ea137\") " pod="openstack/swift-ring-rebalance-h55gd" Dec 03 14:29:09 crc kubenswrapper[4805]: I1203 14:29:09.551225 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-h55gd" Dec 03 14:29:10 crc kubenswrapper[4805]: I1203 14:29:10.049767 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-ring-rebalance-h55gd"] Dec 03 14:29:10 crc kubenswrapper[4805]: I1203 14:29:10.152823 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-698758b865-l46rr" event={"ID":"83e323ed-ad2c-4635-a583-4e5b21d8e268","Type":"ContainerStarted","Data":"55ae092c6099423baa000af1306ee0efd34017bebbf5abd512f35e371752d635"} Dec 03 14:29:10 crc kubenswrapper[4805]: I1203 14:29:10.152961 4805 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-698758b865-l46rr" Dec 03 14:29:10 crc kubenswrapper[4805]: I1203 14:29:10.155698 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-ring-rebalance-h55gd" event={"ID":"acfd02d5-fb0a-4427-95e0-4c9f956ea137","Type":"ContainerStarted","Data":"1bc90614ae492aad000272447edf23d5a1666c8a615ffcd8ddd4f55a33688327"} Dec 03 14:29:10 crc kubenswrapper[4805]: I1203 14:29:10.179146 4805 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-698758b865-l46rr" podStartSLOduration=3.17912462 podStartE2EDuration="3.17912462s" podCreationTimestamp="2025-12-03 14:29:07 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 14:29:10.172233834 +0000 UTC m=+1179.835150767" watchObservedRunningTime="2025-12-03 14:29:10.17912462 +0000 UTC m=+1179.842041543" Dec 03 14:29:10 crc kubenswrapper[4805]: I1203 14:29:10.431905 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/4fbc8d7f-d47f-4d91-915b-504820aac08a-etc-swift\") pod \"swift-storage-0\" (UID: \"4fbc8d7f-d47f-4d91-915b-504820aac08a\") " pod="openstack/swift-storage-0" Dec 03 14:29:10 crc kubenswrapper[4805]: E1203 14:29:10.432137 4805 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Dec 03 14:29:10 crc kubenswrapper[4805]: E1203 14:29:10.432182 4805 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Dec 03 14:29:10 crc kubenswrapper[4805]: E1203 14:29:10.432251 4805 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/4fbc8d7f-d47f-4d91-915b-504820aac08a-etc-swift podName:4fbc8d7f-d47f-4d91-915b-504820aac08a nodeName:}" failed. No retries permitted until 2025-12-03 14:29:12.432222831 +0000 UTC m=+1182.095139754 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/4fbc8d7f-d47f-4d91-915b-504820aac08a-etc-swift") pod "swift-storage-0" (UID: "4fbc8d7f-d47f-4d91-915b-504820aac08a") : configmap "swift-ring-files" not found Dec 03 14:29:10 crc kubenswrapper[4805]: I1203 14:29:10.718044 4805 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e7db8f49-10ce-4818-8d9b-963d6e2e49ce" path="/var/lib/kubelet/pods/e7db8f49-10ce-4818-8d9b-963d6e2e49ce/volumes" Dec 03 14:29:12 crc kubenswrapper[4805]: I1203 14:29:12.471664 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/4fbc8d7f-d47f-4d91-915b-504820aac08a-etc-swift\") pod \"swift-storage-0\" (UID: \"4fbc8d7f-d47f-4d91-915b-504820aac08a\") " pod="openstack/swift-storage-0" Dec 03 14:29:12 crc kubenswrapper[4805]: E1203 14:29:12.471857 4805 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Dec 03 14:29:12 crc kubenswrapper[4805]: E1203 14:29:12.472271 4805 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Dec 03 14:29:12 crc kubenswrapper[4805]: E1203 14:29:12.472332 4805 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/4fbc8d7f-d47f-4d91-915b-504820aac08a-etc-swift podName:4fbc8d7f-d47f-4d91-915b-504820aac08a nodeName:}" failed. No retries permitted until 2025-12-03 14:29:16.472310147 +0000 UTC m=+1186.135227080 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/4fbc8d7f-d47f-4d91-915b-504820aac08a-etc-swift") pod "swift-storage-0" (UID: "4fbc8d7f-d47f-4d91-915b-504820aac08a") : configmap "swift-ring-files" not found Dec 03 14:29:13 crc kubenswrapper[4805]: I1203 14:29:13.690916 4805 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/openstack-galera-0" Dec 03 14:29:13 crc kubenswrapper[4805]: I1203 14:29:13.691560 4805 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/openstack-galera-0" Dec 03 14:29:13 crc kubenswrapper[4805]: I1203 14:29:13.765901 4805 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/openstack-galera-0" Dec 03 14:29:14 crc kubenswrapper[4805]: I1203 14:29:14.186683 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-ring-rebalance-h55gd" event={"ID":"acfd02d5-fb0a-4427-95e0-4c9f956ea137","Type":"ContainerStarted","Data":"e4fc3ae321cc76c9f7df0935f144238ddea8a8483203f25ff4ea0b3906124127"} Dec 03 14:29:14 crc kubenswrapper[4805]: I1203 14:29:14.205914 4805 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/swift-ring-rebalance-h55gd" podStartSLOduration=2.010026223 podStartE2EDuration="5.205895984s" podCreationTimestamp="2025-12-03 14:29:09 +0000 UTC" firstStartedPulling="2025-12-03 14:29:10.062945015 +0000 UTC m=+1179.725861938" lastFinishedPulling="2025-12-03 14:29:13.258814776 +0000 UTC m=+1182.921731699" observedRunningTime="2025-12-03 14:29:14.205342889 +0000 UTC m=+1183.868259832" watchObservedRunningTime="2025-12-03 14:29:14.205895984 +0000 UTC m=+1183.868812907" Dec 03 14:29:14 crc kubenswrapper[4805]: I1203 14:29:14.280762 4805 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/openstack-galera-0" Dec 03 14:29:15 crc kubenswrapper[4805]: I1203 14:29:15.033555 4805 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-db-create-bbs5s"] Dec 03 14:29:15 crc kubenswrapper[4805]: I1203 14:29:15.034942 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-create-bbs5s" Dec 03 14:29:15 crc kubenswrapper[4805]: I1203 14:29:15.042247 4805 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-32eb-account-create-update-jmj7x"] Dec 03 14:29:15 crc kubenswrapper[4805]: I1203 14:29:15.045972 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-32eb-account-create-update-jmj7x" Dec 03 14:29:15 crc kubenswrapper[4805]: I1203 14:29:15.048011 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-db-secret" Dec 03 14:29:15 crc kubenswrapper[4805]: I1203 14:29:15.055628 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-db-create-bbs5s"] Dec 03 14:29:15 crc kubenswrapper[4805]: I1203 14:29:15.065215 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-32eb-account-create-update-jmj7x"] Dec 03 14:29:15 crc kubenswrapper[4805]: I1203 14:29:15.219867 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lz98j\" (UniqueName: \"kubernetes.io/projected/7895c06e-d5d7-4e5b-9f56-442cabf914bc-kube-api-access-lz98j\") pod \"keystone-32eb-account-create-update-jmj7x\" (UID: \"7895c06e-d5d7-4e5b-9f56-442cabf914bc\") " pod="openstack/keystone-32eb-account-create-update-jmj7x" Dec 03 14:29:15 crc kubenswrapper[4805]: I1203 14:29:15.220024 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/7895c06e-d5d7-4e5b-9f56-442cabf914bc-operator-scripts\") pod \"keystone-32eb-account-create-update-jmj7x\" (UID: \"7895c06e-d5d7-4e5b-9f56-442cabf914bc\") " pod="openstack/keystone-32eb-account-create-update-jmj7x" Dec 03 14:29:15 crc kubenswrapper[4805]: I1203 14:29:15.220119 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fz7fn\" (UniqueName: \"kubernetes.io/projected/a43c2a2a-291b-4422-9c06-cb448b722758-kube-api-access-fz7fn\") pod \"keystone-db-create-bbs5s\" (UID: \"a43c2a2a-291b-4422-9c06-cb448b722758\") " pod="openstack/keystone-db-create-bbs5s" Dec 03 14:29:15 crc kubenswrapper[4805]: I1203 14:29:15.220313 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/a43c2a2a-291b-4422-9c06-cb448b722758-operator-scripts\") pod \"keystone-db-create-bbs5s\" (UID: \"a43c2a2a-291b-4422-9c06-cb448b722758\") " pod="openstack/keystone-db-create-bbs5s" Dec 03 14:29:15 crc kubenswrapper[4805]: I1203 14:29:15.268115 4805 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement-db-create-rrmbh"] Dec 03 14:29:15 crc kubenswrapper[4805]: I1203 14:29:15.269449 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-create-rrmbh" Dec 03 14:29:15 crc kubenswrapper[4805]: I1203 14:29:15.286923 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-db-create-rrmbh"] Dec 03 14:29:15 crc kubenswrapper[4805]: I1203 14:29:15.324770 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lz98j\" (UniqueName: \"kubernetes.io/projected/7895c06e-d5d7-4e5b-9f56-442cabf914bc-kube-api-access-lz98j\") pod \"keystone-32eb-account-create-update-jmj7x\" (UID: \"7895c06e-d5d7-4e5b-9f56-442cabf914bc\") " pod="openstack/keystone-32eb-account-create-update-jmj7x" Dec 03 14:29:15 crc kubenswrapper[4805]: I1203 14:29:15.324897 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/7895c06e-d5d7-4e5b-9f56-442cabf914bc-operator-scripts\") pod \"keystone-32eb-account-create-update-jmj7x\" (UID: \"7895c06e-d5d7-4e5b-9f56-442cabf914bc\") " pod="openstack/keystone-32eb-account-create-update-jmj7x" Dec 03 14:29:15 crc kubenswrapper[4805]: I1203 14:29:15.325627 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/7895c06e-d5d7-4e5b-9f56-442cabf914bc-operator-scripts\") pod \"keystone-32eb-account-create-update-jmj7x\" (UID: \"7895c06e-d5d7-4e5b-9f56-442cabf914bc\") " pod="openstack/keystone-32eb-account-create-update-jmj7x" Dec 03 14:29:15 crc kubenswrapper[4805]: I1203 14:29:15.324941 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fz7fn\" (UniqueName: \"kubernetes.io/projected/a43c2a2a-291b-4422-9c06-cb448b722758-kube-api-access-fz7fn\") pod \"keystone-db-create-bbs5s\" (UID: \"a43c2a2a-291b-4422-9c06-cb448b722758\") " pod="openstack/keystone-db-create-bbs5s" Dec 03 14:29:15 crc kubenswrapper[4805]: I1203 14:29:15.330493 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/a43c2a2a-291b-4422-9c06-cb448b722758-operator-scripts\") pod \"keystone-db-create-bbs5s\" (UID: \"a43c2a2a-291b-4422-9c06-cb448b722758\") " pod="openstack/keystone-db-create-bbs5s" Dec 03 14:29:15 crc kubenswrapper[4805]: I1203 14:29:15.331907 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/a43c2a2a-291b-4422-9c06-cb448b722758-operator-scripts\") pod \"keystone-db-create-bbs5s\" (UID: \"a43c2a2a-291b-4422-9c06-cb448b722758\") " pod="openstack/keystone-db-create-bbs5s" Dec 03 14:29:15 crc kubenswrapper[4805]: I1203 14:29:15.346344 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fz7fn\" (UniqueName: \"kubernetes.io/projected/a43c2a2a-291b-4422-9c06-cb448b722758-kube-api-access-fz7fn\") pod \"keystone-db-create-bbs5s\" (UID: \"a43c2a2a-291b-4422-9c06-cb448b722758\") " pod="openstack/keystone-db-create-bbs5s" Dec 03 14:29:15 crc kubenswrapper[4805]: I1203 14:29:15.353412 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lz98j\" (UniqueName: \"kubernetes.io/projected/7895c06e-d5d7-4e5b-9f56-442cabf914bc-kube-api-access-lz98j\") pod \"keystone-32eb-account-create-update-jmj7x\" (UID: \"7895c06e-d5d7-4e5b-9f56-442cabf914bc\") " pod="openstack/keystone-32eb-account-create-update-jmj7x" Dec 03 14:29:15 crc kubenswrapper[4805]: I1203 14:29:15.364088 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-create-bbs5s" Dec 03 14:29:15 crc kubenswrapper[4805]: I1203 14:29:15.370250 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-32eb-account-create-update-jmj7x" Dec 03 14:29:15 crc kubenswrapper[4805]: I1203 14:29:15.389731 4805 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement-7094-account-create-update-c7dm6"] Dec 03 14:29:15 crc kubenswrapper[4805]: I1203 14:29:15.391342 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-7094-account-create-update-c7dm6" Dec 03 14:29:15 crc kubenswrapper[4805]: I1203 14:29:15.393991 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-db-secret" Dec 03 14:29:15 crc kubenswrapper[4805]: I1203 14:29:15.398704 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-7094-account-create-update-c7dm6"] Dec 03 14:29:15 crc kubenswrapper[4805]: I1203 14:29:15.434075 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sxpbf\" (UniqueName: \"kubernetes.io/projected/4c56e040-47cb-4e7e-86d4-613e09cc3243-kube-api-access-sxpbf\") pod \"placement-db-create-rrmbh\" (UID: \"4c56e040-47cb-4e7e-86d4-613e09cc3243\") " pod="openstack/placement-db-create-rrmbh" Dec 03 14:29:15 crc kubenswrapper[4805]: I1203 14:29:15.434142 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/4c56e040-47cb-4e7e-86d4-613e09cc3243-operator-scripts\") pod \"placement-db-create-rrmbh\" (UID: \"4c56e040-47cb-4e7e-86d4-613e09cc3243\") " pod="openstack/placement-db-create-rrmbh" Dec 03 14:29:15 crc kubenswrapper[4805]: I1203 14:29:15.501384 4805 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-db-create-tbhgg"] Dec 03 14:29:15 crc kubenswrapper[4805]: I1203 14:29:15.503409 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-create-tbhgg" Dec 03 14:29:15 crc kubenswrapper[4805]: I1203 14:29:15.512717 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-db-create-tbhgg"] Dec 03 14:29:15 crc kubenswrapper[4805]: I1203 14:29:15.535333 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/26a760ac-b299-4492-ab01-95db5e14ac97-operator-scripts\") pod \"placement-7094-account-create-update-c7dm6\" (UID: \"26a760ac-b299-4492-ab01-95db5e14ac97\") " pod="openstack/placement-7094-account-create-update-c7dm6" Dec 03 14:29:15 crc kubenswrapper[4805]: I1203 14:29:15.535388 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sxpbf\" (UniqueName: \"kubernetes.io/projected/4c56e040-47cb-4e7e-86d4-613e09cc3243-kube-api-access-sxpbf\") pod \"placement-db-create-rrmbh\" (UID: \"4c56e040-47cb-4e7e-86d4-613e09cc3243\") " pod="openstack/placement-db-create-rrmbh" Dec 03 14:29:15 crc kubenswrapper[4805]: I1203 14:29:15.535439 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/4c56e040-47cb-4e7e-86d4-613e09cc3243-operator-scripts\") pod \"placement-db-create-rrmbh\" (UID: \"4c56e040-47cb-4e7e-86d4-613e09cc3243\") " pod="openstack/placement-db-create-rrmbh" Dec 03 14:29:15 crc kubenswrapper[4805]: I1203 14:29:15.535524 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-v85rv\" (UniqueName: \"kubernetes.io/projected/26a760ac-b299-4492-ab01-95db5e14ac97-kube-api-access-v85rv\") pod \"placement-7094-account-create-update-c7dm6\" (UID: \"26a760ac-b299-4492-ab01-95db5e14ac97\") " pod="openstack/placement-7094-account-create-update-c7dm6" Dec 03 14:29:15 crc kubenswrapper[4805]: I1203 14:29:15.536716 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/4c56e040-47cb-4e7e-86d4-613e09cc3243-operator-scripts\") pod \"placement-db-create-rrmbh\" (UID: \"4c56e040-47cb-4e7e-86d4-613e09cc3243\") " pod="openstack/placement-db-create-rrmbh" Dec 03 14:29:15 crc kubenswrapper[4805]: I1203 14:29:15.567736 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sxpbf\" (UniqueName: \"kubernetes.io/projected/4c56e040-47cb-4e7e-86d4-613e09cc3243-kube-api-access-sxpbf\") pod \"placement-db-create-rrmbh\" (UID: \"4c56e040-47cb-4e7e-86d4-613e09cc3243\") " pod="openstack/placement-db-create-rrmbh" Dec 03 14:29:15 crc kubenswrapper[4805]: I1203 14:29:15.589383 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-create-rrmbh" Dec 03 14:29:15 crc kubenswrapper[4805]: I1203 14:29:15.592475 4805 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-de85-account-create-update-8gf95"] Dec 03 14:29:15 crc kubenswrapper[4805]: I1203 14:29:15.594024 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-de85-account-create-update-8gf95" Dec 03 14:29:15 crc kubenswrapper[4805]: I1203 14:29:15.600311 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-db-secret" Dec 03 14:29:15 crc kubenswrapper[4805]: I1203 14:29:15.617748 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-de85-account-create-update-8gf95"] Dec 03 14:29:15 crc kubenswrapper[4805]: I1203 14:29:15.637666 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-r4mpz\" (UniqueName: \"kubernetes.io/projected/3e5d8271-48f9-4bc1-b35c-a37f919eefdd-kube-api-access-r4mpz\") pod \"glance-db-create-tbhgg\" (UID: \"3e5d8271-48f9-4bc1-b35c-a37f919eefdd\") " pod="openstack/glance-db-create-tbhgg" Dec 03 14:29:15 crc kubenswrapper[4805]: I1203 14:29:15.637727 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/3e5d8271-48f9-4bc1-b35c-a37f919eefdd-operator-scripts\") pod \"glance-db-create-tbhgg\" (UID: \"3e5d8271-48f9-4bc1-b35c-a37f919eefdd\") " pod="openstack/glance-db-create-tbhgg" Dec 03 14:29:15 crc kubenswrapper[4805]: I1203 14:29:15.637778 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-v85rv\" (UniqueName: \"kubernetes.io/projected/26a760ac-b299-4492-ab01-95db5e14ac97-kube-api-access-v85rv\") pod \"placement-7094-account-create-update-c7dm6\" (UID: \"26a760ac-b299-4492-ab01-95db5e14ac97\") " pod="openstack/placement-7094-account-create-update-c7dm6" Dec 03 14:29:15 crc kubenswrapper[4805]: I1203 14:29:15.637880 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/26a760ac-b299-4492-ab01-95db5e14ac97-operator-scripts\") pod \"placement-7094-account-create-update-c7dm6\" (UID: \"26a760ac-b299-4492-ab01-95db5e14ac97\") " pod="openstack/placement-7094-account-create-update-c7dm6" Dec 03 14:29:15 crc kubenswrapper[4805]: I1203 14:29:15.641377 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/26a760ac-b299-4492-ab01-95db5e14ac97-operator-scripts\") pod \"placement-7094-account-create-update-c7dm6\" (UID: \"26a760ac-b299-4492-ab01-95db5e14ac97\") " pod="openstack/placement-7094-account-create-update-c7dm6" Dec 03 14:29:15 crc kubenswrapper[4805]: I1203 14:29:15.656427 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-v85rv\" (UniqueName: \"kubernetes.io/projected/26a760ac-b299-4492-ab01-95db5e14ac97-kube-api-access-v85rv\") pod \"placement-7094-account-create-update-c7dm6\" (UID: \"26a760ac-b299-4492-ab01-95db5e14ac97\") " pod="openstack/placement-7094-account-create-update-c7dm6" Dec 03 14:29:15 crc kubenswrapper[4805]: I1203 14:29:15.739329 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/8427fb09-203a-4093-807c-9f03061b6e07-operator-scripts\") pod \"glance-de85-account-create-update-8gf95\" (UID: \"8427fb09-203a-4093-807c-9f03061b6e07\") " pod="openstack/glance-de85-account-create-update-8gf95" Dec 03 14:29:15 crc kubenswrapper[4805]: I1203 14:29:15.739402 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-r4mpz\" (UniqueName: \"kubernetes.io/projected/3e5d8271-48f9-4bc1-b35c-a37f919eefdd-kube-api-access-r4mpz\") pod \"glance-db-create-tbhgg\" (UID: \"3e5d8271-48f9-4bc1-b35c-a37f919eefdd\") " pod="openstack/glance-db-create-tbhgg" Dec 03 14:29:15 crc kubenswrapper[4805]: I1203 14:29:15.739430 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vcrp5\" (UniqueName: \"kubernetes.io/projected/8427fb09-203a-4093-807c-9f03061b6e07-kube-api-access-vcrp5\") pod \"glance-de85-account-create-update-8gf95\" (UID: \"8427fb09-203a-4093-807c-9f03061b6e07\") " pod="openstack/glance-de85-account-create-update-8gf95" Dec 03 14:29:15 crc kubenswrapper[4805]: I1203 14:29:15.739458 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/3e5d8271-48f9-4bc1-b35c-a37f919eefdd-operator-scripts\") pod \"glance-db-create-tbhgg\" (UID: \"3e5d8271-48f9-4bc1-b35c-a37f919eefdd\") " pod="openstack/glance-db-create-tbhgg" Dec 03 14:29:15 crc kubenswrapper[4805]: I1203 14:29:15.741186 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/3e5d8271-48f9-4bc1-b35c-a37f919eefdd-operator-scripts\") pod \"glance-db-create-tbhgg\" (UID: \"3e5d8271-48f9-4bc1-b35c-a37f919eefdd\") " pod="openstack/glance-db-create-tbhgg" Dec 03 14:29:15 crc kubenswrapper[4805]: I1203 14:29:15.757406 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-r4mpz\" (UniqueName: \"kubernetes.io/projected/3e5d8271-48f9-4bc1-b35c-a37f919eefdd-kube-api-access-r4mpz\") pod \"glance-db-create-tbhgg\" (UID: \"3e5d8271-48f9-4bc1-b35c-a37f919eefdd\") " pod="openstack/glance-db-create-tbhgg" Dec 03 14:29:15 crc kubenswrapper[4805]: I1203 14:29:15.841474 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/8427fb09-203a-4093-807c-9f03061b6e07-operator-scripts\") pod \"glance-de85-account-create-update-8gf95\" (UID: \"8427fb09-203a-4093-807c-9f03061b6e07\") " pod="openstack/glance-de85-account-create-update-8gf95" Dec 03 14:29:15 crc kubenswrapper[4805]: I1203 14:29:15.841565 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vcrp5\" (UniqueName: \"kubernetes.io/projected/8427fb09-203a-4093-807c-9f03061b6e07-kube-api-access-vcrp5\") pod \"glance-de85-account-create-update-8gf95\" (UID: \"8427fb09-203a-4093-807c-9f03061b6e07\") " pod="openstack/glance-de85-account-create-update-8gf95" Dec 03 14:29:15 crc kubenswrapper[4805]: I1203 14:29:15.842251 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/8427fb09-203a-4093-807c-9f03061b6e07-operator-scripts\") pod \"glance-de85-account-create-update-8gf95\" (UID: \"8427fb09-203a-4093-807c-9f03061b6e07\") " pod="openstack/glance-de85-account-create-update-8gf95" Dec 03 14:29:15 crc kubenswrapper[4805]: I1203 14:29:15.858401 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vcrp5\" (UniqueName: \"kubernetes.io/projected/8427fb09-203a-4093-807c-9f03061b6e07-kube-api-access-vcrp5\") pod \"glance-de85-account-create-update-8gf95\" (UID: \"8427fb09-203a-4093-807c-9f03061b6e07\") " pod="openstack/glance-de85-account-create-update-8gf95" Dec 03 14:29:15 crc kubenswrapper[4805]: I1203 14:29:15.860251 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-7094-account-create-update-c7dm6" Dec 03 14:29:15 crc kubenswrapper[4805]: I1203 14:29:15.898879 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-32eb-account-create-update-jmj7x"] Dec 03 14:29:15 crc kubenswrapper[4805]: W1203 14:29:15.906696 4805 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod7895c06e_d5d7_4e5b_9f56_442cabf914bc.slice/crio-f8b849472536bd208e0c41f9458d0339551a80b92d85280633b338345be2c3ca WatchSource:0}: Error finding container f8b849472536bd208e0c41f9458d0339551a80b92d85280633b338345be2c3ca: Status 404 returned error can't find the container with id f8b849472536bd208e0c41f9458d0339551a80b92d85280633b338345be2c3ca Dec 03 14:29:15 crc kubenswrapper[4805]: I1203 14:29:15.907724 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-create-tbhgg" Dec 03 14:29:15 crc kubenswrapper[4805]: I1203 14:29:15.935047 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-de85-account-create-update-8gf95" Dec 03 14:29:15 crc kubenswrapper[4805]: I1203 14:29:15.992640 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-db-create-bbs5s"] Dec 03 14:29:16 crc kubenswrapper[4805]: W1203 14:29:16.011111 4805 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poda43c2a2a_291b_4422_9c06_cb448b722758.slice/crio-276e8f06108c78e6cb04674a839349031b4f7a30abbd7a2dbbb52637f50c4090 WatchSource:0}: Error finding container 276e8f06108c78e6cb04674a839349031b4f7a30abbd7a2dbbb52637f50c4090: Status 404 returned error can't find the container with id 276e8f06108c78e6cb04674a839349031b4f7a30abbd7a2dbbb52637f50c4090 Dec 03 14:29:16 crc kubenswrapper[4805]: I1203 14:29:16.067585 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-db-create-rrmbh"] Dec 03 14:29:16 crc kubenswrapper[4805]: I1203 14:29:16.211413 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-create-rrmbh" event={"ID":"4c56e040-47cb-4e7e-86d4-613e09cc3243","Type":"ContainerStarted","Data":"fdfb88f20dcd2c0b764484a6fd63cfda7012223237f0d7e3c1ce139874daf60d"} Dec 03 14:29:16 crc kubenswrapper[4805]: I1203 14:29:16.212812 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-create-bbs5s" event={"ID":"a43c2a2a-291b-4422-9c06-cb448b722758","Type":"ContainerStarted","Data":"276e8f06108c78e6cb04674a839349031b4f7a30abbd7a2dbbb52637f50c4090"} Dec 03 14:29:16 crc kubenswrapper[4805]: I1203 14:29:16.214949 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-32eb-account-create-update-jmj7x" event={"ID":"7895c06e-d5d7-4e5b-9f56-442cabf914bc","Type":"ContainerStarted","Data":"7d87783d81055065efdd933bb95570fe72fbe1fbdce5e19e32bf1b28e8ae8f37"} Dec 03 14:29:16 crc kubenswrapper[4805]: I1203 14:29:16.214992 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-32eb-account-create-update-jmj7x" event={"ID":"7895c06e-d5d7-4e5b-9f56-442cabf914bc","Type":"ContainerStarted","Data":"f8b849472536bd208e0c41f9458d0339551a80b92d85280633b338345be2c3ca"} Dec 03 14:29:16 crc kubenswrapper[4805]: I1203 14:29:16.335944 4805 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-32eb-account-create-update-jmj7x" podStartSLOduration=1.335921483 podStartE2EDuration="1.335921483s" podCreationTimestamp="2025-12-03 14:29:15 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 14:29:16.234277262 +0000 UTC m=+1185.897194195" watchObservedRunningTime="2025-12-03 14:29:16.335921483 +0000 UTC m=+1185.998838406" Dec 03 14:29:16 crc kubenswrapper[4805]: I1203 14:29:16.336575 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-7094-account-create-update-c7dm6"] Dec 03 14:29:16 crc kubenswrapper[4805]: I1203 14:29:16.427163 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-db-create-tbhgg"] Dec 03 14:29:16 crc kubenswrapper[4805]: W1203 14:29:16.430745 4805 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod3e5d8271_48f9_4bc1_b35c_a37f919eefdd.slice/crio-f68520dc10fadc7f524850a3a3cf1ab1b131a291803b74e102024d1c007707f7 WatchSource:0}: Error finding container f68520dc10fadc7f524850a3a3cf1ab1b131a291803b74e102024d1c007707f7: Status 404 returned error can't find the container with id f68520dc10fadc7f524850a3a3cf1ab1b131a291803b74e102024d1c007707f7 Dec 03 14:29:16 crc kubenswrapper[4805]: I1203 14:29:16.512530 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-de85-account-create-update-8gf95"] Dec 03 14:29:16 crc kubenswrapper[4805]: I1203 14:29:16.553564 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/4fbc8d7f-d47f-4d91-915b-504820aac08a-etc-swift\") pod \"swift-storage-0\" (UID: \"4fbc8d7f-d47f-4d91-915b-504820aac08a\") " pod="openstack/swift-storage-0" Dec 03 14:29:16 crc kubenswrapper[4805]: E1203 14:29:16.553798 4805 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Dec 03 14:29:16 crc kubenswrapper[4805]: E1203 14:29:16.553816 4805 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Dec 03 14:29:16 crc kubenswrapper[4805]: E1203 14:29:16.553891 4805 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/4fbc8d7f-d47f-4d91-915b-504820aac08a-etc-swift podName:4fbc8d7f-d47f-4d91-915b-504820aac08a nodeName:}" failed. No retries permitted until 2025-12-03 14:29:24.553873572 +0000 UTC m=+1194.216790495 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/4fbc8d7f-d47f-4d91-915b-504820aac08a-etc-swift") pod "swift-storage-0" (UID: "4fbc8d7f-d47f-4d91-915b-504820aac08a") : configmap "swift-ring-files" not found Dec 03 14:29:17 crc kubenswrapper[4805]: I1203 14:29:17.223065 4805 generic.go:334] "Generic (PLEG): container finished" podID="a43c2a2a-291b-4422-9c06-cb448b722758" containerID="45ec4e94a53ac4236a647cc33770987e9ea8a30d52b48b172a20c30c8795008e" exitCode=0 Dec 03 14:29:17 crc kubenswrapper[4805]: I1203 14:29:17.223120 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-create-bbs5s" event={"ID":"a43c2a2a-291b-4422-9c06-cb448b722758","Type":"ContainerDied","Data":"45ec4e94a53ac4236a647cc33770987e9ea8a30d52b48b172a20c30c8795008e"} Dec 03 14:29:17 crc kubenswrapper[4805]: I1203 14:29:17.224714 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-7094-account-create-update-c7dm6" event={"ID":"26a760ac-b299-4492-ab01-95db5e14ac97","Type":"ContainerStarted","Data":"32d1093218da9d59781f62d4dc392d7e3aea85902b14ff0fc0eec67fef15e41f"} Dec 03 14:29:17 crc kubenswrapper[4805]: I1203 14:29:17.226048 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-de85-account-create-update-8gf95" event={"ID":"8427fb09-203a-4093-807c-9f03061b6e07","Type":"ContainerStarted","Data":"c0fcd01234e8f2624b2874a29e54a0ea991edb93d1437ec645568350ba85ba41"} Dec 03 14:29:17 crc kubenswrapper[4805]: I1203 14:29:17.227531 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-create-tbhgg" event={"ID":"3e5d8271-48f9-4bc1-b35c-a37f919eefdd","Type":"ContainerStarted","Data":"f68520dc10fadc7f524850a3a3cf1ab1b131a291803b74e102024d1c007707f7"} Dec 03 14:29:17 crc kubenswrapper[4805]: I1203 14:29:17.229117 4805 generic.go:334] "Generic (PLEG): container finished" podID="7895c06e-d5d7-4e5b-9f56-442cabf914bc" containerID="7d87783d81055065efdd933bb95570fe72fbe1fbdce5e19e32bf1b28e8ae8f37" exitCode=0 Dec 03 14:29:17 crc kubenswrapper[4805]: I1203 14:29:17.229167 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-32eb-account-create-update-jmj7x" event={"ID":"7895c06e-d5d7-4e5b-9f56-442cabf914bc","Type":"ContainerDied","Data":"7d87783d81055065efdd933bb95570fe72fbe1fbdce5e19e32bf1b28e8ae8f37"} Dec 03 14:29:17 crc kubenswrapper[4805]: I1203 14:29:17.233287 4805 generic.go:334] "Generic (PLEG): container finished" podID="4c56e040-47cb-4e7e-86d4-613e09cc3243" containerID="802333011ddf70752382747945a287b572296e7e1e7e8075bb0350869fd7a9d8" exitCode=0 Dec 03 14:29:17 crc kubenswrapper[4805]: I1203 14:29:17.233319 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-create-rrmbh" event={"ID":"4c56e040-47cb-4e7e-86d4-613e09cc3243","Type":"ContainerDied","Data":"802333011ddf70752382747945a287b572296e7e1e7e8075bb0350869fd7a9d8"} Dec 03 14:29:17 crc kubenswrapper[4805]: I1203 14:29:17.937066 4805 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-698758b865-l46rr" Dec 03 14:29:18 crc kubenswrapper[4805]: I1203 14:29:18.003764 4805 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-57d769cc4f-fthjg"] Dec 03 14:29:18 crc kubenswrapper[4805]: I1203 14:29:18.004183 4805 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-57d769cc4f-fthjg" podUID="2af0398e-b188-49ec-ae9e-391cb1122753" containerName="dnsmasq-dns" containerID="cri-o://b704ee8929e1278fce1a4d68a59d5b34da60bcace6cc40b496d450dce3b2ff98" gracePeriod=10 Dec 03 14:29:18 crc kubenswrapper[4805]: I1203 14:29:18.260718 4805 generic.go:334] "Generic (PLEG): container finished" podID="26a760ac-b299-4492-ab01-95db5e14ac97" containerID="cac18d8710a21e00b23f7447c31eb2589e1fed3f4a88d556913bbe54a0d5d540" exitCode=0 Dec 03 14:29:18 crc kubenswrapper[4805]: I1203 14:29:18.260983 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-7094-account-create-update-c7dm6" event={"ID":"26a760ac-b299-4492-ab01-95db5e14ac97","Type":"ContainerDied","Data":"cac18d8710a21e00b23f7447c31eb2589e1fed3f4a88d556913bbe54a0d5d540"} Dec 03 14:29:18 crc kubenswrapper[4805]: I1203 14:29:18.262790 4805 generic.go:334] "Generic (PLEG): container finished" podID="8427fb09-203a-4093-807c-9f03061b6e07" containerID="cd504bae43e97694691ee6b43212fd15ddf1c14e98f67ec53f5407f32e37e20e" exitCode=0 Dec 03 14:29:18 crc kubenswrapper[4805]: I1203 14:29:18.262830 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-de85-account-create-update-8gf95" event={"ID":"8427fb09-203a-4093-807c-9f03061b6e07","Type":"ContainerDied","Data":"cd504bae43e97694691ee6b43212fd15ddf1c14e98f67ec53f5407f32e37e20e"} Dec 03 14:29:18 crc kubenswrapper[4805]: I1203 14:29:18.264108 4805 generic.go:334] "Generic (PLEG): container finished" podID="3e5d8271-48f9-4bc1-b35c-a37f919eefdd" containerID="3f40241eb2521d3baeb79026fd9c58f88640f341c2ed7f60275e5448221ebc87" exitCode=0 Dec 03 14:29:18 crc kubenswrapper[4805]: I1203 14:29:18.264146 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-create-tbhgg" event={"ID":"3e5d8271-48f9-4bc1-b35c-a37f919eefdd","Type":"ContainerDied","Data":"3f40241eb2521d3baeb79026fd9c58f88640f341c2ed7f60275e5448221ebc87"} Dec 03 14:29:18 crc kubenswrapper[4805]: I1203 14:29:18.265393 4805 generic.go:334] "Generic (PLEG): container finished" podID="2af0398e-b188-49ec-ae9e-391cb1122753" containerID="b704ee8929e1278fce1a4d68a59d5b34da60bcace6cc40b496d450dce3b2ff98" exitCode=0 Dec 03 14:29:18 crc kubenswrapper[4805]: I1203 14:29:18.265431 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-57d769cc4f-fthjg" event={"ID":"2af0398e-b188-49ec-ae9e-391cb1122753","Type":"ContainerDied","Data":"b704ee8929e1278fce1a4d68a59d5b34da60bcace6cc40b496d450dce3b2ff98"} Dec 03 14:29:18 crc kubenswrapper[4805]: I1203 14:29:18.490279 4805 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-57d769cc4f-fthjg" Dec 03 14:29:18 crc kubenswrapper[4805]: I1203 14:29:18.593932 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2af0398e-b188-49ec-ae9e-391cb1122753-config\") pod \"2af0398e-b188-49ec-ae9e-391cb1122753\" (UID: \"2af0398e-b188-49ec-ae9e-391cb1122753\") " Dec 03 14:29:18 crc kubenswrapper[4805]: I1203 14:29:18.594069 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8jl2l\" (UniqueName: \"kubernetes.io/projected/2af0398e-b188-49ec-ae9e-391cb1122753-kube-api-access-8jl2l\") pod \"2af0398e-b188-49ec-ae9e-391cb1122753\" (UID: \"2af0398e-b188-49ec-ae9e-391cb1122753\") " Dec 03 14:29:18 crc kubenswrapper[4805]: I1203 14:29:18.594217 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/2af0398e-b188-49ec-ae9e-391cb1122753-dns-svc\") pod \"2af0398e-b188-49ec-ae9e-391cb1122753\" (UID: \"2af0398e-b188-49ec-ae9e-391cb1122753\") " Dec 03 14:29:18 crc kubenswrapper[4805]: I1203 14:29:18.622808 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2af0398e-b188-49ec-ae9e-391cb1122753-kube-api-access-8jl2l" (OuterVolumeSpecName: "kube-api-access-8jl2l") pod "2af0398e-b188-49ec-ae9e-391cb1122753" (UID: "2af0398e-b188-49ec-ae9e-391cb1122753"). InnerVolumeSpecName "kube-api-access-8jl2l". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 14:29:18 crc kubenswrapper[4805]: I1203 14:29:18.663668 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2af0398e-b188-49ec-ae9e-391cb1122753-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "2af0398e-b188-49ec-ae9e-391cb1122753" (UID: "2af0398e-b188-49ec-ae9e-391cb1122753"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 14:29:18 crc kubenswrapper[4805]: I1203 14:29:18.695591 4805 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/2af0398e-b188-49ec-ae9e-391cb1122753-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 03 14:29:18 crc kubenswrapper[4805]: I1203 14:29:18.695627 4805 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8jl2l\" (UniqueName: \"kubernetes.io/projected/2af0398e-b188-49ec-ae9e-391cb1122753-kube-api-access-8jl2l\") on node \"crc\" DevicePath \"\"" Dec 03 14:29:18 crc kubenswrapper[4805]: I1203 14:29:18.697518 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2af0398e-b188-49ec-ae9e-391cb1122753-config" (OuterVolumeSpecName: "config") pod "2af0398e-b188-49ec-ae9e-391cb1122753" (UID: "2af0398e-b188-49ec-ae9e-391cb1122753"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 14:29:18 crc kubenswrapper[4805]: I1203 14:29:18.722530 4805 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-create-rrmbh" Dec 03 14:29:18 crc kubenswrapper[4805]: I1203 14:29:18.729748 4805 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-create-bbs5s" Dec 03 14:29:18 crc kubenswrapper[4805]: I1203 14:29:18.747553 4805 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-32eb-account-create-update-jmj7x" Dec 03 14:29:18 crc kubenswrapper[4805]: I1203 14:29:18.797500 4805 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2af0398e-b188-49ec-ae9e-391cb1122753-config\") on node \"crc\" DevicePath \"\"" Dec 03 14:29:18 crc kubenswrapper[4805]: I1203 14:29:18.898820 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/7895c06e-d5d7-4e5b-9f56-442cabf914bc-operator-scripts\") pod \"7895c06e-d5d7-4e5b-9f56-442cabf914bc\" (UID: \"7895c06e-d5d7-4e5b-9f56-442cabf914bc\") " Dec 03 14:29:18 crc kubenswrapper[4805]: I1203 14:29:18.899187 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-sxpbf\" (UniqueName: \"kubernetes.io/projected/4c56e040-47cb-4e7e-86d4-613e09cc3243-kube-api-access-sxpbf\") pod \"4c56e040-47cb-4e7e-86d4-613e09cc3243\" (UID: \"4c56e040-47cb-4e7e-86d4-613e09cc3243\") " Dec 03 14:29:18 crc kubenswrapper[4805]: I1203 14:29:18.899214 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/4c56e040-47cb-4e7e-86d4-613e09cc3243-operator-scripts\") pod \"4c56e040-47cb-4e7e-86d4-613e09cc3243\" (UID: \"4c56e040-47cb-4e7e-86d4-613e09cc3243\") " Dec 03 14:29:18 crc kubenswrapper[4805]: I1203 14:29:18.899259 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lz98j\" (UniqueName: \"kubernetes.io/projected/7895c06e-d5d7-4e5b-9f56-442cabf914bc-kube-api-access-lz98j\") pod \"7895c06e-d5d7-4e5b-9f56-442cabf914bc\" (UID: \"7895c06e-d5d7-4e5b-9f56-442cabf914bc\") " Dec 03 14:29:18 crc kubenswrapper[4805]: I1203 14:29:18.899285 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/a43c2a2a-291b-4422-9c06-cb448b722758-operator-scripts\") pod \"a43c2a2a-291b-4422-9c06-cb448b722758\" (UID: \"a43c2a2a-291b-4422-9c06-cb448b722758\") " Dec 03 14:29:18 crc kubenswrapper[4805]: I1203 14:29:18.899333 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fz7fn\" (UniqueName: \"kubernetes.io/projected/a43c2a2a-291b-4422-9c06-cb448b722758-kube-api-access-fz7fn\") pod \"a43c2a2a-291b-4422-9c06-cb448b722758\" (UID: \"a43c2a2a-291b-4422-9c06-cb448b722758\") " Dec 03 14:29:18 crc kubenswrapper[4805]: I1203 14:29:18.901984 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a43c2a2a-291b-4422-9c06-cb448b722758-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "a43c2a2a-291b-4422-9c06-cb448b722758" (UID: "a43c2a2a-291b-4422-9c06-cb448b722758"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 14:29:18 crc kubenswrapper[4805]: I1203 14:29:18.902249 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4c56e040-47cb-4e7e-86d4-613e09cc3243-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "4c56e040-47cb-4e7e-86d4-613e09cc3243" (UID: "4c56e040-47cb-4e7e-86d4-613e09cc3243"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 14:29:18 crc kubenswrapper[4805]: I1203 14:29:18.902528 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7895c06e-d5d7-4e5b-9f56-442cabf914bc-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "7895c06e-d5d7-4e5b-9f56-442cabf914bc" (UID: "7895c06e-d5d7-4e5b-9f56-442cabf914bc"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 14:29:18 crc kubenswrapper[4805]: I1203 14:29:18.904488 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a43c2a2a-291b-4422-9c06-cb448b722758-kube-api-access-fz7fn" (OuterVolumeSpecName: "kube-api-access-fz7fn") pod "a43c2a2a-291b-4422-9c06-cb448b722758" (UID: "a43c2a2a-291b-4422-9c06-cb448b722758"). InnerVolumeSpecName "kube-api-access-fz7fn". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 14:29:18 crc kubenswrapper[4805]: I1203 14:29:18.905235 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7895c06e-d5d7-4e5b-9f56-442cabf914bc-kube-api-access-lz98j" (OuterVolumeSpecName: "kube-api-access-lz98j") pod "7895c06e-d5d7-4e5b-9f56-442cabf914bc" (UID: "7895c06e-d5d7-4e5b-9f56-442cabf914bc"). InnerVolumeSpecName "kube-api-access-lz98j". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 14:29:18 crc kubenswrapper[4805]: I1203 14:29:18.906051 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4c56e040-47cb-4e7e-86d4-613e09cc3243-kube-api-access-sxpbf" (OuterVolumeSpecName: "kube-api-access-sxpbf") pod "4c56e040-47cb-4e7e-86d4-613e09cc3243" (UID: "4c56e040-47cb-4e7e-86d4-613e09cc3243"). InnerVolumeSpecName "kube-api-access-sxpbf". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 14:29:18 crc kubenswrapper[4805]: I1203 14:29:18.973803 4805 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovn-northd-0" Dec 03 14:29:19 crc kubenswrapper[4805]: I1203 14:29:19.002242 4805 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/7895c06e-d5d7-4e5b-9f56-442cabf914bc-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 03 14:29:19 crc kubenswrapper[4805]: I1203 14:29:19.002314 4805 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-sxpbf\" (UniqueName: \"kubernetes.io/projected/4c56e040-47cb-4e7e-86d4-613e09cc3243-kube-api-access-sxpbf\") on node \"crc\" DevicePath \"\"" Dec 03 14:29:19 crc kubenswrapper[4805]: I1203 14:29:19.002332 4805 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/4c56e040-47cb-4e7e-86d4-613e09cc3243-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 03 14:29:19 crc kubenswrapper[4805]: I1203 14:29:19.002345 4805 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lz98j\" (UniqueName: \"kubernetes.io/projected/7895c06e-d5d7-4e5b-9f56-442cabf914bc-kube-api-access-lz98j\") on node \"crc\" DevicePath \"\"" Dec 03 14:29:19 crc kubenswrapper[4805]: I1203 14:29:19.002358 4805 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/a43c2a2a-291b-4422-9c06-cb448b722758-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 03 14:29:19 crc kubenswrapper[4805]: I1203 14:29:19.002371 4805 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fz7fn\" (UniqueName: \"kubernetes.io/projected/a43c2a2a-291b-4422-9c06-cb448b722758-kube-api-access-fz7fn\") on node \"crc\" DevicePath \"\"" Dec 03 14:29:19 crc kubenswrapper[4805]: I1203 14:29:19.273378 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-32eb-account-create-update-jmj7x" event={"ID":"7895c06e-d5d7-4e5b-9f56-442cabf914bc","Type":"ContainerDied","Data":"f8b849472536bd208e0c41f9458d0339551a80b92d85280633b338345be2c3ca"} Dec 03 14:29:19 crc kubenswrapper[4805]: I1203 14:29:19.273420 4805 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="f8b849472536bd208e0c41f9458d0339551a80b92d85280633b338345be2c3ca" Dec 03 14:29:19 crc kubenswrapper[4805]: I1203 14:29:19.273484 4805 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-32eb-account-create-update-jmj7x" Dec 03 14:29:19 crc kubenswrapper[4805]: I1203 14:29:19.284115 4805 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-create-rrmbh" Dec 03 14:29:19 crc kubenswrapper[4805]: I1203 14:29:19.284095 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-create-rrmbh" event={"ID":"4c56e040-47cb-4e7e-86d4-613e09cc3243","Type":"ContainerDied","Data":"fdfb88f20dcd2c0b764484a6fd63cfda7012223237f0d7e3c1ce139874daf60d"} Dec 03 14:29:19 crc kubenswrapper[4805]: I1203 14:29:19.284291 4805 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="fdfb88f20dcd2c0b764484a6fd63cfda7012223237f0d7e3c1ce139874daf60d" Dec 03 14:29:19 crc kubenswrapper[4805]: I1203 14:29:19.286695 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-create-bbs5s" event={"ID":"a43c2a2a-291b-4422-9c06-cb448b722758","Type":"ContainerDied","Data":"276e8f06108c78e6cb04674a839349031b4f7a30abbd7a2dbbb52637f50c4090"} Dec 03 14:29:19 crc kubenswrapper[4805]: I1203 14:29:19.286746 4805 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="276e8f06108c78e6cb04674a839349031b4f7a30abbd7a2dbbb52637f50c4090" Dec 03 14:29:19 crc kubenswrapper[4805]: I1203 14:29:19.286864 4805 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-create-bbs5s" Dec 03 14:29:19 crc kubenswrapper[4805]: I1203 14:29:19.288958 4805 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-57d769cc4f-fthjg" Dec 03 14:29:19 crc kubenswrapper[4805]: I1203 14:29:19.288968 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-57d769cc4f-fthjg" event={"ID":"2af0398e-b188-49ec-ae9e-391cb1122753","Type":"ContainerDied","Data":"f674510e91a8b892788668806a350ee018c948cdc7b18fc4d3d00d606b500b33"} Dec 03 14:29:19 crc kubenswrapper[4805]: I1203 14:29:19.289018 4805 scope.go:117] "RemoveContainer" containerID="b704ee8929e1278fce1a4d68a59d5b34da60bcace6cc40b496d450dce3b2ff98" Dec 03 14:29:19 crc kubenswrapper[4805]: I1203 14:29:19.316806 4805 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-57d769cc4f-fthjg"] Dec 03 14:29:19 crc kubenswrapper[4805]: I1203 14:29:19.324487 4805 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-57d769cc4f-fthjg"] Dec 03 14:29:19 crc kubenswrapper[4805]: I1203 14:29:19.329140 4805 scope.go:117] "RemoveContainer" containerID="8b70115a80a67bf0ded6b6883dda9ff0036d3720338a2ef80b70fcae73736348" Dec 03 14:29:19 crc kubenswrapper[4805]: I1203 14:29:19.831863 4805 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-7094-account-create-update-c7dm6" Dec 03 14:29:19 crc kubenswrapper[4805]: I1203 14:29:19.976190 4805 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-create-tbhgg" Dec 03 14:29:19 crc kubenswrapper[4805]: I1203 14:29:19.982308 4805 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-de85-account-create-update-8gf95" Dec 03 14:29:20 crc kubenswrapper[4805]: I1203 14:29:20.031283 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/26a760ac-b299-4492-ab01-95db5e14ac97-operator-scripts\") pod \"26a760ac-b299-4492-ab01-95db5e14ac97\" (UID: \"26a760ac-b299-4492-ab01-95db5e14ac97\") " Dec 03 14:29:20 crc kubenswrapper[4805]: I1203 14:29:20.031431 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-v85rv\" (UniqueName: \"kubernetes.io/projected/26a760ac-b299-4492-ab01-95db5e14ac97-kube-api-access-v85rv\") pod \"26a760ac-b299-4492-ab01-95db5e14ac97\" (UID: \"26a760ac-b299-4492-ab01-95db5e14ac97\") " Dec 03 14:29:20 crc kubenswrapper[4805]: I1203 14:29:20.033953 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/26a760ac-b299-4492-ab01-95db5e14ac97-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "26a760ac-b299-4492-ab01-95db5e14ac97" (UID: "26a760ac-b299-4492-ab01-95db5e14ac97"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 14:29:20 crc kubenswrapper[4805]: I1203 14:29:20.043538 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/26a760ac-b299-4492-ab01-95db5e14ac97-kube-api-access-v85rv" (OuterVolumeSpecName: "kube-api-access-v85rv") pod "26a760ac-b299-4492-ab01-95db5e14ac97" (UID: "26a760ac-b299-4492-ab01-95db5e14ac97"). InnerVolumeSpecName "kube-api-access-v85rv". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 14:29:20 crc kubenswrapper[4805]: I1203 14:29:20.132626 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/8427fb09-203a-4093-807c-9f03061b6e07-operator-scripts\") pod \"8427fb09-203a-4093-807c-9f03061b6e07\" (UID: \"8427fb09-203a-4093-807c-9f03061b6e07\") " Dec 03 14:29:20 crc kubenswrapper[4805]: I1203 14:29:20.132870 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vcrp5\" (UniqueName: \"kubernetes.io/projected/8427fb09-203a-4093-807c-9f03061b6e07-kube-api-access-vcrp5\") pod \"8427fb09-203a-4093-807c-9f03061b6e07\" (UID: \"8427fb09-203a-4093-807c-9f03061b6e07\") " Dec 03 14:29:20 crc kubenswrapper[4805]: I1203 14:29:20.132968 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-r4mpz\" (UniqueName: \"kubernetes.io/projected/3e5d8271-48f9-4bc1-b35c-a37f919eefdd-kube-api-access-r4mpz\") pod \"3e5d8271-48f9-4bc1-b35c-a37f919eefdd\" (UID: \"3e5d8271-48f9-4bc1-b35c-a37f919eefdd\") " Dec 03 14:29:20 crc kubenswrapper[4805]: I1203 14:29:20.133008 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/3e5d8271-48f9-4bc1-b35c-a37f919eefdd-operator-scripts\") pod \"3e5d8271-48f9-4bc1-b35c-a37f919eefdd\" (UID: \"3e5d8271-48f9-4bc1-b35c-a37f919eefdd\") " Dec 03 14:29:20 crc kubenswrapper[4805]: I1203 14:29:20.133265 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8427fb09-203a-4093-807c-9f03061b6e07-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "8427fb09-203a-4093-807c-9f03061b6e07" (UID: "8427fb09-203a-4093-807c-9f03061b6e07"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 14:29:20 crc kubenswrapper[4805]: I1203 14:29:20.133710 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3e5d8271-48f9-4bc1-b35c-a37f919eefdd-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "3e5d8271-48f9-4bc1-b35c-a37f919eefdd" (UID: "3e5d8271-48f9-4bc1-b35c-a37f919eefdd"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 14:29:20 crc kubenswrapper[4805]: I1203 14:29:20.133984 4805 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/8427fb09-203a-4093-807c-9f03061b6e07-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 03 14:29:20 crc kubenswrapper[4805]: I1203 14:29:20.134011 4805 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-v85rv\" (UniqueName: \"kubernetes.io/projected/26a760ac-b299-4492-ab01-95db5e14ac97-kube-api-access-v85rv\") on node \"crc\" DevicePath \"\"" Dec 03 14:29:20 crc kubenswrapper[4805]: I1203 14:29:20.134023 4805 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/26a760ac-b299-4492-ab01-95db5e14ac97-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 03 14:29:20 crc kubenswrapper[4805]: I1203 14:29:20.134034 4805 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/3e5d8271-48f9-4bc1-b35c-a37f919eefdd-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 03 14:29:20 crc kubenswrapper[4805]: I1203 14:29:20.136939 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3e5d8271-48f9-4bc1-b35c-a37f919eefdd-kube-api-access-r4mpz" (OuterVolumeSpecName: "kube-api-access-r4mpz") pod "3e5d8271-48f9-4bc1-b35c-a37f919eefdd" (UID: "3e5d8271-48f9-4bc1-b35c-a37f919eefdd"). InnerVolumeSpecName "kube-api-access-r4mpz". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 14:29:20 crc kubenswrapper[4805]: I1203 14:29:20.137141 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8427fb09-203a-4093-807c-9f03061b6e07-kube-api-access-vcrp5" (OuterVolumeSpecName: "kube-api-access-vcrp5") pod "8427fb09-203a-4093-807c-9f03061b6e07" (UID: "8427fb09-203a-4093-807c-9f03061b6e07"). InnerVolumeSpecName "kube-api-access-vcrp5". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 14:29:20 crc kubenswrapper[4805]: I1203 14:29:20.235619 4805 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vcrp5\" (UniqueName: \"kubernetes.io/projected/8427fb09-203a-4093-807c-9f03061b6e07-kube-api-access-vcrp5\") on node \"crc\" DevicePath \"\"" Dec 03 14:29:20 crc kubenswrapper[4805]: I1203 14:29:20.235897 4805 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-r4mpz\" (UniqueName: \"kubernetes.io/projected/3e5d8271-48f9-4bc1-b35c-a37f919eefdd-kube-api-access-r4mpz\") on node \"crc\" DevicePath \"\"" Dec 03 14:29:20 crc kubenswrapper[4805]: I1203 14:29:20.299444 4805 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-7094-account-create-update-c7dm6" Dec 03 14:29:20 crc kubenswrapper[4805]: I1203 14:29:20.299379 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-7094-account-create-update-c7dm6" event={"ID":"26a760ac-b299-4492-ab01-95db5e14ac97","Type":"ContainerDied","Data":"32d1093218da9d59781f62d4dc392d7e3aea85902b14ff0fc0eec67fef15e41f"} Dec 03 14:29:20 crc kubenswrapper[4805]: I1203 14:29:20.300759 4805 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="32d1093218da9d59781f62d4dc392d7e3aea85902b14ff0fc0eec67fef15e41f" Dec 03 14:29:20 crc kubenswrapper[4805]: I1203 14:29:20.301099 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-de85-account-create-update-8gf95" event={"ID":"8427fb09-203a-4093-807c-9f03061b6e07","Type":"ContainerDied","Data":"c0fcd01234e8f2624b2874a29e54a0ea991edb93d1437ec645568350ba85ba41"} Dec 03 14:29:20 crc kubenswrapper[4805]: I1203 14:29:20.301131 4805 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="c0fcd01234e8f2624b2874a29e54a0ea991edb93d1437ec645568350ba85ba41" Dec 03 14:29:20 crc kubenswrapper[4805]: I1203 14:29:20.301213 4805 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-de85-account-create-update-8gf95" Dec 03 14:29:20 crc kubenswrapper[4805]: I1203 14:29:20.302781 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-create-tbhgg" event={"ID":"3e5d8271-48f9-4bc1-b35c-a37f919eefdd","Type":"ContainerDied","Data":"f68520dc10fadc7f524850a3a3cf1ab1b131a291803b74e102024d1c007707f7"} Dec 03 14:29:20 crc kubenswrapper[4805]: I1203 14:29:20.302824 4805 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-create-tbhgg" Dec 03 14:29:20 crc kubenswrapper[4805]: I1203 14:29:20.302835 4805 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="f68520dc10fadc7f524850a3a3cf1ab1b131a291803b74e102024d1c007707f7" Dec 03 14:29:20 crc kubenswrapper[4805]: I1203 14:29:20.709288 4805 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2af0398e-b188-49ec-ae9e-391cb1122753" path="/var/lib/kubelet/pods/2af0398e-b188-49ec-ae9e-391cb1122753/volumes" Dec 03 14:29:22 crc kubenswrapper[4805]: I1203 14:29:22.319710 4805 generic.go:334] "Generic (PLEG): container finished" podID="acfd02d5-fb0a-4427-95e0-4c9f956ea137" containerID="e4fc3ae321cc76c9f7df0935f144238ddea8a8483203f25ff4ea0b3906124127" exitCode=0 Dec 03 14:29:22 crc kubenswrapper[4805]: I1203 14:29:22.320015 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-ring-rebalance-h55gd" event={"ID":"acfd02d5-fb0a-4427-95e0-4c9f956ea137","Type":"ContainerDied","Data":"e4fc3ae321cc76c9f7df0935f144238ddea8a8483203f25ff4ea0b3906124127"} Dec 03 14:29:23 crc kubenswrapper[4805]: I1203 14:29:23.674731 4805 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-h55gd" Dec 03 14:29:23 crc kubenswrapper[4805]: I1203 14:29:23.798404 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/acfd02d5-fb0a-4427-95e0-4c9f956ea137-scripts\") pod \"acfd02d5-fb0a-4427-95e0-4c9f956ea137\" (UID: \"acfd02d5-fb0a-4427-95e0-4c9f956ea137\") " Dec 03 14:29:23 crc kubenswrapper[4805]: I1203 14:29:23.798604 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/acfd02d5-fb0a-4427-95e0-4c9f956ea137-etc-swift\") pod \"acfd02d5-fb0a-4427-95e0-4c9f956ea137\" (UID: \"acfd02d5-fb0a-4427-95e0-4c9f956ea137\") " Dec 03 14:29:23 crc kubenswrapper[4805]: I1203 14:29:23.798637 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lx4q5\" (UniqueName: \"kubernetes.io/projected/acfd02d5-fb0a-4427-95e0-4c9f956ea137-kube-api-access-lx4q5\") pod \"acfd02d5-fb0a-4427-95e0-4c9f956ea137\" (UID: \"acfd02d5-fb0a-4427-95e0-4c9f956ea137\") " Dec 03 14:29:23 crc kubenswrapper[4805]: I1203 14:29:23.798672 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/acfd02d5-fb0a-4427-95e0-4c9f956ea137-dispersionconf\") pod \"acfd02d5-fb0a-4427-95e0-4c9f956ea137\" (UID: \"acfd02d5-fb0a-4427-95e0-4c9f956ea137\") " Dec 03 14:29:23 crc kubenswrapper[4805]: I1203 14:29:23.798731 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/acfd02d5-fb0a-4427-95e0-4c9f956ea137-combined-ca-bundle\") pod \"acfd02d5-fb0a-4427-95e0-4c9f956ea137\" (UID: \"acfd02d5-fb0a-4427-95e0-4c9f956ea137\") " Dec 03 14:29:23 crc kubenswrapper[4805]: I1203 14:29:23.798765 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/acfd02d5-fb0a-4427-95e0-4c9f956ea137-swiftconf\") pod \"acfd02d5-fb0a-4427-95e0-4c9f956ea137\" (UID: \"acfd02d5-fb0a-4427-95e0-4c9f956ea137\") " Dec 03 14:29:23 crc kubenswrapper[4805]: I1203 14:29:23.798864 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/acfd02d5-fb0a-4427-95e0-4c9f956ea137-ring-data-devices\") pod \"acfd02d5-fb0a-4427-95e0-4c9f956ea137\" (UID: \"acfd02d5-fb0a-4427-95e0-4c9f956ea137\") " Dec 03 14:29:23 crc kubenswrapper[4805]: I1203 14:29:23.800257 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/acfd02d5-fb0a-4427-95e0-4c9f956ea137-ring-data-devices" (OuterVolumeSpecName: "ring-data-devices") pod "acfd02d5-fb0a-4427-95e0-4c9f956ea137" (UID: "acfd02d5-fb0a-4427-95e0-4c9f956ea137"). InnerVolumeSpecName "ring-data-devices". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 14:29:23 crc kubenswrapper[4805]: I1203 14:29:23.800610 4805 reconciler_common.go:293] "Volume detached for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/acfd02d5-fb0a-4427-95e0-4c9f956ea137-ring-data-devices\") on node \"crc\" DevicePath \"\"" Dec 03 14:29:23 crc kubenswrapper[4805]: I1203 14:29:23.812082 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/acfd02d5-fb0a-4427-95e0-4c9f956ea137-kube-api-access-lx4q5" (OuterVolumeSpecName: "kube-api-access-lx4q5") pod "acfd02d5-fb0a-4427-95e0-4c9f956ea137" (UID: "acfd02d5-fb0a-4427-95e0-4c9f956ea137"). InnerVolumeSpecName "kube-api-access-lx4q5". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 14:29:23 crc kubenswrapper[4805]: I1203 14:29:23.815006 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/acfd02d5-fb0a-4427-95e0-4c9f956ea137-dispersionconf" (OuterVolumeSpecName: "dispersionconf") pod "acfd02d5-fb0a-4427-95e0-4c9f956ea137" (UID: "acfd02d5-fb0a-4427-95e0-4c9f956ea137"). InnerVolumeSpecName "dispersionconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 14:29:23 crc kubenswrapper[4805]: I1203 14:29:23.821917 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/acfd02d5-fb0a-4427-95e0-4c9f956ea137-scripts" (OuterVolumeSpecName: "scripts") pod "acfd02d5-fb0a-4427-95e0-4c9f956ea137" (UID: "acfd02d5-fb0a-4427-95e0-4c9f956ea137"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 14:29:23 crc kubenswrapper[4805]: I1203 14:29:23.827766 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/acfd02d5-fb0a-4427-95e0-4c9f956ea137-swiftconf" (OuterVolumeSpecName: "swiftconf") pod "acfd02d5-fb0a-4427-95e0-4c9f956ea137" (UID: "acfd02d5-fb0a-4427-95e0-4c9f956ea137"). InnerVolumeSpecName "swiftconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 14:29:23 crc kubenswrapper[4805]: I1203 14:29:23.833119 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/acfd02d5-fb0a-4427-95e0-4c9f956ea137-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "acfd02d5-fb0a-4427-95e0-4c9f956ea137" (UID: "acfd02d5-fb0a-4427-95e0-4c9f956ea137"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 14:29:23 crc kubenswrapper[4805]: I1203 14:29:23.901307 4805 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/acfd02d5-fb0a-4427-95e0-4c9f956ea137-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 14:29:23 crc kubenswrapper[4805]: I1203 14:29:23.901362 4805 reconciler_common.go:293] "Volume detached for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/acfd02d5-fb0a-4427-95e0-4c9f956ea137-swiftconf\") on node \"crc\" DevicePath \"\"" Dec 03 14:29:23 crc kubenswrapper[4805]: I1203 14:29:23.901374 4805 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/acfd02d5-fb0a-4427-95e0-4c9f956ea137-scripts\") on node \"crc\" DevicePath \"\"" Dec 03 14:29:23 crc kubenswrapper[4805]: I1203 14:29:23.901386 4805 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lx4q5\" (UniqueName: \"kubernetes.io/projected/acfd02d5-fb0a-4427-95e0-4c9f956ea137-kube-api-access-lx4q5\") on node \"crc\" DevicePath \"\"" Dec 03 14:29:23 crc kubenswrapper[4805]: I1203 14:29:23.901400 4805 reconciler_common.go:293] "Volume detached for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/acfd02d5-fb0a-4427-95e0-4c9f956ea137-dispersionconf\") on node \"crc\" DevicePath \"\"" Dec 03 14:29:24 crc kubenswrapper[4805]: I1203 14:29:24.258187 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/acfd02d5-fb0a-4427-95e0-4c9f956ea137-etc-swift" (OuterVolumeSpecName: "etc-swift") pod "acfd02d5-fb0a-4427-95e0-4c9f956ea137" (UID: "acfd02d5-fb0a-4427-95e0-4c9f956ea137"). InnerVolumeSpecName "etc-swift". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 14:29:24 crc kubenswrapper[4805]: I1203 14:29:24.306865 4805 reconciler_common.go:293] "Volume detached for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/acfd02d5-fb0a-4427-95e0-4c9f956ea137-etc-swift\") on node \"crc\" DevicePath \"\"" Dec 03 14:29:24 crc kubenswrapper[4805]: I1203 14:29:24.334642 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-ring-rebalance-h55gd" event={"ID":"acfd02d5-fb0a-4427-95e0-4c9f956ea137","Type":"ContainerDied","Data":"1bc90614ae492aad000272447edf23d5a1666c8a615ffcd8ddd4f55a33688327"} Dec 03 14:29:24 crc kubenswrapper[4805]: I1203 14:29:24.334684 4805 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="1bc90614ae492aad000272447edf23d5a1666c8a615ffcd8ddd4f55a33688327" Dec 03 14:29:24 crc kubenswrapper[4805]: I1203 14:29:24.334709 4805 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-h55gd" Dec 03 14:29:24 crc kubenswrapper[4805]: I1203 14:29:24.611657 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/4fbc8d7f-d47f-4d91-915b-504820aac08a-etc-swift\") pod \"swift-storage-0\" (UID: \"4fbc8d7f-d47f-4d91-915b-504820aac08a\") " pod="openstack/swift-storage-0" Dec 03 14:29:24 crc kubenswrapper[4805]: I1203 14:29:24.617647 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/4fbc8d7f-d47f-4d91-915b-504820aac08a-etc-swift\") pod \"swift-storage-0\" (UID: \"4fbc8d7f-d47f-4d91-915b-504820aac08a\") " pod="openstack/swift-storage-0" Dec 03 14:29:24 crc kubenswrapper[4805]: I1203 14:29:24.642159 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-storage-0" Dec 03 14:29:25 crc kubenswrapper[4805]: I1203 14:29:25.150832 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-storage-0"] Dec 03 14:29:25 crc kubenswrapper[4805]: W1203 14:29:25.151705 4805 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod4fbc8d7f_d47f_4d91_915b_504820aac08a.slice/crio-6477ff88bfec4c91fdf9ea41eff26166efceceec0b919d506d9016e74238e6b8 WatchSource:0}: Error finding container 6477ff88bfec4c91fdf9ea41eff26166efceceec0b919d506d9016e74238e6b8: Status 404 returned error can't find the container with id 6477ff88bfec4c91fdf9ea41eff26166efceceec0b919d506d9016e74238e6b8 Dec 03 14:29:25 crc kubenswrapper[4805]: I1203 14:29:25.341921 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"4fbc8d7f-d47f-4d91-915b-504820aac08a","Type":"ContainerStarted","Data":"6477ff88bfec4c91fdf9ea41eff26166efceceec0b919d506d9016e74238e6b8"} Dec 03 14:29:25 crc kubenswrapper[4805]: I1203 14:29:25.877664 4805 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-db-sync-27jrz"] Dec 03 14:29:25 crc kubenswrapper[4805]: E1203 14:29:25.878383 4805 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2af0398e-b188-49ec-ae9e-391cb1122753" containerName="init" Dec 03 14:29:25 crc kubenswrapper[4805]: I1203 14:29:25.878408 4805 state_mem.go:107] "Deleted CPUSet assignment" podUID="2af0398e-b188-49ec-ae9e-391cb1122753" containerName="init" Dec 03 14:29:25 crc kubenswrapper[4805]: E1203 14:29:25.878425 4805 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4c56e040-47cb-4e7e-86d4-613e09cc3243" containerName="mariadb-database-create" Dec 03 14:29:25 crc kubenswrapper[4805]: I1203 14:29:25.878434 4805 state_mem.go:107] "Deleted CPUSet assignment" podUID="4c56e040-47cb-4e7e-86d4-613e09cc3243" containerName="mariadb-database-create" Dec 03 14:29:25 crc kubenswrapper[4805]: E1203 14:29:25.878449 4805 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="acfd02d5-fb0a-4427-95e0-4c9f956ea137" containerName="swift-ring-rebalance" Dec 03 14:29:25 crc kubenswrapper[4805]: I1203 14:29:25.878458 4805 state_mem.go:107] "Deleted CPUSet assignment" podUID="acfd02d5-fb0a-4427-95e0-4c9f956ea137" containerName="swift-ring-rebalance" Dec 03 14:29:25 crc kubenswrapper[4805]: E1203 14:29:25.878470 4805 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7895c06e-d5d7-4e5b-9f56-442cabf914bc" containerName="mariadb-account-create-update" Dec 03 14:29:25 crc kubenswrapper[4805]: I1203 14:29:25.878476 4805 state_mem.go:107] "Deleted CPUSet assignment" podUID="7895c06e-d5d7-4e5b-9f56-442cabf914bc" containerName="mariadb-account-create-update" Dec 03 14:29:25 crc kubenswrapper[4805]: E1203 14:29:25.878493 4805 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="26a760ac-b299-4492-ab01-95db5e14ac97" containerName="mariadb-account-create-update" Dec 03 14:29:25 crc kubenswrapper[4805]: I1203 14:29:25.878501 4805 state_mem.go:107] "Deleted CPUSet assignment" podUID="26a760ac-b299-4492-ab01-95db5e14ac97" containerName="mariadb-account-create-update" Dec 03 14:29:25 crc kubenswrapper[4805]: E1203 14:29:25.878517 4805 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3e5d8271-48f9-4bc1-b35c-a37f919eefdd" containerName="mariadb-database-create" Dec 03 14:29:25 crc kubenswrapper[4805]: I1203 14:29:25.878525 4805 state_mem.go:107] "Deleted CPUSet assignment" podUID="3e5d8271-48f9-4bc1-b35c-a37f919eefdd" containerName="mariadb-database-create" Dec 03 14:29:25 crc kubenswrapper[4805]: E1203 14:29:25.878539 4805 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2af0398e-b188-49ec-ae9e-391cb1122753" containerName="dnsmasq-dns" Dec 03 14:29:25 crc kubenswrapper[4805]: I1203 14:29:25.878546 4805 state_mem.go:107] "Deleted CPUSet assignment" podUID="2af0398e-b188-49ec-ae9e-391cb1122753" containerName="dnsmasq-dns" Dec 03 14:29:25 crc kubenswrapper[4805]: E1203 14:29:25.878559 4805 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8427fb09-203a-4093-807c-9f03061b6e07" containerName="mariadb-account-create-update" Dec 03 14:29:25 crc kubenswrapper[4805]: I1203 14:29:25.878567 4805 state_mem.go:107] "Deleted CPUSet assignment" podUID="8427fb09-203a-4093-807c-9f03061b6e07" containerName="mariadb-account-create-update" Dec 03 14:29:25 crc kubenswrapper[4805]: E1203 14:29:25.878585 4805 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a43c2a2a-291b-4422-9c06-cb448b722758" containerName="mariadb-database-create" Dec 03 14:29:25 crc kubenswrapper[4805]: I1203 14:29:25.878593 4805 state_mem.go:107] "Deleted CPUSet assignment" podUID="a43c2a2a-291b-4422-9c06-cb448b722758" containerName="mariadb-database-create" Dec 03 14:29:25 crc kubenswrapper[4805]: I1203 14:29:25.878783 4805 memory_manager.go:354] "RemoveStaleState removing state" podUID="3e5d8271-48f9-4bc1-b35c-a37f919eefdd" containerName="mariadb-database-create" Dec 03 14:29:25 crc kubenswrapper[4805]: I1203 14:29:25.878801 4805 memory_manager.go:354] "RemoveStaleState removing state" podUID="2af0398e-b188-49ec-ae9e-391cb1122753" containerName="dnsmasq-dns" Dec 03 14:29:25 crc kubenswrapper[4805]: I1203 14:29:25.878814 4805 memory_manager.go:354] "RemoveStaleState removing state" podUID="8427fb09-203a-4093-807c-9f03061b6e07" containerName="mariadb-account-create-update" Dec 03 14:29:25 crc kubenswrapper[4805]: I1203 14:29:25.878826 4805 memory_manager.go:354] "RemoveStaleState removing state" podUID="a43c2a2a-291b-4422-9c06-cb448b722758" containerName="mariadb-database-create" Dec 03 14:29:25 crc kubenswrapper[4805]: I1203 14:29:25.878857 4805 memory_manager.go:354] "RemoveStaleState removing state" podUID="7895c06e-d5d7-4e5b-9f56-442cabf914bc" containerName="mariadb-account-create-update" Dec 03 14:29:25 crc kubenswrapper[4805]: I1203 14:29:25.878866 4805 memory_manager.go:354] "RemoveStaleState removing state" podUID="4c56e040-47cb-4e7e-86d4-613e09cc3243" containerName="mariadb-database-create" Dec 03 14:29:25 crc kubenswrapper[4805]: I1203 14:29:25.878876 4805 memory_manager.go:354] "RemoveStaleState removing state" podUID="acfd02d5-fb0a-4427-95e0-4c9f956ea137" containerName="swift-ring-rebalance" Dec 03 14:29:25 crc kubenswrapper[4805]: I1203 14:29:25.878901 4805 memory_manager.go:354] "RemoveStaleState removing state" podUID="26a760ac-b299-4492-ab01-95db5e14ac97" containerName="mariadb-account-create-update" Dec 03 14:29:25 crc kubenswrapper[4805]: I1203 14:29:25.879509 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-sync-27jrz" Dec 03 14:29:25 crc kubenswrapper[4805]: I1203 14:29:25.881856 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-config-data" Dec 03 14:29:25 crc kubenswrapper[4805]: I1203 14:29:25.882099 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-glance-dockercfg-rxb4p" Dec 03 14:29:25 crc kubenswrapper[4805]: I1203 14:29:25.886717 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-db-sync-27jrz"] Dec 03 14:29:25 crc kubenswrapper[4805]: I1203 14:29:25.929904 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fzss2\" (UniqueName: \"kubernetes.io/projected/285c21ec-cbb8-4cb7-b5c2-0b27656f7870-kube-api-access-fzss2\") pod \"glance-db-sync-27jrz\" (UID: \"285c21ec-cbb8-4cb7-b5c2-0b27656f7870\") " pod="openstack/glance-db-sync-27jrz" Dec 03 14:29:25 crc kubenswrapper[4805]: I1203 14:29:25.929944 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/285c21ec-cbb8-4cb7-b5c2-0b27656f7870-combined-ca-bundle\") pod \"glance-db-sync-27jrz\" (UID: \"285c21ec-cbb8-4cb7-b5c2-0b27656f7870\") " pod="openstack/glance-db-sync-27jrz" Dec 03 14:29:25 crc kubenswrapper[4805]: I1203 14:29:25.929994 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/285c21ec-cbb8-4cb7-b5c2-0b27656f7870-config-data\") pod \"glance-db-sync-27jrz\" (UID: \"285c21ec-cbb8-4cb7-b5c2-0b27656f7870\") " pod="openstack/glance-db-sync-27jrz" Dec 03 14:29:25 crc kubenswrapper[4805]: I1203 14:29:25.930057 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/285c21ec-cbb8-4cb7-b5c2-0b27656f7870-db-sync-config-data\") pod \"glance-db-sync-27jrz\" (UID: \"285c21ec-cbb8-4cb7-b5c2-0b27656f7870\") " pod="openstack/glance-db-sync-27jrz" Dec 03 14:29:26 crc kubenswrapper[4805]: I1203 14:29:26.031325 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/285c21ec-cbb8-4cb7-b5c2-0b27656f7870-db-sync-config-data\") pod \"glance-db-sync-27jrz\" (UID: \"285c21ec-cbb8-4cb7-b5c2-0b27656f7870\") " pod="openstack/glance-db-sync-27jrz" Dec 03 14:29:26 crc kubenswrapper[4805]: I1203 14:29:26.031453 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fzss2\" (UniqueName: \"kubernetes.io/projected/285c21ec-cbb8-4cb7-b5c2-0b27656f7870-kube-api-access-fzss2\") pod \"glance-db-sync-27jrz\" (UID: \"285c21ec-cbb8-4cb7-b5c2-0b27656f7870\") " pod="openstack/glance-db-sync-27jrz" Dec 03 14:29:26 crc kubenswrapper[4805]: I1203 14:29:26.031472 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/285c21ec-cbb8-4cb7-b5c2-0b27656f7870-combined-ca-bundle\") pod \"glance-db-sync-27jrz\" (UID: \"285c21ec-cbb8-4cb7-b5c2-0b27656f7870\") " pod="openstack/glance-db-sync-27jrz" Dec 03 14:29:26 crc kubenswrapper[4805]: I1203 14:29:26.031522 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/285c21ec-cbb8-4cb7-b5c2-0b27656f7870-config-data\") pod \"glance-db-sync-27jrz\" (UID: \"285c21ec-cbb8-4cb7-b5c2-0b27656f7870\") " pod="openstack/glance-db-sync-27jrz" Dec 03 14:29:26 crc kubenswrapper[4805]: I1203 14:29:26.037964 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/285c21ec-cbb8-4cb7-b5c2-0b27656f7870-combined-ca-bundle\") pod \"glance-db-sync-27jrz\" (UID: \"285c21ec-cbb8-4cb7-b5c2-0b27656f7870\") " pod="openstack/glance-db-sync-27jrz" Dec 03 14:29:26 crc kubenswrapper[4805]: I1203 14:29:26.038200 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/285c21ec-cbb8-4cb7-b5c2-0b27656f7870-db-sync-config-data\") pod \"glance-db-sync-27jrz\" (UID: \"285c21ec-cbb8-4cb7-b5c2-0b27656f7870\") " pod="openstack/glance-db-sync-27jrz" Dec 03 14:29:26 crc kubenswrapper[4805]: I1203 14:29:26.040997 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/285c21ec-cbb8-4cb7-b5c2-0b27656f7870-config-data\") pod \"glance-db-sync-27jrz\" (UID: \"285c21ec-cbb8-4cb7-b5c2-0b27656f7870\") " pod="openstack/glance-db-sync-27jrz" Dec 03 14:29:26 crc kubenswrapper[4805]: I1203 14:29:26.050300 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fzss2\" (UniqueName: \"kubernetes.io/projected/285c21ec-cbb8-4cb7-b5c2-0b27656f7870-kube-api-access-fzss2\") pod \"glance-db-sync-27jrz\" (UID: \"285c21ec-cbb8-4cb7-b5c2-0b27656f7870\") " pod="openstack/glance-db-sync-27jrz" Dec 03 14:29:26 crc kubenswrapper[4805]: I1203 14:29:26.207537 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-sync-27jrz" Dec 03 14:29:26 crc kubenswrapper[4805]: I1203 14:29:26.617057 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-db-sync-27jrz"] Dec 03 14:29:26 crc kubenswrapper[4805]: W1203 14:29:26.621366 4805 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod285c21ec_cbb8_4cb7_b5c2_0b27656f7870.slice/crio-b516ae9c15859b5e0423b2035fb3df20f22b2377fa4d008f7157d3153132cd8a WatchSource:0}: Error finding container b516ae9c15859b5e0423b2035fb3df20f22b2377fa4d008f7157d3153132cd8a: Status 404 returned error can't find the container with id b516ae9c15859b5e0423b2035fb3df20f22b2377fa4d008f7157d3153132cd8a Dec 03 14:29:27 crc kubenswrapper[4805]: I1203 14:29:27.360411 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-sync-27jrz" event={"ID":"285c21ec-cbb8-4cb7-b5c2-0b27656f7870","Type":"ContainerStarted","Data":"b516ae9c15859b5e0423b2035fb3df20f22b2377fa4d008f7157d3153132cd8a"} Dec 03 14:29:27 crc kubenswrapper[4805]: I1203 14:29:27.368029 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"4fbc8d7f-d47f-4d91-915b-504820aac08a","Type":"ContainerStarted","Data":"82e6434c7fabf37dd38ef757debdecb09ce5fd0754e5e43fa0e4f8af9cb1241a"} Dec 03 14:29:27 crc kubenswrapper[4805]: I1203 14:29:27.368073 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"4fbc8d7f-d47f-4d91-915b-504820aac08a","Type":"ContainerStarted","Data":"15c01a08f849492275d8ad16d92a32464b41b57dcead4b113faf4a1ba79fd366"} Dec 03 14:29:27 crc kubenswrapper[4805]: I1203 14:29:27.368086 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"4fbc8d7f-d47f-4d91-915b-504820aac08a","Type":"ContainerStarted","Data":"4155704657be6b22b994d660df659df756dec20eedee817ac9dced4d064f833a"} Dec 03 14:29:27 crc kubenswrapper[4805]: I1203 14:29:27.368098 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"4fbc8d7f-d47f-4d91-915b-504820aac08a","Type":"ContainerStarted","Data":"0a7adc48ac233b0763b9aa25717c7215b254646ab6e2e5bd670a353f3333da91"} Dec 03 14:29:30 crc kubenswrapper[4805]: I1203 14:29:30.529640 4805 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/ovn-controller-zhnwt" podUID="139abefe-12d4-4d8f-bdc6-43e71c818a65" containerName="ovn-controller" probeResult="failure" output=< Dec 03 14:29:30 crc kubenswrapper[4805]: ERROR - ovn-controller connection status is 'not connected', expecting 'connected' status Dec 03 14:29:30 crc kubenswrapper[4805]: > Dec 03 14:29:30 crc kubenswrapper[4805]: I1203 14:29:30.545728 4805 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovn-controller-ovs-dqwfk" Dec 03 14:29:30 crc kubenswrapper[4805]: I1203 14:29:30.547730 4805 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovn-controller-ovs-dqwfk" Dec 03 14:29:30 crc kubenswrapper[4805]: I1203 14:29:30.775365 4805 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-zhnwt-config-w979k"] Dec 03 14:29:30 crc kubenswrapper[4805]: I1203 14:29:30.776526 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-zhnwt-config-w979k" Dec 03 14:29:30 crc kubenswrapper[4805]: I1203 14:29:30.778857 4805 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovncontroller-extra-scripts" Dec 03 14:29:30 crc kubenswrapper[4805]: I1203 14:29:30.786106 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-zhnwt-config-w979k"] Dec 03 14:29:30 crc kubenswrapper[4805]: I1203 14:29:30.803259 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/e98ed5e3-2df2-4293-8115-cf43a781cd70-var-run\") pod \"ovn-controller-zhnwt-config-w979k\" (UID: \"e98ed5e3-2df2-4293-8115-cf43a781cd70\") " pod="openstack/ovn-controller-zhnwt-config-w979k" Dec 03 14:29:30 crc kubenswrapper[4805]: I1203 14:29:30.803302 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/e98ed5e3-2df2-4293-8115-cf43a781cd70-var-run-ovn\") pod \"ovn-controller-zhnwt-config-w979k\" (UID: \"e98ed5e3-2df2-4293-8115-cf43a781cd70\") " pod="openstack/ovn-controller-zhnwt-config-w979k" Dec 03 14:29:30 crc kubenswrapper[4805]: I1203 14:29:30.803420 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/e98ed5e3-2df2-4293-8115-cf43a781cd70-var-log-ovn\") pod \"ovn-controller-zhnwt-config-w979k\" (UID: \"e98ed5e3-2df2-4293-8115-cf43a781cd70\") " pod="openstack/ovn-controller-zhnwt-config-w979k" Dec 03 14:29:30 crc kubenswrapper[4805]: I1203 14:29:30.803547 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/e98ed5e3-2df2-4293-8115-cf43a781cd70-additional-scripts\") pod \"ovn-controller-zhnwt-config-w979k\" (UID: \"e98ed5e3-2df2-4293-8115-cf43a781cd70\") " pod="openstack/ovn-controller-zhnwt-config-w979k" Dec 03 14:29:30 crc kubenswrapper[4805]: I1203 14:29:30.803755 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/e98ed5e3-2df2-4293-8115-cf43a781cd70-scripts\") pod \"ovn-controller-zhnwt-config-w979k\" (UID: \"e98ed5e3-2df2-4293-8115-cf43a781cd70\") " pod="openstack/ovn-controller-zhnwt-config-w979k" Dec 03 14:29:30 crc kubenswrapper[4805]: I1203 14:29:30.803799 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8cvlt\" (UniqueName: \"kubernetes.io/projected/e98ed5e3-2df2-4293-8115-cf43a781cd70-kube-api-access-8cvlt\") pod \"ovn-controller-zhnwt-config-w979k\" (UID: \"e98ed5e3-2df2-4293-8115-cf43a781cd70\") " pod="openstack/ovn-controller-zhnwt-config-w979k" Dec 03 14:29:30 crc kubenswrapper[4805]: I1203 14:29:30.905803 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/e98ed5e3-2df2-4293-8115-cf43a781cd70-scripts\") pod \"ovn-controller-zhnwt-config-w979k\" (UID: \"e98ed5e3-2df2-4293-8115-cf43a781cd70\") " pod="openstack/ovn-controller-zhnwt-config-w979k" Dec 03 14:29:30 crc kubenswrapper[4805]: I1203 14:29:30.905864 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8cvlt\" (UniqueName: \"kubernetes.io/projected/e98ed5e3-2df2-4293-8115-cf43a781cd70-kube-api-access-8cvlt\") pod \"ovn-controller-zhnwt-config-w979k\" (UID: \"e98ed5e3-2df2-4293-8115-cf43a781cd70\") " pod="openstack/ovn-controller-zhnwt-config-w979k" Dec 03 14:29:30 crc kubenswrapper[4805]: I1203 14:29:30.905913 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/e98ed5e3-2df2-4293-8115-cf43a781cd70-var-run\") pod \"ovn-controller-zhnwt-config-w979k\" (UID: \"e98ed5e3-2df2-4293-8115-cf43a781cd70\") " pod="openstack/ovn-controller-zhnwt-config-w979k" Dec 03 14:29:30 crc kubenswrapper[4805]: I1203 14:29:30.905951 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/e98ed5e3-2df2-4293-8115-cf43a781cd70-var-run-ovn\") pod \"ovn-controller-zhnwt-config-w979k\" (UID: \"e98ed5e3-2df2-4293-8115-cf43a781cd70\") " pod="openstack/ovn-controller-zhnwt-config-w979k" Dec 03 14:29:30 crc kubenswrapper[4805]: I1203 14:29:30.905985 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/e98ed5e3-2df2-4293-8115-cf43a781cd70-var-log-ovn\") pod \"ovn-controller-zhnwt-config-w979k\" (UID: \"e98ed5e3-2df2-4293-8115-cf43a781cd70\") " pod="openstack/ovn-controller-zhnwt-config-w979k" Dec 03 14:29:30 crc kubenswrapper[4805]: I1203 14:29:30.906021 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/e98ed5e3-2df2-4293-8115-cf43a781cd70-additional-scripts\") pod \"ovn-controller-zhnwt-config-w979k\" (UID: \"e98ed5e3-2df2-4293-8115-cf43a781cd70\") " pod="openstack/ovn-controller-zhnwt-config-w979k" Dec 03 14:29:30 crc kubenswrapper[4805]: I1203 14:29:30.906521 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/e98ed5e3-2df2-4293-8115-cf43a781cd70-var-run-ovn\") pod \"ovn-controller-zhnwt-config-w979k\" (UID: \"e98ed5e3-2df2-4293-8115-cf43a781cd70\") " pod="openstack/ovn-controller-zhnwt-config-w979k" Dec 03 14:29:30 crc kubenswrapper[4805]: I1203 14:29:30.906554 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/e98ed5e3-2df2-4293-8115-cf43a781cd70-var-run\") pod \"ovn-controller-zhnwt-config-w979k\" (UID: \"e98ed5e3-2df2-4293-8115-cf43a781cd70\") " pod="openstack/ovn-controller-zhnwt-config-w979k" Dec 03 14:29:30 crc kubenswrapper[4805]: I1203 14:29:30.906571 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/e98ed5e3-2df2-4293-8115-cf43a781cd70-var-log-ovn\") pod \"ovn-controller-zhnwt-config-w979k\" (UID: \"e98ed5e3-2df2-4293-8115-cf43a781cd70\") " pod="openstack/ovn-controller-zhnwt-config-w979k" Dec 03 14:29:30 crc kubenswrapper[4805]: I1203 14:29:30.906966 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/e98ed5e3-2df2-4293-8115-cf43a781cd70-additional-scripts\") pod \"ovn-controller-zhnwt-config-w979k\" (UID: \"e98ed5e3-2df2-4293-8115-cf43a781cd70\") " pod="openstack/ovn-controller-zhnwt-config-w979k" Dec 03 14:29:30 crc kubenswrapper[4805]: I1203 14:29:30.908077 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/e98ed5e3-2df2-4293-8115-cf43a781cd70-scripts\") pod \"ovn-controller-zhnwt-config-w979k\" (UID: \"e98ed5e3-2df2-4293-8115-cf43a781cd70\") " pod="openstack/ovn-controller-zhnwt-config-w979k" Dec 03 14:29:30 crc kubenswrapper[4805]: I1203 14:29:30.926231 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8cvlt\" (UniqueName: \"kubernetes.io/projected/e98ed5e3-2df2-4293-8115-cf43a781cd70-kube-api-access-8cvlt\") pod \"ovn-controller-zhnwt-config-w979k\" (UID: \"e98ed5e3-2df2-4293-8115-cf43a781cd70\") " pod="openstack/ovn-controller-zhnwt-config-w979k" Dec 03 14:29:31 crc kubenswrapper[4805]: I1203 14:29:31.196391 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-zhnwt-config-w979k" Dec 03 14:29:31 crc kubenswrapper[4805]: I1203 14:29:31.403436 4805 generic.go:334] "Generic (PLEG): container finished" podID="1b34fa5c-2867-4dbe-9285-80c327699864" containerID="2e0f381f35c1220d9cac0d2d4c6085458b076fdcc37439c164c7ebd5a322440b" exitCode=0 Dec 03 14:29:31 crc kubenswrapper[4805]: I1203 14:29:31.403521 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"1b34fa5c-2867-4dbe-9285-80c327699864","Type":"ContainerDied","Data":"2e0f381f35c1220d9cac0d2d4c6085458b076fdcc37439c164c7ebd5a322440b"} Dec 03 14:29:31 crc kubenswrapper[4805]: I1203 14:29:31.405871 4805 generic.go:334] "Generic (PLEG): container finished" podID="656f691c-1188-47b9-b8dd-b393a230602a" containerID="0d085016bd6e7629b80b349c936428485e82d12ba8205b13e1f636404a6741cc" exitCode=0 Dec 03 14:29:31 crc kubenswrapper[4805]: I1203 14:29:31.405941 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"656f691c-1188-47b9-b8dd-b393a230602a","Type":"ContainerDied","Data":"0d085016bd6e7629b80b349c936428485e82d12ba8205b13e1f636404a6741cc"} Dec 03 14:29:35 crc kubenswrapper[4805]: I1203 14:29:35.518695 4805 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/ovn-controller-zhnwt" podUID="139abefe-12d4-4d8f-bdc6-43e71c818a65" containerName="ovn-controller" probeResult="failure" output=< Dec 03 14:29:35 crc kubenswrapper[4805]: ERROR - ovn-controller connection status is 'not connected', expecting 'connected' status Dec 03 14:29:35 crc kubenswrapper[4805]: > Dec 03 14:29:38 crc kubenswrapper[4805]: I1203 14:29:38.212146 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-zhnwt-config-w979k"] Dec 03 14:29:38 crc kubenswrapper[4805]: I1203 14:29:38.470004 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-sync-27jrz" event={"ID":"285c21ec-cbb8-4cb7-b5c2-0b27656f7870","Type":"ContainerStarted","Data":"917876c33cc904807afef7d8edc87b2672e031fae006b99501746b08d7943964"} Dec 03 14:29:38 crc kubenswrapper[4805]: I1203 14:29:38.474502 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"1b34fa5c-2867-4dbe-9285-80c327699864","Type":"ContainerStarted","Data":"b834218684c741c79335521d4745edfa20b11d70a0f812104cdab4f0a3846643"} Dec 03 14:29:38 crc kubenswrapper[4805]: I1203 14:29:38.474759 4805 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/rabbitmq-server-0" Dec 03 14:29:38 crc kubenswrapper[4805]: I1203 14:29:38.477361 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-zhnwt-config-w979k" event={"ID":"e98ed5e3-2df2-4293-8115-cf43a781cd70","Type":"ContainerStarted","Data":"9af0ffe65fa24804eda09c0f50ebb849d9543167f896e2cbf921d092bbe9c2bb"} Dec 03 14:29:38 crc kubenswrapper[4805]: I1203 14:29:38.490655 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"4fbc8d7f-d47f-4d91-915b-504820aac08a","Type":"ContainerStarted","Data":"fae449afeac83deee6f1de90703b0c77a2a54dd4e0c222ff04e72987c597f8ee"} Dec 03 14:29:38 crc kubenswrapper[4805]: I1203 14:29:38.490698 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"4fbc8d7f-d47f-4d91-915b-504820aac08a","Type":"ContainerStarted","Data":"265c0c40996a93acbea15d1a451f732a686f67f544ed9bf90f8cf01eed964f96"} Dec 03 14:29:38 crc kubenswrapper[4805]: I1203 14:29:38.490708 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"4fbc8d7f-d47f-4d91-915b-504820aac08a","Type":"ContainerStarted","Data":"b2e7afff247ad92bc9dad4cf4e275b3492307c0e283ec17351148b0f0057071d"} Dec 03 14:29:38 crc kubenswrapper[4805]: I1203 14:29:38.492931 4805 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-db-sync-27jrz" podStartSLOduration=2.263580741 podStartE2EDuration="13.492917186s" podCreationTimestamp="2025-12-03 14:29:25 +0000 UTC" firstStartedPulling="2025-12-03 14:29:26.623523796 +0000 UTC m=+1196.286440719" lastFinishedPulling="2025-12-03 14:29:37.852860241 +0000 UTC m=+1207.515777164" observedRunningTime="2025-12-03 14:29:38.489421842 +0000 UTC m=+1208.152338775" watchObservedRunningTime="2025-12-03 14:29:38.492917186 +0000 UTC m=+1208.155834119" Dec 03 14:29:38 crc kubenswrapper[4805]: I1203 14:29:38.509905 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"656f691c-1188-47b9-b8dd-b393a230602a","Type":"ContainerStarted","Data":"770e52863be49086afe2bea1b59e31fac9f82d73b30ec6fe23a2c1d8d92fe910"} Dec 03 14:29:38 crc kubenswrapper[4805]: I1203 14:29:38.510214 4805 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/rabbitmq-cell1-server-0" Dec 03 14:29:38 crc kubenswrapper[4805]: I1203 14:29:38.523160 4805 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/rabbitmq-server-0" podStartSLOduration=58.106022792 podStartE2EDuration="1m8.523138684s" podCreationTimestamp="2025-12-03 14:28:30 +0000 UTC" firstStartedPulling="2025-12-03 14:28:45.321880358 +0000 UTC m=+1154.984797281" lastFinishedPulling="2025-12-03 14:28:55.73899625 +0000 UTC m=+1165.401913173" observedRunningTime="2025-12-03 14:29:38.519581718 +0000 UTC m=+1208.182498641" watchObservedRunningTime="2025-12-03 14:29:38.523138684 +0000 UTC m=+1208.186055607" Dec 03 14:29:38 crc kubenswrapper[4805]: I1203 14:29:38.544353 4805 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/rabbitmq-cell1-server-0" podStartSLOduration=58.027172762 podStartE2EDuration="1m7.544331928s" podCreationTimestamp="2025-12-03 14:28:31 +0000 UTC" firstStartedPulling="2025-12-03 14:28:46.590188573 +0000 UTC m=+1156.253105496" lastFinishedPulling="2025-12-03 14:28:56.107347729 +0000 UTC m=+1165.770264662" observedRunningTime="2025-12-03 14:29:38.539759604 +0000 UTC m=+1208.202676527" watchObservedRunningTime="2025-12-03 14:29:38.544331928 +0000 UTC m=+1208.207248851" Dec 03 14:29:39 crc kubenswrapper[4805]: I1203 14:29:39.533165 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"4fbc8d7f-d47f-4d91-915b-504820aac08a","Type":"ContainerStarted","Data":"11a4ae54583baad8f8d9af92afbd87ae7639600b83ae6ddf9ce2f44d16aa52bd"} Dec 03 14:29:39 crc kubenswrapper[4805]: I1203 14:29:39.536005 4805 generic.go:334] "Generic (PLEG): container finished" podID="e98ed5e3-2df2-4293-8115-cf43a781cd70" containerID="4aea414289608c29dbb3bf5bd2ab7f74e16d7e4dcfadde7ffd3d078e77fba603" exitCode=0 Dec 03 14:29:39 crc kubenswrapper[4805]: I1203 14:29:39.536036 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-zhnwt-config-w979k" event={"ID":"e98ed5e3-2df2-4293-8115-cf43a781cd70","Type":"ContainerDied","Data":"4aea414289608c29dbb3bf5bd2ab7f74e16d7e4dcfadde7ffd3d078e77fba603"} Dec 03 14:29:40 crc kubenswrapper[4805]: I1203 14:29:40.548385 4805 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovn-controller-zhnwt" Dec 03 14:29:40 crc kubenswrapper[4805]: I1203 14:29:40.559873 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"4fbc8d7f-d47f-4d91-915b-504820aac08a","Type":"ContainerStarted","Data":"91fbc45a8b4aa3fe47ff508403e34c695de2b41c313f1f4632365239187e14c2"} Dec 03 14:29:40 crc kubenswrapper[4805]: I1203 14:29:40.560101 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"4fbc8d7f-d47f-4d91-915b-504820aac08a","Type":"ContainerStarted","Data":"3dcd981c93cc8ef0e00ce8aa59dd74e06a80b7b1c0416dbb91d278b366682cde"} Dec 03 14:29:40 crc kubenswrapper[4805]: I1203 14:29:40.560135 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"4fbc8d7f-d47f-4d91-915b-504820aac08a","Type":"ContainerStarted","Data":"50e721bc9cf15f46f75761ba2838cd206d377989765d6f97fdfe954b0ddeb981"} Dec 03 14:29:40 crc kubenswrapper[4805]: I1203 14:29:40.560150 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"4fbc8d7f-d47f-4d91-915b-504820aac08a","Type":"ContainerStarted","Data":"7df5acbec087e17b8447da69ed9d33882fc91757a0268d1ba3b30a3f15493452"} Dec 03 14:29:40 crc kubenswrapper[4805]: I1203 14:29:40.977413 4805 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-zhnwt-config-w979k" Dec 03 14:29:41 crc kubenswrapper[4805]: I1203 14:29:41.098829 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/e98ed5e3-2df2-4293-8115-cf43a781cd70-additional-scripts\") pod \"e98ed5e3-2df2-4293-8115-cf43a781cd70\" (UID: \"e98ed5e3-2df2-4293-8115-cf43a781cd70\") " Dec 03 14:29:41 crc kubenswrapper[4805]: I1203 14:29:41.099001 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/e98ed5e3-2df2-4293-8115-cf43a781cd70-var-run\") pod \"e98ed5e3-2df2-4293-8115-cf43a781cd70\" (UID: \"e98ed5e3-2df2-4293-8115-cf43a781cd70\") " Dec 03 14:29:41 crc kubenswrapper[4805]: I1203 14:29:41.099116 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/e98ed5e3-2df2-4293-8115-cf43a781cd70-var-run" (OuterVolumeSpecName: "var-run") pod "e98ed5e3-2df2-4293-8115-cf43a781cd70" (UID: "e98ed5e3-2df2-4293-8115-cf43a781cd70"). InnerVolumeSpecName "var-run". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 03 14:29:41 crc kubenswrapper[4805]: I1203 14:29:41.099164 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/e98ed5e3-2df2-4293-8115-cf43a781cd70-var-run-ovn\") pod \"e98ed5e3-2df2-4293-8115-cf43a781cd70\" (UID: \"e98ed5e3-2df2-4293-8115-cf43a781cd70\") " Dec 03 14:29:41 crc kubenswrapper[4805]: I1203 14:29:41.099209 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/e98ed5e3-2df2-4293-8115-cf43a781cd70-var-log-ovn\") pod \"e98ed5e3-2df2-4293-8115-cf43a781cd70\" (UID: \"e98ed5e3-2df2-4293-8115-cf43a781cd70\") " Dec 03 14:29:41 crc kubenswrapper[4805]: I1203 14:29:41.099231 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/e98ed5e3-2df2-4293-8115-cf43a781cd70-var-run-ovn" (OuterVolumeSpecName: "var-run-ovn") pod "e98ed5e3-2df2-4293-8115-cf43a781cd70" (UID: "e98ed5e3-2df2-4293-8115-cf43a781cd70"). InnerVolumeSpecName "var-run-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 03 14:29:41 crc kubenswrapper[4805]: I1203 14:29:41.099299 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/e98ed5e3-2df2-4293-8115-cf43a781cd70-var-log-ovn" (OuterVolumeSpecName: "var-log-ovn") pod "e98ed5e3-2df2-4293-8115-cf43a781cd70" (UID: "e98ed5e3-2df2-4293-8115-cf43a781cd70"). InnerVolumeSpecName "var-log-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 03 14:29:41 crc kubenswrapper[4805]: I1203 14:29:41.099331 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/e98ed5e3-2df2-4293-8115-cf43a781cd70-scripts\") pod \"e98ed5e3-2df2-4293-8115-cf43a781cd70\" (UID: \"e98ed5e3-2df2-4293-8115-cf43a781cd70\") " Dec 03 14:29:41 crc kubenswrapper[4805]: I1203 14:29:41.099476 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e98ed5e3-2df2-4293-8115-cf43a781cd70-additional-scripts" (OuterVolumeSpecName: "additional-scripts") pod "e98ed5e3-2df2-4293-8115-cf43a781cd70" (UID: "e98ed5e3-2df2-4293-8115-cf43a781cd70"). InnerVolumeSpecName "additional-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 14:29:41 crc kubenswrapper[4805]: I1203 14:29:41.099520 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8cvlt\" (UniqueName: \"kubernetes.io/projected/e98ed5e3-2df2-4293-8115-cf43a781cd70-kube-api-access-8cvlt\") pod \"e98ed5e3-2df2-4293-8115-cf43a781cd70\" (UID: \"e98ed5e3-2df2-4293-8115-cf43a781cd70\") " Dec 03 14:29:41 crc kubenswrapper[4805]: I1203 14:29:41.100312 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e98ed5e3-2df2-4293-8115-cf43a781cd70-scripts" (OuterVolumeSpecName: "scripts") pod "e98ed5e3-2df2-4293-8115-cf43a781cd70" (UID: "e98ed5e3-2df2-4293-8115-cf43a781cd70"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 14:29:41 crc kubenswrapper[4805]: I1203 14:29:41.100457 4805 reconciler_common.go:293] "Volume detached for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/e98ed5e3-2df2-4293-8115-cf43a781cd70-additional-scripts\") on node \"crc\" DevicePath \"\"" Dec 03 14:29:41 crc kubenswrapper[4805]: I1203 14:29:41.100480 4805 reconciler_common.go:293] "Volume detached for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/e98ed5e3-2df2-4293-8115-cf43a781cd70-var-run\") on node \"crc\" DevicePath \"\"" Dec 03 14:29:41 crc kubenswrapper[4805]: I1203 14:29:41.100510 4805 reconciler_common.go:293] "Volume detached for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/e98ed5e3-2df2-4293-8115-cf43a781cd70-var-run-ovn\") on node \"crc\" DevicePath \"\"" Dec 03 14:29:41 crc kubenswrapper[4805]: I1203 14:29:41.100522 4805 reconciler_common.go:293] "Volume detached for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/e98ed5e3-2df2-4293-8115-cf43a781cd70-var-log-ovn\") on node \"crc\" DevicePath \"\"" Dec 03 14:29:41 crc kubenswrapper[4805]: I1203 14:29:41.100535 4805 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/e98ed5e3-2df2-4293-8115-cf43a781cd70-scripts\") on node \"crc\" DevicePath \"\"" Dec 03 14:29:41 crc kubenswrapper[4805]: I1203 14:29:41.106117 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e98ed5e3-2df2-4293-8115-cf43a781cd70-kube-api-access-8cvlt" (OuterVolumeSpecName: "kube-api-access-8cvlt") pod "e98ed5e3-2df2-4293-8115-cf43a781cd70" (UID: "e98ed5e3-2df2-4293-8115-cf43a781cd70"). InnerVolumeSpecName "kube-api-access-8cvlt". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 14:29:41 crc kubenswrapper[4805]: I1203 14:29:41.201445 4805 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8cvlt\" (UniqueName: \"kubernetes.io/projected/e98ed5e3-2df2-4293-8115-cf43a781cd70-kube-api-access-8cvlt\") on node \"crc\" DevicePath \"\"" Dec 03 14:29:41 crc kubenswrapper[4805]: I1203 14:29:41.573741 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-zhnwt-config-w979k" event={"ID":"e98ed5e3-2df2-4293-8115-cf43a781cd70","Type":"ContainerDied","Data":"9af0ffe65fa24804eda09c0f50ebb849d9543167f896e2cbf921d092bbe9c2bb"} Dec 03 14:29:41 crc kubenswrapper[4805]: I1203 14:29:41.575030 4805 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="9af0ffe65fa24804eda09c0f50ebb849d9543167f896e2cbf921d092bbe9c2bb" Dec 03 14:29:41 crc kubenswrapper[4805]: I1203 14:29:41.573753 4805 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-zhnwt-config-w979k" Dec 03 14:29:41 crc kubenswrapper[4805]: I1203 14:29:41.592540 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"4fbc8d7f-d47f-4d91-915b-504820aac08a","Type":"ContainerStarted","Data":"6dd9d415c06d7720ff646c2d163cd985f6368b68c7de4af9ae41b383d6ed61e5"} Dec 03 14:29:41 crc kubenswrapper[4805]: I1203 14:29:41.592767 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"4fbc8d7f-d47f-4d91-915b-504820aac08a","Type":"ContainerStarted","Data":"46057090a402eb049dd320f4c202bb99e996df200ff039bf51d295c26da3c4bf"} Dec 03 14:29:41 crc kubenswrapper[4805]: I1203 14:29:41.592853 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"4fbc8d7f-d47f-4d91-915b-504820aac08a","Type":"ContainerStarted","Data":"87bf157da2d7f812828d87a02ee3a3e7c601470d1f3ad3f1fd1444b4ef6fbd87"} Dec 03 14:29:41 crc kubenswrapper[4805]: I1203 14:29:41.645673 4805 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/swift-storage-0" podStartSLOduration=20.033613436 podStartE2EDuration="34.64564629s" podCreationTimestamp="2025-12-03 14:29:07 +0000 UTC" firstStartedPulling="2025-12-03 14:29:25.15415678 +0000 UTC m=+1194.817073713" lastFinishedPulling="2025-12-03 14:29:39.766189634 +0000 UTC m=+1209.429106567" observedRunningTime="2025-12-03 14:29:41.634638462 +0000 UTC m=+1211.297555385" watchObservedRunningTime="2025-12-03 14:29:41.64564629 +0000 UTC m=+1211.308563203" Dec 03 14:29:42 crc kubenswrapper[4805]: I1203 14:29:42.021748 4805 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-77585f5f8c-6gbzs"] Dec 03 14:29:42 crc kubenswrapper[4805]: E1203 14:29:42.022140 4805 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e98ed5e3-2df2-4293-8115-cf43a781cd70" containerName="ovn-config" Dec 03 14:29:42 crc kubenswrapper[4805]: I1203 14:29:42.022156 4805 state_mem.go:107] "Deleted CPUSet assignment" podUID="e98ed5e3-2df2-4293-8115-cf43a781cd70" containerName="ovn-config" Dec 03 14:29:42 crc kubenswrapper[4805]: I1203 14:29:42.022367 4805 memory_manager.go:354] "RemoveStaleState removing state" podUID="e98ed5e3-2df2-4293-8115-cf43a781cd70" containerName="ovn-config" Dec 03 14:29:42 crc kubenswrapper[4805]: I1203 14:29:42.023323 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-77585f5f8c-6gbzs" Dec 03 14:29:42 crc kubenswrapper[4805]: I1203 14:29:42.027452 4805 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"dns-swift-storage-0" Dec 03 14:29:42 crc kubenswrapper[4805]: I1203 14:29:42.045720 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-77585f5f8c-6gbzs"] Dec 03 14:29:42 crc kubenswrapper[4805]: I1203 14:29:42.104472 4805 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ovn-controller-zhnwt-config-w979k"] Dec 03 14:29:42 crc kubenswrapper[4805]: I1203 14:29:42.111356 4805 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ovn-controller-zhnwt-config-w979k"] Dec 03 14:29:42 crc kubenswrapper[4805]: I1203 14:29:42.118921 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/d9710132-a82b-4aa4-82dd-537ae1ea51d6-ovsdbserver-nb\") pod \"dnsmasq-dns-77585f5f8c-6gbzs\" (UID: \"d9710132-a82b-4aa4-82dd-537ae1ea51d6\") " pod="openstack/dnsmasq-dns-77585f5f8c-6gbzs" Dec 03 14:29:42 crc kubenswrapper[4805]: I1203 14:29:42.118972 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/d9710132-a82b-4aa4-82dd-537ae1ea51d6-dns-svc\") pod \"dnsmasq-dns-77585f5f8c-6gbzs\" (UID: \"d9710132-a82b-4aa4-82dd-537ae1ea51d6\") " pod="openstack/dnsmasq-dns-77585f5f8c-6gbzs" Dec 03 14:29:42 crc kubenswrapper[4805]: I1203 14:29:42.118998 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d9710132-a82b-4aa4-82dd-537ae1ea51d6-config\") pod \"dnsmasq-dns-77585f5f8c-6gbzs\" (UID: \"d9710132-a82b-4aa4-82dd-537ae1ea51d6\") " pod="openstack/dnsmasq-dns-77585f5f8c-6gbzs" Dec 03 14:29:42 crc kubenswrapper[4805]: I1203 14:29:42.119035 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/d9710132-a82b-4aa4-82dd-537ae1ea51d6-dns-swift-storage-0\") pod \"dnsmasq-dns-77585f5f8c-6gbzs\" (UID: \"d9710132-a82b-4aa4-82dd-537ae1ea51d6\") " pod="openstack/dnsmasq-dns-77585f5f8c-6gbzs" Dec 03 14:29:42 crc kubenswrapper[4805]: I1203 14:29:42.119191 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/d9710132-a82b-4aa4-82dd-537ae1ea51d6-ovsdbserver-sb\") pod \"dnsmasq-dns-77585f5f8c-6gbzs\" (UID: \"d9710132-a82b-4aa4-82dd-537ae1ea51d6\") " pod="openstack/dnsmasq-dns-77585f5f8c-6gbzs" Dec 03 14:29:42 crc kubenswrapper[4805]: I1203 14:29:42.119367 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kgs8h\" (UniqueName: \"kubernetes.io/projected/d9710132-a82b-4aa4-82dd-537ae1ea51d6-kube-api-access-kgs8h\") pod \"dnsmasq-dns-77585f5f8c-6gbzs\" (UID: \"d9710132-a82b-4aa4-82dd-537ae1ea51d6\") " pod="openstack/dnsmasq-dns-77585f5f8c-6gbzs" Dec 03 14:29:42 crc kubenswrapper[4805]: I1203 14:29:42.147605 4805 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-zhnwt-config-s9ncq"] Dec 03 14:29:42 crc kubenswrapper[4805]: I1203 14:29:42.148968 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-zhnwt-config-s9ncq" Dec 03 14:29:42 crc kubenswrapper[4805]: I1203 14:29:42.151384 4805 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovncontroller-extra-scripts" Dec 03 14:29:42 crc kubenswrapper[4805]: I1203 14:29:42.166414 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-zhnwt-config-s9ncq"] Dec 03 14:29:42 crc kubenswrapper[4805]: I1203 14:29:42.220823 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/d9710132-a82b-4aa4-82dd-537ae1ea51d6-ovsdbserver-nb\") pod \"dnsmasq-dns-77585f5f8c-6gbzs\" (UID: \"d9710132-a82b-4aa4-82dd-537ae1ea51d6\") " pod="openstack/dnsmasq-dns-77585f5f8c-6gbzs" Dec 03 14:29:42 crc kubenswrapper[4805]: I1203 14:29:42.220892 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/d9710132-a82b-4aa4-82dd-537ae1ea51d6-dns-svc\") pod \"dnsmasq-dns-77585f5f8c-6gbzs\" (UID: \"d9710132-a82b-4aa4-82dd-537ae1ea51d6\") " pod="openstack/dnsmasq-dns-77585f5f8c-6gbzs" Dec 03 14:29:42 crc kubenswrapper[4805]: I1203 14:29:42.220927 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d9710132-a82b-4aa4-82dd-537ae1ea51d6-config\") pod \"dnsmasq-dns-77585f5f8c-6gbzs\" (UID: \"d9710132-a82b-4aa4-82dd-537ae1ea51d6\") " pod="openstack/dnsmasq-dns-77585f5f8c-6gbzs" Dec 03 14:29:42 crc kubenswrapper[4805]: I1203 14:29:42.220967 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/d9710132-a82b-4aa4-82dd-537ae1ea51d6-dns-swift-storage-0\") pod \"dnsmasq-dns-77585f5f8c-6gbzs\" (UID: \"d9710132-a82b-4aa4-82dd-537ae1ea51d6\") " pod="openstack/dnsmasq-dns-77585f5f8c-6gbzs" Dec 03 14:29:42 crc kubenswrapper[4805]: I1203 14:29:42.221012 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/d9710132-a82b-4aa4-82dd-537ae1ea51d6-ovsdbserver-sb\") pod \"dnsmasq-dns-77585f5f8c-6gbzs\" (UID: \"d9710132-a82b-4aa4-82dd-537ae1ea51d6\") " pod="openstack/dnsmasq-dns-77585f5f8c-6gbzs" Dec 03 14:29:42 crc kubenswrapper[4805]: I1203 14:29:42.221098 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kgs8h\" (UniqueName: \"kubernetes.io/projected/d9710132-a82b-4aa4-82dd-537ae1ea51d6-kube-api-access-kgs8h\") pod \"dnsmasq-dns-77585f5f8c-6gbzs\" (UID: \"d9710132-a82b-4aa4-82dd-537ae1ea51d6\") " pod="openstack/dnsmasq-dns-77585f5f8c-6gbzs" Dec 03 14:29:42 crc kubenswrapper[4805]: I1203 14:29:42.221802 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/d9710132-a82b-4aa4-82dd-537ae1ea51d6-ovsdbserver-nb\") pod \"dnsmasq-dns-77585f5f8c-6gbzs\" (UID: \"d9710132-a82b-4aa4-82dd-537ae1ea51d6\") " pod="openstack/dnsmasq-dns-77585f5f8c-6gbzs" Dec 03 14:29:42 crc kubenswrapper[4805]: I1203 14:29:42.221961 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/d9710132-a82b-4aa4-82dd-537ae1ea51d6-ovsdbserver-sb\") pod \"dnsmasq-dns-77585f5f8c-6gbzs\" (UID: \"d9710132-a82b-4aa4-82dd-537ae1ea51d6\") " pod="openstack/dnsmasq-dns-77585f5f8c-6gbzs" Dec 03 14:29:42 crc kubenswrapper[4805]: I1203 14:29:42.222122 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d9710132-a82b-4aa4-82dd-537ae1ea51d6-config\") pod \"dnsmasq-dns-77585f5f8c-6gbzs\" (UID: \"d9710132-a82b-4aa4-82dd-537ae1ea51d6\") " pod="openstack/dnsmasq-dns-77585f5f8c-6gbzs" Dec 03 14:29:42 crc kubenswrapper[4805]: I1203 14:29:42.222183 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/d9710132-a82b-4aa4-82dd-537ae1ea51d6-dns-svc\") pod \"dnsmasq-dns-77585f5f8c-6gbzs\" (UID: \"d9710132-a82b-4aa4-82dd-537ae1ea51d6\") " pod="openstack/dnsmasq-dns-77585f5f8c-6gbzs" Dec 03 14:29:42 crc kubenswrapper[4805]: I1203 14:29:42.222348 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/d9710132-a82b-4aa4-82dd-537ae1ea51d6-dns-swift-storage-0\") pod \"dnsmasq-dns-77585f5f8c-6gbzs\" (UID: \"d9710132-a82b-4aa4-82dd-537ae1ea51d6\") " pod="openstack/dnsmasq-dns-77585f5f8c-6gbzs" Dec 03 14:29:42 crc kubenswrapper[4805]: I1203 14:29:42.252220 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kgs8h\" (UniqueName: \"kubernetes.io/projected/d9710132-a82b-4aa4-82dd-537ae1ea51d6-kube-api-access-kgs8h\") pod \"dnsmasq-dns-77585f5f8c-6gbzs\" (UID: \"d9710132-a82b-4aa4-82dd-537ae1ea51d6\") " pod="openstack/dnsmasq-dns-77585f5f8c-6gbzs" Dec 03 14:29:42 crc kubenswrapper[4805]: I1203 14:29:42.322477 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/d8e0e718-99ce-4813-a824-bf2182f8fdf1-scripts\") pod \"ovn-controller-zhnwt-config-s9ncq\" (UID: \"d8e0e718-99ce-4813-a824-bf2182f8fdf1\") " pod="openstack/ovn-controller-zhnwt-config-s9ncq" Dec 03 14:29:42 crc kubenswrapper[4805]: I1203 14:29:42.322570 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/d8e0e718-99ce-4813-a824-bf2182f8fdf1-var-run\") pod \"ovn-controller-zhnwt-config-s9ncq\" (UID: \"d8e0e718-99ce-4813-a824-bf2182f8fdf1\") " pod="openstack/ovn-controller-zhnwt-config-s9ncq" Dec 03 14:29:42 crc kubenswrapper[4805]: I1203 14:29:42.322738 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/d8e0e718-99ce-4813-a824-bf2182f8fdf1-additional-scripts\") pod \"ovn-controller-zhnwt-config-s9ncq\" (UID: \"d8e0e718-99ce-4813-a824-bf2182f8fdf1\") " pod="openstack/ovn-controller-zhnwt-config-s9ncq" Dec 03 14:29:42 crc kubenswrapper[4805]: I1203 14:29:42.322989 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/d8e0e718-99ce-4813-a824-bf2182f8fdf1-var-run-ovn\") pod \"ovn-controller-zhnwt-config-s9ncq\" (UID: \"d8e0e718-99ce-4813-a824-bf2182f8fdf1\") " pod="openstack/ovn-controller-zhnwt-config-s9ncq" Dec 03 14:29:42 crc kubenswrapper[4805]: I1203 14:29:42.323031 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ct5w8\" (UniqueName: \"kubernetes.io/projected/d8e0e718-99ce-4813-a824-bf2182f8fdf1-kube-api-access-ct5w8\") pod \"ovn-controller-zhnwt-config-s9ncq\" (UID: \"d8e0e718-99ce-4813-a824-bf2182f8fdf1\") " pod="openstack/ovn-controller-zhnwt-config-s9ncq" Dec 03 14:29:42 crc kubenswrapper[4805]: I1203 14:29:42.323098 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/d8e0e718-99ce-4813-a824-bf2182f8fdf1-var-log-ovn\") pod \"ovn-controller-zhnwt-config-s9ncq\" (UID: \"d8e0e718-99ce-4813-a824-bf2182f8fdf1\") " pod="openstack/ovn-controller-zhnwt-config-s9ncq" Dec 03 14:29:42 crc kubenswrapper[4805]: I1203 14:29:42.342089 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-77585f5f8c-6gbzs" Dec 03 14:29:42 crc kubenswrapper[4805]: I1203 14:29:42.424830 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/d8e0e718-99ce-4813-a824-bf2182f8fdf1-var-log-ovn\") pod \"ovn-controller-zhnwt-config-s9ncq\" (UID: \"d8e0e718-99ce-4813-a824-bf2182f8fdf1\") " pod="openstack/ovn-controller-zhnwt-config-s9ncq" Dec 03 14:29:42 crc kubenswrapper[4805]: I1203 14:29:42.425090 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/d8e0e718-99ce-4813-a824-bf2182f8fdf1-scripts\") pod \"ovn-controller-zhnwt-config-s9ncq\" (UID: \"d8e0e718-99ce-4813-a824-bf2182f8fdf1\") " pod="openstack/ovn-controller-zhnwt-config-s9ncq" Dec 03 14:29:42 crc kubenswrapper[4805]: I1203 14:29:42.425120 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/d8e0e718-99ce-4813-a824-bf2182f8fdf1-var-run\") pod \"ovn-controller-zhnwt-config-s9ncq\" (UID: \"d8e0e718-99ce-4813-a824-bf2182f8fdf1\") " pod="openstack/ovn-controller-zhnwt-config-s9ncq" Dec 03 14:29:42 crc kubenswrapper[4805]: I1203 14:29:42.425178 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/d8e0e718-99ce-4813-a824-bf2182f8fdf1-additional-scripts\") pod \"ovn-controller-zhnwt-config-s9ncq\" (UID: \"d8e0e718-99ce-4813-a824-bf2182f8fdf1\") " pod="openstack/ovn-controller-zhnwt-config-s9ncq" Dec 03 14:29:42 crc kubenswrapper[4805]: I1203 14:29:42.425255 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/d8e0e718-99ce-4813-a824-bf2182f8fdf1-var-log-ovn\") pod \"ovn-controller-zhnwt-config-s9ncq\" (UID: \"d8e0e718-99ce-4813-a824-bf2182f8fdf1\") " pod="openstack/ovn-controller-zhnwt-config-s9ncq" Dec 03 14:29:42 crc kubenswrapper[4805]: I1203 14:29:42.425263 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/d8e0e718-99ce-4813-a824-bf2182f8fdf1-var-run-ovn\") pod \"ovn-controller-zhnwt-config-s9ncq\" (UID: \"d8e0e718-99ce-4813-a824-bf2182f8fdf1\") " pod="openstack/ovn-controller-zhnwt-config-s9ncq" Dec 03 14:29:42 crc kubenswrapper[4805]: I1203 14:29:42.425345 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/d8e0e718-99ce-4813-a824-bf2182f8fdf1-var-run-ovn\") pod \"ovn-controller-zhnwt-config-s9ncq\" (UID: \"d8e0e718-99ce-4813-a824-bf2182f8fdf1\") " pod="openstack/ovn-controller-zhnwt-config-s9ncq" Dec 03 14:29:42 crc kubenswrapper[4805]: I1203 14:29:42.425395 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ct5w8\" (UniqueName: \"kubernetes.io/projected/d8e0e718-99ce-4813-a824-bf2182f8fdf1-kube-api-access-ct5w8\") pod \"ovn-controller-zhnwt-config-s9ncq\" (UID: \"d8e0e718-99ce-4813-a824-bf2182f8fdf1\") " pod="openstack/ovn-controller-zhnwt-config-s9ncq" Dec 03 14:29:42 crc kubenswrapper[4805]: I1203 14:29:42.425988 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/d8e0e718-99ce-4813-a824-bf2182f8fdf1-var-run\") pod \"ovn-controller-zhnwt-config-s9ncq\" (UID: \"d8e0e718-99ce-4813-a824-bf2182f8fdf1\") " pod="openstack/ovn-controller-zhnwt-config-s9ncq" Dec 03 14:29:42 crc kubenswrapper[4805]: I1203 14:29:42.426697 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/d8e0e718-99ce-4813-a824-bf2182f8fdf1-additional-scripts\") pod \"ovn-controller-zhnwt-config-s9ncq\" (UID: \"d8e0e718-99ce-4813-a824-bf2182f8fdf1\") " pod="openstack/ovn-controller-zhnwt-config-s9ncq" Dec 03 14:29:42 crc kubenswrapper[4805]: I1203 14:29:42.427812 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/d8e0e718-99ce-4813-a824-bf2182f8fdf1-scripts\") pod \"ovn-controller-zhnwt-config-s9ncq\" (UID: \"d8e0e718-99ce-4813-a824-bf2182f8fdf1\") " pod="openstack/ovn-controller-zhnwt-config-s9ncq" Dec 03 14:29:42 crc kubenswrapper[4805]: I1203 14:29:42.447608 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ct5w8\" (UniqueName: \"kubernetes.io/projected/d8e0e718-99ce-4813-a824-bf2182f8fdf1-kube-api-access-ct5w8\") pod \"ovn-controller-zhnwt-config-s9ncq\" (UID: \"d8e0e718-99ce-4813-a824-bf2182f8fdf1\") " pod="openstack/ovn-controller-zhnwt-config-s9ncq" Dec 03 14:29:42 crc kubenswrapper[4805]: I1203 14:29:42.466672 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-zhnwt-config-s9ncq" Dec 03 14:29:42 crc kubenswrapper[4805]: I1203 14:29:42.704391 4805 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e98ed5e3-2df2-4293-8115-cf43a781cd70" path="/var/lib/kubelet/pods/e98ed5e3-2df2-4293-8115-cf43a781cd70/volumes" Dec 03 14:29:42 crc kubenswrapper[4805]: I1203 14:29:42.894410 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-77585f5f8c-6gbzs"] Dec 03 14:29:42 crc kubenswrapper[4805]: W1203 14:29:42.902904 4805 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podd9710132_a82b_4aa4_82dd_537ae1ea51d6.slice/crio-401345d095f6cb7e1da4b63bab60a175de7d4eab8d464225b019cb16242e4ae5 WatchSource:0}: Error finding container 401345d095f6cb7e1da4b63bab60a175de7d4eab8d464225b019cb16242e4ae5: Status 404 returned error can't find the container with id 401345d095f6cb7e1da4b63bab60a175de7d4eab8d464225b019cb16242e4ae5 Dec 03 14:29:42 crc kubenswrapper[4805]: I1203 14:29:42.952166 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-zhnwt-config-s9ncq"] Dec 03 14:29:42 crc kubenswrapper[4805]: W1203 14:29:42.963063 4805 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podd8e0e718_99ce_4813_a824_bf2182f8fdf1.slice/crio-0d07b624db60628a457628a6f2d21f5241fc5d0741ca9b8c3b65c4693d1a3511 WatchSource:0}: Error finding container 0d07b624db60628a457628a6f2d21f5241fc5d0741ca9b8c3b65c4693d1a3511: Status 404 returned error can't find the container with id 0d07b624db60628a457628a6f2d21f5241fc5d0741ca9b8c3b65c4693d1a3511 Dec 03 14:29:43 crc kubenswrapper[4805]: I1203 14:29:43.612639 4805 generic.go:334] "Generic (PLEG): container finished" podID="d9710132-a82b-4aa4-82dd-537ae1ea51d6" containerID="20128c805c35289c7ee1383415b1c93838143347edbb9a9aad4a474420a28b83" exitCode=0 Dec 03 14:29:43 crc kubenswrapper[4805]: I1203 14:29:43.612802 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-77585f5f8c-6gbzs" event={"ID":"d9710132-a82b-4aa4-82dd-537ae1ea51d6","Type":"ContainerDied","Data":"20128c805c35289c7ee1383415b1c93838143347edbb9a9aad4a474420a28b83"} Dec 03 14:29:43 crc kubenswrapper[4805]: I1203 14:29:43.613076 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-77585f5f8c-6gbzs" event={"ID":"d9710132-a82b-4aa4-82dd-537ae1ea51d6","Type":"ContainerStarted","Data":"401345d095f6cb7e1da4b63bab60a175de7d4eab8d464225b019cb16242e4ae5"} Dec 03 14:29:43 crc kubenswrapper[4805]: I1203 14:29:43.615568 4805 generic.go:334] "Generic (PLEG): container finished" podID="d8e0e718-99ce-4813-a824-bf2182f8fdf1" containerID="182cee8becfec97e798fcb0944c192824cf5b87960b0a17801440f1a7101b6df" exitCode=0 Dec 03 14:29:43 crc kubenswrapper[4805]: I1203 14:29:43.615631 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-zhnwt-config-s9ncq" event={"ID":"d8e0e718-99ce-4813-a824-bf2182f8fdf1","Type":"ContainerDied","Data":"182cee8becfec97e798fcb0944c192824cf5b87960b0a17801440f1a7101b6df"} Dec 03 14:29:43 crc kubenswrapper[4805]: I1203 14:29:43.615690 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-zhnwt-config-s9ncq" event={"ID":"d8e0e718-99ce-4813-a824-bf2182f8fdf1","Type":"ContainerStarted","Data":"0d07b624db60628a457628a6f2d21f5241fc5d0741ca9b8c3b65c4693d1a3511"} Dec 03 14:29:43 crc kubenswrapper[4805]: I1203 14:29:43.917621 4805 patch_prober.go:28] interesting pod/machine-config-daemon-gskh4 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 03 14:29:43 crc kubenswrapper[4805]: I1203 14:29:43.918099 4805 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-gskh4" podUID="6098937f-e3f6-45e8-a647-4994a79cd711" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 03 14:29:44 crc kubenswrapper[4805]: I1203 14:29:44.630710 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-77585f5f8c-6gbzs" event={"ID":"d9710132-a82b-4aa4-82dd-537ae1ea51d6","Type":"ContainerStarted","Data":"831f6e175f71916762f0e7a3330fef1491b075f46ef72768ce656b8c3fd9af61"} Dec 03 14:29:44 crc kubenswrapper[4805]: I1203 14:29:44.655779 4805 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-77585f5f8c-6gbzs" podStartSLOduration=3.655758423 podStartE2EDuration="3.655758423s" podCreationTimestamp="2025-12-03 14:29:41 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 14:29:44.650762778 +0000 UTC m=+1214.313679721" watchObservedRunningTime="2025-12-03 14:29:44.655758423 +0000 UTC m=+1214.318675346" Dec 03 14:29:44 crc kubenswrapper[4805]: I1203 14:29:44.966505 4805 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-zhnwt-config-s9ncq" Dec 03 14:29:45 crc kubenswrapper[4805]: I1203 14:29:45.075431 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/d8e0e718-99ce-4813-a824-bf2182f8fdf1-var-run\") pod \"d8e0e718-99ce-4813-a824-bf2182f8fdf1\" (UID: \"d8e0e718-99ce-4813-a824-bf2182f8fdf1\") " Dec 03 14:29:45 crc kubenswrapper[4805]: I1203 14:29:45.075651 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/d8e0e718-99ce-4813-a824-bf2182f8fdf1-scripts\") pod \"d8e0e718-99ce-4813-a824-bf2182f8fdf1\" (UID: \"d8e0e718-99ce-4813-a824-bf2182f8fdf1\") " Dec 03 14:29:45 crc kubenswrapper[4805]: I1203 14:29:45.075717 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/d8e0e718-99ce-4813-a824-bf2182f8fdf1-var-run-ovn\") pod \"d8e0e718-99ce-4813-a824-bf2182f8fdf1\" (UID: \"d8e0e718-99ce-4813-a824-bf2182f8fdf1\") " Dec 03 14:29:45 crc kubenswrapper[4805]: I1203 14:29:45.075748 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/d8e0e718-99ce-4813-a824-bf2182f8fdf1-var-log-ovn\") pod \"d8e0e718-99ce-4813-a824-bf2182f8fdf1\" (UID: \"d8e0e718-99ce-4813-a824-bf2182f8fdf1\") " Dec 03 14:29:45 crc kubenswrapper[4805]: I1203 14:29:45.075888 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ct5w8\" (UniqueName: \"kubernetes.io/projected/d8e0e718-99ce-4813-a824-bf2182f8fdf1-kube-api-access-ct5w8\") pod \"d8e0e718-99ce-4813-a824-bf2182f8fdf1\" (UID: \"d8e0e718-99ce-4813-a824-bf2182f8fdf1\") " Dec 03 14:29:45 crc kubenswrapper[4805]: I1203 14:29:45.075919 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/d8e0e718-99ce-4813-a824-bf2182f8fdf1-additional-scripts\") pod \"d8e0e718-99ce-4813-a824-bf2182f8fdf1\" (UID: \"d8e0e718-99ce-4813-a824-bf2182f8fdf1\") " Dec 03 14:29:45 crc kubenswrapper[4805]: I1203 14:29:45.077083 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/d8e0e718-99ce-4813-a824-bf2182f8fdf1-var-run" (OuterVolumeSpecName: "var-run") pod "d8e0e718-99ce-4813-a824-bf2182f8fdf1" (UID: "d8e0e718-99ce-4813-a824-bf2182f8fdf1"). InnerVolumeSpecName "var-run". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 03 14:29:45 crc kubenswrapper[4805]: I1203 14:29:45.077084 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/d8e0e718-99ce-4813-a824-bf2182f8fdf1-var-run-ovn" (OuterVolumeSpecName: "var-run-ovn") pod "d8e0e718-99ce-4813-a824-bf2182f8fdf1" (UID: "d8e0e718-99ce-4813-a824-bf2182f8fdf1"). InnerVolumeSpecName "var-run-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 03 14:29:45 crc kubenswrapper[4805]: I1203 14:29:45.077203 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/d8e0e718-99ce-4813-a824-bf2182f8fdf1-var-log-ovn" (OuterVolumeSpecName: "var-log-ovn") pod "d8e0e718-99ce-4813-a824-bf2182f8fdf1" (UID: "d8e0e718-99ce-4813-a824-bf2182f8fdf1"). InnerVolumeSpecName "var-log-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 03 14:29:45 crc kubenswrapper[4805]: I1203 14:29:45.077575 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d8e0e718-99ce-4813-a824-bf2182f8fdf1-additional-scripts" (OuterVolumeSpecName: "additional-scripts") pod "d8e0e718-99ce-4813-a824-bf2182f8fdf1" (UID: "d8e0e718-99ce-4813-a824-bf2182f8fdf1"). InnerVolumeSpecName "additional-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 14:29:45 crc kubenswrapper[4805]: I1203 14:29:45.077908 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d8e0e718-99ce-4813-a824-bf2182f8fdf1-scripts" (OuterVolumeSpecName: "scripts") pod "d8e0e718-99ce-4813-a824-bf2182f8fdf1" (UID: "d8e0e718-99ce-4813-a824-bf2182f8fdf1"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 14:29:45 crc kubenswrapper[4805]: I1203 14:29:45.089164 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d8e0e718-99ce-4813-a824-bf2182f8fdf1-kube-api-access-ct5w8" (OuterVolumeSpecName: "kube-api-access-ct5w8") pod "d8e0e718-99ce-4813-a824-bf2182f8fdf1" (UID: "d8e0e718-99ce-4813-a824-bf2182f8fdf1"). InnerVolumeSpecName "kube-api-access-ct5w8". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 14:29:45 crc kubenswrapper[4805]: I1203 14:29:45.178452 4805 reconciler_common.go:293] "Volume detached for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/d8e0e718-99ce-4813-a824-bf2182f8fdf1-var-log-ovn\") on node \"crc\" DevicePath \"\"" Dec 03 14:29:45 crc kubenswrapper[4805]: I1203 14:29:45.178494 4805 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ct5w8\" (UniqueName: \"kubernetes.io/projected/d8e0e718-99ce-4813-a824-bf2182f8fdf1-kube-api-access-ct5w8\") on node \"crc\" DevicePath \"\"" Dec 03 14:29:45 crc kubenswrapper[4805]: I1203 14:29:45.178508 4805 reconciler_common.go:293] "Volume detached for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/d8e0e718-99ce-4813-a824-bf2182f8fdf1-additional-scripts\") on node \"crc\" DevicePath \"\"" Dec 03 14:29:45 crc kubenswrapper[4805]: I1203 14:29:45.178520 4805 reconciler_common.go:293] "Volume detached for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/d8e0e718-99ce-4813-a824-bf2182f8fdf1-var-run\") on node \"crc\" DevicePath \"\"" Dec 03 14:29:45 crc kubenswrapper[4805]: I1203 14:29:45.178531 4805 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/d8e0e718-99ce-4813-a824-bf2182f8fdf1-scripts\") on node \"crc\" DevicePath \"\"" Dec 03 14:29:45 crc kubenswrapper[4805]: I1203 14:29:45.178544 4805 reconciler_common.go:293] "Volume detached for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/d8e0e718-99ce-4813-a824-bf2182f8fdf1-var-run-ovn\") on node \"crc\" DevicePath \"\"" Dec 03 14:29:45 crc kubenswrapper[4805]: I1203 14:29:45.643152 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-zhnwt-config-s9ncq" event={"ID":"d8e0e718-99ce-4813-a824-bf2182f8fdf1","Type":"ContainerDied","Data":"0d07b624db60628a457628a6f2d21f5241fc5d0741ca9b8c3b65c4693d1a3511"} Dec 03 14:29:45 crc kubenswrapper[4805]: I1203 14:29:45.643217 4805 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="0d07b624db60628a457628a6f2d21f5241fc5d0741ca9b8c3b65c4693d1a3511" Dec 03 14:29:45 crc kubenswrapper[4805]: I1203 14:29:45.643172 4805 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-zhnwt-config-s9ncq" Dec 03 14:29:45 crc kubenswrapper[4805]: I1203 14:29:45.643486 4805 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-77585f5f8c-6gbzs" Dec 03 14:29:46 crc kubenswrapper[4805]: I1203 14:29:46.114826 4805 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ovn-controller-zhnwt-config-s9ncq"] Dec 03 14:29:46 crc kubenswrapper[4805]: I1203 14:29:46.123404 4805 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ovn-controller-zhnwt-config-s9ncq"] Dec 03 14:29:46 crc kubenswrapper[4805]: I1203 14:29:46.653862 4805 generic.go:334] "Generic (PLEG): container finished" podID="285c21ec-cbb8-4cb7-b5c2-0b27656f7870" containerID="917876c33cc904807afef7d8edc87b2672e031fae006b99501746b08d7943964" exitCode=0 Dec 03 14:29:46 crc kubenswrapper[4805]: I1203 14:29:46.653957 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-sync-27jrz" event={"ID":"285c21ec-cbb8-4cb7-b5c2-0b27656f7870","Type":"ContainerDied","Data":"917876c33cc904807afef7d8edc87b2672e031fae006b99501746b08d7943964"} Dec 03 14:29:46 crc kubenswrapper[4805]: I1203 14:29:46.705538 4805 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d8e0e718-99ce-4813-a824-bf2182f8fdf1" path="/var/lib/kubelet/pods/d8e0e718-99ce-4813-a824-bf2182f8fdf1/volumes" Dec 03 14:29:48 crc kubenswrapper[4805]: I1203 14:29:48.024829 4805 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-sync-27jrz" Dec 03 14:29:48 crc kubenswrapper[4805]: I1203 14:29:48.125134 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/285c21ec-cbb8-4cb7-b5c2-0b27656f7870-combined-ca-bundle\") pod \"285c21ec-cbb8-4cb7-b5c2-0b27656f7870\" (UID: \"285c21ec-cbb8-4cb7-b5c2-0b27656f7870\") " Dec 03 14:29:48 crc kubenswrapper[4805]: I1203 14:29:48.125238 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fzss2\" (UniqueName: \"kubernetes.io/projected/285c21ec-cbb8-4cb7-b5c2-0b27656f7870-kube-api-access-fzss2\") pod \"285c21ec-cbb8-4cb7-b5c2-0b27656f7870\" (UID: \"285c21ec-cbb8-4cb7-b5c2-0b27656f7870\") " Dec 03 14:29:48 crc kubenswrapper[4805]: I1203 14:29:48.125297 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/285c21ec-cbb8-4cb7-b5c2-0b27656f7870-config-data\") pod \"285c21ec-cbb8-4cb7-b5c2-0b27656f7870\" (UID: \"285c21ec-cbb8-4cb7-b5c2-0b27656f7870\") " Dec 03 14:29:48 crc kubenswrapper[4805]: I1203 14:29:48.125391 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/285c21ec-cbb8-4cb7-b5c2-0b27656f7870-db-sync-config-data\") pod \"285c21ec-cbb8-4cb7-b5c2-0b27656f7870\" (UID: \"285c21ec-cbb8-4cb7-b5c2-0b27656f7870\") " Dec 03 14:29:48 crc kubenswrapper[4805]: I1203 14:29:48.131860 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/285c21ec-cbb8-4cb7-b5c2-0b27656f7870-db-sync-config-data" (OuterVolumeSpecName: "db-sync-config-data") pod "285c21ec-cbb8-4cb7-b5c2-0b27656f7870" (UID: "285c21ec-cbb8-4cb7-b5c2-0b27656f7870"). InnerVolumeSpecName "db-sync-config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 14:29:48 crc kubenswrapper[4805]: I1203 14:29:48.132740 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/285c21ec-cbb8-4cb7-b5c2-0b27656f7870-kube-api-access-fzss2" (OuterVolumeSpecName: "kube-api-access-fzss2") pod "285c21ec-cbb8-4cb7-b5c2-0b27656f7870" (UID: "285c21ec-cbb8-4cb7-b5c2-0b27656f7870"). InnerVolumeSpecName "kube-api-access-fzss2". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 14:29:48 crc kubenswrapper[4805]: I1203 14:29:48.154179 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/285c21ec-cbb8-4cb7-b5c2-0b27656f7870-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "285c21ec-cbb8-4cb7-b5c2-0b27656f7870" (UID: "285c21ec-cbb8-4cb7-b5c2-0b27656f7870"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 14:29:48 crc kubenswrapper[4805]: I1203 14:29:48.173034 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/285c21ec-cbb8-4cb7-b5c2-0b27656f7870-config-data" (OuterVolumeSpecName: "config-data") pod "285c21ec-cbb8-4cb7-b5c2-0b27656f7870" (UID: "285c21ec-cbb8-4cb7-b5c2-0b27656f7870"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 14:29:48 crc kubenswrapper[4805]: I1203 14:29:48.227745 4805 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fzss2\" (UniqueName: \"kubernetes.io/projected/285c21ec-cbb8-4cb7-b5c2-0b27656f7870-kube-api-access-fzss2\") on node \"crc\" DevicePath \"\"" Dec 03 14:29:48 crc kubenswrapper[4805]: I1203 14:29:48.227772 4805 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/285c21ec-cbb8-4cb7-b5c2-0b27656f7870-config-data\") on node \"crc\" DevicePath \"\"" Dec 03 14:29:48 crc kubenswrapper[4805]: I1203 14:29:48.227784 4805 reconciler_common.go:293] "Volume detached for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/285c21ec-cbb8-4cb7-b5c2-0b27656f7870-db-sync-config-data\") on node \"crc\" DevicePath \"\"" Dec 03 14:29:48 crc kubenswrapper[4805]: I1203 14:29:48.227793 4805 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/285c21ec-cbb8-4cb7-b5c2-0b27656f7870-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 14:29:48 crc kubenswrapper[4805]: I1203 14:29:48.674676 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-sync-27jrz" event={"ID":"285c21ec-cbb8-4cb7-b5c2-0b27656f7870","Type":"ContainerDied","Data":"b516ae9c15859b5e0423b2035fb3df20f22b2377fa4d008f7157d3153132cd8a"} Dec 03 14:29:48 crc kubenswrapper[4805]: I1203 14:29:48.674746 4805 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="b516ae9c15859b5e0423b2035fb3df20f22b2377fa4d008f7157d3153132cd8a" Dec 03 14:29:48 crc kubenswrapper[4805]: I1203 14:29:48.674948 4805 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-sync-27jrz" Dec 03 14:29:49 crc kubenswrapper[4805]: I1203 14:29:49.126679 4805 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-77585f5f8c-6gbzs"] Dec 03 14:29:49 crc kubenswrapper[4805]: I1203 14:29:49.126895 4805 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-77585f5f8c-6gbzs" podUID="d9710132-a82b-4aa4-82dd-537ae1ea51d6" containerName="dnsmasq-dns" containerID="cri-o://831f6e175f71916762f0e7a3330fef1491b075f46ef72768ce656b8c3fd9af61" gracePeriod=10 Dec 03 14:29:49 crc kubenswrapper[4805]: I1203 14:29:49.128496 4805 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-77585f5f8c-6gbzs" Dec 03 14:29:49 crc kubenswrapper[4805]: I1203 14:29:49.182680 4805 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-7ff5475cc9-mbdbt"] Dec 03 14:29:49 crc kubenswrapper[4805]: E1203 14:29:49.183064 4805 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="285c21ec-cbb8-4cb7-b5c2-0b27656f7870" containerName="glance-db-sync" Dec 03 14:29:49 crc kubenswrapper[4805]: I1203 14:29:49.183075 4805 state_mem.go:107] "Deleted CPUSet assignment" podUID="285c21ec-cbb8-4cb7-b5c2-0b27656f7870" containerName="glance-db-sync" Dec 03 14:29:49 crc kubenswrapper[4805]: E1203 14:29:49.183097 4805 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d8e0e718-99ce-4813-a824-bf2182f8fdf1" containerName="ovn-config" Dec 03 14:29:49 crc kubenswrapper[4805]: I1203 14:29:49.183103 4805 state_mem.go:107] "Deleted CPUSet assignment" podUID="d8e0e718-99ce-4813-a824-bf2182f8fdf1" containerName="ovn-config" Dec 03 14:29:49 crc kubenswrapper[4805]: I1203 14:29:49.183259 4805 memory_manager.go:354] "RemoveStaleState removing state" podUID="d8e0e718-99ce-4813-a824-bf2182f8fdf1" containerName="ovn-config" Dec 03 14:29:49 crc kubenswrapper[4805]: I1203 14:29:49.183289 4805 memory_manager.go:354] "RemoveStaleState removing state" podUID="285c21ec-cbb8-4cb7-b5c2-0b27656f7870" containerName="glance-db-sync" Dec 03 14:29:49 crc kubenswrapper[4805]: I1203 14:29:49.184211 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7ff5475cc9-mbdbt" Dec 03 14:29:49 crc kubenswrapper[4805]: I1203 14:29:49.205605 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-7ff5475cc9-mbdbt"] Dec 03 14:29:49 crc kubenswrapper[4805]: I1203 14:29:49.244000 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1ca67122-920a-43a6-a434-82ec573d0160-config\") pod \"dnsmasq-dns-7ff5475cc9-mbdbt\" (UID: \"1ca67122-920a-43a6-a434-82ec573d0160\") " pod="openstack/dnsmasq-dns-7ff5475cc9-mbdbt" Dec 03 14:29:49 crc kubenswrapper[4805]: I1203 14:29:49.247890 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/1ca67122-920a-43a6-a434-82ec573d0160-ovsdbserver-nb\") pod \"dnsmasq-dns-7ff5475cc9-mbdbt\" (UID: \"1ca67122-920a-43a6-a434-82ec573d0160\") " pod="openstack/dnsmasq-dns-7ff5475cc9-mbdbt" Dec 03 14:29:49 crc kubenswrapper[4805]: I1203 14:29:49.248288 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/1ca67122-920a-43a6-a434-82ec573d0160-ovsdbserver-sb\") pod \"dnsmasq-dns-7ff5475cc9-mbdbt\" (UID: \"1ca67122-920a-43a6-a434-82ec573d0160\") " pod="openstack/dnsmasq-dns-7ff5475cc9-mbdbt" Dec 03 14:29:49 crc kubenswrapper[4805]: I1203 14:29:49.248383 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/1ca67122-920a-43a6-a434-82ec573d0160-dns-svc\") pod \"dnsmasq-dns-7ff5475cc9-mbdbt\" (UID: \"1ca67122-920a-43a6-a434-82ec573d0160\") " pod="openstack/dnsmasq-dns-7ff5475cc9-mbdbt" Dec 03 14:29:49 crc kubenswrapper[4805]: I1203 14:29:49.248481 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/1ca67122-920a-43a6-a434-82ec573d0160-dns-swift-storage-0\") pod \"dnsmasq-dns-7ff5475cc9-mbdbt\" (UID: \"1ca67122-920a-43a6-a434-82ec573d0160\") " pod="openstack/dnsmasq-dns-7ff5475cc9-mbdbt" Dec 03 14:29:49 crc kubenswrapper[4805]: I1203 14:29:49.248605 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kqtxj\" (UniqueName: \"kubernetes.io/projected/1ca67122-920a-43a6-a434-82ec573d0160-kube-api-access-kqtxj\") pod \"dnsmasq-dns-7ff5475cc9-mbdbt\" (UID: \"1ca67122-920a-43a6-a434-82ec573d0160\") " pod="openstack/dnsmasq-dns-7ff5475cc9-mbdbt" Dec 03 14:29:49 crc kubenswrapper[4805]: I1203 14:29:49.350134 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/1ca67122-920a-43a6-a434-82ec573d0160-ovsdbserver-nb\") pod \"dnsmasq-dns-7ff5475cc9-mbdbt\" (UID: \"1ca67122-920a-43a6-a434-82ec573d0160\") " pod="openstack/dnsmasq-dns-7ff5475cc9-mbdbt" Dec 03 14:29:49 crc kubenswrapper[4805]: I1203 14:29:49.350224 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/1ca67122-920a-43a6-a434-82ec573d0160-ovsdbserver-sb\") pod \"dnsmasq-dns-7ff5475cc9-mbdbt\" (UID: \"1ca67122-920a-43a6-a434-82ec573d0160\") " pod="openstack/dnsmasq-dns-7ff5475cc9-mbdbt" Dec 03 14:29:49 crc kubenswrapper[4805]: I1203 14:29:49.350249 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/1ca67122-920a-43a6-a434-82ec573d0160-dns-svc\") pod \"dnsmasq-dns-7ff5475cc9-mbdbt\" (UID: \"1ca67122-920a-43a6-a434-82ec573d0160\") " pod="openstack/dnsmasq-dns-7ff5475cc9-mbdbt" Dec 03 14:29:49 crc kubenswrapper[4805]: I1203 14:29:49.350280 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/1ca67122-920a-43a6-a434-82ec573d0160-dns-swift-storage-0\") pod \"dnsmasq-dns-7ff5475cc9-mbdbt\" (UID: \"1ca67122-920a-43a6-a434-82ec573d0160\") " pod="openstack/dnsmasq-dns-7ff5475cc9-mbdbt" Dec 03 14:29:49 crc kubenswrapper[4805]: I1203 14:29:49.350298 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kqtxj\" (UniqueName: \"kubernetes.io/projected/1ca67122-920a-43a6-a434-82ec573d0160-kube-api-access-kqtxj\") pod \"dnsmasq-dns-7ff5475cc9-mbdbt\" (UID: \"1ca67122-920a-43a6-a434-82ec573d0160\") " pod="openstack/dnsmasq-dns-7ff5475cc9-mbdbt" Dec 03 14:29:49 crc kubenswrapper[4805]: I1203 14:29:49.350328 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1ca67122-920a-43a6-a434-82ec573d0160-config\") pod \"dnsmasq-dns-7ff5475cc9-mbdbt\" (UID: \"1ca67122-920a-43a6-a434-82ec573d0160\") " pod="openstack/dnsmasq-dns-7ff5475cc9-mbdbt" Dec 03 14:29:49 crc kubenswrapper[4805]: I1203 14:29:49.351399 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1ca67122-920a-43a6-a434-82ec573d0160-config\") pod \"dnsmasq-dns-7ff5475cc9-mbdbt\" (UID: \"1ca67122-920a-43a6-a434-82ec573d0160\") " pod="openstack/dnsmasq-dns-7ff5475cc9-mbdbt" Dec 03 14:29:49 crc kubenswrapper[4805]: I1203 14:29:49.351404 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/1ca67122-920a-43a6-a434-82ec573d0160-ovsdbserver-sb\") pod \"dnsmasq-dns-7ff5475cc9-mbdbt\" (UID: \"1ca67122-920a-43a6-a434-82ec573d0160\") " pod="openstack/dnsmasq-dns-7ff5475cc9-mbdbt" Dec 03 14:29:49 crc kubenswrapper[4805]: I1203 14:29:49.351446 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/1ca67122-920a-43a6-a434-82ec573d0160-dns-swift-storage-0\") pod \"dnsmasq-dns-7ff5475cc9-mbdbt\" (UID: \"1ca67122-920a-43a6-a434-82ec573d0160\") " pod="openstack/dnsmasq-dns-7ff5475cc9-mbdbt" Dec 03 14:29:49 crc kubenswrapper[4805]: I1203 14:29:49.351570 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/1ca67122-920a-43a6-a434-82ec573d0160-dns-svc\") pod \"dnsmasq-dns-7ff5475cc9-mbdbt\" (UID: \"1ca67122-920a-43a6-a434-82ec573d0160\") " pod="openstack/dnsmasq-dns-7ff5475cc9-mbdbt" Dec 03 14:29:49 crc kubenswrapper[4805]: I1203 14:29:49.352005 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/1ca67122-920a-43a6-a434-82ec573d0160-ovsdbserver-nb\") pod \"dnsmasq-dns-7ff5475cc9-mbdbt\" (UID: \"1ca67122-920a-43a6-a434-82ec573d0160\") " pod="openstack/dnsmasq-dns-7ff5475cc9-mbdbt" Dec 03 14:29:49 crc kubenswrapper[4805]: I1203 14:29:49.368431 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kqtxj\" (UniqueName: \"kubernetes.io/projected/1ca67122-920a-43a6-a434-82ec573d0160-kube-api-access-kqtxj\") pod \"dnsmasq-dns-7ff5475cc9-mbdbt\" (UID: \"1ca67122-920a-43a6-a434-82ec573d0160\") " pod="openstack/dnsmasq-dns-7ff5475cc9-mbdbt" Dec 03 14:29:49 crc kubenswrapper[4805]: I1203 14:29:49.508555 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7ff5475cc9-mbdbt" Dec 03 14:29:49 crc kubenswrapper[4805]: I1203 14:29:49.600045 4805 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-77585f5f8c-6gbzs" Dec 03 14:29:49 crc kubenswrapper[4805]: I1203 14:29:49.700823 4805 generic.go:334] "Generic (PLEG): container finished" podID="d9710132-a82b-4aa4-82dd-537ae1ea51d6" containerID="831f6e175f71916762f0e7a3330fef1491b075f46ef72768ce656b8c3fd9af61" exitCode=0 Dec 03 14:29:49 crc kubenswrapper[4805]: I1203 14:29:49.700889 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-77585f5f8c-6gbzs" event={"ID":"d9710132-a82b-4aa4-82dd-537ae1ea51d6","Type":"ContainerDied","Data":"831f6e175f71916762f0e7a3330fef1491b075f46ef72768ce656b8c3fd9af61"} Dec 03 14:29:49 crc kubenswrapper[4805]: I1203 14:29:49.700918 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-77585f5f8c-6gbzs" event={"ID":"d9710132-a82b-4aa4-82dd-537ae1ea51d6","Type":"ContainerDied","Data":"401345d095f6cb7e1da4b63bab60a175de7d4eab8d464225b019cb16242e4ae5"} Dec 03 14:29:49 crc kubenswrapper[4805]: I1203 14:29:49.700938 4805 scope.go:117] "RemoveContainer" containerID="831f6e175f71916762f0e7a3330fef1491b075f46ef72768ce656b8c3fd9af61" Dec 03 14:29:49 crc kubenswrapper[4805]: I1203 14:29:49.701078 4805 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-77585f5f8c-6gbzs" Dec 03 14:29:49 crc kubenswrapper[4805]: I1203 14:29:49.740211 4805 scope.go:117] "RemoveContainer" containerID="20128c805c35289c7ee1383415b1c93838143347edbb9a9aad4a474420a28b83" Dec 03 14:29:49 crc kubenswrapper[4805]: I1203 14:29:49.756709 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/d9710132-a82b-4aa4-82dd-537ae1ea51d6-ovsdbserver-sb\") pod \"d9710132-a82b-4aa4-82dd-537ae1ea51d6\" (UID: \"d9710132-a82b-4aa4-82dd-537ae1ea51d6\") " Dec 03 14:29:49 crc kubenswrapper[4805]: I1203 14:29:49.756852 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kgs8h\" (UniqueName: \"kubernetes.io/projected/d9710132-a82b-4aa4-82dd-537ae1ea51d6-kube-api-access-kgs8h\") pod \"d9710132-a82b-4aa4-82dd-537ae1ea51d6\" (UID: \"d9710132-a82b-4aa4-82dd-537ae1ea51d6\") " Dec 03 14:29:49 crc kubenswrapper[4805]: I1203 14:29:49.756886 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d9710132-a82b-4aa4-82dd-537ae1ea51d6-config\") pod \"d9710132-a82b-4aa4-82dd-537ae1ea51d6\" (UID: \"d9710132-a82b-4aa4-82dd-537ae1ea51d6\") " Dec 03 14:29:49 crc kubenswrapper[4805]: I1203 14:29:49.757019 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/d9710132-a82b-4aa4-82dd-537ae1ea51d6-ovsdbserver-nb\") pod \"d9710132-a82b-4aa4-82dd-537ae1ea51d6\" (UID: \"d9710132-a82b-4aa4-82dd-537ae1ea51d6\") " Dec 03 14:29:49 crc kubenswrapper[4805]: I1203 14:29:49.757038 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/d9710132-a82b-4aa4-82dd-537ae1ea51d6-dns-svc\") pod \"d9710132-a82b-4aa4-82dd-537ae1ea51d6\" (UID: \"d9710132-a82b-4aa4-82dd-537ae1ea51d6\") " Dec 03 14:29:49 crc kubenswrapper[4805]: I1203 14:29:49.757081 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/d9710132-a82b-4aa4-82dd-537ae1ea51d6-dns-swift-storage-0\") pod \"d9710132-a82b-4aa4-82dd-537ae1ea51d6\" (UID: \"d9710132-a82b-4aa4-82dd-537ae1ea51d6\") " Dec 03 14:29:49 crc kubenswrapper[4805]: I1203 14:29:49.762895 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d9710132-a82b-4aa4-82dd-537ae1ea51d6-kube-api-access-kgs8h" (OuterVolumeSpecName: "kube-api-access-kgs8h") pod "d9710132-a82b-4aa4-82dd-537ae1ea51d6" (UID: "d9710132-a82b-4aa4-82dd-537ae1ea51d6"). InnerVolumeSpecName "kube-api-access-kgs8h". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 14:29:49 crc kubenswrapper[4805]: I1203 14:29:49.768609 4805 scope.go:117] "RemoveContainer" containerID="831f6e175f71916762f0e7a3330fef1491b075f46ef72768ce656b8c3fd9af61" Dec 03 14:29:49 crc kubenswrapper[4805]: E1203 14:29:49.769062 4805 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"831f6e175f71916762f0e7a3330fef1491b075f46ef72768ce656b8c3fd9af61\": container with ID starting with 831f6e175f71916762f0e7a3330fef1491b075f46ef72768ce656b8c3fd9af61 not found: ID does not exist" containerID="831f6e175f71916762f0e7a3330fef1491b075f46ef72768ce656b8c3fd9af61" Dec 03 14:29:49 crc kubenswrapper[4805]: I1203 14:29:49.769114 4805 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"831f6e175f71916762f0e7a3330fef1491b075f46ef72768ce656b8c3fd9af61"} err="failed to get container status \"831f6e175f71916762f0e7a3330fef1491b075f46ef72768ce656b8c3fd9af61\": rpc error: code = NotFound desc = could not find container \"831f6e175f71916762f0e7a3330fef1491b075f46ef72768ce656b8c3fd9af61\": container with ID starting with 831f6e175f71916762f0e7a3330fef1491b075f46ef72768ce656b8c3fd9af61 not found: ID does not exist" Dec 03 14:29:49 crc kubenswrapper[4805]: I1203 14:29:49.769141 4805 scope.go:117] "RemoveContainer" containerID="20128c805c35289c7ee1383415b1c93838143347edbb9a9aad4a474420a28b83" Dec 03 14:29:49 crc kubenswrapper[4805]: E1203 14:29:49.769473 4805 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"20128c805c35289c7ee1383415b1c93838143347edbb9a9aad4a474420a28b83\": container with ID starting with 20128c805c35289c7ee1383415b1c93838143347edbb9a9aad4a474420a28b83 not found: ID does not exist" containerID="20128c805c35289c7ee1383415b1c93838143347edbb9a9aad4a474420a28b83" Dec 03 14:29:49 crc kubenswrapper[4805]: I1203 14:29:49.769515 4805 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"20128c805c35289c7ee1383415b1c93838143347edbb9a9aad4a474420a28b83"} err="failed to get container status \"20128c805c35289c7ee1383415b1c93838143347edbb9a9aad4a474420a28b83\": rpc error: code = NotFound desc = could not find container \"20128c805c35289c7ee1383415b1c93838143347edbb9a9aad4a474420a28b83\": container with ID starting with 20128c805c35289c7ee1383415b1c93838143347edbb9a9aad4a474420a28b83 not found: ID does not exist" Dec 03 14:29:49 crc kubenswrapper[4805]: I1203 14:29:49.803966 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d9710132-a82b-4aa4-82dd-537ae1ea51d6-config" (OuterVolumeSpecName: "config") pod "d9710132-a82b-4aa4-82dd-537ae1ea51d6" (UID: "d9710132-a82b-4aa4-82dd-537ae1ea51d6"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 14:29:49 crc kubenswrapper[4805]: I1203 14:29:49.803974 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d9710132-a82b-4aa4-82dd-537ae1ea51d6-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "d9710132-a82b-4aa4-82dd-537ae1ea51d6" (UID: "d9710132-a82b-4aa4-82dd-537ae1ea51d6"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 14:29:49 crc kubenswrapper[4805]: I1203 14:29:49.810561 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d9710132-a82b-4aa4-82dd-537ae1ea51d6-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "d9710132-a82b-4aa4-82dd-537ae1ea51d6" (UID: "d9710132-a82b-4aa4-82dd-537ae1ea51d6"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 14:29:49 crc kubenswrapper[4805]: I1203 14:29:49.814119 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d9710132-a82b-4aa4-82dd-537ae1ea51d6-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "d9710132-a82b-4aa4-82dd-537ae1ea51d6" (UID: "d9710132-a82b-4aa4-82dd-537ae1ea51d6"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 14:29:49 crc kubenswrapper[4805]: I1203 14:29:49.814628 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d9710132-a82b-4aa4-82dd-537ae1ea51d6-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "d9710132-a82b-4aa4-82dd-537ae1ea51d6" (UID: "d9710132-a82b-4aa4-82dd-537ae1ea51d6"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 14:29:49 crc kubenswrapper[4805]: I1203 14:29:49.859513 4805 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/d9710132-a82b-4aa4-82dd-537ae1ea51d6-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Dec 03 14:29:49 crc kubenswrapper[4805]: I1203 14:29:49.859552 4805 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kgs8h\" (UniqueName: \"kubernetes.io/projected/d9710132-a82b-4aa4-82dd-537ae1ea51d6-kube-api-access-kgs8h\") on node \"crc\" DevicePath \"\"" Dec 03 14:29:49 crc kubenswrapper[4805]: I1203 14:29:49.859568 4805 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d9710132-a82b-4aa4-82dd-537ae1ea51d6-config\") on node \"crc\" DevicePath \"\"" Dec 03 14:29:49 crc kubenswrapper[4805]: I1203 14:29:49.859612 4805 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/d9710132-a82b-4aa4-82dd-537ae1ea51d6-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Dec 03 14:29:49 crc kubenswrapper[4805]: I1203 14:29:49.859621 4805 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/d9710132-a82b-4aa4-82dd-537ae1ea51d6-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 03 14:29:49 crc kubenswrapper[4805]: I1203 14:29:49.859629 4805 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/d9710132-a82b-4aa4-82dd-537ae1ea51d6-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Dec 03 14:29:50 crc kubenswrapper[4805]: I1203 14:29:50.003516 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-7ff5475cc9-mbdbt"] Dec 03 14:29:50 crc kubenswrapper[4805]: W1203 14:29:50.012688 4805 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod1ca67122_920a_43a6_a434_82ec573d0160.slice/crio-98681326060f9f03fd0c1123246638539b81883860fd056234e23b2a9170ba45 WatchSource:0}: Error finding container 98681326060f9f03fd0c1123246638539b81883860fd056234e23b2a9170ba45: Status 404 returned error can't find the container with id 98681326060f9f03fd0c1123246638539b81883860fd056234e23b2a9170ba45 Dec 03 14:29:50 crc kubenswrapper[4805]: I1203 14:29:50.043465 4805 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-77585f5f8c-6gbzs"] Dec 03 14:29:50 crc kubenswrapper[4805]: I1203 14:29:50.052759 4805 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-77585f5f8c-6gbzs"] Dec 03 14:29:50 crc kubenswrapper[4805]: I1203 14:29:50.707001 4805 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d9710132-a82b-4aa4-82dd-537ae1ea51d6" path="/var/lib/kubelet/pods/d9710132-a82b-4aa4-82dd-537ae1ea51d6/volumes" Dec 03 14:29:50 crc kubenswrapper[4805]: I1203 14:29:50.720719 4805 generic.go:334] "Generic (PLEG): container finished" podID="1ca67122-920a-43a6-a434-82ec573d0160" containerID="392489bc555f119580c8997d1ea4e8089d6644ef798f7247ca27ec78f95e9855" exitCode=0 Dec 03 14:29:50 crc kubenswrapper[4805]: I1203 14:29:50.720778 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7ff5475cc9-mbdbt" event={"ID":"1ca67122-920a-43a6-a434-82ec573d0160","Type":"ContainerDied","Data":"392489bc555f119580c8997d1ea4e8089d6644ef798f7247ca27ec78f95e9855"} Dec 03 14:29:50 crc kubenswrapper[4805]: I1203 14:29:50.720887 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7ff5475cc9-mbdbt" event={"ID":"1ca67122-920a-43a6-a434-82ec573d0160","Type":"ContainerStarted","Data":"98681326060f9f03fd0c1123246638539b81883860fd056234e23b2a9170ba45"} Dec 03 14:29:51 crc kubenswrapper[4805]: I1203 14:29:51.738229 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7ff5475cc9-mbdbt" event={"ID":"1ca67122-920a-43a6-a434-82ec573d0160","Type":"ContainerStarted","Data":"f9c31ba4a3601f19c3ce752a2e09ba3a90a39dfe55c8ca130f6564acd0157752"} Dec 03 14:29:51 crc kubenswrapper[4805]: I1203 14:29:51.738545 4805 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-7ff5475cc9-mbdbt" Dec 03 14:29:51 crc kubenswrapper[4805]: I1203 14:29:51.759744 4805 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-7ff5475cc9-mbdbt" podStartSLOduration=2.759727346 podStartE2EDuration="2.759727346s" podCreationTimestamp="2025-12-03 14:29:49 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 14:29:51.754213006 +0000 UTC m=+1221.417129939" watchObservedRunningTime="2025-12-03 14:29:51.759727346 +0000 UTC m=+1221.422644259" Dec 03 14:29:52 crc kubenswrapper[4805]: I1203 14:29:52.071119 4805 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/rabbitmq-server-0" Dec 03 14:29:52 crc kubenswrapper[4805]: I1203 14:29:52.365990 4805 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/rabbitmq-cell1-server-0" Dec 03 14:29:52 crc kubenswrapper[4805]: I1203 14:29:52.446979 4805 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-db-create-q7qlk"] Dec 03 14:29:52 crc kubenswrapper[4805]: E1203 14:29:52.447429 4805 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d9710132-a82b-4aa4-82dd-537ae1ea51d6" containerName="init" Dec 03 14:29:52 crc kubenswrapper[4805]: I1203 14:29:52.447446 4805 state_mem.go:107] "Deleted CPUSet assignment" podUID="d9710132-a82b-4aa4-82dd-537ae1ea51d6" containerName="init" Dec 03 14:29:52 crc kubenswrapper[4805]: E1203 14:29:52.447460 4805 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d9710132-a82b-4aa4-82dd-537ae1ea51d6" containerName="dnsmasq-dns" Dec 03 14:29:52 crc kubenswrapper[4805]: I1203 14:29:52.447467 4805 state_mem.go:107] "Deleted CPUSet assignment" podUID="d9710132-a82b-4aa4-82dd-537ae1ea51d6" containerName="dnsmasq-dns" Dec 03 14:29:52 crc kubenswrapper[4805]: I1203 14:29:52.447657 4805 memory_manager.go:354] "RemoveStaleState removing state" podUID="d9710132-a82b-4aa4-82dd-537ae1ea51d6" containerName="dnsmasq-dns" Dec 03 14:29:52 crc kubenswrapper[4805]: I1203 14:29:52.448383 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-create-q7qlk" Dec 03 14:29:52 crc kubenswrapper[4805]: I1203 14:29:52.461720 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-db-create-q7qlk"] Dec 03 14:29:52 crc kubenswrapper[4805]: I1203 14:29:52.522250 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/10163a58-aa6e-4988-8b33-93f8cffa4cc0-operator-scripts\") pod \"cinder-db-create-q7qlk\" (UID: \"10163a58-aa6e-4988-8b33-93f8cffa4cc0\") " pod="openstack/cinder-db-create-q7qlk" Dec 03 14:29:52 crc kubenswrapper[4805]: I1203 14:29:52.522514 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gfgzp\" (UniqueName: \"kubernetes.io/projected/10163a58-aa6e-4988-8b33-93f8cffa4cc0-kube-api-access-gfgzp\") pod \"cinder-db-create-q7qlk\" (UID: \"10163a58-aa6e-4988-8b33-93f8cffa4cc0\") " pod="openstack/cinder-db-create-q7qlk" Dec 03 14:29:52 crc kubenswrapper[4805]: I1203 14:29:52.536911 4805 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-db-create-mhn6v"] Dec 03 14:29:52 crc kubenswrapper[4805]: I1203 14:29:52.538232 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-create-mhn6v" Dec 03 14:29:52 crc kubenswrapper[4805]: I1203 14:29:52.564110 4805 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-7ac1-account-create-update-6jk6w"] Dec 03 14:29:52 crc kubenswrapper[4805]: I1203 14:29:52.565390 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-7ac1-account-create-update-6jk6w" Dec 03 14:29:52 crc kubenswrapper[4805]: I1203 14:29:52.568767 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-db-secret" Dec 03 14:29:52 crc kubenswrapper[4805]: I1203 14:29:52.578083 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-db-create-mhn6v"] Dec 03 14:29:52 crc kubenswrapper[4805]: I1203 14:29:52.589170 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-7ac1-account-create-update-6jk6w"] Dec 03 14:29:52 crc kubenswrapper[4805]: I1203 14:29:52.621003 4805 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-3c27-account-create-update-79pjf"] Dec 03 14:29:52 crc kubenswrapper[4805]: I1203 14:29:52.623191 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-3c27-account-create-update-79pjf" Dec 03 14:29:52 crc kubenswrapper[4805]: I1203 14:29:52.624009 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gfgzp\" (UniqueName: \"kubernetes.io/projected/10163a58-aa6e-4988-8b33-93f8cffa4cc0-kube-api-access-gfgzp\") pod \"cinder-db-create-q7qlk\" (UID: \"10163a58-aa6e-4988-8b33-93f8cffa4cc0\") " pod="openstack/cinder-db-create-q7qlk" Dec 03 14:29:52 crc kubenswrapper[4805]: I1203 14:29:52.624079 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/b27b4fde-29ff-421d-a6be-5d259bfe8016-operator-scripts\") pod \"barbican-7ac1-account-create-update-6jk6w\" (UID: \"b27b4fde-29ff-421d-a6be-5d259bfe8016\") " pod="openstack/barbican-7ac1-account-create-update-6jk6w" Dec 03 14:29:52 crc kubenswrapper[4805]: I1203 14:29:52.624106 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/10163a58-aa6e-4988-8b33-93f8cffa4cc0-operator-scripts\") pod \"cinder-db-create-q7qlk\" (UID: \"10163a58-aa6e-4988-8b33-93f8cffa4cc0\") " pod="openstack/cinder-db-create-q7qlk" Dec 03 14:29:52 crc kubenswrapper[4805]: I1203 14:29:52.624130 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8rz8t\" (UniqueName: \"kubernetes.io/projected/b27b4fde-29ff-421d-a6be-5d259bfe8016-kube-api-access-8rz8t\") pod \"barbican-7ac1-account-create-update-6jk6w\" (UID: \"b27b4fde-29ff-421d-a6be-5d259bfe8016\") " pod="openstack/barbican-7ac1-account-create-update-6jk6w" Dec 03 14:29:52 crc kubenswrapper[4805]: I1203 14:29:52.624150 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-chbc4\" (UniqueName: \"kubernetes.io/projected/aead85c7-af73-44d0-b4dc-5cd98d6439b3-kube-api-access-chbc4\") pod \"barbican-db-create-mhn6v\" (UID: \"aead85c7-af73-44d0-b4dc-5cd98d6439b3\") " pod="openstack/barbican-db-create-mhn6v" Dec 03 14:29:52 crc kubenswrapper[4805]: I1203 14:29:52.624181 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/aead85c7-af73-44d0-b4dc-5cd98d6439b3-operator-scripts\") pod \"barbican-db-create-mhn6v\" (UID: \"aead85c7-af73-44d0-b4dc-5cd98d6439b3\") " pod="openstack/barbican-db-create-mhn6v" Dec 03 14:29:52 crc kubenswrapper[4805]: I1203 14:29:52.625032 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/10163a58-aa6e-4988-8b33-93f8cffa4cc0-operator-scripts\") pod \"cinder-db-create-q7qlk\" (UID: \"10163a58-aa6e-4988-8b33-93f8cffa4cc0\") " pod="openstack/cinder-db-create-q7qlk" Dec 03 14:29:52 crc kubenswrapper[4805]: I1203 14:29:52.633792 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-3c27-account-create-update-79pjf"] Dec 03 14:29:52 crc kubenswrapper[4805]: I1203 14:29:52.635942 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-db-secret" Dec 03 14:29:52 crc kubenswrapper[4805]: I1203 14:29:52.652416 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gfgzp\" (UniqueName: \"kubernetes.io/projected/10163a58-aa6e-4988-8b33-93f8cffa4cc0-kube-api-access-gfgzp\") pod \"cinder-db-create-q7qlk\" (UID: \"10163a58-aa6e-4988-8b33-93f8cffa4cc0\") " pod="openstack/cinder-db-create-q7qlk" Dec 03 14:29:52 crc kubenswrapper[4805]: I1203 14:29:52.707038 4805 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-db-sync-75wdv"] Dec 03 14:29:52 crc kubenswrapper[4805]: I1203 14:29:52.710284 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-sync-75wdv" Dec 03 14:29:52 crc kubenswrapper[4805]: I1203 14:29:52.715740 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-scripts" Dec 03 14:29:52 crc kubenswrapper[4805]: I1203 14:29:52.715879 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-db-sync-75wdv"] Dec 03 14:29:52 crc kubenswrapper[4805]: I1203 14:29:52.715945 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone" Dec 03 14:29:52 crc kubenswrapper[4805]: I1203 14:29:52.716017 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-keystone-dockercfg-5hcz5" Dec 03 14:29:52 crc kubenswrapper[4805]: I1203 14:29:52.716190 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-config-data" Dec 03 14:29:52 crc kubenswrapper[4805]: I1203 14:29:52.725463 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/aead85c7-af73-44d0-b4dc-5cd98d6439b3-operator-scripts\") pod \"barbican-db-create-mhn6v\" (UID: \"aead85c7-af73-44d0-b4dc-5cd98d6439b3\") " pod="openstack/barbican-db-create-mhn6v" Dec 03 14:29:52 crc kubenswrapper[4805]: I1203 14:29:52.725546 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lx9th\" (UniqueName: \"kubernetes.io/projected/eda36a78-7b3a-43f8-858e-fc6b8fdd442a-kube-api-access-lx9th\") pod \"cinder-3c27-account-create-update-79pjf\" (UID: \"eda36a78-7b3a-43f8-858e-fc6b8fdd442a\") " pod="openstack/cinder-3c27-account-create-update-79pjf" Dec 03 14:29:52 crc kubenswrapper[4805]: I1203 14:29:52.725649 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/b27b4fde-29ff-421d-a6be-5d259bfe8016-operator-scripts\") pod \"barbican-7ac1-account-create-update-6jk6w\" (UID: \"b27b4fde-29ff-421d-a6be-5d259bfe8016\") " pod="openstack/barbican-7ac1-account-create-update-6jk6w" Dec 03 14:29:52 crc kubenswrapper[4805]: I1203 14:29:52.725695 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8rz8t\" (UniqueName: \"kubernetes.io/projected/b27b4fde-29ff-421d-a6be-5d259bfe8016-kube-api-access-8rz8t\") pod \"barbican-7ac1-account-create-update-6jk6w\" (UID: \"b27b4fde-29ff-421d-a6be-5d259bfe8016\") " pod="openstack/barbican-7ac1-account-create-update-6jk6w" Dec 03 14:29:52 crc kubenswrapper[4805]: I1203 14:29:52.725729 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/eda36a78-7b3a-43f8-858e-fc6b8fdd442a-operator-scripts\") pod \"cinder-3c27-account-create-update-79pjf\" (UID: \"eda36a78-7b3a-43f8-858e-fc6b8fdd442a\") " pod="openstack/cinder-3c27-account-create-update-79pjf" Dec 03 14:29:52 crc kubenswrapper[4805]: I1203 14:29:52.725756 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-chbc4\" (UniqueName: \"kubernetes.io/projected/aead85c7-af73-44d0-b4dc-5cd98d6439b3-kube-api-access-chbc4\") pod \"barbican-db-create-mhn6v\" (UID: \"aead85c7-af73-44d0-b4dc-5cd98d6439b3\") " pod="openstack/barbican-db-create-mhn6v" Dec 03 14:29:52 crc kubenswrapper[4805]: I1203 14:29:52.726418 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/aead85c7-af73-44d0-b4dc-5cd98d6439b3-operator-scripts\") pod \"barbican-db-create-mhn6v\" (UID: \"aead85c7-af73-44d0-b4dc-5cd98d6439b3\") " pod="openstack/barbican-db-create-mhn6v" Dec 03 14:29:52 crc kubenswrapper[4805]: I1203 14:29:52.726529 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/b27b4fde-29ff-421d-a6be-5d259bfe8016-operator-scripts\") pod \"barbican-7ac1-account-create-update-6jk6w\" (UID: \"b27b4fde-29ff-421d-a6be-5d259bfe8016\") " pod="openstack/barbican-7ac1-account-create-update-6jk6w" Dec 03 14:29:52 crc kubenswrapper[4805]: I1203 14:29:52.729766 4805 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/heat-db-create-ztcqj"] Dec 03 14:29:52 crc kubenswrapper[4805]: I1203 14:29:52.730831 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-db-create-ztcqj" Dec 03 14:29:52 crc kubenswrapper[4805]: I1203 14:29:52.747716 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8rz8t\" (UniqueName: \"kubernetes.io/projected/b27b4fde-29ff-421d-a6be-5d259bfe8016-kube-api-access-8rz8t\") pod \"barbican-7ac1-account-create-update-6jk6w\" (UID: \"b27b4fde-29ff-421d-a6be-5d259bfe8016\") " pod="openstack/barbican-7ac1-account-create-update-6jk6w" Dec 03 14:29:52 crc kubenswrapper[4805]: I1203 14:29:52.748891 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-db-create-ztcqj"] Dec 03 14:29:52 crc kubenswrapper[4805]: I1203 14:29:52.753299 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-chbc4\" (UniqueName: \"kubernetes.io/projected/aead85c7-af73-44d0-b4dc-5cd98d6439b3-kube-api-access-chbc4\") pod \"barbican-db-create-mhn6v\" (UID: \"aead85c7-af73-44d0-b4dc-5cd98d6439b3\") " pod="openstack/barbican-db-create-mhn6v" Dec 03 14:29:52 crc kubenswrapper[4805]: I1203 14:29:52.781332 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-create-q7qlk" Dec 03 14:29:52 crc kubenswrapper[4805]: I1203 14:29:52.828225 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/eda36a78-7b3a-43f8-858e-fc6b8fdd442a-operator-scripts\") pod \"cinder-3c27-account-create-update-79pjf\" (UID: \"eda36a78-7b3a-43f8-858e-fc6b8fdd442a\") " pod="openstack/cinder-3c27-account-create-update-79pjf" Dec 03 14:29:52 crc kubenswrapper[4805]: I1203 14:29:52.828603 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-56b6j\" (UniqueName: \"kubernetes.io/projected/c15336c7-d573-432f-8b4c-ba5a39c82cf8-kube-api-access-56b6j\") pod \"keystone-db-sync-75wdv\" (UID: \"c15336c7-d573-432f-8b4c-ba5a39c82cf8\") " pod="openstack/keystone-db-sync-75wdv" Dec 03 14:29:52 crc kubenswrapper[4805]: I1203 14:29:52.828632 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-b686t\" (UniqueName: \"kubernetes.io/projected/d319f692-2f37-4a01-8e6c-bbf0e7b23c8e-kube-api-access-b686t\") pod \"heat-db-create-ztcqj\" (UID: \"d319f692-2f37-4a01-8e6c-bbf0e7b23c8e\") " pod="openstack/heat-db-create-ztcqj" Dec 03 14:29:52 crc kubenswrapper[4805]: I1203 14:29:52.828719 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lx9th\" (UniqueName: \"kubernetes.io/projected/eda36a78-7b3a-43f8-858e-fc6b8fdd442a-kube-api-access-lx9th\") pod \"cinder-3c27-account-create-update-79pjf\" (UID: \"eda36a78-7b3a-43f8-858e-fc6b8fdd442a\") " pod="openstack/cinder-3c27-account-create-update-79pjf" Dec 03 14:29:52 crc kubenswrapper[4805]: I1203 14:29:52.828804 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c15336c7-d573-432f-8b4c-ba5a39c82cf8-config-data\") pod \"keystone-db-sync-75wdv\" (UID: \"c15336c7-d573-432f-8b4c-ba5a39c82cf8\") " pod="openstack/keystone-db-sync-75wdv" Dec 03 14:29:52 crc kubenswrapper[4805]: I1203 14:29:52.828848 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c15336c7-d573-432f-8b4c-ba5a39c82cf8-combined-ca-bundle\") pod \"keystone-db-sync-75wdv\" (UID: \"c15336c7-d573-432f-8b4c-ba5a39c82cf8\") " pod="openstack/keystone-db-sync-75wdv" Dec 03 14:29:52 crc kubenswrapper[4805]: I1203 14:29:52.828892 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/d319f692-2f37-4a01-8e6c-bbf0e7b23c8e-operator-scripts\") pod \"heat-db-create-ztcqj\" (UID: \"d319f692-2f37-4a01-8e6c-bbf0e7b23c8e\") " pod="openstack/heat-db-create-ztcqj" Dec 03 14:29:52 crc kubenswrapper[4805]: I1203 14:29:52.829321 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/eda36a78-7b3a-43f8-858e-fc6b8fdd442a-operator-scripts\") pod \"cinder-3c27-account-create-update-79pjf\" (UID: \"eda36a78-7b3a-43f8-858e-fc6b8fdd442a\") " pod="openstack/cinder-3c27-account-create-update-79pjf" Dec 03 14:29:52 crc kubenswrapper[4805]: I1203 14:29:52.849290 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lx9th\" (UniqueName: \"kubernetes.io/projected/eda36a78-7b3a-43f8-858e-fc6b8fdd442a-kube-api-access-lx9th\") pod \"cinder-3c27-account-create-update-79pjf\" (UID: \"eda36a78-7b3a-43f8-858e-fc6b8fdd442a\") " pod="openstack/cinder-3c27-account-create-update-79pjf" Dec 03 14:29:52 crc kubenswrapper[4805]: I1203 14:29:52.863082 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-create-mhn6v" Dec 03 14:29:52 crc kubenswrapper[4805]: I1203 14:29:52.885390 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-7ac1-account-create-update-6jk6w" Dec 03 14:29:52 crc kubenswrapper[4805]: I1203 14:29:52.927464 4805 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-db-create-7bqcx"] Dec 03 14:29:52 crc kubenswrapper[4805]: I1203 14:29:52.928755 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-create-7bqcx" Dec 03 14:29:52 crc kubenswrapper[4805]: I1203 14:29:52.930980 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-56b6j\" (UniqueName: \"kubernetes.io/projected/c15336c7-d573-432f-8b4c-ba5a39c82cf8-kube-api-access-56b6j\") pod \"keystone-db-sync-75wdv\" (UID: \"c15336c7-d573-432f-8b4c-ba5a39c82cf8\") " pod="openstack/keystone-db-sync-75wdv" Dec 03 14:29:52 crc kubenswrapper[4805]: I1203 14:29:52.931027 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-b686t\" (UniqueName: \"kubernetes.io/projected/d319f692-2f37-4a01-8e6c-bbf0e7b23c8e-kube-api-access-b686t\") pod \"heat-db-create-ztcqj\" (UID: \"d319f692-2f37-4a01-8e6c-bbf0e7b23c8e\") " pod="openstack/heat-db-create-ztcqj" Dec 03 14:29:52 crc kubenswrapper[4805]: I1203 14:29:52.931124 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c15336c7-d573-432f-8b4c-ba5a39c82cf8-config-data\") pod \"keystone-db-sync-75wdv\" (UID: \"c15336c7-d573-432f-8b4c-ba5a39c82cf8\") " pod="openstack/keystone-db-sync-75wdv" Dec 03 14:29:52 crc kubenswrapper[4805]: I1203 14:29:52.931160 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c15336c7-d573-432f-8b4c-ba5a39c82cf8-combined-ca-bundle\") pod \"keystone-db-sync-75wdv\" (UID: \"c15336c7-d573-432f-8b4c-ba5a39c82cf8\") " pod="openstack/keystone-db-sync-75wdv" Dec 03 14:29:52 crc kubenswrapper[4805]: I1203 14:29:52.931194 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/d319f692-2f37-4a01-8e6c-bbf0e7b23c8e-operator-scripts\") pod \"heat-db-create-ztcqj\" (UID: \"d319f692-2f37-4a01-8e6c-bbf0e7b23c8e\") " pod="openstack/heat-db-create-ztcqj" Dec 03 14:29:52 crc kubenswrapper[4805]: I1203 14:29:52.938216 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c15336c7-d573-432f-8b4c-ba5a39c82cf8-config-data\") pod \"keystone-db-sync-75wdv\" (UID: \"c15336c7-d573-432f-8b4c-ba5a39c82cf8\") " pod="openstack/keystone-db-sync-75wdv" Dec 03 14:29:52 crc kubenswrapper[4805]: I1203 14:29:52.938934 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c15336c7-d573-432f-8b4c-ba5a39c82cf8-combined-ca-bundle\") pod \"keystone-db-sync-75wdv\" (UID: \"c15336c7-d573-432f-8b4c-ba5a39c82cf8\") " pod="openstack/keystone-db-sync-75wdv" Dec 03 14:29:52 crc kubenswrapper[4805]: I1203 14:29:52.939129 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/d319f692-2f37-4a01-8e6c-bbf0e7b23c8e-operator-scripts\") pod \"heat-db-create-ztcqj\" (UID: \"d319f692-2f37-4a01-8e6c-bbf0e7b23c8e\") " pod="openstack/heat-db-create-ztcqj" Dec 03 14:29:52 crc kubenswrapper[4805]: I1203 14:29:52.949031 4805 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/heat-ee1c-account-create-update-g8lk4"] Dec 03 14:29:52 crc kubenswrapper[4805]: I1203 14:29:52.949390 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-3c27-account-create-update-79pjf" Dec 03 14:29:52 crc kubenswrapper[4805]: I1203 14:29:52.950322 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-ee1c-account-create-update-g8lk4" Dec 03 14:29:52 crc kubenswrapper[4805]: I1203 14:29:52.954403 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"heat-db-secret" Dec 03 14:29:52 crc kubenswrapper[4805]: I1203 14:29:52.964815 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-db-create-7bqcx"] Dec 03 14:29:52 crc kubenswrapper[4805]: I1203 14:29:52.967715 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-b686t\" (UniqueName: \"kubernetes.io/projected/d319f692-2f37-4a01-8e6c-bbf0e7b23c8e-kube-api-access-b686t\") pod \"heat-db-create-ztcqj\" (UID: \"d319f692-2f37-4a01-8e6c-bbf0e7b23c8e\") " pod="openstack/heat-db-create-ztcqj" Dec 03 14:29:52 crc kubenswrapper[4805]: I1203 14:29:52.987539 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-56b6j\" (UniqueName: \"kubernetes.io/projected/c15336c7-d573-432f-8b4c-ba5a39c82cf8-kube-api-access-56b6j\") pod \"keystone-db-sync-75wdv\" (UID: \"c15336c7-d573-432f-8b4c-ba5a39c82cf8\") " pod="openstack/keystone-db-sync-75wdv" Dec 03 14:29:53 crc kubenswrapper[4805]: I1203 14:29:53.011807 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-ee1c-account-create-update-g8lk4"] Dec 03 14:29:53 crc kubenswrapper[4805]: I1203 14:29:53.030427 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-sync-75wdv" Dec 03 14:29:53 crc kubenswrapper[4805]: I1203 14:29:53.033631 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8q6b9\" (UniqueName: \"kubernetes.io/projected/99891296-b622-466d-95f7-bcf0e3ebe896-kube-api-access-8q6b9\") pod \"heat-ee1c-account-create-update-g8lk4\" (UID: \"99891296-b622-466d-95f7-bcf0e3ebe896\") " pod="openstack/heat-ee1c-account-create-update-g8lk4" Dec 03 14:29:53 crc kubenswrapper[4805]: I1203 14:29:53.033773 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/99891296-b622-466d-95f7-bcf0e3ebe896-operator-scripts\") pod \"heat-ee1c-account-create-update-g8lk4\" (UID: \"99891296-b622-466d-95f7-bcf0e3ebe896\") " pod="openstack/heat-ee1c-account-create-update-g8lk4" Dec 03 14:29:53 crc kubenswrapper[4805]: I1203 14:29:53.033817 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pm5n9\" (UniqueName: \"kubernetes.io/projected/18be72b0-25d0-41b4-8038-abd0fb38deba-kube-api-access-pm5n9\") pod \"neutron-db-create-7bqcx\" (UID: \"18be72b0-25d0-41b4-8038-abd0fb38deba\") " pod="openstack/neutron-db-create-7bqcx" Dec 03 14:29:53 crc kubenswrapper[4805]: I1203 14:29:53.033878 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/18be72b0-25d0-41b4-8038-abd0fb38deba-operator-scripts\") pod \"neutron-db-create-7bqcx\" (UID: \"18be72b0-25d0-41b4-8038-abd0fb38deba\") " pod="openstack/neutron-db-create-7bqcx" Dec 03 14:29:53 crc kubenswrapper[4805]: I1203 14:29:53.048666 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-db-create-ztcqj" Dec 03 14:29:53 crc kubenswrapper[4805]: I1203 14:29:53.049322 4805 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-77e9-account-create-update-4mkc7"] Dec 03 14:29:53 crc kubenswrapper[4805]: I1203 14:29:53.050657 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-77e9-account-create-update-4mkc7" Dec 03 14:29:53 crc kubenswrapper[4805]: I1203 14:29:53.056530 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-db-secret" Dec 03 14:29:53 crc kubenswrapper[4805]: I1203 14:29:53.063272 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-77e9-account-create-update-4mkc7"] Dec 03 14:29:53 crc kubenswrapper[4805]: I1203 14:29:53.138806 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/99891296-b622-466d-95f7-bcf0e3ebe896-operator-scripts\") pod \"heat-ee1c-account-create-update-g8lk4\" (UID: \"99891296-b622-466d-95f7-bcf0e3ebe896\") " pod="openstack/heat-ee1c-account-create-update-g8lk4" Dec 03 14:29:53 crc kubenswrapper[4805]: I1203 14:29:53.139088 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pm5n9\" (UniqueName: \"kubernetes.io/projected/18be72b0-25d0-41b4-8038-abd0fb38deba-kube-api-access-pm5n9\") pod \"neutron-db-create-7bqcx\" (UID: \"18be72b0-25d0-41b4-8038-abd0fb38deba\") " pod="openstack/neutron-db-create-7bqcx" Dec 03 14:29:53 crc kubenswrapper[4805]: I1203 14:29:53.139127 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/696fec58-2abe-474b-956c-61228970cc74-operator-scripts\") pod \"neutron-77e9-account-create-update-4mkc7\" (UID: \"696fec58-2abe-474b-956c-61228970cc74\") " pod="openstack/neutron-77e9-account-create-update-4mkc7" Dec 03 14:29:53 crc kubenswrapper[4805]: I1203 14:29:53.139146 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/18be72b0-25d0-41b4-8038-abd0fb38deba-operator-scripts\") pod \"neutron-db-create-7bqcx\" (UID: \"18be72b0-25d0-41b4-8038-abd0fb38deba\") " pod="openstack/neutron-db-create-7bqcx" Dec 03 14:29:53 crc kubenswrapper[4805]: I1203 14:29:53.139162 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-f5d87\" (UniqueName: \"kubernetes.io/projected/696fec58-2abe-474b-956c-61228970cc74-kube-api-access-f5d87\") pod \"neutron-77e9-account-create-update-4mkc7\" (UID: \"696fec58-2abe-474b-956c-61228970cc74\") " pod="openstack/neutron-77e9-account-create-update-4mkc7" Dec 03 14:29:53 crc kubenswrapper[4805]: I1203 14:29:53.139228 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8q6b9\" (UniqueName: \"kubernetes.io/projected/99891296-b622-466d-95f7-bcf0e3ebe896-kube-api-access-8q6b9\") pod \"heat-ee1c-account-create-update-g8lk4\" (UID: \"99891296-b622-466d-95f7-bcf0e3ebe896\") " pod="openstack/heat-ee1c-account-create-update-g8lk4" Dec 03 14:29:53 crc kubenswrapper[4805]: I1203 14:29:53.140291 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/99891296-b622-466d-95f7-bcf0e3ebe896-operator-scripts\") pod \"heat-ee1c-account-create-update-g8lk4\" (UID: \"99891296-b622-466d-95f7-bcf0e3ebe896\") " pod="openstack/heat-ee1c-account-create-update-g8lk4" Dec 03 14:29:53 crc kubenswrapper[4805]: I1203 14:29:53.140983 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/18be72b0-25d0-41b4-8038-abd0fb38deba-operator-scripts\") pod \"neutron-db-create-7bqcx\" (UID: \"18be72b0-25d0-41b4-8038-abd0fb38deba\") " pod="openstack/neutron-db-create-7bqcx" Dec 03 14:29:53 crc kubenswrapper[4805]: I1203 14:29:53.162730 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pm5n9\" (UniqueName: \"kubernetes.io/projected/18be72b0-25d0-41b4-8038-abd0fb38deba-kube-api-access-pm5n9\") pod \"neutron-db-create-7bqcx\" (UID: \"18be72b0-25d0-41b4-8038-abd0fb38deba\") " pod="openstack/neutron-db-create-7bqcx" Dec 03 14:29:53 crc kubenswrapper[4805]: I1203 14:29:53.164112 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8q6b9\" (UniqueName: \"kubernetes.io/projected/99891296-b622-466d-95f7-bcf0e3ebe896-kube-api-access-8q6b9\") pod \"heat-ee1c-account-create-update-g8lk4\" (UID: \"99891296-b622-466d-95f7-bcf0e3ebe896\") " pod="openstack/heat-ee1c-account-create-update-g8lk4" Dec 03 14:29:53 crc kubenswrapper[4805]: I1203 14:29:53.240619 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/696fec58-2abe-474b-956c-61228970cc74-operator-scripts\") pod \"neutron-77e9-account-create-update-4mkc7\" (UID: \"696fec58-2abe-474b-956c-61228970cc74\") " pod="openstack/neutron-77e9-account-create-update-4mkc7" Dec 03 14:29:53 crc kubenswrapper[4805]: I1203 14:29:53.240675 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-f5d87\" (UniqueName: \"kubernetes.io/projected/696fec58-2abe-474b-956c-61228970cc74-kube-api-access-f5d87\") pod \"neutron-77e9-account-create-update-4mkc7\" (UID: \"696fec58-2abe-474b-956c-61228970cc74\") " pod="openstack/neutron-77e9-account-create-update-4mkc7" Dec 03 14:29:53 crc kubenswrapper[4805]: I1203 14:29:53.241632 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/696fec58-2abe-474b-956c-61228970cc74-operator-scripts\") pod \"neutron-77e9-account-create-update-4mkc7\" (UID: \"696fec58-2abe-474b-956c-61228970cc74\") " pod="openstack/neutron-77e9-account-create-update-4mkc7" Dec 03 14:29:53 crc kubenswrapper[4805]: I1203 14:29:53.259201 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-f5d87\" (UniqueName: \"kubernetes.io/projected/696fec58-2abe-474b-956c-61228970cc74-kube-api-access-f5d87\") pod \"neutron-77e9-account-create-update-4mkc7\" (UID: \"696fec58-2abe-474b-956c-61228970cc74\") " pod="openstack/neutron-77e9-account-create-update-4mkc7" Dec 03 14:29:53 crc kubenswrapper[4805]: I1203 14:29:53.286435 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-create-7bqcx" Dec 03 14:29:53 crc kubenswrapper[4805]: I1203 14:29:53.312075 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-ee1c-account-create-update-g8lk4" Dec 03 14:29:53 crc kubenswrapper[4805]: I1203 14:29:53.365728 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-77e9-account-create-update-4mkc7" Dec 03 14:29:53 crc kubenswrapper[4805]: I1203 14:29:53.385207 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-db-create-q7qlk"] Dec 03 14:29:53 crc kubenswrapper[4805]: I1203 14:29:53.522621 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-db-create-mhn6v"] Dec 03 14:29:53 crc kubenswrapper[4805]: I1203 14:29:53.538061 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-7ac1-account-create-update-6jk6w"] Dec 03 14:29:53 crc kubenswrapper[4805]: I1203 14:29:53.668650 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-3c27-account-create-update-79pjf"] Dec 03 14:29:53 crc kubenswrapper[4805]: W1203 14:29:53.679060 4805 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podeda36a78_7b3a_43f8_858e_fc6b8fdd442a.slice/crio-e1ab8f10f90b7a1e654ccd10ecbb0aa26354fef41c2d67856d65ff258654f189 WatchSource:0}: Error finding container e1ab8f10f90b7a1e654ccd10ecbb0aa26354fef41c2d67856d65ff258654f189: Status 404 returned error can't find the container with id e1ab8f10f90b7a1e654ccd10ecbb0aa26354fef41c2d67856d65ff258654f189 Dec 03 14:29:53 crc kubenswrapper[4805]: I1203 14:29:53.756975 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-create-mhn6v" event={"ID":"aead85c7-af73-44d0-b4dc-5cd98d6439b3","Type":"ContainerStarted","Data":"2bbb4a8bcf7fbb3ad29ccee5dc8fe35c4195e01ac79fd458e8fc7754ff9d7ec4"} Dec 03 14:29:53 crc kubenswrapper[4805]: I1203 14:29:53.759084 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-7ac1-account-create-update-6jk6w" event={"ID":"b27b4fde-29ff-421d-a6be-5d259bfe8016","Type":"ContainerStarted","Data":"24787f091e8092a60f5efcf2725d0b36c7fc01ff06f25a31609e6a5134f01129"} Dec 03 14:29:53 crc kubenswrapper[4805]: I1203 14:29:53.761588 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-create-q7qlk" event={"ID":"10163a58-aa6e-4988-8b33-93f8cffa4cc0","Type":"ContainerStarted","Data":"64ee2b5cb6e8d04eee8d54c5d44eafe2560f188049882dbf285590b259ff3918"} Dec 03 14:29:53 crc kubenswrapper[4805]: I1203 14:29:53.763081 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-3c27-account-create-update-79pjf" event={"ID":"eda36a78-7b3a-43f8-858e-fc6b8fdd442a","Type":"ContainerStarted","Data":"e1ab8f10f90b7a1e654ccd10ecbb0aa26354fef41c2d67856d65ff258654f189"} Dec 03 14:29:53 crc kubenswrapper[4805]: I1203 14:29:53.840561 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-db-sync-75wdv"] Dec 03 14:29:53 crc kubenswrapper[4805]: W1203 14:29:53.846974 4805 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podc15336c7_d573_432f_8b4c_ba5a39c82cf8.slice/crio-1c2b207a6035d84e65253753a49163fba7166f10f38bf68f08db84370b2265b0 WatchSource:0}: Error finding container 1c2b207a6035d84e65253753a49163fba7166f10f38bf68f08db84370b2265b0: Status 404 returned error can't find the container with id 1c2b207a6035d84e65253753a49163fba7166f10f38bf68f08db84370b2265b0 Dec 03 14:29:53 crc kubenswrapper[4805]: I1203 14:29:53.855625 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-db-create-ztcqj"] Dec 03 14:29:53 crc kubenswrapper[4805]: I1203 14:29:53.996621 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-ee1c-account-create-update-g8lk4"] Dec 03 14:29:54 crc kubenswrapper[4805]: I1203 14:29:54.011529 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-db-create-7bqcx"] Dec 03 14:29:54 crc kubenswrapper[4805]: I1203 14:29:54.188715 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-77e9-account-create-update-4mkc7"] Dec 03 14:29:54 crc kubenswrapper[4805]: I1203 14:29:54.772137 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-77e9-account-create-update-4mkc7" event={"ID":"696fec58-2abe-474b-956c-61228970cc74","Type":"ContainerStarted","Data":"700bc221d6da3381dbfac425b3a950c183eff8f50efb87cb2ffbfba5db6d5988"} Dec 03 14:29:54 crc kubenswrapper[4805]: I1203 14:29:54.775502 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-create-7bqcx" event={"ID":"18be72b0-25d0-41b4-8038-abd0fb38deba","Type":"ContainerStarted","Data":"fe21d11f1355e18c0356ed25699196c90403209283a96305edc8bb9126f6ed82"} Dec 03 14:29:54 crc kubenswrapper[4805]: I1203 14:29:54.778407 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-db-create-ztcqj" event={"ID":"d319f692-2f37-4a01-8e6c-bbf0e7b23c8e","Type":"ContainerStarted","Data":"c13dbf4257e929b12f2e17148b30a65f6caabc0ce1dd669c97496025d93df359"} Dec 03 14:29:54 crc kubenswrapper[4805]: I1203 14:29:54.780202 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-sync-75wdv" event={"ID":"c15336c7-d573-432f-8b4c-ba5a39c82cf8","Type":"ContainerStarted","Data":"1c2b207a6035d84e65253753a49163fba7166f10f38bf68f08db84370b2265b0"} Dec 03 14:29:54 crc kubenswrapper[4805]: I1203 14:29:54.781830 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-ee1c-account-create-update-g8lk4" event={"ID":"99891296-b622-466d-95f7-bcf0e3ebe896","Type":"ContainerStarted","Data":"1d7fb38a5fc314642d90e6aeb765a23fee3c2623efb21d4c5ceedcbf6b4e2e2c"} Dec 03 14:29:57 crc kubenswrapper[4805]: I1203 14:29:57.805859 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-create-mhn6v" event={"ID":"aead85c7-af73-44d0-b4dc-5cd98d6439b3","Type":"ContainerStarted","Data":"60d8a6dccfa76964665d7813e9a0bde148a8bb28ab2dce1beae33824fbdf6004"} Dec 03 14:29:57 crc kubenswrapper[4805]: I1203 14:29:57.808920 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-ee1c-account-create-update-g8lk4" event={"ID":"99891296-b622-466d-95f7-bcf0e3ebe896","Type":"ContainerStarted","Data":"1d8a116dbb3024b6dc8a70a7377f7cb462d152825caf107dd604cf92bafb4e4d"} Dec 03 14:29:57 crc kubenswrapper[4805]: I1203 14:29:57.811705 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-create-7bqcx" event={"ID":"18be72b0-25d0-41b4-8038-abd0fb38deba","Type":"ContainerStarted","Data":"9eb2ebbbba3ba07290543694db5cafa019298a402693fbd060c1c233ad4f8274"} Dec 03 14:29:57 crc kubenswrapper[4805]: I1203 14:29:57.813611 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-77e9-account-create-update-4mkc7" event={"ID":"696fec58-2abe-474b-956c-61228970cc74","Type":"ContainerStarted","Data":"a71a86043818a7d764e21262d2b172946552b42e4e3f1bcaed5a22daee349963"} Dec 03 14:29:57 crc kubenswrapper[4805]: I1203 14:29:57.818949 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-7ac1-account-create-update-6jk6w" event={"ID":"b27b4fde-29ff-421d-a6be-5d259bfe8016","Type":"ContainerStarted","Data":"d881e747656aad5e3ca4bcf5fa4c04aa2b595cfd284ad8b751619771f81ad1f0"} Dec 03 14:29:57 crc kubenswrapper[4805]: I1203 14:29:57.821011 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-create-q7qlk" event={"ID":"10163a58-aa6e-4988-8b33-93f8cffa4cc0","Type":"ContainerStarted","Data":"f78221c81b7e774055adb1e6d9260b350822233e80e328702784f13d9d72a46d"} Dec 03 14:29:57 crc kubenswrapper[4805]: I1203 14:29:57.825216 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-3c27-account-create-update-79pjf" event={"ID":"eda36a78-7b3a-43f8-858e-fc6b8fdd442a","Type":"ContainerStarted","Data":"183d5ec2c5b2105cc148606bc4cb45cdcd4f8f29170a59cfd7ac84260c358bec"} Dec 03 14:29:57 crc kubenswrapper[4805]: I1203 14:29:57.828230 4805 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-db-create-mhn6v" podStartSLOduration=5.828212167 podStartE2EDuration="5.828212167s" podCreationTimestamp="2025-12-03 14:29:52 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 14:29:57.822459512 +0000 UTC m=+1227.485376435" watchObservedRunningTime="2025-12-03 14:29:57.828212167 +0000 UTC m=+1227.491129100" Dec 03 14:29:57 crc kubenswrapper[4805]: I1203 14:29:57.828917 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-db-create-ztcqj" event={"ID":"d319f692-2f37-4a01-8e6c-bbf0e7b23c8e","Type":"ContainerStarted","Data":"67b68344596f7862edfa448581c05599565990196ca30c2e7b72a99ec85c5a87"} Dec 03 14:29:57 crc kubenswrapper[4805]: I1203 14:29:57.840657 4805 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-db-create-7bqcx" podStartSLOduration=5.840640364 podStartE2EDuration="5.840640364s" podCreationTimestamp="2025-12-03 14:29:52 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 14:29:57.835377681 +0000 UTC m=+1227.498294604" watchObservedRunningTime="2025-12-03 14:29:57.840640364 +0000 UTC m=+1227.503557287" Dec 03 14:29:57 crc kubenswrapper[4805]: I1203 14:29:57.852712 4805 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-77e9-account-create-update-4mkc7" podStartSLOduration=4.85269625 podStartE2EDuration="4.85269625s" podCreationTimestamp="2025-12-03 14:29:53 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 14:29:57.849019361 +0000 UTC m=+1227.511936284" watchObservedRunningTime="2025-12-03 14:29:57.85269625 +0000 UTC m=+1227.515613173" Dec 03 14:29:57 crc kubenswrapper[4805]: I1203 14:29:57.902315 4805 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-db-create-q7qlk" podStartSLOduration=5.902292803 podStartE2EDuration="5.902292803s" podCreationTimestamp="2025-12-03 14:29:52 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 14:29:57.891532341 +0000 UTC m=+1227.554449264" watchObservedRunningTime="2025-12-03 14:29:57.902292803 +0000 UTC m=+1227.565209726" Dec 03 14:29:57 crc kubenswrapper[4805]: I1203 14:29:57.909727 4805 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/heat-ee1c-account-create-update-g8lk4" podStartSLOduration=5.909707623 podStartE2EDuration="5.909707623s" podCreationTimestamp="2025-12-03 14:29:52 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 14:29:57.904321528 +0000 UTC m=+1227.567238451" watchObservedRunningTime="2025-12-03 14:29:57.909707623 +0000 UTC m=+1227.572624546" Dec 03 14:29:57 crc kubenswrapper[4805]: I1203 14:29:57.944785 4805 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-7ac1-account-create-update-6jk6w" podStartSLOduration=5.944599398 podStartE2EDuration="5.944599398s" podCreationTimestamp="2025-12-03 14:29:52 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 14:29:57.921158593 +0000 UTC m=+1227.584075536" watchObservedRunningTime="2025-12-03 14:29:57.944599398 +0000 UTC m=+1227.607516321" Dec 03 14:29:57 crc kubenswrapper[4805]: I1203 14:29:57.947776 4805 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-3c27-account-create-update-79pjf" podStartSLOduration=5.947761733 podStartE2EDuration="5.947761733s" podCreationTimestamp="2025-12-03 14:29:52 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 14:29:57.934301229 +0000 UTC m=+1227.597218152" watchObservedRunningTime="2025-12-03 14:29:57.947761733 +0000 UTC m=+1227.610678656" Dec 03 14:29:57 crc kubenswrapper[4805]: I1203 14:29:57.956437 4805 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/heat-db-create-ztcqj" podStartSLOduration=5.956422638 podStartE2EDuration="5.956422638s" podCreationTimestamp="2025-12-03 14:29:52 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 14:29:57.945212125 +0000 UTC m=+1227.608129048" watchObservedRunningTime="2025-12-03 14:29:57.956422638 +0000 UTC m=+1227.619339561" Dec 03 14:29:58 crc kubenswrapper[4805]: E1203 14:29:58.336910 4805 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podb27b4fde_29ff_421d_a6be_5d259bfe8016.slice/crio-conmon-d881e747656aad5e3ca4bcf5fa4c04aa2b595cfd284ad8b751619771f81ad1f0.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podeda36a78_7b3a_43f8_858e_fc6b8fdd442a.slice/crio-183d5ec2c5b2105cc148606bc4cb45cdcd4f8f29170a59cfd7ac84260c358bec.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podeda36a78_7b3a_43f8_858e_fc6b8fdd442a.slice/crio-conmon-183d5ec2c5b2105cc148606bc4cb45cdcd4f8f29170a59cfd7ac84260c358bec.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podb27b4fde_29ff_421d_a6be_5d259bfe8016.slice/crio-d881e747656aad5e3ca4bcf5fa4c04aa2b595cfd284ad8b751619771f81ad1f0.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podaead85c7_af73_44d0_b4dc_5cd98d6439b3.slice/crio-conmon-60d8a6dccfa76964665d7813e9a0bde148a8bb28ab2dce1beae33824fbdf6004.scope\": RecentStats: unable to find data in memory cache]" Dec 03 14:29:58 crc kubenswrapper[4805]: I1203 14:29:58.844646 4805 generic.go:334] "Generic (PLEG): container finished" podID="10163a58-aa6e-4988-8b33-93f8cffa4cc0" containerID="f78221c81b7e774055adb1e6d9260b350822233e80e328702784f13d9d72a46d" exitCode=0 Dec 03 14:29:58 crc kubenswrapper[4805]: I1203 14:29:58.844747 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-create-q7qlk" event={"ID":"10163a58-aa6e-4988-8b33-93f8cffa4cc0","Type":"ContainerDied","Data":"f78221c81b7e774055adb1e6d9260b350822233e80e328702784f13d9d72a46d"} Dec 03 14:29:58 crc kubenswrapper[4805]: I1203 14:29:58.846589 4805 generic.go:334] "Generic (PLEG): container finished" podID="eda36a78-7b3a-43f8-858e-fc6b8fdd442a" containerID="183d5ec2c5b2105cc148606bc4cb45cdcd4f8f29170a59cfd7ac84260c358bec" exitCode=0 Dec 03 14:29:58 crc kubenswrapper[4805]: I1203 14:29:58.846616 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-3c27-account-create-update-79pjf" event={"ID":"eda36a78-7b3a-43f8-858e-fc6b8fdd442a","Type":"ContainerDied","Data":"183d5ec2c5b2105cc148606bc4cb45cdcd4f8f29170a59cfd7ac84260c358bec"} Dec 03 14:29:58 crc kubenswrapper[4805]: I1203 14:29:58.848363 4805 generic.go:334] "Generic (PLEG): container finished" podID="d319f692-2f37-4a01-8e6c-bbf0e7b23c8e" containerID="67b68344596f7862edfa448581c05599565990196ca30c2e7b72a99ec85c5a87" exitCode=0 Dec 03 14:29:58 crc kubenswrapper[4805]: I1203 14:29:58.848431 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-db-create-ztcqj" event={"ID":"d319f692-2f37-4a01-8e6c-bbf0e7b23c8e","Type":"ContainerDied","Data":"67b68344596f7862edfa448581c05599565990196ca30c2e7b72a99ec85c5a87"} Dec 03 14:29:58 crc kubenswrapper[4805]: I1203 14:29:58.850127 4805 generic.go:334] "Generic (PLEG): container finished" podID="aead85c7-af73-44d0-b4dc-5cd98d6439b3" containerID="60d8a6dccfa76964665d7813e9a0bde148a8bb28ab2dce1beae33824fbdf6004" exitCode=0 Dec 03 14:29:58 crc kubenswrapper[4805]: I1203 14:29:58.850148 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-create-mhn6v" event={"ID":"aead85c7-af73-44d0-b4dc-5cd98d6439b3","Type":"ContainerDied","Data":"60d8a6dccfa76964665d7813e9a0bde148a8bb28ab2dce1beae33824fbdf6004"} Dec 03 14:29:58 crc kubenswrapper[4805]: I1203 14:29:58.854616 4805 generic.go:334] "Generic (PLEG): container finished" podID="99891296-b622-466d-95f7-bcf0e3ebe896" containerID="1d8a116dbb3024b6dc8a70a7377f7cb462d152825caf107dd604cf92bafb4e4d" exitCode=0 Dec 03 14:29:58 crc kubenswrapper[4805]: I1203 14:29:58.854733 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-ee1c-account-create-update-g8lk4" event={"ID":"99891296-b622-466d-95f7-bcf0e3ebe896","Type":"ContainerDied","Data":"1d8a116dbb3024b6dc8a70a7377f7cb462d152825caf107dd604cf92bafb4e4d"} Dec 03 14:29:58 crc kubenswrapper[4805]: I1203 14:29:58.864384 4805 generic.go:334] "Generic (PLEG): container finished" podID="696fec58-2abe-474b-956c-61228970cc74" containerID="a71a86043818a7d764e21262d2b172946552b42e4e3f1bcaed5a22daee349963" exitCode=0 Dec 03 14:29:58 crc kubenswrapper[4805]: I1203 14:29:58.864489 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-77e9-account-create-update-4mkc7" event={"ID":"696fec58-2abe-474b-956c-61228970cc74","Type":"ContainerDied","Data":"a71a86043818a7d764e21262d2b172946552b42e4e3f1bcaed5a22daee349963"} Dec 03 14:29:58 crc kubenswrapper[4805]: I1203 14:29:58.868392 4805 generic.go:334] "Generic (PLEG): container finished" podID="18be72b0-25d0-41b4-8038-abd0fb38deba" containerID="9eb2ebbbba3ba07290543694db5cafa019298a402693fbd060c1c233ad4f8274" exitCode=0 Dec 03 14:29:58 crc kubenswrapper[4805]: I1203 14:29:58.868462 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-create-7bqcx" event={"ID":"18be72b0-25d0-41b4-8038-abd0fb38deba","Type":"ContainerDied","Data":"9eb2ebbbba3ba07290543694db5cafa019298a402693fbd060c1c233ad4f8274"} Dec 03 14:29:58 crc kubenswrapper[4805]: I1203 14:29:58.869683 4805 generic.go:334] "Generic (PLEG): container finished" podID="b27b4fde-29ff-421d-a6be-5d259bfe8016" containerID="d881e747656aad5e3ca4bcf5fa4c04aa2b595cfd284ad8b751619771f81ad1f0" exitCode=0 Dec 03 14:29:58 crc kubenswrapper[4805]: I1203 14:29:58.869717 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-7ac1-account-create-update-6jk6w" event={"ID":"b27b4fde-29ff-421d-a6be-5d259bfe8016","Type":"ContainerDied","Data":"d881e747656aad5e3ca4bcf5fa4c04aa2b595cfd284ad8b751619771f81ad1f0"} Dec 03 14:29:59 crc kubenswrapper[4805]: I1203 14:29:59.510721 4805 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-7ff5475cc9-mbdbt" Dec 03 14:29:59 crc kubenswrapper[4805]: I1203 14:29:59.573786 4805 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-698758b865-l46rr"] Dec 03 14:29:59 crc kubenswrapper[4805]: I1203 14:29:59.574012 4805 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-698758b865-l46rr" podUID="83e323ed-ad2c-4635-a583-4e5b21d8e268" containerName="dnsmasq-dns" containerID="cri-o://55ae092c6099423baa000af1306ee0efd34017bebbf5abd512f35e371752d635" gracePeriod=10 Dec 03 14:29:59 crc kubenswrapper[4805]: I1203 14:29:59.912528 4805 generic.go:334] "Generic (PLEG): container finished" podID="83e323ed-ad2c-4635-a583-4e5b21d8e268" containerID="55ae092c6099423baa000af1306ee0efd34017bebbf5abd512f35e371752d635" exitCode=0 Dec 03 14:29:59 crc kubenswrapper[4805]: I1203 14:29:59.912781 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-698758b865-l46rr" event={"ID":"83e323ed-ad2c-4635-a583-4e5b21d8e268","Type":"ContainerDied","Data":"55ae092c6099423baa000af1306ee0efd34017bebbf5abd512f35e371752d635"} Dec 03 14:30:00 crc kubenswrapper[4805]: I1203 14:30:00.138638 4805 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29412870-46tbm"] Dec 03 14:30:00 crc kubenswrapper[4805]: I1203 14:30:00.140295 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29412870-46tbm" Dec 03 14:30:00 crc kubenswrapper[4805]: I1203 14:30:00.144344 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Dec 03 14:30:00 crc kubenswrapper[4805]: I1203 14:30:00.144636 4805 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Dec 03 14:30:00 crc kubenswrapper[4805]: I1203 14:30:00.149427 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29412870-46tbm"] Dec 03 14:30:00 crc kubenswrapper[4805]: I1203 14:30:00.271473 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dw6kr\" (UniqueName: \"kubernetes.io/projected/5aae8010-7566-4f0b-ba0f-8d905bc9e455-kube-api-access-dw6kr\") pod \"collect-profiles-29412870-46tbm\" (UID: \"5aae8010-7566-4f0b-ba0f-8d905bc9e455\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29412870-46tbm" Dec 03 14:30:00 crc kubenswrapper[4805]: I1203 14:30:00.271576 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/5aae8010-7566-4f0b-ba0f-8d905bc9e455-config-volume\") pod \"collect-profiles-29412870-46tbm\" (UID: \"5aae8010-7566-4f0b-ba0f-8d905bc9e455\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29412870-46tbm" Dec 03 14:30:00 crc kubenswrapper[4805]: I1203 14:30:00.271612 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/5aae8010-7566-4f0b-ba0f-8d905bc9e455-secret-volume\") pod \"collect-profiles-29412870-46tbm\" (UID: \"5aae8010-7566-4f0b-ba0f-8d905bc9e455\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29412870-46tbm" Dec 03 14:30:00 crc kubenswrapper[4805]: I1203 14:30:00.374283 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dw6kr\" (UniqueName: \"kubernetes.io/projected/5aae8010-7566-4f0b-ba0f-8d905bc9e455-kube-api-access-dw6kr\") pod \"collect-profiles-29412870-46tbm\" (UID: \"5aae8010-7566-4f0b-ba0f-8d905bc9e455\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29412870-46tbm" Dec 03 14:30:00 crc kubenswrapper[4805]: I1203 14:30:00.374348 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/5aae8010-7566-4f0b-ba0f-8d905bc9e455-config-volume\") pod \"collect-profiles-29412870-46tbm\" (UID: \"5aae8010-7566-4f0b-ba0f-8d905bc9e455\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29412870-46tbm" Dec 03 14:30:00 crc kubenswrapper[4805]: I1203 14:30:00.374393 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/5aae8010-7566-4f0b-ba0f-8d905bc9e455-secret-volume\") pod \"collect-profiles-29412870-46tbm\" (UID: \"5aae8010-7566-4f0b-ba0f-8d905bc9e455\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29412870-46tbm" Dec 03 14:30:00 crc kubenswrapper[4805]: I1203 14:30:00.375645 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/5aae8010-7566-4f0b-ba0f-8d905bc9e455-config-volume\") pod \"collect-profiles-29412870-46tbm\" (UID: \"5aae8010-7566-4f0b-ba0f-8d905bc9e455\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29412870-46tbm" Dec 03 14:30:00 crc kubenswrapper[4805]: I1203 14:30:00.390209 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/5aae8010-7566-4f0b-ba0f-8d905bc9e455-secret-volume\") pod \"collect-profiles-29412870-46tbm\" (UID: \"5aae8010-7566-4f0b-ba0f-8d905bc9e455\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29412870-46tbm" Dec 03 14:30:00 crc kubenswrapper[4805]: I1203 14:30:00.390599 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dw6kr\" (UniqueName: \"kubernetes.io/projected/5aae8010-7566-4f0b-ba0f-8d905bc9e455-kube-api-access-dw6kr\") pod \"collect-profiles-29412870-46tbm\" (UID: \"5aae8010-7566-4f0b-ba0f-8d905bc9e455\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29412870-46tbm" Dec 03 14:30:00 crc kubenswrapper[4805]: I1203 14:30:00.473632 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29412870-46tbm" Dec 03 14:30:01 crc kubenswrapper[4805]: I1203 14:30:01.934943 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-3c27-account-create-update-79pjf" event={"ID":"eda36a78-7b3a-43f8-858e-fc6b8fdd442a","Type":"ContainerDied","Data":"e1ab8f10f90b7a1e654ccd10ecbb0aa26354fef41c2d67856d65ff258654f189"} Dec 03 14:30:01 crc kubenswrapper[4805]: I1203 14:30:01.935464 4805 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="e1ab8f10f90b7a1e654ccd10ecbb0aa26354fef41c2d67856d65ff258654f189" Dec 03 14:30:01 crc kubenswrapper[4805]: I1203 14:30:01.936883 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-db-create-ztcqj" event={"ID":"d319f692-2f37-4a01-8e6c-bbf0e7b23c8e","Type":"ContainerDied","Data":"c13dbf4257e929b12f2e17148b30a65f6caabc0ce1dd669c97496025d93df359"} Dec 03 14:30:01 crc kubenswrapper[4805]: I1203 14:30:01.936904 4805 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="c13dbf4257e929b12f2e17148b30a65f6caabc0ce1dd669c97496025d93df359" Dec 03 14:30:01 crc kubenswrapper[4805]: I1203 14:30:01.940736 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-create-mhn6v" event={"ID":"aead85c7-af73-44d0-b4dc-5cd98d6439b3","Type":"ContainerDied","Data":"2bbb4a8bcf7fbb3ad29ccee5dc8fe35c4195e01ac79fd458e8fc7754ff9d7ec4"} Dec 03 14:30:01 crc kubenswrapper[4805]: I1203 14:30:01.940764 4805 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="2bbb4a8bcf7fbb3ad29ccee5dc8fe35c4195e01ac79fd458e8fc7754ff9d7ec4" Dec 03 14:30:01 crc kubenswrapper[4805]: I1203 14:30:01.947990 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-ee1c-account-create-update-g8lk4" event={"ID":"99891296-b622-466d-95f7-bcf0e3ebe896","Type":"ContainerDied","Data":"1d7fb38a5fc314642d90e6aeb765a23fee3c2623efb21d4c5ceedcbf6b4e2e2c"} Dec 03 14:30:01 crc kubenswrapper[4805]: I1203 14:30:01.948025 4805 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="1d7fb38a5fc314642d90e6aeb765a23fee3c2623efb21d4c5ceedcbf6b4e2e2c" Dec 03 14:30:01 crc kubenswrapper[4805]: I1203 14:30:01.950110 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-77e9-account-create-update-4mkc7" event={"ID":"696fec58-2abe-474b-956c-61228970cc74","Type":"ContainerDied","Data":"700bc221d6da3381dbfac425b3a950c183eff8f50efb87cb2ffbfba5db6d5988"} Dec 03 14:30:01 crc kubenswrapper[4805]: I1203 14:30:01.950130 4805 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="700bc221d6da3381dbfac425b3a950c183eff8f50efb87cb2ffbfba5db6d5988" Dec 03 14:30:01 crc kubenswrapper[4805]: I1203 14:30:01.951713 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-create-7bqcx" event={"ID":"18be72b0-25d0-41b4-8038-abd0fb38deba","Type":"ContainerDied","Data":"fe21d11f1355e18c0356ed25699196c90403209283a96305edc8bb9126f6ed82"} Dec 03 14:30:01 crc kubenswrapper[4805]: I1203 14:30:01.951728 4805 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="fe21d11f1355e18c0356ed25699196c90403209283a96305edc8bb9126f6ed82" Dec 03 14:30:01 crc kubenswrapper[4805]: I1203 14:30:01.956143 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-7ac1-account-create-update-6jk6w" event={"ID":"b27b4fde-29ff-421d-a6be-5d259bfe8016","Type":"ContainerDied","Data":"24787f091e8092a60f5efcf2725d0b36c7fc01ff06f25a31609e6a5134f01129"} Dec 03 14:30:01 crc kubenswrapper[4805]: I1203 14:30:01.956168 4805 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="24787f091e8092a60f5efcf2725d0b36c7fc01ff06f25a31609e6a5134f01129" Dec 03 14:30:01 crc kubenswrapper[4805]: I1203 14:30:01.963072 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-create-q7qlk" event={"ID":"10163a58-aa6e-4988-8b33-93f8cffa4cc0","Type":"ContainerDied","Data":"64ee2b5cb6e8d04eee8d54c5d44eafe2560f188049882dbf285590b259ff3918"} Dec 03 14:30:01 crc kubenswrapper[4805]: I1203 14:30:01.963098 4805 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="64ee2b5cb6e8d04eee8d54c5d44eafe2560f188049882dbf285590b259ff3918" Dec 03 14:30:01 crc kubenswrapper[4805]: I1203 14:30:01.988481 4805 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-77e9-account-create-update-4mkc7" Dec 03 14:30:02 crc kubenswrapper[4805]: I1203 14:30:02.009009 4805 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-create-7bqcx" Dec 03 14:30:02 crc kubenswrapper[4805]: I1203 14:30:02.025110 4805 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-create-mhn6v" Dec 03 14:30:02 crc kubenswrapper[4805]: I1203 14:30:02.037632 4805 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/heat-ee1c-account-create-update-g8lk4" Dec 03 14:30:02 crc kubenswrapper[4805]: I1203 14:30:02.058686 4805 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/heat-db-create-ztcqj" Dec 03 14:30:02 crc kubenswrapper[4805]: I1203 14:30:02.070813 4805 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-3c27-account-create-update-79pjf" Dec 03 14:30:02 crc kubenswrapper[4805]: I1203 14:30:02.088226 4805 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-7ac1-account-create-update-6jk6w" Dec 03 14:30:02 crc kubenswrapper[4805]: I1203 14:30:02.093693 4805 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-create-q7qlk" Dec 03 14:30:02 crc kubenswrapper[4805]: I1203 14:30:02.099370 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8q6b9\" (UniqueName: \"kubernetes.io/projected/99891296-b622-466d-95f7-bcf0e3ebe896-kube-api-access-8q6b9\") pod \"99891296-b622-466d-95f7-bcf0e3ebe896\" (UID: \"99891296-b622-466d-95f7-bcf0e3ebe896\") " Dec 03 14:30:02 crc kubenswrapper[4805]: I1203 14:30:02.099417 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-f5d87\" (UniqueName: \"kubernetes.io/projected/696fec58-2abe-474b-956c-61228970cc74-kube-api-access-f5d87\") pod \"696fec58-2abe-474b-956c-61228970cc74\" (UID: \"696fec58-2abe-474b-956c-61228970cc74\") " Dec 03 14:30:02 crc kubenswrapper[4805]: I1203 14:30:02.099464 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pm5n9\" (UniqueName: \"kubernetes.io/projected/18be72b0-25d0-41b4-8038-abd0fb38deba-kube-api-access-pm5n9\") pod \"18be72b0-25d0-41b4-8038-abd0fb38deba\" (UID: \"18be72b0-25d0-41b4-8038-abd0fb38deba\") " Dec 03 14:30:02 crc kubenswrapper[4805]: I1203 14:30:02.099498 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/aead85c7-af73-44d0-b4dc-5cd98d6439b3-operator-scripts\") pod \"aead85c7-af73-44d0-b4dc-5cd98d6439b3\" (UID: \"aead85c7-af73-44d0-b4dc-5cd98d6439b3\") " Dec 03 14:30:02 crc kubenswrapper[4805]: I1203 14:30:02.099534 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/99891296-b622-466d-95f7-bcf0e3ebe896-operator-scripts\") pod \"99891296-b622-466d-95f7-bcf0e3ebe896\" (UID: \"99891296-b622-466d-95f7-bcf0e3ebe896\") " Dec 03 14:30:02 crc kubenswrapper[4805]: I1203 14:30:02.099551 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-chbc4\" (UniqueName: \"kubernetes.io/projected/aead85c7-af73-44d0-b4dc-5cd98d6439b3-kube-api-access-chbc4\") pod \"aead85c7-af73-44d0-b4dc-5cd98d6439b3\" (UID: \"aead85c7-af73-44d0-b4dc-5cd98d6439b3\") " Dec 03 14:30:02 crc kubenswrapper[4805]: I1203 14:30:02.099568 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-b686t\" (UniqueName: \"kubernetes.io/projected/d319f692-2f37-4a01-8e6c-bbf0e7b23c8e-kube-api-access-b686t\") pod \"d319f692-2f37-4a01-8e6c-bbf0e7b23c8e\" (UID: \"d319f692-2f37-4a01-8e6c-bbf0e7b23c8e\") " Dec 03 14:30:02 crc kubenswrapper[4805]: I1203 14:30:02.099597 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/d319f692-2f37-4a01-8e6c-bbf0e7b23c8e-operator-scripts\") pod \"d319f692-2f37-4a01-8e6c-bbf0e7b23c8e\" (UID: \"d319f692-2f37-4a01-8e6c-bbf0e7b23c8e\") " Dec 03 14:30:02 crc kubenswrapper[4805]: I1203 14:30:02.099641 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/18be72b0-25d0-41b4-8038-abd0fb38deba-operator-scripts\") pod \"18be72b0-25d0-41b4-8038-abd0fb38deba\" (UID: \"18be72b0-25d0-41b4-8038-abd0fb38deba\") " Dec 03 14:30:02 crc kubenswrapper[4805]: I1203 14:30:02.099665 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/696fec58-2abe-474b-956c-61228970cc74-operator-scripts\") pod \"696fec58-2abe-474b-956c-61228970cc74\" (UID: \"696fec58-2abe-474b-956c-61228970cc74\") " Dec 03 14:30:02 crc kubenswrapper[4805]: I1203 14:30:02.100608 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/696fec58-2abe-474b-956c-61228970cc74-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "696fec58-2abe-474b-956c-61228970cc74" (UID: "696fec58-2abe-474b-956c-61228970cc74"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 14:30:02 crc kubenswrapper[4805]: I1203 14:30:02.100986 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/aead85c7-af73-44d0-b4dc-5cd98d6439b3-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "aead85c7-af73-44d0-b4dc-5cd98d6439b3" (UID: "aead85c7-af73-44d0-b4dc-5cd98d6439b3"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 14:30:02 crc kubenswrapper[4805]: I1203 14:30:02.101322 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/99891296-b622-466d-95f7-bcf0e3ebe896-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "99891296-b622-466d-95f7-bcf0e3ebe896" (UID: "99891296-b622-466d-95f7-bcf0e3ebe896"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 14:30:02 crc kubenswrapper[4805]: I1203 14:30:02.101827 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d319f692-2f37-4a01-8e6c-bbf0e7b23c8e-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "d319f692-2f37-4a01-8e6c-bbf0e7b23c8e" (UID: "d319f692-2f37-4a01-8e6c-bbf0e7b23c8e"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 14:30:02 crc kubenswrapper[4805]: I1203 14:30:02.102163 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/18be72b0-25d0-41b4-8038-abd0fb38deba-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "18be72b0-25d0-41b4-8038-abd0fb38deba" (UID: "18be72b0-25d0-41b4-8038-abd0fb38deba"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 14:30:02 crc kubenswrapper[4805]: I1203 14:30:02.102357 4805 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-698758b865-l46rr" Dec 03 14:30:02 crc kubenswrapper[4805]: I1203 14:30:02.108562 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/696fec58-2abe-474b-956c-61228970cc74-kube-api-access-f5d87" (OuterVolumeSpecName: "kube-api-access-f5d87") pod "696fec58-2abe-474b-956c-61228970cc74" (UID: "696fec58-2abe-474b-956c-61228970cc74"). InnerVolumeSpecName "kube-api-access-f5d87". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 14:30:02 crc kubenswrapper[4805]: I1203 14:30:02.109416 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/aead85c7-af73-44d0-b4dc-5cd98d6439b3-kube-api-access-chbc4" (OuterVolumeSpecName: "kube-api-access-chbc4") pod "aead85c7-af73-44d0-b4dc-5cd98d6439b3" (UID: "aead85c7-af73-44d0-b4dc-5cd98d6439b3"). InnerVolumeSpecName "kube-api-access-chbc4". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 14:30:02 crc kubenswrapper[4805]: I1203 14:30:02.109606 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/99891296-b622-466d-95f7-bcf0e3ebe896-kube-api-access-8q6b9" (OuterVolumeSpecName: "kube-api-access-8q6b9") pod "99891296-b622-466d-95f7-bcf0e3ebe896" (UID: "99891296-b622-466d-95f7-bcf0e3ebe896"). InnerVolumeSpecName "kube-api-access-8q6b9". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 14:30:02 crc kubenswrapper[4805]: I1203 14:30:02.112942 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d319f692-2f37-4a01-8e6c-bbf0e7b23c8e-kube-api-access-b686t" (OuterVolumeSpecName: "kube-api-access-b686t") pod "d319f692-2f37-4a01-8e6c-bbf0e7b23c8e" (UID: "d319f692-2f37-4a01-8e6c-bbf0e7b23c8e"). InnerVolumeSpecName "kube-api-access-b686t". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 14:30:02 crc kubenswrapper[4805]: I1203 14:30:02.113729 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/18be72b0-25d0-41b4-8038-abd0fb38deba-kube-api-access-pm5n9" (OuterVolumeSpecName: "kube-api-access-pm5n9") pod "18be72b0-25d0-41b4-8038-abd0fb38deba" (UID: "18be72b0-25d0-41b4-8038-abd0fb38deba"). InnerVolumeSpecName "kube-api-access-pm5n9". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 14:30:02 crc kubenswrapper[4805]: I1203 14:30:02.200880 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8rz8t\" (UniqueName: \"kubernetes.io/projected/b27b4fde-29ff-421d-a6be-5d259bfe8016-kube-api-access-8rz8t\") pod \"b27b4fde-29ff-421d-a6be-5d259bfe8016\" (UID: \"b27b4fde-29ff-421d-a6be-5d259bfe8016\") " Dec 03 14:30:02 crc kubenswrapper[4805]: I1203 14:30:02.200965 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lx9th\" (UniqueName: \"kubernetes.io/projected/eda36a78-7b3a-43f8-858e-fc6b8fdd442a-kube-api-access-lx9th\") pod \"eda36a78-7b3a-43f8-858e-fc6b8fdd442a\" (UID: \"eda36a78-7b3a-43f8-858e-fc6b8fdd442a\") " Dec 03 14:30:02 crc kubenswrapper[4805]: I1203 14:30:02.201024 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/b27b4fde-29ff-421d-a6be-5d259bfe8016-operator-scripts\") pod \"b27b4fde-29ff-421d-a6be-5d259bfe8016\" (UID: \"b27b4fde-29ff-421d-a6be-5d259bfe8016\") " Dec 03 14:30:02 crc kubenswrapper[4805]: I1203 14:30:02.201045 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gfgzp\" (UniqueName: \"kubernetes.io/projected/10163a58-aa6e-4988-8b33-93f8cffa4cc0-kube-api-access-gfgzp\") pod \"10163a58-aa6e-4988-8b33-93f8cffa4cc0\" (UID: \"10163a58-aa6e-4988-8b33-93f8cffa4cc0\") " Dec 03 14:30:02 crc kubenswrapper[4805]: I1203 14:30:02.201067 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/83e323ed-ad2c-4635-a583-4e5b21d8e268-config\") pod \"83e323ed-ad2c-4635-a583-4e5b21d8e268\" (UID: \"83e323ed-ad2c-4635-a583-4e5b21d8e268\") " Dec 03 14:30:02 crc kubenswrapper[4805]: I1203 14:30:02.201125 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/eda36a78-7b3a-43f8-858e-fc6b8fdd442a-operator-scripts\") pod \"eda36a78-7b3a-43f8-858e-fc6b8fdd442a\" (UID: \"eda36a78-7b3a-43f8-858e-fc6b8fdd442a\") " Dec 03 14:30:02 crc kubenswrapper[4805]: I1203 14:30:02.201168 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qj4w8\" (UniqueName: \"kubernetes.io/projected/83e323ed-ad2c-4635-a583-4e5b21d8e268-kube-api-access-qj4w8\") pod \"83e323ed-ad2c-4635-a583-4e5b21d8e268\" (UID: \"83e323ed-ad2c-4635-a583-4e5b21d8e268\") " Dec 03 14:30:02 crc kubenswrapper[4805]: I1203 14:30:02.201188 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/83e323ed-ad2c-4635-a583-4e5b21d8e268-ovsdbserver-nb\") pod \"83e323ed-ad2c-4635-a583-4e5b21d8e268\" (UID: \"83e323ed-ad2c-4635-a583-4e5b21d8e268\") " Dec 03 14:30:02 crc kubenswrapper[4805]: I1203 14:30:02.201223 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/10163a58-aa6e-4988-8b33-93f8cffa4cc0-operator-scripts\") pod \"10163a58-aa6e-4988-8b33-93f8cffa4cc0\" (UID: \"10163a58-aa6e-4988-8b33-93f8cffa4cc0\") " Dec 03 14:30:02 crc kubenswrapper[4805]: I1203 14:30:02.201241 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/83e323ed-ad2c-4635-a583-4e5b21d8e268-ovsdbserver-sb\") pod \"83e323ed-ad2c-4635-a583-4e5b21d8e268\" (UID: \"83e323ed-ad2c-4635-a583-4e5b21d8e268\") " Dec 03 14:30:02 crc kubenswrapper[4805]: I1203 14:30:02.201274 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/83e323ed-ad2c-4635-a583-4e5b21d8e268-dns-svc\") pod \"83e323ed-ad2c-4635-a583-4e5b21d8e268\" (UID: \"83e323ed-ad2c-4635-a583-4e5b21d8e268\") " Dec 03 14:30:02 crc kubenswrapper[4805]: I1203 14:30:02.201576 4805 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pm5n9\" (UniqueName: \"kubernetes.io/projected/18be72b0-25d0-41b4-8038-abd0fb38deba-kube-api-access-pm5n9\") on node \"crc\" DevicePath \"\"" Dec 03 14:30:02 crc kubenswrapper[4805]: I1203 14:30:02.201593 4805 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/aead85c7-af73-44d0-b4dc-5cd98d6439b3-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 03 14:30:02 crc kubenswrapper[4805]: I1203 14:30:02.201602 4805 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/99891296-b622-466d-95f7-bcf0e3ebe896-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 03 14:30:02 crc kubenswrapper[4805]: I1203 14:30:02.201611 4805 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-chbc4\" (UniqueName: \"kubernetes.io/projected/aead85c7-af73-44d0-b4dc-5cd98d6439b3-kube-api-access-chbc4\") on node \"crc\" DevicePath \"\"" Dec 03 14:30:02 crc kubenswrapper[4805]: I1203 14:30:02.201623 4805 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-b686t\" (UniqueName: \"kubernetes.io/projected/d319f692-2f37-4a01-8e6c-bbf0e7b23c8e-kube-api-access-b686t\") on node \"crc\" DevicePath \"\"" Dec 03 14:30:02 crc kubenswrapper[4805]: I1203 14:30:02.201635 4805 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/d319f692-2f37-4a01-8e6c-bbf0e7b23c8e-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 03 14:30:02 crc kubenswrapper[4805]: I1203 14:30:02.201644 4805 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/18be72b0-25d0-41b4-8038-abd0fb38deba-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 03 14:30:02 crc kubenswrapper[4805]: I1203 14:30:02.201653 4805 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/696fec58-2abe-474b-956c-61228970cc74-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 03 14:30:02 crc kubenswrapper[4805]: I1203 14:30:02.201661 4805 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8q6b9\" (UniqueName: \"kubernetes.io/projected/99891296-b622-466d-95f7-bcf0e3ebe896-kube-api-access-8q6b9\") on node \"crc\" DevicePath \"\"" Dec 03 14:30:02 crc kubenswrapper[4805]: I1203 14:30:02.201669 4805 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-f5d87\" (UniqueName: \"kubernetes.io/projected/696fec58-2abe-474b-956c-61228970cc74-kube-api-access-f5d87\") on node \"crc\" DevicePath \"\"" Dec 03 14:30:02 crc kubenswrapper[4805]: I1203 14:30:02.202394 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/10163a58-aa6e-4988-8b33-93f8cffa4cc0-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "10163a58-aa6e-4988-8b33-93f8cffa4cc0" (UID: "10163a58-aa6e-4988-8b33-93f8cffa4cc0"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 14:30:02 crc kubenswrapper[4805]: I1203 14:30:02.202595 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b27b4fde-29ff-421d-a6be-5d259bfe8016-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "b27b4fde-29ff-421d-a6be-5d259bfe8016" (UID: "b27b4fde-29ff-421d-a6be-5d259bfe8016"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 14:30:02 crc kubenswrapper[4805]: I1203 14:30:02.202771 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/eda36a78-7b3a-43f8-858e-fc6b8fdd442a-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "eda36a78-7b3a-43f8-858e-fc6b8fdd442a" (UID: "eda36a78-7b3a-43f8-858e-fc6b8fdd442a"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 14:30:02 crc kubenswrapper[4805]: I1203 14:30:02.205107 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b27b4fde-29ff-421d-a6be-5d259bfe8016-kube-api-access-8rz8t" (OuterVolumeSpecName: "kube-api-access-8rz8t") pod "b27b4fde-29ff-421d-a6be-5d259bfe8016" (UID: "b27b4fde-29ff-421d-a6be-5d259bfe8016"). InnerVolumeSpecName "kube-api-access-8rz8t". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 14:30:02 crc kubenswrapper[4805]: I1203 14:30:02.205258 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/10163a58-aa6e-4988-8b33-93f8cffa4cc0-kube-api-access-gfgzp" (OuterVolumeSpecName: "kube-api-access-gfgzp") pod "10163a58-aa6e-4988-8b33-93f8cffa4cc0" (UID: "10163a58-aa6e-4988-8b33-93f8cffa4cc0"). InnerVolumeSpecName "kube-api-access-gfgzp". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 14:30:02 crc kubenswrapper[4805]: I1203 14:30:02.207017 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/83e323ed-ad2c-4635-a583-4e5b21d8e268-kube-api-access-qj4w8" (OuterVolumeSpecName: "kube-api-access-qj4w8") pod "83e323ed-ad2c-4635-a583-4e5b21d8e268" (UID: "83e323ed-ad2c-4635-a583-4e5b21d8e268"). InnerVolumeSpecName "kube-api-access-qj4w8". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 14:30:02 crc kubenswrapper[4805]: I1203 14:30:02.207533 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/eda36a78-7b3a-43f8-858e-fc6b8fdd442a-kube-api-access-lx9th" (OuterVolumeSpecName: "kube-api-access-lx9th") pod "eda36a78-7b3a-43f8-858e-fc6b8fdd442a" (UID: "eda36a78-7b3a-43f8-858e-fc6b8fdd442a"). InnerVolumeSpecName "kube-api-access-lx9th". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 14:30:02 crc kubenswrapper[4805]: I1203 14:30:02.240231 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/83e323ed-ad2c-4635-a583-4e5b21d8e268-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "83e323ed-ad2c-4635-a583-4e5b21d8e268" (UID: "83e323ed-ad2c-4635-a583-4e5b21d8e268"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 14:30:02 crc kubenswrapper[4805]: I1203 14:30:02.241337 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/83e323ed-ad2c-4635-a583-4e5b21d8e268-config" (OuterVolumeSpecName: "config") pod "83e323ed-ad2c-4635-a583-4e5b21d8e268" (UID: "83e323ed-ad2c-4635-a583-4e5b21d8e268"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 14:30:02 crc kubenswrapper[4805]: I1203 14:30:02.243611 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/83e323ed-ad2c-4635-a583-4e5b21d8e268-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "83e323ed-ad2c-4635-a583-4e5b21d8e268" (UID: "83e323ed-ad2c-4635-a583-4e5b21d8e268"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 14:30:02 crc kubenswrapper[4805]: I1203 14:30:02.247443 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/83e323ed-ad2c-4635-a583-4e5b21d8e268-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "83e323ed-ad2c-4635-a583-4e5b21d8e268" (UID: "83e323ed-ad2c-4635-a583-4e5b21d8e268"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 14:30:02 crc kubenswrapper[4805]: I1203 14:30:02.266730 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29412870-46tbm"] Dec 03 14:30:02 crc kubenswrapper[4805]: I1203 14:30:02.302763 4805 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/eda36a78-7b3a-43f8-858e-fc6b8fdd442a-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 03 14:30:02 crc kubenswrapper[4805]: I1203 14:30:02.302813 4805 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qj4w8\" (UniqueName: \"kubernetes.io/projected/83e323ed-ad2c-4635-a583-4e5b21d8e268-kube-api-access-qj4w8\") on node \"crc\" DevicePath \"\"" Dec 03 14:30:02 crc kubenswrapper[4805]: I1203 14:30:02.302828 4805 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/83e323ed-ad2c-4635-a583-4e5b21d8e268-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Dec 03 14:30:02 crc kubenswrapper[4805]: I1203 14:30:02.302849 4805 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/10163a58-aa6e-4988-8b33-93f8cffa4cc0-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 03 14:30:02 crc kubenswrapper[4805]: I1203 14:30:02.302858 4805 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/83e323ed-ad2c-4635-a583-4e5b21d8e268-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Dec 03 14:30:02 crc kubenswrapper[4805]: I1203 14:30:02.302900 4805 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/83e323ed-ad2c-4635-a583-4e5b21d8e268-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 03 14:30:02 crc kubenswrapper[4805]: I1203 14:30:02.302910 4805 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8rz8t\" (UniqueName: \"kubernetes.io/projected/b27b4fde-29ff-421d-a6be-5d259bfe8016-kube-api-access-8rz8t\") on node \"crc\" DevicePath \"\"" Dec 03 14:30:02 crc kubenswrapper[4805]: I1203 14:30:02.302919 4805 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lx9th\" (UniqueName: \"kubernetes.io/projected/eda36a78-7b3a-43f8-858e-fc6b8fdd442a-kube-api-access-lx9th\") on node \"crc\" DevicePath \"\"" Dec 03 14:30:02 crc kubenswrapper[4805]: I1203 14:30:02.302927 4805 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/b27b4fde-29ff-421d-a6be-5d259bfe8016-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 03 14:30:02 crc kubenswrapper[4805]: I1203 14:30:02.302935 4805 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gfgzp\" (UniqueName: \"kubernetes.io/projected/10163a58-aa6e-4988-8b33-93f8cffa4cc0-kube-api-access-gfgzp\") on node \"crc\" DevicePath \"\"" Dec 03 14:30:02 crc kubenswrapper[4805]: I1203 14:30:02.302945 4805 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/83e323ed-ad2c-4635-a583-4e5b21d8e268-config\") on node \"crc\" DevicePath \"\"" Dec 03 14:30:02 crc kubenswrapper[4805]: I1203 14:30:02.970792 4805 generic.go:334] "Generic (PLEG): container finished" podID="5aae8010-7566-4f0b-ba0f-8d905bc9e455" containerID="36e5b2355468d0b356654a149f8c3bcfd0ca25fcf904e7d3354aa72bd9a641ec" exitCode=0 Dec 03 14:30:02 crc kubenswrapper[4805]: I1203 14:30:02.971195 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29412870-46tbm" event={"ID":"5aae8010-7566-4f0b-ba0f-8d905bc9e455","Type":"ContainerDied","Data":"36e5b2355468d0b356654a149f8c3bcfd0ca25fcf904e7d3354aa72bd9a641ec"} Dec 03 14:30:02 crc kubenswrapper[4805]: I1203 14:30:02.971219 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29412870-46tbm" event={"ID":"5aae8010-7566-4f0b-ba0f-8d905bc9e455","Type":"ContainerStarted","Data":"4382eca18d3c0f1ed472c23accffda1824e8aa1afebefd3128e603c701c46737"} Dec 03 14:30:02 crc kubenswrapper[4805]: I1203 14:30:02.973363 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-sync-75wdv" event={"ID":"c15336c7-d573-432f-8b4c-ba5a39c82cf8","Type":"ContainerStarted","Data":"f85ba51232525929e730722508afdb428d6d624935f15d634a7f4ad1b9ec47b9"} Dec 03 14:30:02 crc kubenswrapper[4805]: I1203 14:30:02.975368 4805 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-7ac1-account-create-update-6jk6w" Dec 03 14:30:02 crc kubenswrapper[4805]: I1203 14:30:02.975375 4805 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-create-q7qlk" Dec 03 14:30:02 crc kubenswrapper[4805]: I1203 14:30:02.975391 4805 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/heat-ee1c-account-create-update-g8lk4" Dec 03 14:30:02 crc kubenswrapper[4805]: I1203 14:30:02.975414 4805 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-698758b865-l46rr" Dec 03 14:30:02 crc kubenswrapper[4805]: I1203 14:30:02.975414 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-698758b865-l46rr" event={"ID":"83e323ed-ad2c-4635-a583-4e5b21d8e268","Type":"ContainerDied","Data":"38e9567f9c900cc2c54caf706a009e55dc558787efb3fb6195719b107e2a2247"} Dec 03 14:30:02 crc kubenswrapper[4805]: I1203 14:30:02.975479 4805 scope.go:117] "RemoveContainer" containerID="55ae092c6099423baa000af1306ee0efd34017bebbf5abd512f35e371752d635" Dec 03 14:30:02 crc kubenswrapper[4805]: I1203 14:30:02.975498 4805 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-3c27-account-create-update-79pjf" Dec 03 14:30:02 crc kubenswrapper[4805]: I1203 14:30:02.975519 4805 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-77e9-account-create-update-4mkc7" Dec 03 14:30:02 crc kubenswrapper[4805]: I1203 14:30:02.975532 4805 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/heat-db-create-ztcqj" Dec 03 14:30:02 crc kubenswrapper[4805]: I1203 14:30:02.975564 4805 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-create-mhn6v" Dec 03 14:30:02 crc kubenswrapper[4805]: I1203 14:30:02.975626 4805 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-create-7bqcx" Dec 03 14:30:03 crc kubenswrapper[4805]: I1203 14:30:03.004750 4805 scope.go:117] "RemoveContainer" containerID="4d555253ea8f012425b96f097443fb842d8d499049efb2c8bb321d17e8ee0788" Dec 03 14:30:03 crc kubenswrapper[4805]: I1203 14:30:03.012151 4805 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-db-sync-75wdv" podStartSLOduration=3.030915815 podStartE2EDuration="11.012128515s" podCreationTimestamp="2025-12-03 14:29:52 +0000 UTC" firstStartedPulling="2025-12-03 14:29:53.858178179 +0000 UTC m=+1223.521095102" lastFinishedPulling="2025-12-03 14:30:01.839390879 +0000 UTC m=+1231.502307802" observedRunningTime="2025-12-03 14:30:03.008222339 +0000 UTC m=+1232.671139262" watchObservedRunningTime="2025-12-03 14:30:03.012128515 +0000 UTC m=+1232.675045468" Dec 03 14:30:03 crc kubenswrapper[4805]: I1203 14:30:03.068723 4805 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-698758b865-l46rr"] Dec 03 14:30:03 crc kubenswrapper[4805]: I1203 14:30:03.075762 4805 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-698758b865-l46rr"] Dec 03 14:30:04 crc kubenswrapper[4805]: I1203 14:30:04.318598 4805 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29412870-46tbm" Dec 03 14:30:04 crc kubenswrapper[4805]: I1203 14:30:04.448261 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/5aae8010-7566-4f0b-ba0f-8d905bc9e455-config-volume\") pod \"5aae8010-7566-4f0b-ba0f-8d905bc9e455\" (UID: \"5aae8010-7566-4f0b-ba0f-8d905bc9e455\") " Dec 03 14:30:04 crc kubenswrapper[4805]: I1203 14:30:04.448366 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dw6kr\" (UniqueName: \"kubernetes.io/projected/5aae8010-7566-4f0b-ba0f-8d905bc9e455-kube-api-access-dw6kr\") pod \"5aae8010-7566-4f0b-ba0f-8d905bc9e455\" (UID: \"5aae8010-7566-4f0b-ba0f-8d905bc9e455\") " Dec 03 14:30:04 crc kubenswrapper[4805]: I1203 14:30:04.448400 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/5aae8010-7566-4f0b-ba0f-8d905bc9e455-secret-volume\") pod \"5aae8010-7566-4f0b-ba0f-8d905bc9e455\" (UID: \"5aae8010-7566-4f0b-ba0f-8d905bc9e455\") " Dec 03 14:30:04 crc kubenswrapper[4805]: I1203 14:30:04.448931 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5aae8010-7566-4f0b-ba0f-8d905bc9e455-config-volume" (OuterVolumeSpecName: "config-volume") pod "5aae8010-7566-4f0b-ba0f-8d905bc9e455" (UID: "5aae8010-7566-4f0b-ba0f-8d905bc9e455"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 14:30:04 crc kubenswrapper[4805]: I1203 14:30:04.455118 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5aae8010-7566-4f0b-ba0f-8d905bc9e455-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "5aae8010-7566-4f0b-ba0f-8d905bc9e455" (UID: "5aae8010-7566-4f0b-ba0f-8d905bc9e455"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 14:30:04 crc kubenswrapper[4805]: I1203 14:30:04.456052 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5aae8010-7566-4f0b-ba0f-8d905bc9e455-kube-api-access-dw6kr" (OuterVolumeSpecName: "kube-api-access-dw6kr") pod "5aae8010-7566-4f0b-ba0f-8d905bc9e455" (UID: "5aae8010-7566-4f0b-ba0f-8d905bc9e455"). InnerVolumeSpecName "kube-api-access-dw6kr". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 14:30:04 crc kubenswrapper[4805]: I1203 14:30:04.550065 4805 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/5aae8010-7566-4f0b-ba0f-8d905bc9e455-config-volume\") on node \"crc\" DevicePath \"\"" Dec 03 14:30:04 crc kubenswrapper[4805]: I1203 14:30:04.550095 4805 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dw6kr\" (UniqueName: \"kubernetes.io/projected/5aae8010-7566-4f0b-ba0f-8d905bc9e455-kube-api-access-dw6kr\") on node \"crc\" DevicePath \"\"" Dec 03 14:30:04 crc kubenswrapper[4805]: I1203 14:30:04.550109 4805 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/5aae8010-7566-4f0b-ba0f-8d905bc9e455-secret-volume\") on node \"crc\" DevicePath \"\"" Dec 03 14:30:04 crc kubenswrapper[4805]: I1203 14:30:04.707563 4805 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="83e323ed-ad2c-4635-a583-4e5b21d8e268" path="/var/lib/kubelet/pods/83e323ed-ad2c-4635-a583-4e5b21d8e268/volumes" Dec 03 14:30:04 crc kubenswrapper[4805]: I1203 14:30:04.993290 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29412870-46tbm" event={"ID":"5aae8010-7566-4f0b-ba0f-8d905bc9e455","Type":"ContainerDied","Data":"4382eca18d3c0f1ed472c23accffda1824e8aa1afebefd3128e603c701c46737"} Dec 03 14:30:04 crc kubenswrapper[4805]: I1203 14:30:04.993640 4805 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="4382eca18d3c0f1ed472c23accffda1824e8aa1afebefd3128e603c701c46737" Dec 03 14:30:04 crc kubenswrapper[4805]: I1203 14:30:04.993309 4805 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29412870-46tbm" Dec 03 14:30:04 crc kubenswrapper[4805]: I1203 14:30:04.998136 4805 generic.go:334] "Generic (PLEG): container finished" podID="c15336c7-d573-432f-8b4c-ba5a39c82cf8" containerID="f85ba51232525929e730722508afdb428d6d624935f15d634a7f4ad1b9ec47b9" exitCode=0 Dec 03 14:30:04 crc kubenswrapper[4805]: I1203 14:30:04.998180 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-sync-75wdv" event={"ID":"c15336c7-d573-432f-8b4c-ba5a39c82cf8","Type":"ContainerDied","Data":"f85ba51232525929e730722508afdb428d6d624935f15d634a7f4ad1b9ec47b9"} Dec 03 14:30:06 crc kubenswrapper[4805]: I1203 14:30:06.363504 4805 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-sync-75wdv" Dec 03 14:30:06 crc kubenswrapper[4805]: I1203 14:30:06.501212 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c15336c7-d573-432f-8b4c-ba5a39c82cf8-combined-ca-bundle\") pod \"c15336c7-d573-432f-8b4c-ba5a39c82cf8\" (UID: \"c15336c7-d573-432f-8b4c-ba5a39c82cf8\") " Dec 03 14:30:06 crc kubenswrapper[4805]: I1203 14:30:06.501329 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-56b6j\" (UniqueName: \"kubernetes.io/projected/c15336c7-d573-432f-8b4c-ba5a39c82cf8-kube-api-access-56b6j\") pod \"c15336c7-d573-432f-8b4c-ba5a39c82cf8\" (UID: \"c15336c7-d573-432f-8b4c-ba5a39c82cf8\") " Dec 03 14:30:06 crc kubenswrapper[4805]: I1203 14:30:06.501453 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c15336c7-d573-432f-8b4c-ba5a39c82cf8-config-data\") pod \"c15336c7-d573-432f-8b4c-ba5a39c82cf8\" (UID: \"c15336c7-d573-432f-8b4c-ba5a39c82cf8\") " Dec 03 14:30:06 crc kubenswrapper[4805]: I1203 14:30:06.506577 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c15336c7-d573-432f-8b4c-ba5a39c82cf8-kube-api-access-56b6j" (OuterVolumeSpecName: "kube-api-access-56b6j") pod "c15336c7-d573-432f-8b4c-ba5a39c82cf8" (UID: "c15336c7-d573-432f-8b4c-ba5a39c82cf8"). InnerVolumeSpecName "kube-api-access-56b6j". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 14:30:06 crc kubenswrapper[4805]: I1203 14:30:06.528144 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c15336c7-d573-432f-8b4c-ba5a39c82cf8-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "c15336c7-d573-432f-8b4c-ba5a39c82cf8" (UID: "c15336c7-d573-432f-8b4c-ba5a39c82cf8"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 14:30:06 crc kubenswrapper[4805]: I1203 14:30:06.549200 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c15336c7-d573-432f-8b4c-ba5a39c82cf8-config-data" (OuterVolumeSpecName: "config-data") pod "c15336c7-d573-432f-8b4c-ba5a39c82cf8" (UID: "c15336c7-d573-432f-8b4c-ba5a39c82cf8"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 14:30:06 crc kubenswrapper[4805]: I1203 14:30:06.603567 4805 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-56b6j\" (UniqueName: \"kubernetes.io/projected/c15336c7-d573-432f-8b4c-ba5a39c82cf8-kube-api-access-56b6j\") on node \"crc\" DevicePath \"\"" Dec 03 14:30:06 crc kubenswrapper[4805]: I1203 14:30:06.603815 4805 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c15336c7-d573-432f-8b4c-ba5a39c82cf8-config-data\") on node \"crc\" DevicePath \"\"" Dec 03 14:30:06 crc kubenswrapper[4805]: I1203 14:30:06.603824 4805 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c15336c7-d573-432f-8b4c-ba5a39c82cf8-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 14:30:07 crc kubenswrapper[4805]: I1203 14:30:07.025823 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-sync-75wdv" event={"ID":"c15336c7-d573-432f-8b4c-ba5a39c82cf8","Type":"ContainerDied","Data":"1c2b207a6035d84e65253753a49163fba7166f10f38bf68f08db84370b2265b0"} Dec 03 14:30:07 crc kubenswrapper[4805]: I1203 14:30:07.025967 4805 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="1c2b207a6035d84e65253753a49163fba7166f10f38bf68f08db84370b2265b0" Dec 03 14:30:07 crc kubenswrapper[4805]: I1203 14:30:07.026077 4805 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-sync-75wdv" Dec 03 14:30:07 crc kubenswrapper[4805]: I1203 14:30:07.648075 4805 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-5c5cc7c5ff-kxgt6"] Dec 03 14:30:07 crc kubenswrapper[4805]: E1203 14:30:07.648636 4805 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c15336c7-d573-432f-8b4c-ba5a39c82cf8" containerName="keystone-db-sync" Dec 03 14:30:07 crc kubenswrapper[4805]: I1203 14:30:07.648660 4805 state_mem.go:107] "Deleted CPUSet assignment" podUID="c15336c7-d573-432f-8b4c-ba5a39c82cf8" containerName="keystone-db-sync" Dec 03 14:30:07 crc kubenswrapper[4805]: E1203 14:30:07.648675 4805 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="83e323ed-ad2c-4635-a583-4e5b21d8e268" containerName="dnsmasq-dns" Dec 03 14:30:07 crc kubenswrapper[4805]: I1203 14:30:07.648687 4805 state_mem.go:107] "Deleted CPUSet assignment" podUID="83e323ed-ad2c-4635-a583-4e5b21d8e268" containerName="dnsmasq-dns" Dec 03 14:30:07 crc kubenswrapper[4805]: E1203 14:30:07.648702 4805 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d319f692-2f37-4a01-8e6c-bbf0e7b23c8e" containerName="mariadb-database-create" Dec 03 14:30:07 crc kubenswrapper[4805]: I1203 14:30:07.648714 4805 state_mem.go:107] "Deleted CPUSet assignment" podUID="d319f692-2f37-4a01-8e6c-bbf0e7b23c8e" containerName="mariadb-database-create" Dec 03 14:30:07 crc kubenswrapper[4805]: E1203 14:30:07.648741 4805 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5aae8010-7566-4f0b-ba0f-8d905bc9e455" containerName="collect-profiles" Dec 03 14:30:07 crc kubenswrapper[4805]: I1203 14:30:07.648752 4805 state_mem.go:107] "Deleted CPUSet assignment" podUID="5aae8010-7566-4f0b-ba0f-8d905bc9e455" containerName="collect-profiles" Dec 03 14:30:07 crc kubenswrapper[4805]: E1203 14:30:07.648768 4805 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="696fec58-2abe-474b-956c-61228970cc74" containerName="mariadb-account-create-update" Dec 03 14:30:07 crc kubenswrapper[4805]: I1203 14:30:07.648778 4805 state_mem.go:107] "Deleted CPUSet assignment" podUID="696fec58-2abe-474b-956c-61228970cc74" containerName="mariadb-account-create-update" Dec 03 14:30:07 crc kubenswrapper[4805]: E1203 14:30:07.648798 4805 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="10163a58-aa6e-4988-8b33-93f8cffa4cc0" containerName="mariadb-database-create" Dec 03 14:30:07 crc kubenswrapper[4805]: I1203 14:30:07.648808 4805 state_mem.go:107] "Deleted CPUSet assignment" podUID="10163a58-aa6e-4988-8b33-93f8cffa4cc0" containerName="mariadb-database-create" Dec 03 14:30:07 crc kubenswrapper[4805]: E1203 14:30:07.648826 4805 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="99891296-b622-466d-95f7-bcf0e3ebe896" containerName="mariadb-account-create-update" Dec 03 14:30:07 crc kubenswrapper[4805]: I1203 14:30:07.648859 4805 state_mem.go:107] "Deleted CPUSet assignment" podUID="99891296-b622-466d-95f7-bcf0e3ebe896" containerName="mariadb-account-create-update" Dec 03 14:30:07 crc kubenswrapper[4805]: E1203 14:30:07.648882 4805 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="eda36a78-7b3a-43f8-858e-fc6b8fdd442a" containerName="mariadb-account-create-update" Dec 03 14:30:07 crc kubenswrapper[4805]: I1203 14:30:07.648894 4805 state_mem.go:107] "Deleted CPUSet assignment" podUID="eda36a78-7b3a-43f8-858e-fc6b8fdd442a" containerName="mariadb-account-create-update" Dec 03 14:30:07 crc kubenswrapper[4805]: E1203 14:30:07.648914 4805 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b27b4fde-29ff-421d-a6be-5d259bfe8016" containerName="mariadb-account-create-update" Dec 03 14:30:07 crc kubenswrapper[4805]: I1203 14:30:07.648925 4805 state_mem.go:107] "Deleted CPUSet assignment" podUID="b27b4fde-29ff-421d-a6be-5d259bfe8016" containerName="mariadb-account-create-update" Dec 03 14:30:07 crc kubenswrapper[4805]: E1203 14:30:07.648936 4805 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="18be72b0-25d0-41b4-8038-abd0fb38deba" containerName="mariadb-database-create" Dec 03 14:30:07 crc kubenswrapper[4805]: I1203 14:30:07.648946 4805 state_mem.go:107] "Deleted CPUSet assignment" podUID="18be72b0-25d0-41b4-8038-abd0fb38deba" containerName="mariadb-database-create" Dec 03 14:30:07 crc kubenswrapper[4805]: E1203 14:30:07.648968 4805 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="83e323ed-ad2c-4635-a583-4e5b21d8e268" containerName="init" Dec 03 14:30:07 crc kubenswrapper[4805]: I1203 14:30:07.648980 4805 state_mem.go:107] "Deleted CPUSet assignment" podUID="83e323ed-ad2c-4635-a583-4e5b21d8e268" containerName="init" Dec 03 14:30:07 crc kubenswrapper[4805]: E1203 14:30:07.649007 4805 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="aead85c7-af73-44d0-b4dc-5cd98d6439b3" containerName="mariadb-database-create" Dec 03 14:30:07 crc kubenswrapper[4805]: I1203 14:30:07.649017 4805 state_mem.go:107] "Deleted CPUSet assignment" podUID="aead85c7-af73-44d0-b4dc-5cd98d6439b3" containerName="mariadb-database-create" Dec 03 14:30:07 crc kubenswrapper[4805]: I1203 14:30:07.649267 4805 memory_manager.go:354] "RemoveStaleState removing state" podUID="83e323ed-ad2c-4635-a583-4e5b21d8e268" containerName="dnsmasq-dns" Dec 03 14:30:07 crc kubenswrapper[4805]: I1203 14:30:07.649284 4805 memory_manager.go:354] "RemoveStaleState removing state" podUID="eda36a78-7b3a-43f8-858e-fc6b8fdd442a" containerName="mariadb-account-create-update" Dec 03 14:30:07 crc kubenswrapper[4805]: I1203 14:30:07.649301 4805 memory_manager.go:354] "RemoveStaleState removing state" podUID="99891296-b622-466d-95f7-bcf0e3ebe896" containerName="mariadb-account-create-update" Dec 03 14:30:07 crc kubenswrapper[4805]: I1203 14:30:07.649321 4805 memory_manager.go:354] "RemoveStaleState removing state" podUID="c15336c7-d573-432f-8b4c-ba5a39c82cf8" containerName="keystone-db-sync" Dec 03 14:30:07 crc kubenswrapper[4805]: I1203 14:30:07.649341 4805 memory_manager.go:354] "RemoveStaleState removing state" podUID="696fec58-2abe-474b-956c-61228970cc74" containerName="mariadb-account-create-update" Dec 03 14:30:07 crc kubenswrapper[4805]: I1203 14:30:07.649357 4805 memory_manager.go:354] "RemoveStaleState removing state" podUID="10163a58-aa6e-4988-8b33-93f8cffa4cc0" containerName="mariadb-database-create" Dec 03 14:30:07 crc kubenswrapper[4805]: I1203 14:30:07.649374 4805 memory_manager.go:354] "RemoveStaleState removing state" podUID="b27b4fde-29ff-421d-a6be-5d259bfe8016" containerName="mariadb-account-create-update" Dec 03 14:30:07 crc kubenswrapper[4805]: I1203 14:30:07.649392 4805 memory_manager.go:354] "RemoveStaleState removing state" podUID="18be72b0-25d0-41b4-8038-abd0fb38deba" containerName="mariadb-database-create" Dec 03 14:30:07 crc kubenswrapper[4805]: I1203 14:30:07.649408 4805 memory_manager.go:354] "RemoveStaleState removing state" podUID="5aae8010-7566-4f0b-ba0f-8d905bc9e455" containerName="collect-profiles" Dec 03 14:30:07 crc kubenswrapper[4805]: I1203 14:30:07.649427 4805 memory_manager.go:354] "RemoveStaleState removing state" podUID="aead85c7-af73-44d0-b4dc-5cd98d6439b3" containerName="mariadb-database-create" Dec 03 14:30:07 crc kubenswrapper[4805]: I1203 14:30:07.649446 4805 memory_manager.go:354] "RemoveStaleState removing state" podUID="d319f692-2f37-4a01-8e6c-bbf0e7b23c8e" containerName="mariadb-database-create" Dec 03 14:30:07 crc kubenswrapper[4805]: I1203 14:30:07.651085 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5c5cc7c5ff-kxgt6" Dec 03 14:30:07 crc kubenswrapper[4805]: I1203 14:30:07.670965 4805 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-bootstrap-qk4m4"] Dec 03 14:30:07 crc kubenswrapper[4805]: I1203 14:30:07.672288 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-qk4m4" Dec 03 14:30:07 crc kubenswrapper[4805]: I1203 14:30:07.679529 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-config-data" Dec 03 14:30:07 crc kubenswrapper[4805]: I1203 14:30:07.679756 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"osp-secret" Dec 03 14:30:07 crc kubenswrapper[4805]: I1203 14:30:07.679918 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-keystone-dockercfg-5hcz5" Dec 03 14:30:07 crc kubenswrapper[4805]: I1203 14:30:07.680069 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-scripts" Dec 03 14:30:07 crc kubenswrapper[4805]: I1203 14:30:07.680255 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone" Dec 03 14:30:07 crc kubenswrapper[4805]: I1203 14:30:07.686629 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-bootstrap-qk4m4"] Dec 03 14:30:07 crc kubenswrapper[4805]: I1203 14:30:07.709112 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5c5cc7c5ff-kxgt6"] Dec 03 14:30:07 crc kubenswrapper[4805]: I1203 14:30:07.721668 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/d1b43e5a-9dc5-4b92-a5dd-a22beff9f447-dns-swift-storage-0\") pod \"dnsmasq-dns-5c5cc7c5ff-kxgt6\" (UID: \"d1b43e5a-9dc5-4b92-a5dd-a22beff9f447\") " pod="openstack/dnsmasq-dns-5c5cc7c5ff-kxgt6" Dec 03 14:30:07 crc kubenswrapper[4805]: I1203 14:30:07.721750 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/d1b43e5a-9dc5-4b92-a5dd-a22beff9f447-dns-svc\") pod \"dnsmasq-dns-5c5cc7c5ff-kxgt6\" (UID: \"d1b43e5a-9dc5-4b92-a5dd-a22beff9f447\") " pod="openstack/dnsmasq-dns-5c5cc7c5ff-kxgt6" Dec 03 14:30:07 crc kubenswrapper[4805]: I1203 14:30:07.721773 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/d1b43e5a-9dc5-4b92-a5dd-a22beff9f447-ovsdbserver-sb\") pod \"dnsmasq-dns-5c5cc7c5ff-kxgt6\" (UID: \"d1b43e5a-9dc5-4b92-a5dd-a22beff9f447\") " pod="openstack/dnsmasq-dns-5c5cc7c5ff-kxgt6" Dec 03 14:30:07 crc kubenswrapper[4805]: I1203 14:30:07.721804 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2k58p\" (UniqueName: \"kubernetes.io/projected/d1b43e5a-9dc5-4b92-a5dd-a22beff9f447-kube-api-access-2k58p\") pod \"dnsmasq-dns-5c5cc7c5ff-kxgt6\" (UID: \"d1b43e5a-9dc5-4b92-a5dd-a22beff9f447\") " pod="openstack/dnsmasq-dns-5c5cc7c5ff-kxgt6" Dec 03 14:30:07 crc kubenswrapper[4805]: I1203 14:30:07.721833 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/d1b43e5a-9dc5-4b92-a5dd-a22beff9f447-ovsdbserver-nb\") pod \"dnsmasq-dns-5c5cc7c5ff-kxgt6\" (UID: \"d1b43e5a-9dc5-4b92-a5dd-a22beff9f447\") " pod="openstack/dnsmasq-dns-5c5cc7c5ff-kxgt6" Dec 03 14:30:07 crc kubenswrapper[4805]: I1203 14:30:07.721886 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d1b43e5a-9dc5-4b92-a5dd-a22beff9f447-config\") pod \"dnsmasq-dns-5c5cc7c5ff-kxgt6\" (UID: \"d1b43e5a-9dc5-4b92-a5dd-a22beff9f447\") " pod="openstack/dnsmasq-dns-5c5cc7c5ff-kxgt6" Dec 03 14:30:07 crc kubenswrapper[4805]: I1203 14:30:07.823790 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/d1b43e5a-9dc5-4b92-a5dd-a22beff9f447-dns-swift-storage-0\") pod \"dnsmasq-dns-5c5cc7c5ff-kxgt6\" (UID: \"d1b43e5a-9dc5-4b92-a5dd-a22beff9f447\") " pod="openstack/dnsmasq-dns-5c5cc7c5ff-kxgt6" Dec 03 14:30:07 crc kubenswrapper[4805]: I1203 14:30:07.823920 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/d1b43e5a-9dc5-4b92-a5dd-a22beff9f447-dns-svc\") pod \"dnsmasq-dns-5c5cc7c5ff-kxgt6\" (UID: \"d1b43e5a-9dc5-4b92-a5dd-a22beff9f447\") " pod="openstack/dnsmasq-dns-5c5cc7c5ff-kxgt6" Dec 03 14:30:07 crc kubenswrapper[4805]: I1203 14:30:07.823955 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/d1b43e5a-9dc5-4b92-a5dd-a22beff9f447-ovsdbserver-sb\") pod \"dnsmasq-dns-5c5cc7c5ff-kxgt6\" (UID: \"d1b43e5a-9dc5-4b92-a5dd-a22beff9f447\") " pod="openstack/dnsmasq-dns-5c5cc7c5ff-kxgt6" Dec 03 14:30:07 crc kubenswrapper[4805]: I1203 14:30:07.823980 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/16dbab06-8ff8-4a5a-9fe8-4967c6838a4a-fernet-keys\") pod \"keystone-bootstrap-qk4m4\" (UID: \"16dbab06-8ff8-4a5a-9fe8-4967c6838a4a\") " pod="openstack/keystone-bootstrap-qk4m4" Dec 03 14:30:07 crc kubenswrapper[4805]: I1203 14:30:07.824012 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kbxbj\" (UniqueName: \"kubernetes.io/projected/16dbab06-8ff8-4a5a-9fe8-4967c6838a4a-kube-api-access-kbxbj\") pod \"keystone-bootstrap-qk4m4\" (UID: \"16dbab06-8ff8-4a5a-9fe8-4967c6838a4a\") " pod="openstack/keystone-bootstrap-qk4m4" Dec 03 14:30:07 crc kubenswrapper[4805]: I1203 14:30:07.824059 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2k58p\" (UniqueName: \"kubernetes.io/projected/d1b43e5a-9dc5-4b92-a5dd-a22beff9f447-kube-api-access-2k58p\") pod \"dnsmasq-dns-5c5cc7c5ff-kxgt6\" (UID: \"d1b43e5a-9dc5-4b92-a5dd-a22beff9f447\") " pod="openstack/dnsmasq-dns-5c5cc7c5ff-kxgt6" Dec 03 14:30:07 crc kubenswrapper[4805]: I1203 14:30:07.824099 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/16dbab06-8ff8-4a5a-9fe8-4967c6838a4a-credential-keys\") pod \"keystone-bootstrap-qk4m4\" (UID: \"16dbab06-8ff8-4a5a-9fe8-4967c6838a4a\") " pod="openstack/keystone-bootstrap-qk4m4" Dec 03 14:30:07 crc kubenswrapper[4805]: I1203 14:30:07.824148 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/d1b43e5a-9dc5-4b92-a5dd-a22beff9f447-ovsdbserver-nb\") pod \"dnsmasq-dns-5c5cc7c5ff-kxgt6\" (UID: \"d1b43e5a-9dc5-4b92-a5dd-a22beff9f447\") " pod="openstack/dnsmasq-dns-5c5cc7c5ff-kxgt6" Dec 03 14:30:07 crc kubenswrapper[4805]: I1203 14:30:07.824218 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/16dbab06-8ff8-4a5a-9fe8-4967c6838a4a-combined-ca-bundle\") pod \"keystone-bootstrap-qk4m4\" (UID: \"16dbab06-8ff8-4a5a-9fe8-4967c6838a4a\") " pod="openstack/keystone-bootstrap-qk4m4" Dec 03 14:30:07 crc kubenswrapper[4805]: I1203 14:30:07.824246 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/16dbab06-8ff8-4a5a-9fe8-4967c6838a4a-config-data\") pod \"keystone-bootstrap-qk4m4\" (UID: \"16dbab06-8ff8-4a5a-9fe8-4967c6838a4a\") " pod="openstack/keystone-bootstrap-qk4m4" Dec 03 14:30:07 crc kubenswrapper[4805]: I1203 14:30:07.824274 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d1b43e5a-9dc5-4b92-a5dd-a22beff9f447-config\") pod \"dnsmasq-dns-5c5cc7c5ff-kxgt6\" (UID: \"d1b43e5a-9dc5-4b92-a5dd-a22beff9f447\") " pod="openstack/dnsmasq-dns-5c5cc7c5ff-kxgt6" Dec 03 14:30:07 crc kubenswrapper[4805]: I1203 14:30:07.824297 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/16dbab06-8ff8-4a5a-9fe8-4967c6838a4a-scripts\") pod \"keystone-bootstrap-qk4m4\" (UID: \"16dbab06-8ff8-4a5a-9fe8-4967c6838a4a\") " pod="openstack/keystone-bootstrap-qk4m4" Dec 03 14:30:07 crc kubenswrapper[4805]: I1203 14:30:07.825277 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/d1b43e5a-9dc5-4b92-a5dd-a22beff9f447-dns-swift-storage-0\") pod \"dnsmasq-dns-5c5cc7c5ff-kxgt6\" (UID: \"d1b43e5a-9dc5-4b92-a5dd-a22beff9f447\") " pod="openstack/dnsmasq-dns-5c5cc7c5ff-kxgt6" Dec 03 14:30:07 crc kubenswrapper[4805]: I1203 14:30:07.826453 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/d1b43e5a-9dc5-4b92-a5dd-a22beff9f447-dns-svc\") pod \"dnsmasq-dns-5c5cc7c5ff-kxgt6\" (UID: \"d1b43e5a-9dc5-4b92-a5dd-a22beff9f447\") " pod="openstack/dnsmasq-dns-5c5cc7c5ff-kxgt6" Dec 03 14:30:07 crc kubenswrapper[4805]: I1203 14:30:07.826722 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d1b43e5a-9dc5-4b92-a5dd-a22beff9f447-config\") pod \"dnsmasq-dns-5c5cc7c5ff-kxgt6\" (UID: \"d1b43e5a-9dc5-4b92-a5dd-a22beff9f447\") " pod="openstack/dnsmasq-dns-5c5cc7c5ff-kxgt6" Dec 03 14:30:07 crc kubenswrapper[4805]: I1203 14:30:07.827084 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/d1b43e5a-9dc5-4b92-a5dd-a22beff9f447-ovsdbserver-nb\") pod \"dnsmasq-dns-5c5cc7c5ff-kxgt6\" (UID: \"d1b43e5a-9dc5-4b92-a5dd-a22beff9f447\") " pod="openstack/dnsmasq-dns-5c5cc7c5ff-kxgt6" Dec 03 14:30:07 crc kubenswrapper[4805]: I1203 14:30:07.827152 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/d1b43e5a-9dc5-4b92-a5dd-a22beff9f447-ovsdbserver-sb\") pod \"dnsmasq-dns-5c5cc7c5ff-kxgt6\" (UID: \"d1b43e5a-9dc5-4b92-a5dd-a22beff9f447\") " pod="openstack/dnsmasq-dns-5c5cc7c5ff-kxgt6" Dec 03 14:30:07 crc kubenswrapper[4805]: I1203 14:30:07.866420 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2k58p\" (UniqueName: \"kubernetes.io/projected/d1b43e5a-9dc5-4b92-a5dd-a22beff9f447-kube-api-access-2k58p\") pod \"dnsmasq-dns-5c5cc7c5ff-kxgt6\" (UID: \"d1b43e5a-9dc5-4b92-a5dd-a22beff9f447\") " pod="openstack/dnsmasq-dns-5c5cc7c5ff-kxgt6" Dec 03 14:30:07 crc kubenswrapper[4805]: I1203 14:30:07.896031 4805 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Dec 03 14:30:07 crc kubenswrapper[4805]: I1203 14:30:07.898056 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 03 14:30:07 crc kubenswrapper[4805]: I1203 14:30:07.899938 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Dec 03 14:30:07 crc kubenswrapper[4805]: I1203 14:30:07.900175 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Dec 03 14:30:07 crc kubenswrapper[4805]: I1203 14:30:07.925754 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/16dbab06-8ff8-4a5a-9fe8-4967c6838a4a-combined-ca-bundle\") pod \"keystone-bootstrap-qk4m4\" (UID: \"16dbab06-8ff8-4a5a-9fe8-4967c6838a4a\") " pod="openstack/keystone-bootstrap-qk4m4" Dec 03 14:30:07 crc kubenswrapper[4805]: I1203 14:30:07.925810 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/16dbab06-8ff8-4a5a-9fe8-4967c6838a4a-config-data\") pod \"keystone-bootstrap-qk4m4\" (UID: \"16dbab06-8ff8-4a5a-9fe8-4967c6838a4a\") " pod="openstack/keystone-bootstrap-qk4m4" Dec 03 14:30:07 crc kubenswrapper[4805]: I1203 14:30:07.925893 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/16dbab06-8ff8-4a5a-9fe8-4967c6838a4a-scripts\") pod \"keystone-bootstrap-qk4m4\" (UID: \"16dbab06-8ff8-4a5a-9fe8-4967c6838a4a\") " pod="openstack/keystone-bootstrap-qk4m4" Dec 03 14:30:07 crc kubenswrapper[4805]: I1203 14:30:07.925986 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/16dbab06-8ff8-4a5a-9fe8-4967c6838a4a-fernet-keys\") pod \"keystone-bootstrap-qk4m4\" (UID: \"16dbab06-8ff8-4a5a-9fe8-4967c6838a4a\") " pod="openstack/keystone-bootstrap-qk4m4" Dec 03 14:30:07 crc kubenswrapper[4805]: I1203 14:30:07.926009 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kbxbj\" (UniqueName: \"kubernetes.io/projected/16dbab06-8ff8-4a5a-9fe8-4967c6838a4a-kube-api-access-kbxbj\") pod \"keystone-bootstrap-qk4m4\" (UID: \"16dbab06-8ff8-4a5a-9fe8-4967c6838a4a\") " pod="openstack/keystone-bootstrap-qk4m4" Dec 03 14:30:07 crc kubenswrapper[4805]: I1203 14:30:07.926055 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/16dbab06-8ff8-4a5a-9fe8-4967c6838a4a-credential-keys\") pod \"keystone-bootstrap-qk4m4\" (UID: \"16dbab06-8ff8-4a5a-9fe8-4967c6838a4a\") " pod="openstack/keystone-bootstrap-qk4m4" Dec 03 14:30:07 crc kubenswrapper[4805]: I1203 14:30:07.926681 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 03 14:30:07 crc kubenswrapper[4805]: I1203 14:30:07.953619 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/16dbab06-8ff8-4a5a-9fe8-4967c6838a4a-fernet-keys\") pod \"keystone-bootstrap-qk4m4\" (UID: \"16dbab06-8ff8-4a5a-9fe8-4967c6838a4a\") " pod="openstack/keystone-bootstrap-qk4m4" Dec 03 14:30:07 crc kubenswrapper[4805]: I1203 14:30:07.953739 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/16dbab06-8ff8-4a5a-9fe8-4967c6838a4a-combined-ca-bundle\") pod \"keystone-bootstrap-qk4m4\" (UID: \"16dbab06-8ff8-4a5a-9fe8-4967c6838a4a\") " pod="openstack/keystone-bootstrap-qk4m4" Dec 03 14:30:07 crc kubenswrapper[4805]: I1203 14:30:07.954330 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/16dbab06-8ff8-4a5a-9fe8-4967c6838a4a-scripts\") pod \"keystone-bootstrap-qk4m4\" (UID: \"16dbab06-8ff8-4a5a-9fe8-4967c6838a4a\") " pod="openstack/keystone-bootstrap-qk4m4" Dec 03 14:30:07 crc kubenswrapper[4805]: I1203 14:30:07.954521 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/16dbab06-8ff8-4a5a-9fe8-4967c6838a4a-config-data\") pod \"keystone-bootstrap-qk4m4\" (UID: \"16dbab06-8ff8-4a5a-9fe8-4967c6838a4a\") " pod="openstack/keystone-bootstrap-qk4m4" Dec 03 14:30:07 crc kubenswrapper[4805]: I1203 14:30:07.970290 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/16dbab06-8ff8-4a5a-9fe8-4967c6838a4a-credential-keys\") pod \"keystone-bootstrap-qk4m4\" (UID: \"16dbab06-8ff8-4a5a-9fe8-4967c6838a4a\") " pod="openstack/keystone-bootstrap-qk4m4" Dec 03 14:30:07 crc kubenswrapper[4805]: I1203 14:30:07.971139 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kbxbj\" (UniqueName: \"kubernetes.io/projected/16dbab06-8ff8-4a5a-9fe8-4967c6838a4a-kube-api-access-kbxbj\") pod \"keystone-bootstrap-qk4m4\" (UID: \"16dbab06-8ff8-4a5a-9fe8-4967c6838a4a\") " pod="openstack/keystone-bootstrap-qk4m4" Dec 03 14:30:07 crc kubenswrapper[4805]: I1203 14:30:07.974387 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5c5cc7c5ff-kxgt6" Dec 03 14:30:08 crc kubenswrapper[4805]: I1203 14:30:08.002111 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-qk4m4" Dec 03 14:30:08 crc kubenswrapper[4805]: I1203 14:30:08.027614 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8533d470-22f2-401a-a39d-c37391920d81-config-data\") pod \"ceilometer-0\" (UID: \"8533d470-22f2-401a-a39d-c37391920d81\") " pod="openstack/ceilometer-0" Dec 03 14:30:08 crc kubenswrapper[4805]: I1203 14:30:08.027666 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kvkpr\" (UniqueName: \"kubernetes.io/projected/8533d470-22f2-401a-a39d-c37391920d81-kube-api-access-kvkpr\") pod \"ceilometer-0\" (UID: \"8533d470-22f2-401a-a39d-c37391920d81\") " pod="openstack/ceilometer-0" Dec 03 14:30:08 crc kubenswrapper[4805]: I1203 14:30:08.027695 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8533d470-22f2-401a-a39d-c37391920d81-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"8533d470-22f2-401a-a39d-c37391920d81\") " pod="openstack/ceilometer-0" Dec 03 14:30:08 crc kubenswrapper[4805]: I1203 14:30:08.027718 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8533d470-22f2-401a-a39d-c37391920d81-scripts\") pod \"ceilometer-0\" (UID: \"8533d470-22f2-401a-a39d-c37391920d81\") " pod="openstack/ceilometer-0" Dec 03 14:30:08 crc kubenswrapper[4805]: I1203 14:30:08.027742 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/8533d470-22f2-401a-a39d-c37391920d81-log-httpd\") pod \"ceilometer-0\" (UID: \"8533d470-22f2-401a-a39d-c37391920d81\") " pod="openstack/ceilometer-0" Dec 03 14:30:08 crc kubenswrapper[4805]: I1203 14:30:08.027785 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/8533d470-22f2-401a-a39d-c37391920d81-run-httpd\") pod \"ceilometer-0\" (UID: \"8533d470-22f2-401a-a39d-c37391920d81\") " pod="openstack/ceilometer-0" Dec 03 14:30:08 crc kubenswrapper[4805]: I1203 14:30:08.027807 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/8533d470-22f2-401a-a39d-c37391920d81-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"8533d470-22f2-401a-a39d-c37391920d81\") " pod="openstack/ceilometer-0" Dec 03 14:30:08 crc kubenswrapper[4805]: I1203 14:30:08.077967 4805 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5c5cc7c5ff-kxgt6"] Dec 03 14:30:08 crc kubenswrapper[4805]: I1203 14:30:08.125152 4805 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-db-sync-74tz4"] Dec 03 14:30:08 crc kubenswrapper[4805]: I1203 14:30:08.126198 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-sync-74tz4" Dec 03 14:30:08 crc kubenswrapper[4805]: I1203 14:30:08.128773 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8533d470-22f2-401a-a39d-c37391920d81-scripts\") pod \"ceilometer-0\" (UID: \"8533d470-22f2-401a-a39d-c37391920d81\") " pod="openstack/ceilometer-0" Dec 03 14:30:08 crc kubenswrapper[4805]: I1203 14:30:08.134370 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/8533d470-22f2-401a-a39d-c37391920d81-log-httpd\") pod \"ceilometer-0\" (UID: \"8533d470-22f2-401a-a39d-c37391920d81\") " pod="openstack/ceilometer-0" Dec 03 14:30:08 crc kubenswrapper[4805]: I1203 14:30:08.134569 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/8533d470-22f2-401a-a39d-c37391920d81-run-httpd\") pod \"ceilometer-0\" (UID: \"8533d470-22f2-401a-a39d-c37391920d81\") " pod="openstack/ceilometer-0" Dec 03 14:30:08 crc kubenswrapper[4805]: I1203 14:30:08.134674 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/8533d470-22f2-401a-a39d-c37391920d81-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"8533d470-22f2-401a-a39d-c37391920d81\") " pod="openstack/ceilometer-0" Dec 03 14:30:08 crc kubenswrapper[4805]: I1203 14:30:08.134910 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8533d470-22f2-401a-a39d-c37391920d81-config-data\") pod \"ceilometer-0\" (UID: \"8533d470-22f2-401a-a39d-c37391920d81\") " pod="openstack/ceilometer-0" Dec 03 14:30:08 crc kubenswrapper[4805]: I1203 14:30:08.135007 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kvkpr\" (UniqueName: \"kubernetes.io/projected/8533d470-22f2-401a-a39d-c37391920d81-kube-api-access-kvkpr\") pod \"ceilometer-0\" (UID: \"8533d470-22f2-401a-a39d-c37391920d81\") " pod="openstack/ceilometer-0" Dec 03 14:30:08 crc kubenswrapper[4805]: I1203 14:30:08.135136 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8533d470-22f2-401a-a39d-c37391920d81-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"8533d470-22f2-401a-a39d-c37391920d81\") " pod="openstack/ceilometer-0" Dec 03 14:30:08 crc kubenswrapper[4805]: I1203 14:30:08.140366 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/8533d470-22f2-401a-a39d-c37391920d81-run-httpd\") pod \"ceilometer-0\" (UID: \"8533d470-22f2-401a-a39d-c37391920d81\") " pod="openstack/ceilometer-0" Dec 03 14:30:08 crc kubenswrapper[4805]: I1203 14:30:08.132397 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-config-data" Dec 03 14:30:08 crc kubenswrapper[4805]: I1203 14:30:08.140920 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/8533d470-22f2-401a-a39d-c37391920d81-log-httpd\") pod \"ceilometer-0\" (UID: \"8533d470-22f2-401a-a39d-c37391920d81\") " pod="openstack/ceilometer-0" Dec 03 14:30:08 crc kubenswrapper[4805]: I1203 14:30:08.132431 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-barbican-dockercfg-kkm6j" Dec 03 14:30:08 crc kubenswrapper[4805]: I1203 14:30:08.141750 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8533d470-22f2-401a-a39d-c37391920d81-scripts\") pod \"ceilometer-0\" (UID: \"8533d470-22f2-401a-a39d-c37391920d81\") " pod="openstack/ceilometer-0" Dec 03 14:30:08 crc kubenswrapper[4805]: I1203 14:30:08.148151 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8533d470-22f2-401a-a39d-c37391920d81-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"8533d470-22f2-401a-a39d-c37391920d81\") " pod="openstack/ceilometer-0" Dec 03 14:30:08 crc kubenswrapper[4805]: I1203 14:30:08.148576 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/8533d470-22f2-401a-a39d-c37391920d81-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"8533d470-22f2-401a-a39d-c37391920d81\") " pod="openstack/ceilometer-0" Dec 03 14:30:08 crc kubenswrapper[4805]: I1203 14:30:08.151287 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8533d470-22f2-401a-a39d-c37391920d81-config-data\") pod \"ceilometer-0\" (UID: \"8533d470-22f2-401a-a39d-c37391920d81\") " pod="openstack/ceilometer-0" Dec 03 14:30:08 crc kubenswrapper[4805]: I1203 14:30:08.176916 4805 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-db-sync-fh9c2"] Dec 03 14:30:08 crc kubenswrapper[4805]: I1203 14:30:08.186789 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-sync-fh9c2" Dec 03 14:30:08 crc kubenswrapper[4805]: I1203 14:30:08.195187 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-cinder-dockercfg-r5964" Dec 03 14:30:08 crc kubenswrapper[4805]: I1203 14:30:08.202124 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-config-data" Dec 03 14:30:08 crc kubenswrapper[4805]: I1203 14:30:08.205831 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-scripts" Dec 03 14:30:08 crc kubenswrapper[4805]: I1203 14:30:08.212296 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kvkpr\" (UniqueName: \"kubernetes.io/projected/8533d470-22f2-401a-a39d-c37391920d81-kube-api-access-kvkpr\") pod \"ceilometer-0\" (UID: \"8533d470-22f2-401a-a39d-c37391920d81\") " pod="openstack/ceilometer-0" Dec 03 14:30:08 crc kubenswrapper[4805]: I1203 14:30:08.225129 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 03 14:30:08 crc kubenswrapper[4805]: I1203 14:30:08.253796 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-c5jlg\" (UniqueName: \"kubernetes.io/projected/13ba03f0-acc2-4fb8-90a9-583b4f8694cb-kube-api-access-c5jlg\") pod \"barbican-db-sync-74tz4\" (UID: \"13ba03f0-acc2-4fb8-90a9-583b4f8694cb\") " pod="openstack/barbican-db-sync-74tz4" Dec 03 14:30:08 crc kubenswrapper[4805]: I1203 14:30:08.253879 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/13ba03f0-acc2-4fb8-90a9-583b4f8694cb-db-sync-config-data\") pod \"barbican-db-sync-74tz4\" (UID: \"13ba03f0-acc2-4fb8-90a9-583b4f8694cb\") " pod="openstack/barbican-db-sync-74tz4" Dec 03 14:30:08 crc kubenswrapper[4805]: I1203 14:30:08.253914 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/13ba03f0-acc2-4fb8-90a9-583b4f8694cb-combined-ca-bundle\") pod \"barbican-db-sync-74tz4\" (UID: \"13ba03f0-acc2-4fb8-90a9-583b4f8694cb\") " pod="openstack/barbican-db-sync-74tz4" Dec 03 14:30:08 crc kubenswrapper[4805]: I1203 14:30:08.257518 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-db-sync-74tz4"] Dec 03 14:30:08 crc kubenswrapper[4805]: I1203 14:30:08.269235 4805 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement-db-sync-sj4wt"] Dec 03 14:30:08 crc kubenswrapper[4805]: I1203 14:30:08.272643 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-sync-sj4wt" Dec 03 14:30:08 crc kubenswrapper[4805]: I1203 14:30:08.279041 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-scripts" Dec 03 14:30:08 crc kubenswrapper[4805]: I1203 14:30:08.279420 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-placement-dockercfg-nmchz" Dec 03 14:30:08 crc kubenswrapper[4805]: I1203 14:30:08.279530 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-config-data" Dec 03 14:30:08 crc kubenswrapper[4805]: I1203 14:30:08.297855 4805 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-8b5c85b87-bvb8w"] Dec 03 14:30:08 crc kubenswrapper[4805]: I1203 14:30:08.300466 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-8b5c85b87-bvb8w" Dec 03 14:30:08 crc kubenswrapper[4805]: I1203 14:30:08.317947 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-8b5c85b87-bvb8w"] Dec 03 14:30:08 crc kubenswrapper[4805]: I1203 14:30:08.337056 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-db-sync-fh9c2"] Dec 03 14:30:08 crc kubenswrapper[4805]: I1203 14:30:08.340106 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-db-sync-sj4wt"] Dec 03 14:30:08 crc kubenswrapper[4805]: I1203 14:30:08.360221 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/a8299a7f-082a-4a51-becd-7c393a6f4b51-ovsdbserver-sb\") pod \"dnsmasq-dns-8b5c85b87-bvb8w\" (UID: \"a8299a7f-082a-4a51-becd-7c393a6f4b51\") " pod="openstack/dnsmasq-dns-8b5c85b87-bvb8w" Dec 03 14:30:08 crc kubenswrapper[4805]: I1203 14:30:08.360264 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/260717ec-f898-4ccd-8a4f-c7d6b14ade6e-combined-ca-bundle\") pod \"cinder-db-sync-fh9c2\" (UID: \"260717ec-f898-4ccd-8a4f-c7d6b14ade6e\") " pod="openstack/cinder-db-sync-fh9c2" Dec 03 14:30:08 crc kubenswrapper[4805]: I1203 14:30:08.360483 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7c1b5525-de7b-4a8c-9062-bb9a68336989-combined-ca-bundle\") pod \"placement-db-sync-sj4wt\" (UID: \"7c1b5525-de7b-4a8c-9062-bb9a68336989\") " pod="openstack/placement-db-sync-sj4wt" Dec 03 14:30:08 crc kubenswrapper[4805]: I1203 14:30:08.360508 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-svwdk\" (UniqueName: \"kubernetes.io/projected/7c1b5525-de7b-4a8c-9062-bb9a68336989-kube-api-access-svwdk\") pod \"placement-db-sync-sj4wt\" (UID: \"7c1b5525-de7b-4a8c-9062-bb9a68336989\") " pod="openstack/placement-db-sync-sj4wt" Dec 03 14:30:08 crc kubenswrapper[4805]: I1203 14:30:08.360565 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/260717ec-f898-4ccd-8a4f-c7d6b14ade6e-db-sync-config-data\") pod \"cinder-db-sync-fh9c2\" (UID: \"260717ec-f898-4ccd-8a4f-c7d6b14ade6e\") " pod="openstack/cinder-db-sync-fh9c2" Dec 03 14:30:08 crc kubenswrapper[4805]: I1203 14:30:08.360592 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7c1b5525-de7b-4a8c-9062-bb9a68336989-config-data\") pod \"placement-db-sync-sj4wt\" (UID: \"7c1b5525-de7b-4a8c-9062-bb9a68336989\") " pod="openstack/placement-db-sync-sj4wt" Dec 03 14:30:08 crc kubenswrapper[4805]: I1203 14:30:08.360647 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7c1b5525-de7b-4a8c-9062-bb9a68336989-scripts\") pod \"placement-db-sync-sj4wt\" (UID: \"7c1b5525-de7b-4a8c-9062-bb9a68336989\") " pod="openstack/placement-db-sync-sj4wt" Dec 03 14:30:08 crc kubenswrapper[4805]: I1203 14:30:08.360684 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xxm8t\" (UniqueName: \"kubernetes.io/projected/260717ec-f898-4ccd-8a4f-c7d6b14ade6e-kube-api-access-xxm8t\") pod \"cinder-db-sync-fh9c2\" (UID: \"260717ec-f898-4ccd-8a4f-c7d6b14ade6e\") " pod="openstack/cinder-db-sync-fh9c2" Dec 03 14:30:08 crc kubenswrapper[4805]: I1203 14:30:08.360723 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/260717ec-f898-4ccd-8a4f-c7d6b14ade6e-config-data\") pod \"cinder-db-sync-fh9c2\" (UID: \"260717ec-f898-4ccd-8a4f-c7d6b14ade6e\") " pod="openstack/cinder-db-sync-fh9c2" Dec 03 14:30:08 crc kubenswrapper[4805]: I1203 14:30:08.360745 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/260717ec-f898-4ccd-8a4f-c7d6b14ade6e-etc-machine-id\") pod \"cinder-db-sync-fh9c2\" (UID: \"260717ec-f898-4ccd-8a4f-c7d6b14ade6e\") " pod="openstack/cinder-db-sync-fh9c2" Dec 03 14:30:08 crc kubenswrapper[4805]: I1203 14:30:08.360761 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/a8299a7f-082a-4a51-becd-7c393a6f4b51-dns-swift-storage-0\") pod \"dnsmasq-dns-8b5c85b87-bvb8w\" (UID: \"a8299a7f-082a-4a51-becd-7c393a6f4b51\") " pod="openstack/dnsmasq-dns-8b5c85b87-bvb8w" Dec 03 14:30:08 crc kubenswrapper[4805]: I1203 14:30:08.360789 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-c5jlg\" (UniqueName: \"kubernetes.io/projected/13ba03f0-acc2-4fb8-90a9-583b4f8694cb-kube-api-access-c5jlg\") pod \"barbican-db-sync-74tz4\" (UID: \"13ba03f0-acc2-4fb8-90a9-583b4f8694cb\") " pod="openstack/barbican-db-sync-74tz4" Dec 03 14:30:08 crc kubenswrapper[4805]: I1203 14:30:08.360805 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/a8299a7f-082a-4a51-becd-7c393a6f4b51-ovsdbserver-nb\") pod \"dnsmasq-dns-8b5c85b87-bvb8w\" (UID: \"a8299a7f-082a-4a51-becd-7c393a6f4b51\") " pod="openstack/dnsmasq-dns-8b5c85b87-bvb8w" Dec 03 14:30:08 crc kubenswrapper[4805]: I1203 14:30:08.360829 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/13ba03f0-acc2-4fb8-90a9-583b4f8694cb-db-sync-config-data\") pod \"barbican-db-sync-74tz4\" (UID: \"13ba03f0-acc2-4fb8-90a9-583b4f8694cb\") " pod="openstack/barbican-db-sync-74tz4" Dec 03 14:30:08 crc kubenswrapper[4805]: I1203 14:30:08.360867 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/7c1b5525-de7b-4a8c-9062-bb9a68336989-logs\") pod \"placement-db-sync-sj4wt\" (UID: \"7c1b5525-de7b-4a8c-9062-bb9a68336989\") " pod="openstack/placement-db-sync-sj4wt" Dec 03 14:30:08 crc kubenswrapper[4805]: I1203 14:30:08.360894 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/13ba03f0-acc2-4fb8-90a9-583b4f8694cb-combined-ca-bundle\") pod \"barbican-db-sync-74tz4\" (UID: \"13ba03f0-acc2-4fb8-90a9-583b4f8694cb\") " pod="openstack/barbican-db-sync-74tz4" Dec 03 14:30:08 crc kubenswrapper[4805]: I1203 14:30:08.361510 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/260717ec-f898-4ccd-8a4f-c7d6b14ade6e-scripts\") pod \"cinder-db-sync-fh9c2\" (UID: \"260717ec-f898-4ccd-8a4f-c7d6b14ade6e\") " pod="openstack/cinder-db-sync-fh9c2" Dec 03 14:30:08 crc kubenswrapper[4805]: I1203 14:30:08.361534 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a8299a7f-082a-4a51-becd-7c393a6f4b51-config\") pod \"dnsmasq-dns-8b5c85b87-bvb8w\" (UID: \"a8299a7f-082a-4a51-becd-7c393a6f4b51\") " pod="openstack/dnsmasq-dns-8b5c85b87-bvb8w" Dec 03 14:30:08 crc kubenswrapper[4805]: I1203 14:30:08.361565 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wkmfm\" (UniqueName: \"kubernetes.io/projected/a8299a7f-082a-4a51-becd-7c393a6f4b51-kube-api-access-wkmfm\") pod \"dnsmasq-dns-8b5c85b87-bvb8w\" (UID: \"a8299a7f-082a-4a51-becd-7c393a6f4b51\") " pod="openstack/dnsmasq-dns-8b5c85b87-bvb8w" Dec 03 14:30:08 crc kubenswrapper[4805]: I1203 14:30:08.361581 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/a8299a7f-082a-4a51-becd-7c393a6f4b51-dns-svc\") pod \"dnsmasq-dns-8b5c85b87-bvb8w\" (UID: \"a8299a7f-082a-4a51-becd-7c393a6f4b51\") " pod="openstack/dnsmasq-dns-8b5c85b87-bvb8w" Dec 03 14:30:08 crc kubenswrapper[4805]: I1203 14:30:08.368948 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/13ba03f0-acc2-4fb8-90a9-583b4f8694cb-db-sync-config-data\") pod \"barbican-db-sync-74tz4\" (UID: \"13ba03f0-acc2-4fb8-90a9-583b4f8694cb\") " pod="openstack/barbican-db-sync-74tz4" Dec 03 14:30:08 crc kubenswrapper[4805]: I1203 14:30:08.380695 4805 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/heat-db-sync-jvqmk"] Dec 03 14:30:08 crc kubenswrapper[4805]: I1203 14:30:08.381371 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/13ba03f0-acc2-4fb8-90a9-583b4f8694cb-combined-ca-bundle\") pod \"barbican-db-sync-74tz4\" (UID: \"13ba03f0-acc2-4fb8-90a9-583b4f8694cb\") " pod="openstack/barbican-db-sync-74tz4" Dec 03 14:30:08 crc kubenswrapper[4805]: I1203 14:30:08.384145 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-db-sync-jvqmk" Dec 03 14:30:08 crc kubenswrapper[4805]: I1203 14:30:08.383684 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-c5jlg\" (UniqueName: \"kubernetes.io/projected/13ba03f0-acc2-4fb8-90a9-583b4f8694cb-kube-api-access-c5jlg\") pod \"barbican-db-sync-74tz4\" (UID: \"13ba03f0-acc2-4fb8-90a9-583b4f8694cb\") " pod="openstack/barbican-db-sync-74tz4" Dec 03 14:30:08 crc kubenswrapper[4805]: I1203 14:30:08.386446 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"heat-config-data" Dec 03 14:30:08 crc kubenswrapper[4805]: I1203 14:30:08.386548 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"heat-heat-dockercfg-2wxds" Dec 03 14:30:08 crc kubenswrapper[4805]: I1203 14:30:08.401072 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-db-sync-jvqmk"] Dec 03 14:30:08 crc kubenswrapper[4805]: I1203 14:30:08.437543 4805 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-db-sync-b6k7w"] Dec 03 14:30:08 crc kubenswrapper[4805]: I1203 14:30:08.439202 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-sync-b6k7w" Dec 03 14:30:08 crc kubenswrapper[4805]: I1203 14:30:08.443360 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-config" Dec 03 14:30:08 crc kubenswrapper[4805]: I1203 14:30:08.443527 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-neutron-dockercfg-68v4l" Dec 03 14:30:08 crc kubenswrapper[4805]: I1203 14:30:08.447104 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-httpd-config" Dec 03 14:30:08 crc kubenswrapper[4805]: I1203 14:30:08.462699 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/a8299a7f-082a-4a51-becd-7c393a6f4b51-ovsdbserver-sb\") pod \"dnsmasq-dns-8b5c85b87-bvb8w\" (UID: \"a8299a7f-082a-4a51-becd-7c393a6f4b51\") " pod="openstack/dnsmasq-dns-8b5c85b87-bvb8w" Dec 03 14:30:08 crc kubenswrapper[4805]: I1203 14:30:08.462744 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/260717ec-f898-4ccd-8a4f-c7d6b14ade6e-combined-ca-bundle\") pod \"cinder-db-sync-fh9c2\" (UID: \"260717ec-f898-4ccd-8a4f-c7d6b14ade6e\") " pod="openstack/cinder-db-sync-fh9c2" Dec 03 14:30:08 crc kubenswrapper[4805]: I1203 14:30:08.462775 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zh8k9\" (UniqueName: \"kubernetes.io/projected/14c327e3-e880-4b49-b366-364a0395cc55-kube-api-access-zh8k9\") pod \"heat-db-sync-jvqmk\" (UID: \"14c327e3-e880-4b49-b366-364a0395cc55\") " pod="openstack/heat-db-sync-jvqmk" Dec 03 14:30:08 crc kubenswrapper[4805]: I1203 14:30:08.462796 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7c1b5525-de7b-4a8c-9062-bb9a68336989-combined-ca-bundle\") pod \"placement-db-sync-sj4wt\" (UID: \"7c1b5525-de7b-4a8c-9062-bb9a68336989\") " pod="openstack/placement-db-sync-sj4wt" Dec 03 14:30:08 crc kubenswrapper[4805]: I1203 14:30:08.462812 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-svwdk\" (UniqueName: \"kubernetes.io/projected/7c1b5525-de7b-4a8c-9062-bb9a68336989-kube-api-access-svwdk\") pod \"placement-db-sync-sj4wt\" (UID: \"7c1b5525-de7b-4a8c-9062-bb9a68336989\") " pod="openstack/placement-db-sync-sj4wt" Dec 03 14:30:08 crc kubenswrapper[4805]: I1203 14:30:08.462849 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/260717ec-f898-4ccd-8a4f-c7d6b14ade6e-db-sync-config-data\") pod \"cinder-db-sync-fh9c2\" (UID: \"260717ec-f898-4ccd-8a4f-c7d6b14ade6e\") " pod="openstack/cinder-db-sync-fh9c2" Dec 03 14:30:08 crc kubenswrapper[4805]: I1203 14:30:08.462871 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7c1b5525-de7b-4a8c-9062-bb9a68336989-config-data\") pod \"placement-db-sync-sj4wt\" (UID: \"7c1b5525-de7b-4a8c-9062-bb9a68336989\") " pod="openstack/placement-db-sync-sj4wt" Dec 03 14:30:08 crc kubenswrapper[4805]: I1203 14:30:08.462899 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7c1b5525-de7b-4a8c-9062-bb9a68336989-scripts\") pod \"placement-db-sync-sj4wt\" (UID: \"7c1b5525-de7b-4a8c-9062-bb9a68336989\") " pod="openstack/placement-db-sync-sj4wt" Dec 03 14:30:08 crc kubenswrapper[4805]: I1203 14:30:08.462921 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xxm8t\" (UniqueName: \"kubernetes.io/projected/260717ec-f898-4ccd-8a4f-c7d6b14ade6e-kube-api-access-xxm8t\") pod \"cinder-db-sync-fh9c2\" (UID: \"260717ec-f898-4ccd-8a4f-c7d6b14ade6e\") " pod="openstack/cinder-db-sync-fh9c2" Dec 03 14:30:08 crc kubenswrapper[4805]: I1203 14:30:08.462938 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/14c327e3-e880-4b49-b366-364a0395cc55-combined-ca-bundle\") pod \"heat-db-sync-jvqmk\" (UID: \"14c327e3-e880-4b49-b366-364a0395cc55\") " pod="openstack/heat-db-sync-jvqmk" Dec 03 14:30:08 crc kubenswrapper[4805]: I1203 14:30:08.462961 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/260717ec-f898-4ccd-8a4f-c7d6b14ade6e-config-data\") pod \"cinder-db-sync-fh9c2\" (UID: \"260717ec-f898-4ccd-8a4f-c7d6b14ade6e\") " pod="openstack/cinder-db-sync-fh9c2" Dec 03 14:30:08 crc kubenswrapper[4805]: I1203 14:30:08.462980 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/260717ec-f898-4ccd-8a4f-c7d6b14ade6e-etc-machine-id\") pod \"cinder-db-sync-fh9c2\" (UID: \"260717ec-f898-4ccd-8a4f-c7d6b14ade6e\") " pod="openstack/cinder-db-sync-fh9c2" Dec 03 14:30:08 crc kubenswrapper[4805]: I1203 14:30:08.462996 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/a8299a7f-082a-4a51-becd-7c393a6f4b51-dns-swift-storage-0\") pod \"dnsmasq-dns-8b5c85b87-bvb8w\" (UID: \"a8299a7f-082a-4a51-becd-7c393a6f4b51\") " pod="openstack/dnsmasq-dns-8b5c85b87-bvb8w" Dec 03 14:30:08 crc kubenswrapper[4805]: I1203 14:30:08.463016 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/a8299a7f-082a-4a51-becd-7c393a6f4b51-ovsdbserver-nb\") pod \"dnsmasq-dns-8b5c85b87-bvb8w\" (UID: \"a8299a7f-082a-4a51-becd-7c393a6f4b51\") " pod="openstack/dnsmasq-dns-8b5c85b87-bvb8w" Dec 03 14:30:08 crc kubenswrapper[4805]: I1203 14:30:08.463045 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/7c1b5525-de7b-4a8c-9062-bb9a68336989-logs\") pod \"placement-db-sync-sj4wt\" (UID: \"7c1b5525-de7b-4a8c-9062-bb9a68336989\") " pod="openstack/placement-db-sync-sj4wt" Dec 03 14:30:08 crc kubenswrapper[4805]: I1203 14:30:08.463063 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/14c327e3-e880-4b49-b366-364a0395cc55-config-data\") pod \"heat-db-sync-jvqmk\" (UID: \"14c327e3-e880-4b49-b366-364a0395cc55\") " pod="openstack/heat-db-sync-jvqmk" Dec 03 14:30:08 crc kubenswrapper[4805]: I1203 14:30:08.463442 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/260717ec-f898-4ccd-8a4f-c7d6b14ade6e-scripts\") pod \"cinder-db-sync-fh9c2\" (UID: \"260717ec-f898-4ccd-8a4f-c7d6b14ade6e\") " pod="openstack/cinder-db-sync-fh9c2" Dec 03 14:30:08 crc kubenswrapper[4805]: I1203 14:30:08.463462 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a8299a7f-082a-4a51-becd-7c393a6f4b51-config\") pod \"dnsmasq-dns-8b5c85b87-bvb8w\" (UID: \"a8299a7f-082a-4a51-becd-7c393a6f4b51\") " pod="openstack/dnsmasq-dns-8b5c85b87-bvb8w" Dec 03 14:30:08 crc kubenswrapper[4805]: I1203 14:30:08.463483 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wkmfm\" (UniqueName: \"kubernetes.io/projected/a8299a7f-082a-4a51-becd-7c393a6f4b51-kube-api-access-wkmfm\") pod \"dnsmasq-dns-8b5c85b87-bvb8w\" (UID: \"a8299a7f-082a-4a51-becd-7c393a6f4b51\") " pod="openstack/dnsmasq-dns-8b5c85b87-bvb8w" Dec 03 14:30:08 crc kubenswrapper[4805]: I1203 14:30:08.463503 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/a8299a7f-082a-4a51-becd-7c393a6f4b51-dns-svc\") pod \"dnsmasq-dns-8b5c85b87-bvb8w\" (UID: \"a8299a7f-082a-4a51-becd-7c393a6f4b51\") " pod="openstack/dnsmasq-dns-8b5c85b87-bvb8w" Dec 03 14:30:08 crc kubenswrapper[4805]: I1203 14:30:08.464240 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/a8299a7f-082a-4a51-becd-7c393a6f4b51-dns-svc\") pod \"dnsmasq-dns-8b5c85b87-bvb8w\" (UID: \"a8299a7f-082a-4a51-becd-7c393a6f4b51\") " pod="openstack/dnsmasq-dns-8b5c85b87-bvb8w" Dec 03 14:30:08 crc kubenswrapper[4805]: I1203 14:30:08.464880 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/260717ec-f898-4ccd-8a4f-c7d6b14ade6e-etc-machine-id\") pod \"cinder-db-sync-fh9c2\" (UID: \"260717ec-f898-4ccd-8a4f-c7d6b14ade6e\") " pod="openstack/cinder-db-sync-fh9c2" Dec 03 14:30:08 crc kubenswrapper[4805]: I1203 14:30:08.466554 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a8299a7f-082a-4a51-becd-7c393a6f4b51-config\") pod \"dnsmasq-dns-8b5c85b87-bvb8w\" (UID: \"a8299a7f-082a-4a51-becd-7c393a6f4b51\") " pod="openstack/dnsmasq-dns-8b5c85b87-bvb8w" Dec 03 14:30:08 crc kubenswrapper[4805]: I1203 14:30:08.466970 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/a8299a7f-082a-4a51-becd-7c393a6f4b51-ovsdbserver-sb\") pod \"dnsmasq-dns-8b5c85b87-bvb8w\" (UID: \"a8299a7f-082a-4a51-becd-7c393a6f4b51\") " pod="openstack/dnsmasq-dns-8b5c85b87-bvb8w" Dec 03 14:30:08 crc kubenswrapper[4805]: I1203 14:30:08.467217 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/260717ec-f898-4ccd-8a4f-c7d6b14ade6e-db-sync-config-data\") pod \"cinder-db-sync-fh9c2\" (UID: \"260717ec-f898-4ccd-8a4f-c7d6b14ade6e\") " pod="openstack/cinder-db-sync-fh9c2" Dec 03 14:30:08 crc kubenswrapper[4805]: I1203 14:30:08.467637 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/7c1b5525-de7b-4a8c-9062-bb9a68336989-logs\") pod \"placement-db-sync-sj4wt\" (UID: \"7c1b5525-de7b-4a8c-9062-bb9a68336989\") " pod="openstack/placement-db-sync-sj4wt" Dec 03 14:30:08 crc kubenswrapper[4805]: I1203 14:30:08.468678 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/260717ec-f898-4ccd-8a4f-c7d6b14ade6e-combined-ca-bundle\") pod \"cinder-db-sync-fh9c2\" (UID: \"260717ec-f898-4ccd-8a4f-c7d6b14ade6e\") " pod="openstack/cinder-db-sync-fh9c2" Dec 03 14:30:08 crc kubenswrapper[4805]: I1203 14:30:08.470363 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/a8299a7f-082a-4a51-becd-7c393a6f4b51-ovsdbserver-nb\") pod \"dnsmasq-dns-8b5c85b87-bvb8w\" (UID: \"a8299a7f-082a-4a51-becd-7c393a6f4b51\") " pod="openstack/dnsmasq-dns-8b5c85b87-bvb8w" Dec 03 14:30:08 crc kubenswrapper[4805]: I1203 14:30:08.471817 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/a8299a7f-082a-4a51-becd-7c393a6f4b51-dns-swift-storage-0\") pod \"dnsmasq-dns-8b5c85b87-bvb8w\" (UID: \"a8299a7f-082a-4a51-becd-7c393a6f4b51\") " pod="openstack/dnsmasq-dns-8b5c85b87-bvb8w" Dec 03 14:30:08 crc kubenswrapper[4805]: I1203 14:30:08.472706 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/260717ec-f898-4ccd-8a4f-c7d6b14ade6e-config-data\") pod \"cinder-db-sync-fh9c2\" (UID: \"260717ec-f898-4ccd-8a4f-c7d6b14ade6e\") " pod="openstack/cinder-db-sync-fh9c2" Dec 03 14:30:08 crc kubenswrapper[4805]: I1203 14:30:08.474732 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7c1b5525-de7b-4a8c-9062-bb9a68336989-config-data\") pod \"placement-db-sync-sj4wt\" (UID: \"7c1b5525-de7b-4a8c-9062-bb9a68336989\") " pod="openstack/placement-db-sync-sj4wt" Dec 03 14:30:08 crc kubenswrapper[4805]: I1203 14:30:08.476287 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7c1b5525-de7b-4a8c-9062-bb9a68336989-scripts\") pod \"placement-db-sync-sj4wt\" (UID: \"7c1b5525-de7b-4a8c-9062-bb9a68336989\") " pod="openstack/placement-db-sync-sj4wt" Dec 03 14:30:08 crc kubenswrapper[4805]: I1203 14:30:08.476663 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7c1b5525-de7b-4a8c-9062-bb9a68336989-combined-ca-bundle\") pod \"placement-db-sync-sj4wt\" (UID: \"7c1b5525-de7b-4a8c-9062-bb9a68336989\") " pod="openstack/placement-db-sync-sj4wt" Dec 03 14:30:08 crc kubenswrapper[4805]: I1203 14:30:08.483815 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wkmfm\" (UniqueName: \"kubernetes.io/projected/a8299a7f-082a-4a51-becd-7c393a6f4b51-kube-api-access-wkmfm\") pod \"dnsmasq-dns-8b5c85b87-bvb8w\" (UID: \"a8299a7f-082a-4a51-becd-7c393a6f4b51\") " pod="openstack/dnsmasq-dns-8b5c85b87-bvb8w" Dec 03 14:30:08 crc kubenswrapper[4805]: I1203 14:30:08.485956 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-svwdk\" (UniqueName: \"kubernetes.io/projected/7c1b5525-de7b-4a8c-9062-bb9a68336989-kube-api-access-svwdk\") pod \"placement-db-sync-sj4wt\" (UID: \"7c1b5525-de7b-4a8c-9062-bb9a68336989\") " pod="openstack/placement-db-sync-sj4wt" Dec 03 14:30:08 crc kubenswrapper[4805]: I1203 14:30:08.486207 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/260717ec-f898-4ccd-8a4f-c7d6b14ade6e-scripts\") pod \"cinder-db-sync-fh9c2\" (UID: \"260717ec-f898-4ccd-8a4f-c7d6b14ade6e\") " pod="openstack/cinder-db-sync-fh9c2" Dec 03 14:30:08 crc kubenswrapper[4805]: I1203 14:30:08.487342 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xxm8t\" (UniqueName: \"kubernetes.io/projected/260717ec-f898-4ccd-8a4f-c7d6b14ade6e-kube-api-access-xxm8t\") pod \"cinder-db-sync-fh9c2\" (UID: \"260717ec-f898-4ccd-8a4f-c7d6b14ade6e\") " pod="openstack/cinder-db-sync-fh9c2" Dec 03 14:30:08 crc kubenswrapper[4805]: I1203 14:30:08.491201 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-db-sync-b6k7w"] Dec 03 14:30:08 crc kubenswrapper[4805]: I1203 14:30:08.491738 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-sync-74tz4" Dec 03 14:30:08 crc kubenswrapper[4805]: I1203 14:30:08.566819 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/14c327e3-e880-4b49-b366-364a0395cc55-combined-ca-bundle\") pod \"heat-db-sync-jvqmk\" (UID: \"14c327e3-e880-4b49-b366-364a0395cc55\") " pod="openstack/heat-db-sync-jvqmk" Dec 03 14:30:08 crc kubenswrapper[4805]: I1203 14:30:08.566901 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/32920f09-a728-4b35-a5d7-65dae2d70834-config\") pod \"neutron-db-sync-b6k7w\" (UID: \"32920f09-a728-4b35-a5d7-65dae2d70834\") " pod="openstack/neutron-db-sync-b6k7w" Dec 03 14:30:08 crc kubenswrapper[4805]: I1203 14:30:08.566952 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/14c327e3-e880-4b49-b366-364a0395cc55-config-data\") pod \"heat-db-sync-jvqmk\" (UID: \"14c327e3-e880-4b49-b366-364a0395cc55\") " pod="openstack/heat-db-sync-jvqmk" Dec 03 14:30:08 crc kubenswrapper[4805]: I1203 14:30:08.567020 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/32920f09-a728-4b35-a5d7-65dae2d70834-combined-ca-bundle\") pod \"neutron-db-sync-b6k7w\" (UID: \"32920f09-a728-4b35-a5d7-65dae2d70834\") " pod="openstack/neutron-db-sync-b6k7w" Dec 03 14:30:08 crc kubenswrapper[4805]: I1203 14:30:08.567053 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zh8k9\" (UniqueName: \"kubernetes.io/projected/14c327e3-e880-4b49-b366-364a0395cc55-kube-api-access-zh8k9\") pod \"heat-db-sync-jvqmk\" (UID: \"14c327e3-e880-4b49-b366-364a0395cc55\") " pod="openstack/heat-db-sync-jvqmk" Dec 03 14:30:08 crc kubenswrapper[4805]: I1203 14:30:08.567090 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-v5gd6\" (UniqueName: \"kubernetes.io/projected/32920f09-a728-4b35-a5d7-65dae2d70834-kube-api-access-v5gd6\") pod \"neutron-db-sync-b6k7w\" (UID: \"32920f09-a728-4b35-a5d7-65dae2d70834\") " pod="openstack/neutron-db-sync-b6k7w" Dec 03 14:30:08 crc kubenswrapper[4805]: I1203 14:30:08.576478 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/14c327e3-e880-4b49-b366-364a0395cc55-config-data\") pod \"heat-db-sync-jvqmk\" (UID: \"14c327e3-e880-4b49-b366-364a0395cc55\") " pod="openstack/heat-db-sync-jvqmk" Dec 03 14:30:08 crc kubenswrapper[4805]: I1203 14:30:08.577077 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/14c327e3-e880-4b49-b366-364a0395cc55-combined-ca-bundle\") pod \"heat-db-sync-jvqmk\" (UID: \"14c327e3-e880-4b49-b366-364a0395cc55\") " pod="openstack/heat-db-sync-jvqmk" Dec 03 14:30:08 crc kubenswrapper[4805]: I1203 14:30:08.581248 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-sync-fh9c2" Dec 03 14:30:08 crc kubenswrapper[4805]: I1203 14:30:08.588726 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zh8k9\" (UniqueName: \"kubernetes.io/projected/14c327e3-e880-4b49-b366-364a0395cc55-kube-api-access-zh8k9\") pod \"heat-db-sync-jvqmk\" (UID: \"14c327e3-e880-4b49-b366-364a0395cc55\") " pod="openstack/heat-db-sync-jvqmk" Dec 03 14:30:08 crc kubenswrapper[4805]: I1203 14:30:08.640055 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-sync-sj4wt" Dec 03 14:30:08 crc kubenswrapper[4805]: I1203 14:30:08.669259 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-v5gd6\" (UniqueName: \"kubernetes.io/projected/32920f09-a728-4b35-a5d7-65dae2d70834-kube-api-access-v5gd6\") pod \"neutron-db-sync-b6k7w\" (UID: \"32920f09-a728-4b35-a5d7-65dae2d70834\") " pod="openstack/neutron-db-sync-b6k7w" Dec 03 14:30:08 crc kubenswrapper[4805]: I1203 14:30:08.669341 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/32920f09-a728-4b35-a5d7-65dae2d70834-config\") pod \"neutron-db-sync-b6k7w\" (UID: \"32920f09-a728-4b35-a5d7-65dae2d70834\") " pod="openstack/neutron-db-sync-b6k7w" Dec 03 14:30:08 crc kubenswrapper[4805]: I1203 14:30:08.669547 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/32920f09-a728-4b35-a5d7-65dae2d70834-combined-ca-bundle\") pod \"neutron-db-sync-b6k7w\" (UID: \"32920f09-a728-4b35-a5d7-65dae2d70834\") " pod="openstack/neutron-db-sync-b6k7w" Dec 03 14:30:08 crc kubenswrapper[4805]: I1203 14:30:08.672888 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-8b5c85b87-bvb8w" Dec 03 14:30:08 crc kubenswrapper[4805]: I1203 14:30:08.679957 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/32920f09-a728-4b35-a5d7-65dae2d70834-combined-ca-bundle\") pod \"neutron-db-sync-b6k7w\" (UID: \"32920f09-a728-4b35-a5d7-65dae2d70834\") " pod="openstack/neutron-db-sync-b6k7w" Dec 03 14:30:08 crc kubenswrapper[4805]: I1203 14:30:08.692425 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/secret/32920f09-a728-4b35-a5d7-65dae2d70834-config\") pod \"neutron-db-sync-b6k7w\" (UID: \"32920f09-a728-4b35-a5d7-65dae2d70834\") " pod="openstack/neutron-db-sync-b6k7w" Dec 03 14:30:08 crc kubenswrapper[4805]: I1203 14:30:08.699905 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-v5gd6\" (UniqueName: \"kubernetes.io/projected/32920f09-a728-4b35-a5d7-65dae2d70834-kube-api-access-v5gd6\") pod \"neutron-db-sync-b6k7w\" (UID: \"32920f09-a728-4b35-a5d7-65dae2d70834\") " pod="openstack/neutron-db-sync-b6k7w" Dec 03 14:30:08 crc kubenswrapper[4805]: I1203 14:30:08.719581 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-bootstrap-qk4m4"] Dec 03 14:30:08 crc kubenswrapper[4805]: I1203 14:30:08.723490 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-db-sync-jvqmk" Dec 03 14:30:08 crc kubenswrapper[4805]: I1203 14:30:08.751131 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-db-sync-74tz4"] Dec 03 14:30:08 crc kubenswrapper[4805]: I1203 14:30:08.773113 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-sync-b6k7w" Dec 03 14:30:08 crc kubenswrapper[4805]: I1203 14:30:08.795890 4805 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-external-api-0"] Dec 03 14:30:08 crc kubenswrapper[4805]: I1203 14:30:08.797824 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Dec 03 14:30:08 crc kubenswrapper[4805]: I1203 14:30:08.800308 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-scripts" Dec 03 14:30:08 crc kubenswrapper[4805]: I1203 14:30:08.800783 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-external-config-data" Dec 03 14:30:08 crc kubenswrapper[4805]: I1203 14:30:08.803588 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-public-svc" Dec 03 14:30:08 crc kubenswrapper[4805]: I1203 14:30:08.804047 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-glance-dockercfg-rxb4p" Dec 03 14:30:08 crc kubenswrapper[4805]: I1203 14:30:08.809310 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Dec 03 14:30:08 crc kubenswrapper[4805]: I1203 14:30:08.818511 4805 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5c5cc7c5ff-kxgt6"] Dec 03 14:30:08 crc kubenswrapper[4805]: I1203 14:30:08.869163 4805 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 03 14:30:08 crc kubenswrapper[4805]: I1203 14:30:08.871805 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Dec 03 14:30:08 crc kubenswrapper[4805]: I1203 14:30:08.877896 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/10156ee9-4b49-4912-aece-1949d1cc3318-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"10156ee9-4b49-4912-aece-1949d1cc3318\") " pod="openstack/glance-default-external-api-0" Dec 03 14:30:08 crc kubenswrapper[4805]: I1203 14:30:08.877951 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/10156ee9-4b49-4912-aece-1949d1cc3318-config-data\") pod \"glance-default-external-api-0\" (UID: \"10156ee9-4b49-4912-aece-1949d1cc3318\") " pod="openstack/glance-default-external-api-0" Dec 03 14:30:08 crc kubenswrapper[4805]: I1203 14:30:08.877997 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/10156ee9-4b49-4912-aece-1949d1cc3318-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"10156ee9-4b49-4912-aece-1949d1cc3318\") " pod="openstack/glance-default-external-api-0" Dec 03 14:30:08 crc kubenswrapper[4805]: I1203 14:30:08.878153 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"glance-default-external-api-0\" (UID: \"10156ee9-4b49-4912-aece-1949d1cc3318\") " pod="openstack/glance-default-external-api-0" Dec 03 14:30:08 crc kubenswrapper[4805]: I1203 14:30:08.878246 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/10156ee9-4b49-4912-aece-1949d1cc3318-scripts\") pod \"glance-default-external-api-0\" (UID: \"10156ee9-4b49-4912-aece-1949d1cc3318\") " pod="openstack/glance-default-external-api-0" Dec 03 14:30:08 crc kubenswrapper[4805]: I1203 14:30:08.878282 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/10156ee9-4b49-4912-aece-1949d1cc3318-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"10156ee9-4b49-4912-aece-1949d1cc3318\") " pod="openstack/glance-default-external-api-0" Dec 03 14:30:08 crc kubenswrapper[4805]: I1203 14:30:08.878451 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-f2lqz\" (UniqueName: \"kubernetes.io/projected/10156ee9-4b49-4912-aece-1949d1cc3318-kube-api-access-f2lqz\") pod \"glance-default-external-api-0\" (UID: \"10156ee9-4b49-4912-aece-1949d1cc3318\") " pod="openstack/glance-default-external-api-0" Dec 03 14:30:08 crc kubenswrapper[4805]: I1203 14:30:08.878499 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/10156ee9-4b49-4912-aece-1949d1cc3318-logs\") pod \"glance-default-external-api-0\" (UID: \"10156ee9-4b49-4912-aece-1949d1cc3318\") " pod="openstack/glance-default-external-api-0" Dec 03 14:30:08 crc kubenswrapper[4805]: I1203 14:30:08.880959 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-internal-svc" Dec 03 14:30:08 crc kubenswrapper[4805]: I1203 14:30:08.881181 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-internal-config-data" Dec 03 14:30:08 crc kubenswrapper[4805]: I1203 14:30:08.884483 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 03 14:30:08 crc kubenswrapper[4805]: I1203 14:30:08.907808 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 03 14:30:08 crc kubenswrapper[4805]: I1203 14:30:08.983501 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"glance-default-external-api-0\" (UID: \"10156ee9-4b49-4912-aece-1949d1cc3318\") " pod="openstack/glance-default-external-api-0" Dec 03 14:30:08 crc kubenswrapper[4805]: I1203 14:30:08.987072 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/10156ee9-4b49-4912-aece-1949d1cc3318-scripts\") pod \"glance-default-external-api-0\" (UID: \"10156ee9-4b49-4912-aece-1949d1cc3318\") " pod="openstack/glance-default-external-api-0" Dec 03 14:30:08 crc kubenswrapper[4805]: I1203 14:30:08.987206 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/10156ee9-4b49-4912-aece-1949d1cc3318-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"10156ee9-4b49-4912-aece-1949d1cc3318\") " pod="openstack/glance-default-external-api-0" Dec 03 14:30:08 crc kubenswrapper[4805]: I1203 14:30:08.987294 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"glance-default-internal-api-0\" (UID: \"a5c9bc3e-ad1c-428d-bd2e-0730f12b50a8\") " pod="openstack/glance-default-internal-api-0" Dec 03 14:30:08 crc kubenswrapper[4805]: I1203 14:30:08.987414 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a5c9bc3e-ad1c-428d-bd2e-0730f12b50a8-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"a5c9bc3e-ad1c-428d-bd2e-0730f12b50a8\") " pod="openstack/glance-default-internal-api-0" Dec 03 14:30:08 crc kubenswrapper[4805]: I1203 14:30:08.987452 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/a5c9bc3e-ad1c-428d-bd2e-0730f12b50a8-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"a5c9bc3e-ad1c-428d-bd2e-0730f12b50a8\") " pod="openstack/glance-default-internal-api-0" Dec 03 14:30:08 crc kubenswrapper[4805]: I1203 14:30:08.987507 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-f2lqz\" (UniqueName: \"kubernetes.io/projected/10156ee9-4b49-4912-aece-1949d1cc3318-kube-api-access-f2lqz\") pod \"glance-default-external-api-0\" (UID: \"10156ee9-4b49-4912-aece-1949d1cc3318\") " pod="openstack/glance-default-external-api-0" Dec 03 14:30:08 crc kubenswrapper[4805]: I1203 14:30:08.987556 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/10156ee9-4b49-4912-aece-1949d1cc3318-logs\") pod \"glance-default-external-api-0\" (UID: \"10156ee9-4b49-4912-aece-1949d1cc3318\") " pod="openstack/glance-default-external-api-0" Dec 03 14:30:08 crc kubenswrapper[4805]: I1203 14:30:08.987598 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-x88ts\" (UniqueName: \"kubernetes.io/projected/a5c9bc3e-ad1c-428d-bd2e-0730f12b50a8-kube-api-access-x88ts\") pod \"glance-default-internal-api-0\" (UID: \"a5c9bc3e-ad1c-428d-bd2e-0730f12b50a8\") " pod="openstack/glance-default-internal-api-0" Dec 03 14:30:08 crc kubenswrapper[4805]: I1203 14:30:08.987637 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/10156ee9-4b49-4912-aece-1949d1cc3318-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"10156ee9-4b49-4912-aece-1949d1cc3318\") " pod="openstack/glance-default-external-api-0" Dec 03 14:30:08 crc kubenswrapper[4805]: I1203 14:30:08.987692 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/a5c9bc3e-ad1c-428d-bd2e-0730f12b50a8-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"a5c9bc3e-ad1c-428d-bd2e-0730f12b50a8\") " pod="openstack/glance-default-internal-api-0" Dec 03 14:30:08 crc kubenswrapper[4805]: I1203 14:30:08.987726 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/10156ee9-4b49-4912-aece-1949d1cc3318-config-data\") pod \"glance-default-external-api-0\" (UID: \"10156ee9-4b49-4912-aece-1949d1cc3318\") " pod="openstack/glance-default-external-api-0" Dec 03 14:30:08 crc kubenswrapper[4805]: I1203 14:30:08.987818 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a5c9bc3e-ad1c-428d-bd2e-0730f12b50a8-logs\") pod \"glance-default-internal-api-0\" (UID: \"a5c9bc3e-ad1c-428d-bd2e-0730f12b50a8\") " pod="openstack/glance-default-internal-api-0" Dec 03 14:30:08 crc kubenswrapper[4805]: I1203 14:30:08.987869 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/10156ee9-4b49-4912-aece-1949d1cc3318-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"10156ee9-4b49-4912-aece-1949d1cc3318\") " pod="openstack/glance-default-external-api-0" Dec 03 14:30:08 crc kubenswrapper[4805]: I1203 14:30:08.987923 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a5c9bc3e-ad1c-428d-bd2e-0730f12b50a8-config-data\") pod \"glance-default-internal-api-0\" (UID: \"a5c9bc3e-ad1c-428d-bd2e-0730f12b50a8\") " pod="openstack/glance-default-internal-api-0" Dec 03 14:30:08 crc kubenswrapper[4805]: I1203 14:30:08.988013 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a5c9bc3e-ad1c-428d-bd2e-0730f12b50a8-scripts\") pod \"glance-default-internal-api-0\" (UID: \"a5c9bc3e-ad1c-428d-bd2e-0730f12b50a8\") " pod="openstack/glance-default-internal-api-0" Dec 03 14:30:08 crc kubenswrapper[4805]: I1203 14:30:08.984444 4805 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"glance-default-external-api-0\" (UID: \"10156ee9-4b49-4912-aece-1949d1cc3318\") device mount path \"/mnt/openstack/pv07\"" pod="openstack/glance-default-external-api-0" Dec 03 14:30:08 crc kubenswrapper[4805]: I1203 14:30:08.989176 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/10156ee9-4b49-4912-aece-1949d1cc3318-logs\") pod \"glance-default-external-api-0\" (UID: \"10156ee9-4b49-4912-aece-1949d1cc3318\") " pod="openstack/glance-default-external-api-0" Dec 03 14:30:08 crc kubenswrapper[4805]: I1203 14:30:08.989422 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/10156ee9-4b49-4912-aece-1949d1cc3318-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"10156ee9-4b49-4912-aece-1949d1cc3318\") " pod="openstack/glance-default-external-api-0" Dec 03 14:30:08 crc kubenswrapper[4805]: I1203 14:30:08.994823 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/10156ee9-4b49-4912-aece-1949d1cc3318-scripts\") pod \"glance-default-external-api-0\" (UID: \"10156ee9-4b49-4912-aece-1949d1cc3318\") " pod="openstack/glance-default-external-api-0" Dec 03 14:30:08 crc kubenswrapper[4805]: I1203 14:30:08.996512 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/10156ee9-4b49-4912-aece-1949d1cc3318-config-data\") pod \"glance-default-external-api-0\" (UID: \"10156ee9-4b49-4912-aece-1949d1cc3318\") " pod="openstack/glance-default-external-api-0" Dec 03 14:30:09 crc kubenswrapper[4805]: W1203 14:30:09.006562 4805 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod13ba03f0_acc2_4fb8_90a9_583b4f8694cb.slice/crio-6c6724ca7aa2fbf1c806e79de571dbe6e1f2da3605e0f56f04d421afcbd569f6 WatchSource:0}: Error finding container 6c6724ca7aa2fbf1c806e79de571dbe6e1f2da3605e0f56f04d421afcbd569f6: Status 404 returned error can't find the container with id 6c6724ca7aa2fbf1c806e79de571dbe6e1f2da3605e0f56f04d421afcbd569f6 Dec 03 14:30:09 crc kubenswrapper[4805]: I1203 14:30:09.007221 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/10156ee9-4b49-4912-aece-1949d1cc3318-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"10156ee9-4b49-4912-aece-1949d1cc3318\") " pod="openstack/glance-default-external-api-0" Dec 03 14:30:09 crc kubenswrapper[4805]: I1203 14:30:09.017491 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/10156ee9-4b49-4912-aece-1949d1cc3318-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"10156ee9-4b49-4912-aece-1949d1cc3318\") " pod="openstack/glance-default-external-api-0" Dec 03 14:30:09 crc kubenswrapper[4805]: I1203 14:30:09.042268 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-f2lqz\" (UniqueName: \"kubernetes.io/projected/10156ee9-4b49-4912-aece-1949d1cc3318-kube-api-access-f2lqz\") pod \"glance-default-external-api-0\" (UID: \"10156ee9-4b49-4912-aece-1949d1cc3318\") " pod="openstack/glance-default-external-api-0" Dec 03 14:30:09 crc kubenswrapper[4805]: I1203 14:30:09.095971 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"glance-default-external-api-0\" (UID: \"10156ee9-4b49-4912-aece-1949d1cc3318\") " pod="openstack/glance-default-external-api-0" Dec 03 14:30:09 crc kubenswrapper[4805]: I1203 14:30:09.097472 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"glance-default-internal-api-0\" (UID: \"a5c9bc3e-ad1c-428d-bd2e-0730f12b50a8\") " pod="openstack/glance-default-internal-api-0" Dec 03 14:30:09 crc kubenswrapper[4805]: I1203 14:30:09.097774 4805 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"glance-default-internal-api-0\" (UID: \"a5c9bc3e-ad1c-428d-bd2e-0730f12b50a8\") device mount path \"/mnt/openstack/pv11\"" pod="openstack/glance-default-internal-api-0" Dec 03 14:30:09 crc kubenswrapper[4805]: I1203 14:30:09.101706 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a5c9bc3e-ad1c-428d-bd2e-0730f12b50a8-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"a5c9bc3e-ad1c-428d-bd2e-0730f12b50a8\") " pod="openstack/glance-default-internal-api-0" Dec 03 14:30:09 crc kubenswrapper[4805]: I1203 14:30:09.103221 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/a5c9bc3e-ad1c-428d-bd2e-0730f12b50a8-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"a5c9bc3e-ad1c-428d-bd2e-0730f12b50a8\") " pod="openstack/glance-default-internal-api-0" Dec 03 14:30:09 crc kubenswrapper[4805]: I1203 14:30:09.105615 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-x88ts\" (UniqueName: \"kubernetes.io/projected/a5c9bc3e-ad1c-428d-bd2e-0730f12b50a8-kube-api-access-x88ts\") pod \"glance-default-internal-api-0\" (UID: \"a5c9bc3e-ad1c-428d-bd2e-0730f12b50a8\") " pod="openstack/glance-default-internal-api-0" Dec 03 14:30:09 crc kubenswrapper[4805]: I1203 14:30:09.105927 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/a5c9bc3e-ad1c-428d-bd2e-0730f12b50a8-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"a5c9bc3e-ad1c-428d-bd2e-0730f12b50a8\") " pod="openstack/glance-default-internal-api-0" Dec 03 14:30:09 crc kubenswrapper[4805]: I1203 14:30:09.106176 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a5c9bc3e-ad1c-428d-bd2e-0730f12b50a8-logs\") pod \"glance-default-internal-api-0\" (UID: \"a5c9bc3e-ad1c-428d-bd2e-0730f12b50a8\") " pod="openstack/glance-default-internal-api-0" Dec 03 14:30:09 crc kubenswrapper[4805]: I1203 14:30:09.106357 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a5c9bc3e-ad1c-428d-bd2e-0730f12b50a8-config-data\") pod \"glance-default-internal-api-0\" (UID: \"a5c9bc3e-ad1c-428d-bd2e-0730f12b50a8\") " pod="openstack/glance-default-internal-api-0" Dec 03 14:30:09 crc kubenswrapper[4805]: I1203 14:30:09.106513 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a5c9bc3e-ad1c-428d-bd2e-0730f12b50a8-scripts\") pod \"glance-default-internal-api-0\" (UID: \"a5c9bc3e-ad1c-428d-bd2e-0730f12b50a8\") " pod="openstack/glance-default-internal-api-0" Dec 03 14:30:09 crc kubenswrapper[4805]: I1203 14:30:09.117254 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a5c9bc3e-ad1c-428d-bd2e-0730f12b50a8-logs\") pod \"glance-default-internal-api-0\" (UID: \"a5c9bc3e-ad1c-428d-bd2e-0730f12b50a8\") " pod="openstack/glance-default-internal-api-0" Dec 03 14:30:09 crc kubenswrapper[4805]: I1203 14:30:09.119038 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/a5c9bc3e-ad1c-428d-bd2e-0730f12b50a8-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"a5c9bc3e-ad1c-428d-bd2e-0730f12b50a8\") " pod="openstack/glance-default-internal-api-0" Dec 03 14:30:09 crc kubenswrapper[4805]: I1203 14:30:09.120123 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a5c9bc3e-ad1c-428d-bd2e-0730f12b50a8-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"a5c9bc3e-ad1c-428d-bd2e-0730f12b50a8\") " pod="openstack/glance-default-internal-api-0" Dec 03 14:30:09 crc kubenswrapper[4805]: I1203 14:30:09.122618 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-x88ts\" (UniqueName: \"kubernetes.io/projected/a5c9bc3e-ad1c-428d-bd2e-0730f12b50a8-kube-api-access-x88ts\") pod \"glance-default-internal-api-0\" (UID: \"a5c9bc3e-ad1c-428d-bd2e-0730f12b50a8\") " pod="openstack/glance-default-internal-api-0" Dec 03 14:30:09 crc kubenswrapper[4805]: I1203 14:30:09.123168 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a5c9bc3e-ad1c-428d-bd2e-0730f12b50a8-config-data\") pod \"glance-default-internal-api-0\" (UID: \"a5c9bc3e-ad1c-428d-bd2e-0730f12b50a8\") " pod="openstack/glance-default-internal-api-0" Dec 03 14:30:09 crc kubenswrapper[4805]: I1203 14:30:09.131608 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a5c9bc3e-ad1c-428d-bd2e-0730f12b50a8-scripts\") pod \"glance-default-internal-api-0\" (UID: \"a5c9bc3e-ad1c-428d-bd2e-0730f12b50a8\") " pod="openstack/glance-default-internal-api-0" Dec 03 14:30:09 crc kubenswrapper[4805]: I1203 14:30:09.136277 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-qk4m4" event={"ID":"16dbab06-8ff8-4a5a-9fe8-4967c6838a4a","Type":"ContainerStarted","Data":"48d1faafacba3dd6eb3f10ec93a9f796e52e3079e0837e1e2ce22d0318284457"} Dec 03 14:30:09 crc kubenswrapper[4805]: I1203 14:30:09.145737 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/a5c9bc3e-ad1c-428d-bd2e-0730f12b50a8-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"a5c9bc3e-ad1c-428d-bd2e-0730f12b50a8\") " pod="openstack/glance-default-internal-api-0" Dec 03 14:30:09 crc kubenswrapper[4805]: I1203 14:30:09.151273 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5c5cc7c5ff-kxgt6" event={"ID":"d1b43e5a-9dc5-4b92-a5dd-a22beff9f447","Type":"ContainerStarted","Data":"2fbcde76b59b73d3ace51b056a59c1d578b46372b5506e18da85b00b657f64bd"} Dec 03 14:30:09 crc kubenswrapper[4805]: I1203 14:30:09.160112 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"8533d470-22f2-401a-a39d-c37391920d81","Type":"ContainerStarted","Data":"bc0c877e9184355cf11dd77131fca5e9b83afff5f85a897efcbb6fe87c45a6c1"} Dec 03 14:30:09 crc kubenswrapper[4805]: I1203 14:30:09.161068 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-sync-74tz4" event={"ID":"13ba03f0-acc2-4fb8-90a9-583b4f8694cb","Type":"ContainerStarted","Data":"6c6724ca7aa2fbf1c806e79de571dbe6e1f2da3605e0f56f04d421afcbd569f6"} Dec 03 14:30:09 crc kubenswrapper[4805]: I1203 14:30:09.191097 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"glance-default-internal-api-0\" (UID: \"a5c9bc3e-ad1c-428d-bd2e-0730f12b50a8\") " pod="openstack/glance-default-internal-api-0" Dec 03 14:30:09 crc kubenswrapper[4805]: I1203 14:30:09.231738 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Dec 03 14:30:09 crc kubenswrapper[4805]: I1203 14:30:09.240947 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Dec 03 14:30:09 crc kubenswrapper[4805]: I1203 14:30:09.391045 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-db-sync-fh9c2"] Dec 03 14:30:09 crc kubenswrapper[4805]: I1203 14:30:09.400874 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-db-sync-sj4wt"] Dec 03 14:30:09 crc kubenswrapper[4805]: W1203 14:30:09.408458 4805 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod260717ec_f898_4ccd_8a4f_c7d6b14ade6e.slice/crio-d577be3bf13b8fce168f412977177621057fb7ad146b5d216cd41ffc22be9462 WatchSource:0}: Error finding container d577be3bf13b8fce168f412977177621057fb7ad146b5d216cd41ffc22be9462: Status 404 returned error can't find the container with id d577be3bf13b8fce168f412977177621057fb7ad146b5d216cd41ffc22be9462 Dec 03 14:30:09 crc kubenswrapper[4805]: W1203 14:30:09.410810 4805 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod7c1b5525_de7b_4a8c_9062_bb9a68336989.slice/crio-438e761f1425123da258ad8b5ac54c3ef29d62c8d78ab6103c5ecee84a795ff2 WatchSource:0}: Error finding container 438e761f1425123da258ad8b5ac54c3ef29d62c8d78ab6103c5ecee84a795ff2: Status 404 returned error can't find the container with id 438e761f1425123da258ad8b5ac54c3ef29d62c8d78ab6103c5ecee84a795ff2 Dec 03 14:30:09 crc kubenswrapper[4805]: I1203 14:30:09.555821 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-8b5c85b87-bvb8w"] Dec 03 14:30:09 crc kubenswrapper[4805]: W1203 14:30:09.558475 4805 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poda8299a7f_082a_4a51_becd_7c393a6f4b51.slice/crio-ee5acaf567b19c4061d7d5cc37b1adc10bd8c89de090f9b575df0e9c5c942ea9 WatchSource:0}: Error finding container ee5acaf567b19c4061d7d5cc37b1adc10bd8c89de090f9b575df0e9c5c942ea9: Status 404 returned error can't find the container with id ee5acaf567b19c4061d7d5cc37b1adc10bd8c89de090f9b575df0e9c5c942ea9 Dec 03 14:30:09 crc kubenswrapper[4805]: I1203 14:30:09.652309 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-db-sync-jvqmk"] Dec 03 14:30:09 crc kubenswrapper[4805]: I1203 14:30:09.677327 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-db-sync-b6k7w"] Dec 03 14:30:09 crc kubenswrapper[4805]: W1203 14:30:09.680310 4805 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod32920f09_a728_4b35_a5d7_65dae2d70834.slice/crio-e197e76c91b87d530bb2b8c17e87b85b444572fd3c9af48e537d98f418a87313 WatchSource:0}: Error finding container e197e76c91b87d530bb2b8c17e87b85b444572fd3c9af48e537d98f418a87313: Status 404 returned error can't find the container with id e197e76c91b87d530bb2b8c17e87b85b444572fd3c9af48e537d98f418a87313 Dec 03 14:30:09 crc kubenswrapper[4805]: I1203 14:30:09.934681 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Dec 03 14:30:09 crc kubenswrapper[4805]: I1203 14:30:09.983986 4805 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 03 14:30:10 crc kubenswrapper[4805]: I1203 14:30:10.115123 4805 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 03 14:30:10 crc kubenswrapper[4805]: I1203 14:30:10.178892 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-8b5c85b87-bvb8w" event={"ID":"a8299a7f-082a-4a51-becd-7c393a6f4b51","Type":"ContainerStarted","Data":"ee5acaf567b19c4061d7d5cc37b1adc10bd8c89de090f9b575df0e9c5c942ea9"} Dec 03 14:30:10 crc kubenswrapper[4805]: I1203 14:30:10.180480 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-sync-fh9c2" event={"ID":"260717ec-f898-4ccd-8a4f-c7d6b14ade6e","Type":"ContainerStarted","Data":"d577be3bf13b8fce168f412977177621057fb7ad146b5d216cd41ffc22be9462"} Dec 03 14:30:10 crc kubenswrapper[4805]: I1203 14:30:10.184921 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-sync-sj4wt" event={"ID":"7c1b5525-de7b-4a8c-9062-bb9a68336989","Type":"ContainerStarted","Data":"438e761f1425123da258ad8b5ac54c3ef29d62c8d78ab6103c5ecee84a795ff2"} Dec 03 14:30:10 crc kubenswrapper[4805]: I1203 14:30:10.195907 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"10156ee9-4b49-4912-aece-1949d1cc3318","Type":"ContainerStarted","Data":"500782b9560ad99b333bf78fa3b221bad4289ac4f9a8e64de6d7e2c5d7a45816"} Dec 03 14:30:10 crc kubenswrapper[4805]: I1203 14:30:10.199269 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-db-sync-jvqmk" event={"ID":"14c327e3-e880-4b49-b366-364a0395cc55","Type":"ContainerStarted","Data":"1a8457670ac140f0c542b92f1e3130ca93dd7a523ed000d606773dd4046ddee3"} Dec 03 14:30:10 crc kubenswrapper[4805]: I1203 14:30:10.199559 4805 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Dec 03 14:30:10 crc kubenswrapper[4805]: I1203 14:30:10.207071 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-sync-b6k7w" event={"ID":"32920f09-a728-4b35-a5d7-65dae2d70834","Type":"ContainerStarted","Data":"e197e76c91b87d530bb2b8c17e87b85b444572fd3c9af48e537d98f418a87313"} Dec 03 14:30:10 crc kubenswrapper[4805]: I1203 14:30:10.663533 4805 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 03 14:30:10 crc kubenswrapper[4805]: W1203 14:30:10.666336 4805 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poda5c9bc3e_ad1c_428d_bd2e_0730f12b50a8.slice/crio-10dfc8c10ea0cb8d67b76038eb994ba9d3cced592669a9182639acd5d197ae45 WatchSource:0}: Error finding container 10dfc8c10ea0cb8d67b76038eb994ba9d3cced592669a9182639acd5d197ae45: Status 404 returned error can't find the container with id 10dfc8c10ea0cb8d67b76038eb994ba9d3cced592669a9182639acd5d197ae45 Dec 03 14:30:11 crc kubenswrapper[4805]: I1203 14:30:11.318691 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-sync-b6k7w" event={"ID":"32920f09-a728-4b35-a5d7-65dae2d70834","Type":"ContainerStarted","Data":"c8040c823d6c387af893c10d68844e6a3f47022e19a18567b34538fe04cfa697"} Dec 03 14:30:11 crc kubenswrapper[4805]: I1203 14:30:11.331367 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-qk4m4" event={"ID":"16dbab06-8ff8-4a5a-9fe8-4967c6838a4a","Type":"ContainerStarted","Data":"794a23f04151f6d0d061c884ab2a4ed2e4ba2f4abbcb5f8271c84b63773854dc"} Dec 03 14:30:11 crc kubenswrapper[4805]: I1203 14:30:11.339736 4805 generic.go:334] "Generic (PLEG): container finished" podID="a8299a7f-082a-4a51-becd-7c393a6f4b51" containerID="30eba44f20563e0ae1a1e3b18ccb56d1b633b85b21a348ad72c59383ee93daef" exitCode=0 Dec 03 14:30:11 crc kubenswrapper[4805]: I1203 14:30:11.339876 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-8b5c85b87-bvb8w" event={"ID":"a8299a7f-082a-4a51-becd-7c393a6f4b51","Type":"ContainerDied","Data":"30eba44f20563e0ae1a1e3b18ccb56d1b633b85b21a348ad72c59383ee93daef"} Dec 03 14:30:11 crc kubenswrapper[4805]: I1203 14:30:11.346315 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"a5c9bc3e-ad1c-428d-bd2e-0730f12b50a8","Type":"ContainerStarted","Data":"10dfc8c10ea0cb8d67b76038eb994ba9d3cced592669a9182639acd5d197ae45"} Dec 03 14:30:11 crc kubenswrapper[4805]: I1203 14:30:11.355713 4805 generic.go:334] "Generic (PLEG): container finished" podID="d1b43e5a-9dc5-4b92-a5dd-a22beff9f447" containerID="c6267b207a140cc4a91e085609f68968d0abb44d3f734ef6753b160a63f9ca6c" exitCode=0 Dec 03 14:30:11 crc kubenswrapper[4805]: I1203 14:30:11.355998 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5c5cc7c5ff-kxgt6" event={"ID":"d1b43e5a-9dc5-4b92-a5dd-a22beff9f447","Type":"ContainerDied","Data":"c6267b207a140cc4a91e085609f68968d0abb44d3f734ef6753b160a63f9ca6c"} Dec 03 14:30:11 crc kubenswrapper[4805]: I1203 14:30:11.387252 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"10156ee9-4b49-4912-aece-1949d1cc3318","Type":"ContainerStarted","Data":"cce0b3eb1a191554e4667b6b4e39dee3d15e0713c7fd79b5fda0181e0d7cbeac"} Dec 03 14:30:11 crc kubenswrapper[4805]: I1203 14:30:11.394552 4805 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-db-sync-b6k7w" podStartSLOduration=3.394531213 podStartE2EDuration="3.394531213s" podCreationTimestamp="2025-12-03 14:30:08 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 14:30:11.36597405 +0000 UTC m=+1241.028890973" watchObservedRunningTime="2025-12-03 14:30:11.394531213 +0000 UTC m=+1241.057448136" Dec 03 14:30:11 crc kubenswrapper[4805]: I1203 14:30:11.436780 4805 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-bootstrap-qk4m4" podStartSLOduration=4.436758566 podStartE2EDuration="4.436758566s" podCreationTimestamp="2025-12-03 14:30:07 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 14:30:11.386630139 +0000 UTC m=+1241.049547062" watchObservedRunningTime="2025-12-03 14:30:11.436758566 +0000 UTC m=+1241.099675489" Dec 03 14:30:11 crc kubenswrapper[4805]: I1203 14:30:11.883482 4805 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5c5cc7c5ff-kxgt6" Dec 03 14:30:12 crc kubenswrapper[4805]: I1203 14:30:12.013876 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/d1b43e5a-9dc5-4b92-a5dd-a22beff9f447-ovsdbserver-nb\") pod \"d1b43e5a-9dc5-4b92-a5dd-a22beff9f447\" (UID: \"d1b43e5a-9dc5-4b92-a5dd-a22beff9f447\") " Dec 03 14:30:12 crc kubenswrapper[4805]: I1203 14:30:12.013937 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/d1b43e5a-9dc5-4b92-a5dd-a22beff9f447-dns-svc\") pod \"d1b43e5a-9dc5-4b92-a5dd-a22beff9f447\" (UID: \"d1b43e5a-9dc5-4b92-a5dd-a22beff9f447\") " Dec 03 14:30:12 crc kubenswrapper[4805]: I1203 14:30:12.013987 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2k58p\" (UniqueName: \"kubernetes.io/projected/d1b43e5a-9dc5-4b92-a5dd-a22beff9f447-kube-api-access-2k58p\") pod \"d1b43e5a-9dc5-4b92-a5dd-a22beff9f447\" (UID: \"d1b43e5a-9dc5-4b92-a5dd-a22beff9f447\") " Dec 03 14:30:12 crc kubenswrapper[4805]: I1203 14:30:12.014040 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/d1b43e5a-9dc5-4b92-a5dd-a22beff9f447-ovsdbserver-sb\") pod \"d1b43e5a-9dc5-4b92-a5dd-a22beff9f447\" (UID: \"d1b43e5a-9dc5-4b92-a5dd-a22beff9f447\") " Dec 03 14:30:12 crc kubenswrapper[4805]: I1203 14:30:12.014079 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/d1b43e5a-9dc5-4b92-a5dd-a22beff9f447-dns-swift-storage-0\") pod \"d1b43e5a-9dc5-4b92-a5dd-a22beff9f447\" (UID: \"d1b43e5a-9dc5-4b92-a5dd-a22beff9f447\") " Dec 03 14:30:12 crc kubenswrapper[4805]: I1203 14:30:12.015356 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d1b43e5a-9dc5-4b92-a5dd-a22beff9f447-config\") pod \"d1b43e5a-9dc5-4b92-a5dd-a22beff9f447\" (UID: \"d1b43e5a-9dc5-4b92-a5dd-a22beff9f447\") " Dec 03 14:30:12 crc kubenswrapper[4805]: I1203 14:30:12.041815 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d1b43e5a-9dc5-4b92-a5dd-a22beff9f447-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "d1b43e5a-9dc5-4b92-a5dd-a22beff9f447" (UID: "d1b43e5a-9dc5-4b92-a5dd-a22beff9f447"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 14:30:12 crc kubenswrapper[4805]: I1203 14:30:12.041833 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d1b43e5a-9dc5-4b92-a5dd-a22beff9f447-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "d1b43e5a-9dc5-4b92-a5dd-a22beff9f447" (UID: "d1b43e5a-9dc5-4b92-a5dd-a22beff9f447"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 14:30:12 crc kubenswrapper[4805]: I1203 14:30:12.047640 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d1b43e5a-9dc5-4b92-a5dd-a22beff9f447-kube-api-access-2k58p" (OuterVolumeSpecName: "kube-api-access-2k58p") pod "d1b43e5a-9dc5-4b92-a5dd-a22beff9f447" (UID: "d1b43e5a-9dc5-4b92-a5dd-a22beff9f447"). InnerVolumeSpecName "kube-api-access-2k58p". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 14:30:12 crc kubenswrapper[4805]: I1203 14:30:12.058722 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d1b43e5a-9dc5-4b92-a5dd-a22beff9f447-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "d1b43e5a-9dc5-4b92-a5dd-a22beff9f447" (UID: "d1b43e5a-9dc5-4b92-a5dd-a22beff9f447"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 14:30:12 crc kubenswrapper[4805]: I1203 14:30:12.065394 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d1b43e5a-9dc5-4b92-a5dd-a22beff9f447-config" (OuterVolumeSpecName: "config") pod "d1b43e5a-9dc5-4b92-a5dd-a22beff9f447" (UID: "d1b43e5a-9dc5-4b92-a5dd-a22beff9f447"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 14:30:12 crc kubenswrapper[4805]: I1203 14:30:12.071271 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d1b43e5a-9dc5-4b92-a5dd-a22beff9f447-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "d1b43e5a-9dc5-4b92-a5dd-a22beff9f447" (UID: "d1b43e5a-9dc5-4b92-a5dd-a22beff9f447"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 14:30:12 crc kubenswrapper[4805]: I1203 14:30:12.119222 4805 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d1b43e5a-9dc5-4b92-a5dd-a22beff9f447-config\") on node \"crc\" DevicePath \"\"" Dec 03 14:30:12 crc kubenswrapper[4805]: I1203 14:30:12.119253 4805 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/d1b43e5a-9dc5-4b92-a5dd-a22beff9f447-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Dec 03 14:30:12 crc kubenswrapper[4805]: I1203 14:30:12.119265 4805 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/d1b43e5a-9dc5-4b92-a5dd-a22beff9f447-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 03 14:30:12 crc kubenswrapper[4805]: I1203 14:30:12.119274 4805 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2k58p\" (UniqueName: \"kubernetes.io/projected/d1b43e5a-9dc5-4b92-a5dd-a22beff9f447-kube-api-access-2k58p\") on node \"crc\" DevicePath \"\"" Dec 03 14:30:12 crc kubenswrapper[4805]: I1203 14:30:12.119283 4805 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/d1b43e5a-9dc5-4b92-a5dd-a22beff9f447-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Dec 03 14:30:12 crc kubenswrapper[4805]: I1203 14:30:12.119291 4805 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/d1b43e5a-9dc5-4b92-a5dd-a22beff9f447-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Dec 03 14:30:12 crc kubenswrapper[4805]: I1203 14:30:12.399543 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"10156ee9-4b49-4912-aece-1949d1cc3318","Type":"ContainerStarted","Data":"da921af3e5e34a16996892cb141982be6389d56c3d7bc6ca0b3979933ddb9aa4"} Dec 03 14:30:12 crc kubenswrapper[4805]: I1203 14:30:12.400091 4805 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-external-api-0" podUID="10156ee9-4b49-4912-aece-1949d1cc3318" containerName="glance-log" containerID="cri-o://cce0b3eb1a191554e4667b6b4e39dee3d15e0713c7fd79b5fda0181e0d7cbeac" gracePeriod=30 Dec 03 14:30:12 crc kubenswrapper[4805]: I1203 14:30:12.400713 4805 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-external-api-0" podUID="10156ee9-4b49-4912-aece-1949d1cc3318" containerName="glance-httpd" containerID="cri-o://da921af3e5e34a16996892cb141982be6389d56c3d7bc6ca0b3979933ddb9aa4" gracePeriod=30 Dec 03 14:30:12 crc kubenswrapper[4805]: I1203 14:30:12.410096 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-8b5c85b87-bvb8w" event={"ID":"a8299a7f-082a-4a51-becd-7c393a6f4b51","Type":"ContainerStarted","Data":"58d07340f3e9637400fe3855db1f26dfcf2ec01f9006fa34bbfe5a88846c9ff2"} Dec 03 14:30:12 crc kubenswrapper[4805]: I1203 14:30:12.410216 4805 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-8b5c85b87-bvb8w" Dec 03 14:30:12 crc kubenswrapper[4805]: I1203 14:30:12.412728 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"a5c9bc3e-ad1c-428d-bd2e-0730f12b50a8","Type":"ContainerStarted","Data":"188ca93940f09e5d5f1a3e68344b1ff578bfc6d8a437936fc9d054967cfa3ed2"} Dec 03 14:30:12 crc kubenswrapper[4805]: I1203 14:30:12.419376 4805 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5c5cc7c5ff-kxgt6" Dec 03 14:30:12 crc kubenswrapper[4805]: I1203 14:30:12.424919 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5c5cc7c5ff-kxgt6" event={"ID":"d1b43e5a-9dc5-4b92-a5dd-a22beff9f447","Type":"ContainerDied","Data":"2fbcde76b59b73d3ace51b056a59c1d578b46372b5506e18da85b00b657f64bd"} Dec 03 14:30:12 crc kubenswrapper[4805]: I1203 14:30:12.425265 4805 scope.go:117] "RemoveContainer" containerID="c6267b207a140cc4a91e085609f68968d0abb44d3f734ef6753b160a63f9ca6c" Dec 03 14:30:12 crc kubenswrapper[4805]: I1203 14:30:12.447077 4805 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-external-api-0" podStartSLOduration=5.447058155 podStartE2EDuration="5.447058155s" podCreationTimestamp="2025-12-03 14:30:07 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 14:30:12.425252245 +0000 UTC m=+1242.088169178" watchObservedRunningTime="2025-12-03 14:30:12.447058155 +0000 UTC m=+1242.109975078" Dec 03 14:30:12 crc kubenswrapper[4805]: I1203 14:30:12.465297 4805 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-8b5c85b87-bvb8w" podStartSLOduration=4.465273178 podStartE2EDuration="4.465273178s" podCreationTimestamp="2025-12-03 14:30:08 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 14:30:12.44021651 +0000 UTC m=+1242.103133433" watchObservedRunningTime="2025-12-03 14:30:12.465273178 +0000 UTC m=+1242.128190101" Dec 03 14:30:12 crc kubenswrapper[4805]: I1203 14:30:12.521032 4805 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5c5cc7c5ff-kxgt6"] Dec 03 14:30:12 crc kubenswrapper[4805]: I1203 14:30:12.533420 4805 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-5c5cc7c5ff-kxgt6"] Dec 03 14:30:12 crc kubenswrapper[4805]: I1203 14:30:12.709052 4805 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d1b43e5a-9dc5-4b92-a5dd-a22beff9f447" path="/var/lib/kubelet/pods/d1b43e5a-9dc5-4b92-a5dd-a22beff9f447/volumes" Dec 03 14:30:13 crc kubenswrapper[4805]: I1203 14:30:13.134553 4805 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Dec 03 14:30:13 crc kubenswrapper[4805]: I1203 14:30:13.258769 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"10156ee9-4b49-4912-aece-1949d1cc3318\" (UID: \"10156ee9-4b49-4912-aece-1949d1cc3318\") " Dec 03 14:30:13 crc kubenswrapper[4805]: I1203 14:30:13.259455 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/10156ee9-4b49-4912-aece-1949d1cc3318-config-data\") pod \"10156ee9-4b49-4912-aece-1949d1cc3318\" (UID: \"10156ee9-4b49-4912-aece-1949d1cc3318\") " Dec 03 14:30:13 crc kubenswrapper[4805]: I1203 14:30:13.259499 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/10156ee9-4b49-4912-aece-1949d1cc3318-scripts\") pod \"10156ee9-4b49-4912-aece-1949d1cc3318\" (UID: \"10156ee9-4b49-4912-aece-1949d1cc3318\") " Dec 03 14:30:13 crc kubenswrapper[4805]: I1203 14:30:13.259750 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/10156ee9-4b49-4912-aece-1949d1cc3318-combined-ca-bundle\") pod \"10156ee9-4b49-4912-aece-1949d1cc3318\" (UID: \"10156ee9-4b49-4912-aece-1949d1cc3318\") " Dec 03 14:30:13 crc kubenswrapper[4805]: I1203 14:30:13.259904 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/10156ee9-4b49-4912-aece-1949d1cc3318-logs\") pod \"10156ee9-4b49-4912-aece-1949d1cc3318\" (UID: \"10156ee9-4b49-4912-aece-1949d1cc3318\") " Dec 03 14:30:13 crc kubenswrapper[4805]: I1203 14:30:13.260635 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/10156ee9-4b49-4912-aece-1949d1cc3318-httpd-run\") pod \"10156ee9-4b49-4912-aece-1949d1cc3318\" (UID: \"10156ee9-4b49-4912-aece-1949d1cc3318\") " Dec 03 14:30:13 crc kubenswrapper[4805]: I1203 14:30:13.260669 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/10156ee9-4b49-4912-aece-1949d1cc3318-public-tls-certs\") pod \"10156ee9-4b49-4912-aece-1949d1cc3318\" (UID: \"10156ee9-4b49-4912-aece-1949d1cc3318\") " Dec 03 14:30:13 crc kubenswrapper[4805]: I1203 14:30:13.260697 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-f2lqz\" (UniqueName: \"kubernetes.io/projected/10156ee9-4b49-4912-aece-1949d1cc3318-kube-api-access-f2lqz\") pod \"10156ee9-4b49-4912-aece-1949d1cc3318\" (UID: \"10156ee9-4b49-4912-aece-1949d1cc3318\") " Dec 03 14:30:13 crc kubenswrapper[4805]: I1203 14:30:13.261045 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/10156ee9-4b49-4912-aece-1949d1cc3318-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "10156ee9-4b49-4912-aece-1949d1cc3318" (UID: "10156ee9-4b49-4912-aece-1949d1cc3318"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 14:30:13 crc kubenswrapper[4805]: I1203 14:30:13.261162 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/10156ee9-4b49-4912-aece-1949d1cc3318-logs" (OuterVolumeSpecName: "logs") pod "10156ee9-4b49-4912-aece-1949d1cc3318" (UID: "10156ee9-4b49-4912-aece-1949d1cc3318"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 14:30:13 crc kubenswrapper[4805]: I1203 14:30:13.261515 4805 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/10156ee9-4b49-4912-aece-1949d1cc3318-logs\") on node \"crc\" DevicePath \"\"" Dec 03 14:30:13 crc kubenswrapper[4805]: I1203 14:30:13.261635 4805 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/10156ee9-4b49-4912-aece-1949d1cc3318-httpd-run\") on node \"crc\" DevicePath \"\"" Dec 03 14:30:13 crc kubenswrapper[4805]: I1203 14:30:13.267948 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage07-crc" (OuterVolumeSpecName: "glance") pod "10156ee9-4b49-4912-aece-1949d1cc3318" (UID: "10156ee9-4b49-4912-aece-1949d1cc3318"). InnerVolumeSpecName "local-storage07-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Dec 03 14:30:13 crc kubenswrapper[4805]: I1203 14:30:13.269091 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/10156ee9-4b49-4912-aece-1949d1cc3318-kube-api-access-f2lqz" (OuterVolumeSpecName: "kube-api-access-f2lqz") pod "10156ee9-4b49-4912-aece-1949d1cc3318" (UID: "10156ee9-4b49-4912-aece-1949d1cc3318"). InnerVolumeSpecName "kube-api-access-f2lqz". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 14:30:13 crc kubenswrapper[4805]: I1203 14:30:13.279098 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/10156ee9-4b49-4912-aece-1949d1cc3318-scripts" (OuterVolumeSpecName: "scripts") pod "10156ee9-4b49-4912-aece-1949d1cc3318" (UID: "10156ee9-4b49-4912-aece-1949d1cc3318"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 14:30:13 crc kubenswrapper[4805]: I1203 14:30:13.313347 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/10156ee9-4b49-4912-aece-1949d1cc3318-config-data" (OuterVolumeSpecName: "config-data") pod "10156ee9-4b49-4912-aece-1949d1cc3318" (UID: "10156ee9-4b49-4912-aece-1949d1cc3318"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 14:30:13 crc kubenswrapper[4805]: I1203 14:30:13.314027 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/10156ee9-4b49-4912-aece-1949d1cc3318-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "10156ee9-4b49-4912-aece-1949d1cc3318" (UID: "10156ee9-4b49-4912-aece-1949d1cc3318"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 14:30:13 crc kubenswrapper[4805]: I1203 14:30:13.314331 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/10156ee9-4b49-4912-aece-1949d1cc3318-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "10156ee9-4b49-4912-aece-1949d1cc3318" (UID: "10156ee9-4b49-4912-aece-1949d1cc3318"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 14:30:13 crc kubenswrapper[4805]: I1203 14:30:13.363458 4805 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/10156ee9-4b49-4912-aece-1949d1cc3318-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 14:30:13 crc kubenswrapper[4805]: I1203 14:30:13.363803 4805 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/10156ee9-4b49-4912-aece-1949d1cc3318-public-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 03 14:30:13 crc kubenswrapper[4805]: I1203 14:30:13.363816 4805 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-f2lqz\" (UniqueName: \"kubernetes.io/projected/10156ee9-4b49-4912-aece-1949d1cc3318-kube-api-access-f2lqz\") on node \"crc\" DevicePath \"\"" Dec 03 14:30:13 crc kubenswrapper[4805]: I1203 14:30:13.363864 4805 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") on node \"crc\" " Dec 03 14:30:13 crc kubenswrapper[4805]: I1203 14:30:13.363887 4805 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/10156ee9-4b49-4912-aece-1949d1cc3318-config-data\") on node \"crc\" DevicePath \"\"" Dec 03 14:30:13 crc kubenswrapper[4805]: I1203 14:30:13.363899 4805 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/10156ee9-4b49-4912-aece-1949d1cc3318-scripts\") on node \"crc\" DevicePath \"\"" Dec 03 14:30:13 crc kubenswrapper[4805]: I1203 14:30:13.436761 4805 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage07-crc" (UniqueName: "kubernetes.io/local-volume/local-storage07-crc") on node "crc" Dec 03 14:30:13 crc kubenswrapper[4805]: I1203 14:30:13.445598 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"a5c9bc3e-ad1c-428d-bd2e-0730f12b50a8","Type":"ContainerStarted","Data":"1398a8bb8b40a97438fe6d0fbcc36517488187b6d66a96af990ed301eeaf05fd"} Dec 03 14:30:13 crc kubenswrapper[4805]: I1203 14:30:13.445828 4805 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-internal-api-0" podUID="a5c9bc3e-ad1c-428d-bd2e-0730f12b50a8" containerName="glance-log" containerID="cri-o://188ca93940f09e5d5f1a3e68344b1ff578bfc6d8a437936fc9d054967cfa3ed2" gracePeriod=30 Dec 03 14:30:13 crc kubenswrapper[4805]: I1203 14:30:13.445973 4805 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-internal-api-0" podUID="a5c9bc3e-ad1c-428d-bd2e-0730f12b50a8" containerName="glance-httpd" containerID="cri-o://1398a8bb8b40a97438fe6d0fbcc36517488187b6d66a96af990ed301eeaf05fd" gracePeriod=30 Dec 03 14:30:13 crc kubenswrapper[4805]: I1203 14:30:13.473263 4805 generic.go:334] "Generic (PLEG): container finished" podID="10156ee9-4b49-4912-aece-1949d1cc3318" containerID="da921af3e5e34a16996892cb141982be6389d56c3d7bc6ca0b3979933ddb9aa4" exitCode=143 Dec 03 14:30:13 crc kubenswrapper[4805]: I1203 14:30:13.473311 4805 generic.go:334] "Generic (PLEG): container finished" podID="10156ee9-4b49-4912-aece-1949d1cc3318" containerID="cce0b3eb1a191554e4667b6b4e39dee3d15e0713c7fd79b5fda0181e0d7cbeac" exitCode=143 Dec 03 14:30:13 crc kubenswrapper[4805]: I1203 14:30:13.473578 4805 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Dec 03 14:30:13 crc kubenswrapper[4805]: I1203 14:30:13.473912 4805 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-internal-api-0" podStartSLOduration=6.473895761 podStartE2EDuration="6.473895761s" podCreationTimestamp="2025-12-03 14:30:07 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 14:30:13.471915868 +0000 UTC m=+1243.134832791" watchObservedRunningTime="2025-12-03 14:30:13.473895761 +0000 UTC m=+1243.136812694" Dec 03 14:30:13 crc kubenswrapper[4805]: I1203 14:30:13.473567 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"10156ee9-4b49-4912-aece-1949d1cc3318","Type":"ContainerDied","Data":"da921af3e5e34a16996892cb141982be6389d56c3d7bc6ca0b3979933ddb9aa4"} Dec 03 14:30:13 crc kubenswrapper[4805]: I1203 14:30:13.474276 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"10156ee9-4b49-4912-aece-1949d1cc3318","Type":"ContainerDied","Data":"cce0b3eb1a191554e4667b6b4e39dee3d15e0713c7fd79b5fda0181e0d7cbeac"} Dec 03 14:30:13 crc kubenswrapper[4805]: I1203 14:30:13.474293 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"10156ee9-4b49-4912-aece-1949d1cc3318","Type":"ContainerDied","Data":"500782b9560ad99b333bf78fa3b221bad4289ac4f9a8e64de6d7e2c5d7a45816"} Dec 03 14:30:13 crc kubenswrapper[4805]: I1203 14:30:13.474369 4805 scope.go:117] "RemoveContainer" containerID="da921af3e5e34a16996892cb141982be6389d56c3d7bc6ca0b3979933ddb9aa4" Dec 03 14:30:13 crc kubenswrapper[4805]: I1203 14:30:13.474587 4805 reconciler_common.go:293] "Volume detached for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") on node \"crc\" DevicePath \"\"" Dec 03 14:30:13 crc kubenswrapper[4805]: I1203 14:30:13.508782 4805 scope.go:117] "RemoveContainer" containerID="cce0b3eb1a191554e4667b6b4e39dee3d15e0713c7fd79b5fda0181e0d7cbeac" Dec 03 14:30:13 crc kubenswrapper[4805]: I1203 14:30:13.519414 4805 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Dec 03 14:30:13 crc kubenswrapper[4805]: I1203 14:30:13.526234 4805 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-default-external-api-0"] Dec 03 14:30:13 crc kubenswrapper[4805]: I1203 14:30:13.556905 4805 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-external-api-0"] Dec 03 14:30:13 crc kubenswrapper[4805]: E1203 14:30:13.557317 4805 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d1b43e5a-9dc5-4b92-a5dd-a22beff9f447" containerName="init" Dec 03 14:30:13 crc kubenswrapper[4805]: I1203 14:30:13.557331 4805 state_mem.go:107] "Deleted CPUSet assignment" podUID="d1b43e5a-9dc5-4b92-a5dd-a22beff9f447" containerName="init" Dec 03 14:30:13 crc kubenswrapper[4805]: E1203 14:30:13.557348 4805 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="10156ee9-4b49-4912-aece-1949d1cc3318" containerName="glance-httpd" Dec 03 14:30:13 crc kubenswrapper[4805]: I1203 14:30:13.557354 4805 state_mem.go:107] "Deleted CPUSet assignment" podUID="10156ee9-4b49-4912-aece-1949d1cc3318" containerName="glance-httpd" Dec 03 14:30:13 crc kubenswrapper[4805]: E1203 14:30:13.557373 4805 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="10156ee9-4b49-4912-aece-1949d1cc3318" containerName="glance-log" Dec 03 14:30:13 crc kubenswrapper[4805]: I1203 14:30:13.557379 4805 state_mem.go:107] "Deleted CPUSet assignment" podUID="10156ee9-4b49-4912-aece-1949d1cc3318" containerName="glance-log" Dec 03 14:30:13 crc kubenswrapper[4805]: I1203 14:30:13.557571 4805 memory_manager.go:354] "RemoveStaleState removing state" podUID="d1b43e5a-9dc5-4b92-a5dd-a22beff9f447" containerName="init" Dec 03 14:30:13 crc kubenswrapper[4805]: I1203 14:30:13.557585 4805 memory_manager.go:354] "RemoveStaleState removing state" podUID="10156ee9-4b49-4912-aece-1949d1cc3318" containerName="glance-httpd" Dec 03 14:30:13 crc kubenswrapper[4805]: I1203 14:30:13.557594 4805 memory_manager.go:354] "RemoveStaleState removing state" podUID="10156ee9-4b49-4912-aece-1949d1cc3318" containerName="glance-log" Dec 03 14:30:13 crc kubenswrapper[4805]: I1203 14:30:13.558480 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Dec 03 14:30:13 crc kubenswrapper[4805]: I1203 14:30:13.563590 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-public-svc" Dec 03 14:30:13 crc kubenswrapper[4805]: I1203 14:30:13.563750 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-external-config-data" Dec 03 14:30:13 crc kubenswrapper[4805]: I1203 14:30:13.588383 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Dec 03 14:30:13 crc kubenswrapper[4805]: I1203 14:30:13.685795 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/9ed6dfc0-82db-4d86-ad92-cfe94fa370a2-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"9ed6dfc0-82db-4d86-ad92-cfe94fa370a2\") " pod="openstack/glance-default-external-api-0" Dec 03 14:30:13 crc kubenswrapper[4805]: I1203 14:30:13.685860 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9ed6dfc0-82db-4d86-ad92-cfe94fa370a2-config-data\") pod \"glance-default-external-api-0\" (UID: \"9ed6dfc0-82db-4d86-ad92-cfe94fa370a2\") " pod="openstack/glance-default-external-api-0" Dec 03 14:30:13 crc kubenswrapper[4805]: I1203 14:30:13.685918 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9ed6dfc0-82db-4d86-ad92-cfe94fa370a2-scripts\") pod \"glance-default-external-api-0\" (UID: \"9ed6dfc0-82db-4d86-ad92-cfe94fa370a2\") " pod="openstack/glance-default-external-api-0" Dec 03 14:30:13 crc kubenswrapper[4805]: I1203 14:30:13.685953 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-c5677\" (UniqueName: \"kubernetes.io/projected/9ed6dfc0-82db-4d86-ad92-cfe94fa370a2-kube-api-access-c5677\") pod \"glance-default-external-api-0\" (UID: \"9ed6dfc0-82db-4d86-ad92-cfe94fa370a2\") " pod="openstack/glance-default-external-api-0" Dec 03 14:30:13 crc kubenswrapper[4805]: I1203 14:30:13.686042 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"glance-default-external-api-0\" (UID: \"9ed6dfc0-82db-4d86-ad92-cfe94fa370a2\") " pod="openstack/glance-default-external-api-0" Dec 03 14:30:13 crc kubenswrapper[4805]: I1203 14:30:13.686085 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/9ed6dfc0-82db-4d86-ad92-cfe94fa370a2-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"9ed6dfc0-82db-4d86-ad92-cfe94fa370a2\") " pod="openstack/glance-default-external-api-0" Dec 03 14:30:13 crc kubenswrapper[4805]: I1203 14:30:13.686106 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/9ed6dfc0-82db-4d86-ad92-cfe94fa370a2-logs\") pod \"glance-default-external-api-0\" (UID: \"9ed6dfc0-82db-4d86-ad92-cfe94fa370a2\") " pod="openstack/glance-default-external-api-0" Dec 03 14:30:13 crc kubenswrapper[4805]: I1203 14:30:13.686123 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9ed6dfc0-82db-4d86-ad92-cfe94fa370a2-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"9ed6dfc0-82db-4d86-ad92-cfe94fa370a2\") " pod="openstack/glance-default-external-api-0" Dec 03 14:30:13 crc kubenswrapper[4805]: I1203 14:30:13.728099 4805 scope.go:117] "RemoveContainer" containerID="da921af3e5e34a16996892cb141982be6389d56c3d7bc6ca0b3979933ddb9aa4" Dec 03 14:30:13 crc kubenswrapper[4805]: E1203 14:30:13.728612 4805 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"da921af3e5e34a16996892cb141982be6389d56c3d7bc6ca0b3979933ddb9aa4\": container with ID starting with da921af3e5e34a16996892cb141982be6389d56c3d7bc6ca0b3979933ddb9aa4 not found: ID does not exist" containerID="da921af3e5e34a16996892cb141982be6389d56c3d7bc6ca0b3979933ddb9aa4" Dec 03 14:30:13 crc kubenswrapper[4805]: I1203 14:30:13.728646 4805 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"da921af3e5e34a16996892cb141982be6389d56c3d7bc6ca0b3979933ddb9aa4"} err="failed to get container status \"da921af3e5e34a16996892cb141982be6389d56c3d7bc6ca0b3979933ddb9aa4\": rpc error: code = NotFound desc = could not find container \"da921af3e5e34a16996892cb141982be6389d56c3d7bc6ca0b3979933ddb9aa4\": container with ID starting with da921af3e5e34a16996892cb141982be6389d56c3d7bc6ca0b3979933ddb9aa4 not found: ID does not exist" Dec 03 14:30:13 crc kubenswrapper[4805]: I1203 14:30:13.728684 4805 scope.go:117] "RemoveContainer" containerID="cce0b3eb1a191554e4667b6b4e39dee3d15e0713c7fd79b5fda0181e0d7cbeac" Dec 03 14:30:13 crc kubenswrapper[4805]: E1203 14:30:13.728963 4805 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"cce0b3eb1a191554e4667b6b4e39dee3d15e0713c7fd79b5fda0181e0d7cbeac\": container with ID starting with cce0b3eb1a191554e4667b6b4e39dee3d15e0713c7fd79b5fda0181e0d7cbeac not found: ID does not exist" containerID="cce0b3eb1a191554e4667b6b4e39dee3d15e0713c7fd79b5fda0181e0d7cbeac" Dec 03 14:30:13 crc kubenswrapper[4805]: I1203 14:30:13.728987 4805 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"cce0b3eb1a191554e4667b6b4e39dee3d15e0713c7fd79b5fda0181e0d7cbeac"} err="failed to get container status \"cce0b3eb1a191554e4667b6b4e39dee3d15e0713c7fd79b5fda0181e0d7cbeac\": rpc error: code = NotFound desc = could not find container \"cce0b3eb1a191554e4667b6b4e39dee3d15e0713c7fd79b5fda0181e0d7cbeac\": container with ID starting with cce0b3eb1a191554e4667b6b4e39dee3d15e0713c7fd79b5fda0181e0d7cbeac not found: ID does not exist" Dec 03 14:30:13 crc kubenswrapper[4805]: I1203 14:30:13.729003 4805 scope.go:117] "RemoveContainer" containerID="da921af3e5e34a16996892cb141982be6389d56c3d7bc6ca0b3979933ddb9aa4" Dec 03 14:30:13 crc kubenswrapper[4805]: I1203 14:30:13.729215 4805 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"da921af3e5e34a16996892cb141982be6389d56c3d7bc6ca0b3979933ddb9aa4"} err="failed to get container status \"da921af3e5e34a16996892cb141982be6389d56c3d7bc6ca0b3979933ddb9aa4\": rpc error: code = NotFound desc = could not find container \"da921af3e5e34a16996892cb141982be6389d56c3d7bc6ca0b3979933ddb9aa4\": container with ID starting with da921af3e5e34a16996892cb141982be6389d56c3d7bc6ca0b3979933ddb9aa4 not found: ID does not exist" Dec 03 14:30:13 crc kubenswrapper[4805]: I1203 14:30:13.729234 4805 scope.go:117] "RemoveContainer" containerID="cce0b3eb1a191554e4667b6b4e39dee3d15e0713c7fd79b5fda0181e0d7cbeac" Dec 03 14:30:13 crc kubenswrapper[4805]: I1203 14:30:13.729606 4805 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"cce0b3eb1a191554e4667b6b4e39dee3d15e0713c7fd79b5fda0181e0d7cbeac"} err="failed to get container status \"cce0b3eb1a191554e4667b6b4e39dee3d15e0713c7fd79b5fda0181e0d7cbeac\": rpc error: code = NotFound desc = could not find container \"cce0b3eb1a191554e4667b6b4e39dee3d15e0713c7fd79b5fda0181e0d7cbeac\": container with ID starting with cce0b3eb1a191554e4667b6b4e39dee3d15e0713c7fd79b5fda0181e0d7cbeac not found: ID does not exist" Dec 03 14:30:13 crc kubenswrapper[4805]: I1203 14:30:13.787788 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9ed6dfc0-82db-4d86-ad92-cfe94fa370a2-scripts\") pod \"glance-default-external-api-0\" (UID: \"9ed6dfc0-82db-4d86-ad92-cfe94fa370a2\") " pod="openstack/glance-default-external-api-0" Dec 03 14:30:13 crc kubenswrapper[4805]: I1203 14:30:13.787870 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-c5677\" (UniqueName: \"kubernetes.io/projected/9ed6dfc0-82db-4d86-ad92-cfe94fa370a2-kube-api-access-c5677\") pod \"glance-default-external-api-0\" (UID: \"9ed6dfc0-82db-4d86-ad92-cfe94fa370a2\") " pod="openstack/glance-default-external-api-0" Dec 03 14:30:13 crc kubenswrapper[4805]: I1203 14:30:13.787903 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"glance-default-external-api-0\" (UID: \"9ed6dfc0-82db-4d86-ad92-cfe94fa370a2\") " pod="openstack/glance-default-external-api-0" Dec 03 14:30:13 crc kubenswrapper[4805]: I1203 14:30:13.787946 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/9ed6dfc0-82db-4d86-ad92-cfe94fa370a2-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"9ed6dfc0-82db-4d86-ad92-cfe94fa370a2\") " pod="openstack/glance-default-external-api-0" Dec 03 14:30:13 crc kubenswrapper[4805]: I1203 14:30:13.787966 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/9ed6dfc0-82db-4d86-ad92-cfe94fa370a2-logs\") pod \"glance-default-external-api-0\" (UID: \"9ed6dfc0-82db-4d86-ad92-cfe94fa370a2\") " pod="openstack/glance-default-external-api-0" Dec 03 14:30:13 crc kubenswrapper[4805]: I1203 14:30:13.787983 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9ed6dfc0-82db-4d86-ad92-cfe94fa370a2-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"9ed6dfc0-82db-4d86-ad92-cfe94fa370a2\") " pod="openstack/glance-default-external-api-0" Dec 03 14:30:13 crc kubenswrapper[4805]: I1203 14:30:13.788032 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/9ed6dfc0-82db-4d86-ad92-cfe94fa370a2-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"9ed6dfc0-82db-4d86-ad92-cfe94fa370a2\") " pod="openstack/glance-default-external-api-0" Dec 03 14:30:13 crc kubenswrapper[4805]: I1203 14:30:13.788050 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9ed6dfc0-82db-4d86-ad92-cfe94fa370a2-config-data\") pod \"glance-default-external-api-0\" (UID: \"9ed6dfc0-82db-4d86-ad92-cfe94fa370a2\") " pod="openstack/glance-default-external-api-0" Dec 03 14:30:13 crc kubenswrapper[4805]: I1203 14:30:13.788160 4805 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"glance-default-external-api-0\" (UID: \"9ed6dfc0-82db-4d86-ad92-cfe94fa370a2\") device mount path \"/mnt/openstack/pv07\"" pod="openstack/glance-default-external-api-0" Dec 03 14:30:13 crc kubenswrapper[4805]: I1203 14:30:13.789401 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/9ed6dfc0-82db-4d86-ad92-cfe94fa370a2-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"9ed6dfc0-82db-4d86-ad92-cfe94fa370a2\") " pod="openstack/glance-default-external-api-0" Dec 03 14:30:13 crc kubenswrapper[4805]: I1203 14:30:13.791681 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/9ed6dfc0-82db-4d86-ad92-cfe94fa370a2-logs\") pod \"glance-default-external-api-0\" (UID: \"9ed6dfc0-82db-4d86-ad92-cfe94fa370a2\") " pod="openstack/glance-default-external-api-0" Dec 03 14:30:13 crc kubenswrapper[4805]: I1203 14:30:13.793316 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9ed6dfc0-82db-4d86-ad92-cfe94fa370a2-scripts\") pod \"glance-default-external-api-0\" (UID: \"9ed6dfc0-82db-4d86-ad92-cfe94fa370a2\") " pod="openstack/glance-default-external-api-0" Dec 03 14:30:13 crc kubenswrapper[4805]: I1203 14:30:13.793692 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9ed6dfc0-82db-4d86-ad92-cfe94fa370a2-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"9ed6dfc0-82db-4d86-ad92-cfe94fa370a2\") " pod="openstack/glance-default-external-api-0" Dec 03 14:30:13 crc kubenswrapper[4805]: I1203 14:30:13.793856 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9ed6dfc0-82db-4d86-ad92-cfe94fa370a2-config-data\") pod \"glance-default-external-api-0\" (UID: \"9ed6dfc0-82db-4d86-ad92-cfe94fa370a2\") " pod="openstack/glance-default-external-api-0" Dec 03 14:30:13 crc kubenswrapper[4805]: I1203 14:30:13.795442 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/9ed6dfc0-82db-4d86-ad92-cfe94fa370a2-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"9ed6dfc0-82db-4d86-ad92-cfe94fa370a2\") " pod="openstack/glance-default-external-api-0" Dec 03 14:30:13 crc kubenswrapper[4805]: I1203 14:30:13.808564 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-c5677\" (UniqueName: \"kubernetes.io/projected/9ed6dfc0-82db-4d86-ad92-cfe94fa370a2-kube-api-access-c5677\") pod \"glance-default-external-api-0\" (UID: \"9ed6dfc0-82db-4d86-ad92-cfe94fa370a2\") " pod="openstack/glance-default-external-api-0" Dec 03 14:30:13 crc kubenswrapper[4805]: I1203 14:30:13.816333 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"glance-default-external-api-0\" (UID: \"9ed6dfc0-82db-4d86-ad92-cfe94fa370a2\") " pod="openstack/glance-default-external-api-0" Dec 03 14:30:13 crc kubenswrapper[4805]: I1203 14:30:13.896958 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Dec 03 14:30:13 crc kubenswrapper[4805]: I1203 14:30:13.918119 4805 patch_prober.go:28] interesting pod/machine-config-daemon-gskh4 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 03 14:30:13 crc kubenswrapper[4805]: I1203 14:30:13.918183 4805 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-gskh4" podUID="6098937f-e3f6-45e8-a647-4994a79cd711" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 03 14:30:14 crc kubenswrapper[4805]: I1203 14:30:14.496496 4805 generic.go:334] "Generic (PLEG): container finished" podID="a5c9bc3e-ad1c-428d-bd2e-0730f12b50a8" containerID="1398a8bb8b40a97438fe6d0fbcc36517488187b6d66a96af990ed301eeaf05fd" exitCode=0 Dec 03 14:30:14 crc kubenswrapper[4805]: I1203 14:30:14.496866 4805 generic.go:334] "Generic (PLEG): container finished" podID="a5c9bc3e-ad1c-428d-bd2e-0730f12b50a8" containerID="188ca93940f09e5d5f1a3e68344b1ff578bfc6d8a437936fc9d054967cfa3ed2" exitCode=143 Dec 03 14:30:14 crc kubenswrapper[4805]: I1203 14:30:14.496582 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"a5c9bc3e-ad1c-428d-bd2e-0730f12b50a8","Type":"ContainerDied","Data":"1398a8bb8b40a97438fe6d0fbcc36517488187b6d66a96af990ed301eeaf05fd"} Dec 03 14:30:14 crc kubenswrapper[4805]: I1203 14:30:14.496953 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"a5c9bc3e-ad1c-428d-bd2e-0730f12b50a8","Type":"ContainerDied","Data":"188ca93940f09e5d5f1a3e68344b1ff578bfc6d8a437936fc9d054967cfa3ed2"} Dec 03 14:30:14 crc kubenswrapper[4805]: I1203 14:30:14.709747 4805 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="10156ee9-4b49-4912-aece-1949d1cc3318" path="/var/lib/kubelet/pods/10156ee9-4b49-4912-aece-1949d1cc3318/volumes" Dec 03 14:30:15 crc kubenswrapper[4805]: I1203 14:30:15.510189 4805 generic.go:334] "Generic (PLEG): container finished" podID="16dbab06-8ff8-4a5a-9fe8-4967c6838a4a" containerID="794a23f04151f6d0d061c884ab2a4ed2e4ba2f4abbcb5f8271c84b63773854dc" exitCode=0 Dec 03 14:30:15 crc kubenswrapper[4805]: I1203 14:30:15.510252 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-qk4m4" event={"ID":"16dbab06-8ff8-4a5a-9fe8-4967c6838a4a","Type":"ContainerDied","Data":"794a23f04151f6d0d061c884ab2a4ed2e4ba2f4abbcb5f8271c84b63773854dc"} Dec 03 14:30:16 crc kubenswrapper[4805]: I1203 14:30:16.247353 4805 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Dec 03 14:30:16 crc kubenswrapper[4805]: I1203 14:30:16.331518 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a5c9bc3e-ad1c-428d-bd2e-0730f12b50a8-scripts\") pod \"a5c9bc3e-ad1c-428d-bd2e-0730f12b50a8\" (UID: \"a5c9bc3e-ad1c-428d-bd2e-0730f12b50a8\") " Dec 03 14:30:16 crc kubenswrapper[4805]: I1203 14:30:16.331639 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a5c9bc3e-ad1c-428d-bd2e-0730f12b50a8-combined-ca-bundle\") pod \"a5c9bc3e-ad1c-428d-bd2e-0730f12b50a8\" (UID: \"a5c9bc3e-ad1c-428d-bd2e-0730f12b50a8\") " Dec 03 14:30:16 crc kubenswrapper[4805]: I1203 14:30:16.331697 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/a5c9bc3e-ad1c-428d-bd2e-0730f12b50a8-httpd-run\") pod \"a5c9bc3e-ad1c-428d-bd2e-0730f12b50a8\" (UID: \"a5c9bc3e-ad1c-428d-bd2e-0730f12b50a8\") " Dec 03 14:30:16 crc kubenswrapper[4805]: I1203 14:30:16.331737 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"a5c9bc3e-ad1c-428d-bd2e-0730f12b50a8\" (UID: \"a5c9bc3e-ad1c-428d-bd2e-0730f12b50a8\") " Dec 03 14:30:16 crc kubenswrapper[4805]: I1203 14:30:16.331790 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x88ts\" (UniqueName: \"kubernetes.io/projected/a5c9bc3e-ad1c-428d-bd2e-0730f12b50a8-kube-api-access-x88ts\") pod \"a5c9bc3e-ad1c-428d-bd2e-0730f12b50a8\" (UID: \"a5c9bc3e-ad1c-428d-bd2e-0730f12b50a8\") " Dec 03 14:30:16 crc kubenswrapper[4805]: I1203 14:30:16.331829 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/a5c9bc3e-ad1c-428d-bd2e-0730f12b50a8-internal-tls-certs\") pod \"a5c9bc3e-ad1c-428d-bd2e-0730f12b50a8\" (UID: \"a5c9bc3e-ad1c-428d-bd2e-0730f12b50a8\") " Dec 03 14:30:16 crc kubenswrapper[4805]: I1203 14:30:16.331940 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a5c9bc3e-ad1c-428d-bd2e-0730f12b50a8-logs\") pod \"a5c9bc3e-ad1c-428d-bd2e-0730f12b50a8\" (UID: \"a5c9bc3e-ad1c-428d-bd2e-0730f12b50a8\") " Dec 03 14:30:16 crc kubenswrapper[4805]: I1203 14:30:16.331989 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a5c9bc3e-ad1c-428d-bd2e-0730f12b50a8-config-data\") pod \"a5c9bc3e-ad1c-428d-bd2e-0730f12b50a8\" (UID: \"a5c9bc3e-ad1c-428d-bd2e-0730f12b50a8\") " Dec 03 14:30:16 crc kubenswrapper[4805]: I1203 14:30:16.332233 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a5c9bc3e-ad1c-428d-bd2e-0730f12b50a8-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "a5c9bc3e-ad1c-428d-bd2e-0730f12b50a8" (UID: "a5c9bc3e-ad1c-428d-bd2e-0730f12b50a8"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 14:30:16 crc kubenswrapper[4805]: I1203 14:30:16.332331 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a5c9bc3e-ad1c-428d-bd2e-0730f12b50a8-logs" (OuterVolumeSpecName: "logs") pod "a5c9bc3e-ad1c-428d-bd2e-0730f12b50a8" (UID: "a5c9bc3e-ad1c-428d-bd2e-0730f12b50a8"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 14:30:16 crc kubenswrapper[4805]: I1203 14:30:16.332414 4805 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/a5c9bc3e-ad1c-428d-bd2e-0730f12b50a8-httpd-run\") on node \"crc\" DevicePath \"\"" Dec 03 14:30:16 crc kubenswrapper[4805]: I1203 14:30:16.353025 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a5c9bc3e-ad1c-428d-bd2e-0730f12b50a8-scripts" (OuterVolumeSpecName: "scripts") pod "a5c9bc3e-ad1c-428d-bd2e-0730f12b50a8" (UID: "a5c9bc3e-ad1c-428d-bd2e-0730f12b50a8"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 14:30:16 crc kubenswrapper[4805]: I1203 14:30:16.353109 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage11-crc" (OuterVolumeSpecName: "glance") pod "a5c9bc3e-ad1c-428d-bd2e-0730f12b50a8" (UID: "a5c9bc3e-ad1c-428d-bd2e-0730f12b50a8"). InnerVolumeSpecName "local-storage11-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Dec 03 14:30:16 crc kubenswrapper[4805]: I1203 14:30:16.356418 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a5c9bc3e-ad1c-428d-bd2e-0730f12b50a8-kube-api-access-x88ts" (OuterVolumeSpecName: "kube-api-access-x88ts") pod "a5c9bc3e-ad1c-428d-bd2e-0730f12b50a8" (UID: "a5c9bc3e-ad1c-428d-bd2e-0730f12b50a8"). InnerVolumeSpecName "kube-api-access-x88ts". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 14:30:16 crc kubenswrapper[4805]: I1203 14:30:16.363254 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a5c9bc3e-ad1c-428d-bd2e-0730f12b50a8-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "a5c9bc3e-ad1c-428d-bd2e-0730f12b50a8" (UID: "a5c9bc3e-ad1c-428d-bd2e-0730f12b50a8"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 14:30:16 crc kubenswrapper[4805]: I1203 14:30:16.385440 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a5c9bc3e-ad1c-428d-bd2e-0730f12b50a8-config-data" (OuterVolumeSpecName: "config-data") pod "a5c9bc3e-ad1c-428d-bd2e-0730f12b50a8" (UID: "a5c9bc3e-ad1c-428d-bd2e-0730f12b50a8"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 14:30:16 crc kubenswrapper[4805]: I1203 14:30:16.408777 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a5c9bc3e-ad1c-428d-bd2e-0730f12b50a8-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "a5c9bc3e-ad1c-428d-bd2e-0730f12b50a8" (UID: "a5c9bc3e-ad1c-428d-bd2e-0730f12b50a8"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 14:30:16 crc kubenswrapper[4805]: I1203 14:30:16.434191 4805 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a5c9bc3e-ad1c-428d-bd2e-0730f12b50a8-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 14:30:16 crc kubenswrapper[4805]: I1203 14:30:16.434254 4805 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") on node \"crc\" " Dec 03 14:30:16 crc kubenswrapper[4805]: I1203 14:30:16.434268 4805 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x88ts\" (UniqueName: \"kubernetes.io/projected/a5c9bc3e-ad1c-428d-bd2e-0730f12b50a8-kube-api-access-x88ts\") on node \"crc\" DevicePath \"\"" Dec 03 14:30:16 crc kubenswrapper[4805]: I1203 14:30:16.434280 4805 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/a5c9bc3e-ad1c-428d-bd2e-0730f12b50a8-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 03 14:30:16 crc kubenswrapper[4805]: I1203 14:30:16.434290 4805 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a5c9bc3e-ad1c-428d-bd2e-0730f12b50a8-logs\") on node \"crc\" DevicePath \"\"" Dec 03 14:30:16 crc kubenswrapper[4805]: I1203 14:30:16.434298 4805 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a5c9bc3e-ad1c-428d-bd2e-0730f12b50a8-config-data\") on node \"crc\" DevicePath \"\"" Dec 03 14:30:16 crc kubenswrapper[4805]: I1203 14:30:16.434306 4805 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a5c9bc3e-ad1c-428d-bd2e-0730f12b50a8-scripts\") on node \"crc\" DevicePath \"\"" Dec 03 14:30:16 crc kubenswrapper[4805]: I1203 14:30:16.454698 4805 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage11-crc" (UniqueName: "kubernetes.io/local-volume/local-storage11-crc") on node "crc" Dec 03 14:30:16 crc kubenswrapper[4805]: I1203 14:30:16.532337 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"a5c9bc3e-ad1c-428d-bd2e-0730f12b50a8","Type":"ContainerDied","Data":"10dfc8c10ea0cb8d67b76038eb994ba9d3cced592669a9182639acd5d197ae45"} Dec 03 14:30:16 crc kubenswrapper[4805]: I1203 14:30:16.532404 4805 scope.go:117] "RemoveContainer" containerID="1398a8bb8b40a97438fe6d0fbcc36517488187b6d66a96af990ed301eeaf05fd" Dec 03 14:30:16 crc kubenswrapper[4805]: I1203 14:30:16.532505 4805 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Dec 03 14:30:16 crc kubenswrapper[4805]: I1203 14:30:16.535878 4805 reconciler_common.go:293] "Volume detached for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") on node \"crc\" DevicePath \"\"" Dec 03 14:30:16 crc kubenswrapper[4805]: I1203 14:30:16.570095 4805 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 03 14:30:16 crc kubenswrapper[4805]: I1203 14:30:16.583746 4805 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 03 14:30:16 crc kubenswrapper[4805]: I1203 14:30:16.616470 4805 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 03 14:30:16 crc kubenswrapper[4805]: E1203 14:30:16.618097 4805 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a5c9bc3e-ad1c-428d-bd2e-0730f12b50a8" containerName="glance-httpd" Dec 03 14:30:16 crc kubenswrapper[4805]: I1203 14:30:16.618163 4805 state_mem.go:107] "Deleted CPUSet assignment" podUID="a5c9bc3e-ad1c-428d-bd2e-0730f12b50a8" containerName="glance-httpd" Dec 03 14:30:16 crc kubenswrapper[4805]: E1203 14:30:16.618208 4805 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a5c9bc3e-ad1c-428d-bd2e-0730f12b50a8" containerName="glance-log" Dec 03 14:30:16 crc kubenswrapper[4805]: I1203 14:30:16.618218 4805 state_mem.go:107] "Deleted CPUSet assignment" podUID="a5c9bc3e-ad1c-428d-bd2e-0730f12b50a8" containerName="glance-log" Dec 03 14:30:16 crc kubenswrapper[4805]: I1203 14:30:16.619313 4805 memory_manager.go:354] "RemoveStaleState removing state" podUID="a5c9bc3e-ad1c-428d-bd2e-0730f12b50a8" containerName="glance-log" Dec 03 14:30:16 crc kubenswrapper[4805]: I1203 14:30:16.619358 4805 memory_manager.go:354] "RemoveStaleState removing state" podUID="a5c9bc3e-ad1c-428d-bd2e-0730f12b50a8" containerName="glance-httpd" Dec 03 14:30:16 crc kubenswrapper[4805]: I1203 14:30:16.622041 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Dec 03 14:30:16 crc kubenswrapper[4805]: I1203 14:30:16.623910 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-internal-svc" Dec 03 14:30:16 crc kubenswrapper[4805]: I1203 14:30:16.626563 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-internal-config-data" Dec 03 14:30:16 crc kubenswrapper[4805]: I1203 14:30:16.657109 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 03 14:30:16 crc kubenswrapper[4805]: I1203 14:30:16.709659 4805 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a5c9bc3e-ad1c-428d-bd2e-0730f12b50a8" path="/var/lib/kubelet/pods/a5c9bc3e-ad1c-428d-bd2e-0730f12b50a8/volumes" Dec 03 14:30:16 crc kubenswrapper[4805]: I1203 14:30:16.751492 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"glance-default-internal-api-0\" (UID: \"10877f19-dd2a-4eb2-99e0-2d0cf82d5ad5\") " pod="openstack/glance-default-internal-api-0" Dec 03 14:30:16 crc kubenswrapper[4805]: I1203 14:30:16.751558 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mf6bl\" (UniqueName: \"kubernetes.io/projected/10877f19-dd2a-4eb2-99e0-2d0cf82d5ad5-kube-api-access-mf6bl\") pod \"glance-default-internal-api-0\" (UID: \"10877f19-dd2a-4eb2-99e0-2d0cf82d5ad5\") " pod="openstack/glance-default-internal-api-0" Dec 03 14:30:16 crc kubenswrapper[4805]: I1203 14:30:16.751621 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/10877f19-dd2a-4eb2-99e0-2d0cf82d5ad5-logs\") pod \"glance-default-internal-api-0\" (UID: \"10877f19-dd2a-4eb2-99e0-2d0cf82d5ad5\") " pod="openstack/glance-default-internal-api-0" Dec 03 14:30:16 crc kubenswrapper[4805]: I1203 14:30:16.751642 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/10877f19-dd2a-4eb2-99e0-2d0cf82d5ad5-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"10877f19-dd2a-4eb2-99e0-2d0cf82d5ad5\") " pod="openstack/glance-default-internal-api-0" Dec 03 14:30:16 crc kubenswrapper[4805]: I1203 14:30:16.751675 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/10877f19-dd2a-4eb2-99e0-2d0cf82d5ad5-scripts\") pod \"glance-default-internal-api-0\" (UID: \"10877f19-dd2a-4eb2-99e0-2d0cf82d5ad5\") " pod="openstack/glance-default-internal-api-0" Dec 03 14:30:16 crc kubenswrapper[4805]: I1203 14:30:16.751720 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/10877f19-dd2a-4eb2-99e0-2d0cf82d5ad5-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"10877f19-dd2a-4eb2-99e0-2d0cf82d5ad5\") " pod="openstack/glance-default-internal-api-0" Dec 03 14:30:16 crc kubenswrapper[4805]: I1203 14:30:16.751745 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/10877f19-dd2a-4eb2-99e0-2d0cf82d5ad5-config-data\") pod \"glance-default-internal-api-0\" (UID: \"10877f19-dd2a-4eb2-99e0-2d0cf82d5ad5\") " pod="openstack/glance-default-internal-api-0" Dec 03 14:30:16 crc kubenswrapper[4805]: I1203 14:30:16.751765 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/10877f19-dd2a-4eb2-99e0-2d0cf82d5ad5-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"10877f19-dd2a-4eb2-99e0-2d0cf82d5ad5\") " pod="openstack/glance-default-internal-api-0" Dec 03 14:30:16 crc kubenswrapper[4805]: I1203 14:30:16.853481 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mf6bl\" (UniqueName: \"kubernetes.io/projected/10877f19-dd2a-4eb2-99e0-2d0cf82d5ad5-kube-api-access-mf6bl\") pod \"glance-default-internal-api-0\" (UID: \"10877f19-dd2a-4eb2-99e0-2d0cf82d5ad5\") " pod="openstack/glance-default-internal-api-0" Dec 03 14:30:16 crc kubenswrapper[4805]: I1203 14:30:16.853581 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/10877f19-dd2a-4eb2-99e0-2d0cf82d5ad5-logs\") pod \"glance-default-internal-api-0\" (UID: \"10877f19-dd2a-4eb2-99e0-2d0cf82d5ad5\") " pod="openstack/glance-default-internal-api-0" Dec 03 14:30:16 crc kubenswrapper[4805]: I1203 14:30:16.853599 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/10877f19-dd2a-4eb2-99e0-2d0cf82d5ad5-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"10877f19-dd2a-4eb2-99e0-2d0cf82d5ad5\") " pod="openstack/glance-default-internal-api-0" Dec 03 14:30:16 crc kubenswrapper[4805]: I1203 14:30:16.853627 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/10877f19-dd2a-4eb2-99e0-2d0cf82d5ad5-scripts\") pod \"glance-default-internal-api-0\" (UID: \"10877f19-dd2a-4eb2-99e0-2d0cf82d5ad5\") " pod="openstack/glance-default-internal-api-0" Dec 03 14:30:16 crc kubenswrapper[4805]: I1203 14:30:16.853673 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/10877f19-dd2a-4eb2-99e0-2d0cf82d5ad5-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"10877f19-dd2a-4eb2-99e0-2d0cf82d5ad5\") " pod="openstack/glance-default-internal-api-0" Dec 03 14:30:16 crc kubenswrapper[4805]: I1203 14:30:16.853720 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/10877f19-dd2a-4eb2-99e0-2d0cf82d5ad5-config-data\") pod \"glance-default-internal-api-0\" (UID: \"10877f19-dd2a-4eb2-99e0-2d0cf82d5ad5\") " pod="openstack/glance-default-internal-api-0" Dec 03 14:30:16 crc kubenswrapper[4805]: I1203 14:30:16.853748 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/10877f19-dd2a-4eb2-99e0-2d0cf82d5ad5-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"10877f19-dd2a-4eb2-99e0-2d0cf82d5ad5\") " pod="openstack/glance-default-internal-api-0" Dec 03 14:30:16 crc kubenswrapper[4805]: I1203 14:30:16.853888 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"glance-default-internal-api-0\" (UID: \"10877f19-dd2a-4eb2-99e0-2d0cf82d5ad5\") " pod="openstack/glance-default-internal-api-0" Dec 03 14:30:16 crc kubenswrapper[4805]: I1203 14:30:16.854559 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/10877f19-dd2a-4eb2-99e0-2d0cf82d5ad5-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"10877f19-dd2a-4eb2-99e0-2d0cf82d5ad5\") " pod="openstack/glance-default-internal-api-0" Dec 03 14:30:16 crc kubenswrapper[4805]: I1203 14:30:16.854808 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/10877f19-dd2a-4eb2-99e0-2d0cf82d5ad5-logs\") pod \"glance-default-internal-api-0\" (UID: \"10877f19-dd2a-4eb2-99e0-2d0cf82d5ad5\") " pod="openstack/glance-default-internal-api-0" Dec 03 14:30:16 crc kubenswrapper[4805]: I1203 14:30:16.855590 4805 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"glance-default-internal-api-0\" (UID: \"10877f19-dd2a-4eb2-99e0-2d0cf82d5ad5\") device mount path \"/mnt/openstack/pv11\"" pod="openstack/glance-default-internal-api-0" Dec 03 14:30:16 crc kubenswrapper[4805]: I1203 14:30:16.859125 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/10877f19-dd2a-4eb2-99e0-2d0cf82d5ad5-scripts\") pod \"glance-default-internal-api-0\" (UID: \"10877f19-dd2a-4eb2-99e0-2d0cf82d5ad5\") " pod="openstack/glance-default-internal-api-0" Dec 03 14:30:16 crc kubenswrapper[4805]: I1203 14:30:16.859533 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/10877f19-dd2a-4eb2-99e0-2d0cf82d5ad5-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"10877f19-dd2a-4eb2-99e0-2d0cf82d5ad5\") " pod="openstack/glance-default-internal-api-0" Dec 03 14:30:16 crc kubenswrapper[4805]: I1203 14:30:16.859590 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/10877f19-dd2a-4eb2-99e0-2d0cf82d5ad5-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"10877f19-dd2a-4eb2-99e0-2d0cf82d5ad5\") " pod="openstack/glance-default-internal-api-0" Dec 03 14:30:16 crc kubenswrapper[4805]: I1203 14:30:16.861953 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/10877f19-dd2a-4eb2-99e0-2d0cf82d5ad5-config-data\") pod \"glance-default-internal-api-0\" (UID: \"10877f19-dd2a-4eb2-99e0-2d0cf82d5ad5\") " pod="openstack/glance-default-internal-api-0" Dec 03 14:30:16 crc kubenswrapper[4805]: I1203 14:30:16.874528 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mf6bl\" (UniqueName: \"kubernetes.io/projected/10877f19-dd2a-4eb2-99e0-2d0cf82d5ad5-kube-api-access-mf6bl\") pod \"glance-default-internal-api-0\" (UID: \"10877f19-dd2a-4eb2-99e0-2d0cf82d5ad5\") " pod="openstack/glance-default-internal-api-0" Dec 03 14:30:16 crc kubenswrapper[4805]: I1203 14:30:16.887205 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"glance-default-internal-api-0\" (UID: \"10877f19-dd2a-4eb2-99e0-2d0cf82d5ad5\") " pod="openstack/glance-default-internal-api-0" Dec 03 14:30:16 crc kubenswrapper[4805]: I1203 14:30:16.950067 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Dec 03 14:30:18 crc kubenswrapper[4805]: I1203 14:30:18.674070 4805 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-8b5c85b87-bvb8w" Dec 03 14:30:18 crc kubenswrapper[4805]: I1203 14:30:18.747276 4805 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-7ff5475cc9-mbdbt"] Dec 03 14:30:18 crc kubenswrapper[4805]: I1203 14:30:18.747790 4805 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-7ff5475cc9-mbdbt" podUID="1ca67122-920a-43a6-a434-82ec573d0160" containerName="dnsmasq-dns" containerID="cri-o://f9c31ba4a3601f19c3ce752a2e09ba3a90a39dfe55c8ca130f6564acd0157752" gracePeriod=10 Dec 03 14:30:19 crc kubenswrapper[4805]: E1203 14:30:19.254313 4805 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod1ca67122_920a_43a6_a434_82ec573d0160.slice/crio-conmon-f9c31ba4a3601f19c3ce752a2e09ba3a90a39dfe55c8ca130f6564acd0157752.scope\": RecentStats: unable to find data in memory cache]" Dec 03 14:30:19 crc kubenswrapper[4805]: I1203 14:30:19.510458 4805 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-7ff5475cc9-mbdbt" podUID="1ca67122-920a-43a6-a434-82ec573d0160" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.125:5353: connect: connection refused" Dec 03 14:30:19 crc kubenswrapper[4805]: I1203 14:30:19.560657 4805 generic.go:334] "Generic (PLEG): container finished" podID="1ca67122-920a-43a6-a434-82ec573d0160" containerID="f9c31ba4a3601f19c3ce752a2e09ba3a90a39dfe55c8ca130f6564acd0157752" exitCode=0 Dec 03 14:30:19 crc kubenswrapper[4805]: I1203 14:30:19.560931 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7ff5475cc9-mbdbt" event={"ID":"1ca67122-920a-43a6-a434-82ec573d0160","Type":"ContainerDied","Data":"f9c31ba4a3601f19c3ce752a2e09ba3a90a39dfe55c8ca130f6564acd0157752"} Dec 03 14:30:24 crc kubenswrapper[4805]: E1203 14:30:24.942640 4805 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-heat-engine:current-podified" Dec 03 14:30:24 crc kubenswrapper[4805]: E1203 14:30:24.943472 4805 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:heat-db-sync,Image:quay.io/podified-antelope-centos9/openstack-heat-engine:current-podified,Command:[/bin/bash],Args:[-c /usr/bin/heat-manage --config-dir /etc/heat/heat.conf.d db_sync],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:KOLLA_BOOTSTRAP,Value:true,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config-data,ReadOnly:true,MountPath:/etc/heat/heat.conf.d/00-default.conf,SubPath:00-default.conf,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:false,MountPath:/etc/heat/heat.conf.d/01-custom.conf,SubPath:01-custom.conf,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/etc/my.cnf,SubPath:my.cnf,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:combined-ca-bundle,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-zh8k9,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*42418,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:*42418,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod heat-db-sync-jvqmk_openstack(14c327e3-e880-4b49-b366-364a0395cc55): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 03 14:30:24 crc kubenswrapper[4805]: E1203 14:30:24.944693 4805 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"heat-db-sync\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/heat-db-sync-jvqmk" podUID="14c327e3-e880-4b49-b366-364a0395cc55" Dec 03 14:30:25 crc kubenswrapper[4805]: E1203 14:30:25.613670 4805 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"heat-db-sync\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-heat-engine:current-podified\\\"\"" pod="openstack/heat-db-sync-jvqmk" podUID="14c327e3-e880-4b49-b366-364a0395cc55" Dec 03 14:30:27 crc kubenswrapper[4805]: I1203 14:30:27.100648 4805 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-qk4m4" Dec 03 14:30:27 crc kubenswrapper[4805]: I1203 14:30:27.210194 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/16dbab06-8ff8-4a5a-9fe8-4967c6838a4a-combined-ca-bundle\") pod \"16dbab06-8ff8-4a5a-9fe8-4967c6838a4a\" (UID: \"16dbab06-8ff8-4a5a-9fe8-4967c6838a4a\") " Dec 03 14:30:27 crc kubenswrapper[4805]: I1203 14:30:27.210341 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kbxbj\" (UniqueName: \"kubernetes.io/projected/16dbab06-8ff8-4a5a-9fe8-4967c6838a4a-kube-api-access-kbxbj\") pod \"16dbab06-8ff8-4a5a-9fe8-4967c6838a4a\" (UID: \"16dbab06-8ff8-4a5a-9fe8-4967c6838a4a\") " Dec 03 14:30:27 crc kubenswrapper[4805]: I1203 14:30:27.210385 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/16dbab06-8ff8-4a5a-9fe8-4967c6838a4a-scripts\") pod \"16dbab06-8ff8-4a5a-9fe8-4967c6838a4a\" (UID: \"16dbab06-8ff8-4a5a-9fe8-4967c6838a4a\") " Dec 03 14:30:27 crc kubenswrapper[4805]: I1203 14:30:27.210524 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/16dbab06-8ff8-4a5a-9fe8-4967c6838a4a-fernet-keys\") pod \"16dbab06-8ff8-4a5a-9fe8-4967c6838a4a\" (UID: \"16dbab06-8ff8-4a5a-9fe8-4967c6838a4a\") " Dec 03 14:30:27 crc kubenswrapper[4805]: I1203 14:30:27.211428 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/16dbab06-8ff8-4a5a-9fe8-4967c6838a4a-config-data\") pod \"16dbab06-8ff8-4a5a-9fe8-4967c6838a4a\" (UID: \"16dbab06-8ff8-4a5a-9fe8-4967c6838a4a\") " Dec 03 14:30:27 crc kubenswrapper[4805]: I1203 14:30:27.211552 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/16dbab06-8ff8-4a5a-9fe8-4967c6838a4a-credential-keys\") pod \"16dbab06-8ff8-4a5a-9fe8-4967c6838a4a\" (UID: \"16dbab06-8ff8-4a5a-9fe8-4967c6838a4a\") " Dec 03 14:30:27 crc kubenswrapper[4805]: I1203 14:30:27.217377 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/16dbab06-8ff8-4a5a-9fe8-4967c6838a4a-scripts" (OuterVolumeSpecName: "scripts") pod "16dbab06-8ff8-4a5a-9fe8-4967c6838a4a" (UID: "16dbab06-8ff8-4a5a-9fe8-4967c6838a4a"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 14:30:27 crc kubenswrapper[4805]: I1203 14:30:27.217986 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/16dbab06-8ff8-4a5a-9fe8-4967c6838a4a-kube-api-access-kbxbj" (OuterVolumeSpecName: "kube-api-access-kbxbj") pod "16dbab06-8ff8-4a5a-9fe8-4967c6838a4a" (UID: "16dbab06-8ff8-4a5a-9fe8-4967c6838a4a"). InnerVolumeSpecName "kube-api-access-kbxbj". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 14:30:27 crc kubenswrapper[4805]: I1203 14:30:27.218103 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/16dbab06-8ff8-4a5a-9fe8-4967c6838a4a-credential-keys" (OuterVolumeSpecName: "credential-keys") pod "16dbab06-8ff8-4a5a-9fe8-4967c6838a4a" (UID: "16dbab06-8ff8-4a5a-9fe8-4967c6838a4a"). InnerVolumeSpecName "credential-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 14:30:27 crc kubenswrapper[4805]: I1203 14:30:27.228876 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/16dbab06-8ff8-4a5a-9fe8-4967c6838a4a-fernet-keys" (OuterVolumeSpecName: "fernet-keys") pod "16dbab06-8ff8-4a5a-9fe8-4967c6838a4a" (UID: "16dbab06-8ff8-4a5a-9fe8-4967c6838a4a"). InnerVolumeSpecName "fernet-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 14:30:27 crc kubenswrapper[4805]: I1203 14:30:27.242887 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/16dbab06-8ff8-4a5a-9fe8-4967c6838a4a-config-data" (OuterVolumeSpecName: "config-data") pod "16dbab06-8ff8-4a5a-9fe8-4967c6838a4a" (UID: "16dbab06-8ff8-4a5a-9fe8-4967c6838a4a"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 14:30:27 crc kubenswrapper[4805]: I1203 14:30:27.242919 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/16dbab06-8ff8-4a5a-9fe8-4967c6838a4a-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "16dbab06-8ff8-4a5a-9fe8-4967c6838a4a" (UID: "16dbab06-8ff8-4a5a-9fe8-4967c6838a4a"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 14:30:27 crc kubenswrapper[4805]: I1203 14:30:27.313674 4805 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/16dbab06-8ff8-4a5a-9fe8-4967c6838a4a-config-data\") on node \"crc\" DevicePath \"\"" Dec 03 14:30:27 crc kubenswrapper[4805]: I1203 14:30:27.313716 4805 reconciler_common.go:293] "Volume detached for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/16dbab06-8ff8-4a5a-9fe8-4967c6838a4a-credential-keys\") on node \"crc\" DevicePath \"\"" Dec 03 14:30:27 crc kubenswrapper[4805]: I1203 14:30:27.313726 4805 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/16dbab06-8ff8-4a5a-9fe8-4967c6838a4a-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 14:30:27 crc kubenswrapper[4805]: I1203 14:30:27.313736 4805 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kbxbj\" (UniqueName: \"kubernetes.io/projected/16dbab06-8ff8-4a5a-9fe8-4967c6838a4a-kube-api-access-kbxbj\") on node \"crc\" DevicePath \"\"" Dec 03 14:30:27 crc kubenswrapper[4805]: I1203 14:30:27.313746 4805 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/16dbab06-8ff8-4a5a-9fe8-4967c6838a4a-scripts\") on node \"crc\" DevicePath \"\"" Dec 03 14:30:27 crc kubenswrapper[4805]: I1203 14:30:27.313754 4805 reconciler_common.go:293] "Volume detached for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/16dbab06-8ff8-4a5a-9fe8-4967c6838a4a-fernet-keys\") on node \"crc\" DevicePath \"\"" Dec 03 14:30:27 crc kubenswrapper[4805]: I1203 14:30:27.628401 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-qk4m4" event={"ID":"16dbab06-8ff8-4a5a-9fe8-4967c6838a4a","Type":"ContainerDied","Data":"48d1faafacba3dd6eb3f10ec93a9f796e52e3079e0837e1e2ce22d0318284457"} Dec 03 14:30:27 crc kubenswrapper[4805]: I1203 14:30:27.628431 4805 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-qk4m4" Dec 03 14:30:27 crc kubenswrapper[4805]: I1203 14:30:27.628446 4805 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="48d1faafacba3dd6eb3f10ec93a9f796e52e3079e0837e1e2ce22d0318284457" Dec 03 14:30:28 crc kubenswrapper[4805]: I1203 14:30:28.182830 4805 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-bootstrap-qk4m4"] Dec 03 14:30:28 crc kubenswrapper[4805]: I1203 14:30:28.188924 4805 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-bootstrap-qk4m4"] Dec 03 14:30:28 crc kubenswrapper[4805]: I1203 14:30:28.303959 4805 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-bootstrap-74pdd"] Dec 03 14:30:28 crc kubenswrapper[4805]: E1203 14:30:28.304578 4805 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="16dbab06-8ff8-4a5a-9fe8-4967c6838a4a" containerName="keystone-bootstrap" Dec 03 14:30:28 crc kubenswrapper[4805]: I1203 14:30:28.304595 4805 state_mem.go:107] "Deleted CPUSet assignment" podUID="16dbab06-8ff8-4a5a-9fe8-4967c6838a4a" containerName="keystone-bootstrap" Dec 03 14:30:28 crc kubenswrapper[4805]: I1203 14:30:28.304991 4805 memory_manager.go:354] "RemoveStaleState removing state" podUID="16dbab06-8ff8-4a5a-9fe8-4967c6838a4a" containerName="keystone-bootstrap" Dec 03 14:30:28 crc kubenswrapper[4805]: I1203 14:30:28.305750 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-74pdd" Dec 03 14:30:28 crc kubenswrapper[4805]: I1203 14:30:28.310185 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-keystone-dockercfg-5hcz5" Dec 03 14:30:28 crc kubenswrapper[4805]: I1203 14:30:28.310437 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-config-data" Dec 03 14:30:28 crc kubenswrapper[4805]: I1203 14:30:28.310584 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone" Dec 03 14:30:28 crc kubenswrapper[4805]: I1203 14:30:28.310695 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-scripts" Dec 03 14:30:28 crc kubenswrapper[4805]: I1203 14:30:28.310852 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"osp-secret" Dec 03 14:30:28 crc kubenswrapper[4805]: I1203 14:30:28.313489 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-bootstrap-74pdd"] Dec 03 14:30:28 crc kubenswrapper[4805]: I1203 14:30:28.433450 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/b4e050ed-47f5-4d9f-a288-b6d2a7ab294a-credential-keys\") pod \"keystone-bootstrap-74pdd\" (UID: \"b4e050ed-47f5-4d9f-a288-b6d2a7ab294a\") " pod="openstack/keystone-bootstrap-74pdd" Dec 03 14:30:28 crc kubenswrapper[4805]: I1203 14:30:28.433531 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b4e050ed-47f5-4d9f-a288-b6d2a7ab294a-combined-ca-bundle\") pod \"keystone-bootstrap-74pdd\" (UID: \"b4e050ed-47f5-4d9f-a288-b6d2a7ab294a\") " pod="openstack/keystone-bootstrap-74pdd" Dec 03 14:30:28 crc kubenswrapper[4805]: I1203 14:30:28.433557 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b4e050ed-47f5-4d9f-a288-b6d2a7ab294a-scripts\") pod \"keystone-bootstrap-74pdd\" (UID: \"b4e050ed-47f5-4d9f-a288-b6d2a7ab294a\") " pod="openstack/keystone-bootstrap-74pdd" Dec 03 14:30:28 crc kubenswrapper[4805]: I1203 14:30:28.433622 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b4e050ed-47f5-4d9f-a288-b6d2a7ab294a-config-data\") pod \"keystone-bootstrap-74pdd\" (UID: \"b4e050ed-47f5-4d9f-a288-b6d2a7ab294a\") " pod="openstack/keystone-bootstrap-74pdd" Dec 03 14:30:28 crc kubenswrapper[4805]: I1203 14:30:28.433646 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-w7ppx\" (UniqueName: \"kubernetes.io/projected/b4e050ed-47f5-4d9f-a288-b6d2a7ab294a-kube-api-access-w7ppx\") pod \"keystone-bootstrap-74pdd\" (UID: \"b4e050ed-47f5-4d9f-a288-b6d2a7ab294a\") " pod="openstack/keystone-bootstrap-74pdd" Dec 03 14:30:28 crc kubenswrapper[4805]: I1203 14:30:28.433918 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/b4e050ed-47f5-4d9f-a288-b6d2a7ab294a-fernet-keys\") pod \"keystone-bootstrap-74pdd\" (UID: \"b4e050ed-47f5-4d9f-a288-b6d2a7ab294a\") " pod="openstack/keystone-bootstrap-74pdd" Dec 03 14:30:28 crc kubenswrapper[4805]: I1203 14:30:28.535775 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b4e050ed-47f5-4d9f-a288-b6d2a7ab294a-combined-ca-bundle\") pod \"keystone-bootstrap-74pdd\" (UID: \"b4e050ed-47f5-4d9f-a288-b6d2a7ab294a\") " pod="openstack/keystone-bootstrap-74pdd" Dec 03 14:30:28 crc kubenswrapper[4805]: I1203 14:30:28.535819 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b4e050ed-47f5-4d9f-a288-b6d2a7ab294a-scripts\") pod \"keystone-bootstrap-74pdd\" (UID: \"b4e050ed-47f5-4d9f-a288-b6d2a7ab294a\") " pod="openstack/keystone-bootstrap-74pdd" Dec 03 14:30:28 crc kubenswrapper[4805]: I1203 14:30:28.535908 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b4e050ed-47f5-4d9f-a288-b6d2a7ab294a-config-data\") pod \"keystone-bootstrap-74pdd\" (UID: \"b4e050ed-47f5-4d9f-a288-b6d2a7ab294a\") " pod="openstack/keystone-bootstrap-74pdd" Dec 03 14:30:28 crc kubenswrapper[4805]: I1203 14:30:28.535936 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-w7ppx\" (UniqueName: \"kubernetes.io/projected/b4e050ed-47f5-4d9f-a288-b6d2a7ab294a-kube-api-access-w7ppx\") pod \"keystone-bootstrap-74pdd\" (UID: \"b4e050ed-47f5-4d9f-a288-b6d2a7ab294a\") " pod="openstack/keystone-bootstrap-74pdd" Dec 03 14:30:28 crc kubenswrapper[4805]: I1203 14:30:28.536006 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/b4e050ed-47f5-4d9f-a288-b6d2a7ab294a-fernet-keys\") pod \"keystone-bootstrap-74pdd\" (UID: \"b4e050ed-47f5-4d9f-a288-b6d2a7ab294a\") " pod="openstack/keystone-bootstrap-74pdd" Dec 03 14:30:28 crc kubenswrapper[4805]: I1203 14:30:28.536059 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/b4e050ed-47f5-4d9f-a288-b6d2a7ab294a-credential-keys\") pod \"keystone-bootstrap-74pdd\" (UID: \"b4e050ed-47f5-4d9f-a288-b6d2a7ab294a\") " pod="openstack/keystone-bootstrap-74pdd" Dec 03 14:30:28 crc kubenswrapper[4805]: I1203 14:30:28.540681 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/b4e050ed-47f5-4d9f-a288-b6d2a7ab294a-credential-keys\") pod \"keystone-bootstrap-74pdd\" (UID: \"b4e050ed-47f5-4d9f-a288-b6d2a7ab294a\") " pod="openstack/keystone-bootstrap-74pdd" Dec 03 14:30:28 crc kubenswrapper[4805]: I1203 14:30:28.540723 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/b4e050ed-47f5-4d9f-a288-b6d2a7ab294a-fernet-keys\") pod \"keystone-bootstrap-74pdd\" (UID: \"b4e050ed-47f5-4d9f-a288-b6d2a7ab294a\") " pod="openstack/keystone-bootstrap-74pdd" Dec 03 14:30:28 crc kubenswrapper[4805]: I1203 14:30:28.541028 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b4e050ed-47f5-4d9f-a288-b6d2a7ab294a-config-data\") pod \"keystone-bootstrap-74pdd\" (UID: \"b4e050ed-47f5-4d9f-a288-b6d2a7ab294a\") " pod="openstack/keystone-bootstrap-74pdd" Dec 03 14:30:28 crc kubenswrapper[4805]: I1203 14:30:28.542753 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b4e050ed-47f5-4d9f-a288-b6d2a7ab294a-scripts\") pod \"keystone-bootstrap-74pdd\" (UID: \"b4e050ed-47f5-4d9f-a288-b6d2a7ab294a\") " pod="openstack/keystone-bootstrap-74pdd" Dec 03 14:30:28 crc kubenswrapper[4805]: I1203 14:30:28.543028 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b4e050ed-47f5-4d9f-a288-b6d2a7ab294a-combined-ca-bundle\") pod \"keystone-bootstrap-74pdd\" (UID: \"b4e050ed-47f5-4d9f-a288-b6d2a7ab294a\") " pod="openstack/keystone-bootstrap-74pdd" Dec 03 14:30:28 crc kubenswrapper[4805]: I1203 14:30:28.553028 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-w7ppx\" (UniqueName: \"kubernetes.io/projected/b4e050ed-47f5-4d9f-a288-b6d2a7ab294a-kube-api-access-w7ppx\") pod \"keystone-bootstrap-74pdd\" (UID: \"b4e050ed-47f5-4d9f-a288-b6d2a7ab294a\") " pod="openstack/keystone-bootstrap-74pdd" Dec 03 14:30:28 crc kubenswrapper[4805]: I1203 14:30:28.630200 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-74pdd" Dec 03 14:30:28 crc kubenswrapper[4805]: I1203 14:30:28.712300 4805 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="16dbab06-8ff8-4a5a-9fe8-4967c6838a4a" path="/var/lib/kubelet/pods/16dbab06-8ff8-4a5a-9fe8-4967c6838a4a/volumes" Dec 03 14:30:29 crc kubenswrapper[4805]: I1203 14:30:29.509591 4805 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-7ff5475cc9-mbdbt" podUID="1ca67122-920a-43a6-a434-82ec573d0160" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.125:5353: i/o timeout" Dec 03 14:30:34 crc kubenswrapper[4805]: I1203 14:30:34.510732 4805 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-7ff5475cc9-mbdbt" podUID="1ca67122-920a-43a6-a434-82ec573d0160" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.125:5353: i/o timeout" Dec 03 14:30:34 crc kubenswrapper[4805]: I1203 14:30:34.511900 4805 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-7ff5475cc9-mbdbt" Dec 03 14:30:35 crc kubenswrapper[4805]: I1203 14:30:35.309495 4805 scope.go:117] "RemoveContainer" containerID="188ca93940f09e5d5f1a3e68344b1ff578bfc6d8a437936fc9d054967cfa3ed2" Dec 03 14:30:35 crc kubenswrapper[4805]: I1203 14:30:35.400997 4805 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7ff5475cc9-mbdbt" Dec 03 14:30:35 crc kubenswrapper[4805]: I1203 14:30:35.553981 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/1ca67122-920a-43a6-a434-82ec573d0160-dns-swift-storage-0\") pod \"1ca67122-920a-43a6-a434-82ec573d0160\" (UID: \"1ca67122-920a-43a6-a434-82ec573d0160\") " Dec 03 14:30:35 crc kubenswrapper[4805]: I1203 14:30:35.554035 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/1ca67122-920a-43a6-a434-82ec573d0160-dns-svc\") pod \"1ca67122-920a-43a6-a434-82ec573d0160\" (UID: \"1ca67122-920a-43a6-a434-82ec573d0160\") " Dec 03 14:30:35 crc kubenswrapper[4805]: I1203 14:30:35.554069 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/1ca67122-920a-43a6-a434-82ec573d0160-ovsdbserver-sb\") pod \"1ca67122-920a-43a6-a434-82ec573d0160\" (UID: \"1ca67122-920a-43a6-a434-82ec573d0160\") " Dec 03 14:30:35 crc kubenswrapper[4805]: I1203 14:30:35.554105 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kqtxj\" (UniqueName: \"kubernetes.io/projected/1ca67122-920a-43a6-a434-82ec573d0160-kube-api-access-kqtxj\") pod \"1ca67122-920a-43a6-a434-82ec573d0160\" (UID: \"1ca67122-920a-43a6-a434-82ec573d0160\") " Dec 03 14:30:35 crc kubenswrapper[4805]: I1203 14:30:35.554122 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1ca67122-920a-43a6-a434-82ec573d0160-config\") pod \"1ca67122-920a-43a6-a434-82ec573d0160\" (UID: \"1ca67122-920a-43a6-a434-82ec573d0160\") " Dec 03 14:30:35 crc kubenswrapper[4805]: I1203 14:30:35.554191 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/1ca67122-920a-43a6-a434-82ec573d0160-ovsdbserver-nb\") pod \"1ca67122-920a-43a6-a434-82ec573d0160\" (UID: \"1ca67122-920a-43a6-a434-82ec573d0160\") " Dec 03 14:30:35 crc kubenswrapper[4805]: I1203 14:30:35.560350 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1ca67122-920a-43a6-a434-82ec573d0160-kube-api-access-kqtxj" (OuterVolumeSpecName: "kube-api-access-kqtxj") pod "1ca67122-920a-43a6-a434-82ec573d0160" (UID: "1ca67122-920a-43a6-a434-82ec573d0160"). InnerVolumeSpecName "kube-api-access-kqtxj". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 14:30:35 crc kubenswrapper[4805]: I1203 14:30:35.601439 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1ca67122-920a-43a6-a434-82ec573d0160-config" (OuterVolumeSpecName: "config") pod "1ca67122-920a-43a6-a434-82ec573d0160" (UID: "1ca67122-920a-43a6-a434-82ec573d0160"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 14:30:35 crc kubenswrapper[4805]: I1203 14:30:35.603678 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1ca67122-920a-43a6-a434-82ec573d0160-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "1ca67122-920a-43a6-a434-82ec573d0160" (UID: "1ca67122-920a-43a6-a434-82ec573d0160"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 14:30:35 crc kubenswrapper[4805]: I1203 14:30:35.611598 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1ca67122-920a-43a6-a434-82ec573d0160-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "1ca67122-920a-43a6-a434-82ec573d0160" (UID: "1ca67122-920a-43a6-a434-82ec573d0160"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 14:30:35 crc kubenswrapper[4805]: I1203 14:30:35.622017 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1ca67122-920a-43a6-a434-82ec573d0160-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "1ca67122-920a-43a6-a434-82ec573d0160" (UID: "1ca67122-920a-43a6-a434-82ec573d0160"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 14:30:35 crc kubenswrapper[4805]: I1203 14:30:35.642776 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1ca67122-920a-43a6-a434-82ec573d0160-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "1ca67122-920a-43a6-a434-82ec573d0160" (UID: "1ca67122-920a-43a6-a434-82ec573d0160"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 14:30:35 crc kubenswrapper[4805]: I1203 14:30:35.657486 4805 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/1ca67122-920a-43a6-a434-82ec573d0160-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Dec 03 14:30:35 crc kubenswrapper[4805]: I1203 14:30:35.657522 4805 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/1ca67122-920a-43a6-a434-82ec573d0160-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 03 14:30:35 crc kubenswrapper[4805]: I1203 14:30:35.657532 4805 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/1ca67122-920a-43a6-a434-82ec573d0160-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Dec 03 14:30:35 crc kubenswrapper[4805]: I1203 14:30:35.657541 4805 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kqtxj\" (UniqueName: \"kubernetes.io/projected/1ca67122-920a-43a6-a434-82ec573d0160-kube-api-access-kqtxj\") on node \"crc\" DevicePath \"\"" Dec 03 14:30:35 crc kubenswrapper[4805]: I1203 14:30:35.657552 4805 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1ca67122-920a-43a6-a434-82ec573d0160-config\") on node \"crc\" DevicePath \"\"" Dec 03 14:30:35 crc kubenswrapper[4805]: I1203 14:30:35.657560 4805 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/1ca67122-920a-43a6-a434-82ec573d0160-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Dec 03 14:30:35 crc kubenswrapper[4805]: I1203 14:30:35.701794 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7ff5475cc9-mbdbt" event={"ID":"1ca67122-920a-43a6-a434-82ec573d0160","Type":"ContainerDied","Data":"98681326060f9f03fd0c1123246638539b81883860fd056234e23b2a9170ba45"} Dec 03 14:30:35 crc kubenswrapper[4805]: I1203 14:30:35.701911 4805 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7ff5475cc9-mbdbt" Dec 03 14:30:35 crc kubenswrapper[4805]: I1203 14:30:35.734731 4805 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-7ff5475cc9-mbdbt"] Dec 03 14:30:35 crc kubenswrapper[4805]: I1203 14:30:35.741424 4805 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-7ff5475cc9-mbdbt"] Dec 03 14:30:35 crc kubenswrapper[4805]: E1203 14:30:35.858004 4805 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-barbican-api:current-podified" Dec 03 14:30:35 crc kubenswrapper[4805]: E1203 14:30:35.858266 4805 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:barbican-db-sync,Image:quay.io/podified-antelope-centos9/openstack-barbican-api:current-podified,Command:[/bin/bash],Args:[-c barbican-manage db upgrade],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:KOLLA_BOOTSTRAP,Value:TRUE,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:db-sync-config-data,ReadOnly:true,MountPath:/etc/barbican/barbican.conf.d,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:combined-ca-bundle,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-c5jlg,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*42403,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:*42403,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod barbican-db-sync-74tz4_openstack(13ba03f0-acc2-4fb8-90a9-583b4f8694cb): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 03 14:30:35 crc kubenswrapper[4805]: E1203 14:30:35.860473 4805 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"barbican-db-sync\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/barbican-db-sync-74tz4" podUID="13ba03f0-acc2-4fb8-90a9-583b4f8694cb" Dec 03 14:30:36 crc kubenswrapper[4805]: I1203 14:30:36.719638 4805 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1ca67122-920a-43a6-a434-82ec573d0160" path="/var/lib/kubelet/pods/1ca67122-920a-43a6-a434-82ec573d0160/volumes" Dec 03 14:30:36 crc kubenswrapper[4805]: E1203 14:30:36.724991 4805 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"barbican-db-sync\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-barbican-api:current-podified\\\"\"" pod="openstack/barbican-db-sync-74tz4" podUID="13ba03f0-acc2-4fb8-90a9-583b4f8694cb" Dec 03 14:30:36 crc kubenswrapper[4805]: E1203 14:30:36.821145 4805 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-cinder-api:current-podified" Dec 03 14:30:36 crc kubenswrapper[4805]: E1203 14:30:36.821391 4805 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:cinder-db-sync,Image:quay.io/podified-antelope-centos9/openstack-cinder-api:current-podified,Command:[/bin/bash],Args:[-c /usr/local/bin/kolla_set_configs && /usr/local/bin/kolla_start],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:KOLLA_BOOTSTRAP,Value:TRUE,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:etc-machine-id,ReadOnly:true,MountPath:/etc/machine-id,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:scripts,ReadOnly:true,MountPath:/usr/local/bin/container-scripts,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/var/lib/config-data/merged,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/etc/my.cnf,SubPath:my.cnf,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:db-sync-config-data,ReadOnly:true,MountPath:/etc/cinder/cinder.conf.d,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/var/lib/kolla/config_files/config.json,SubPath:db-sync-config.json,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:combined-ca-bundle,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-xxm8t,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:*0,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod cinder-db-sync-fh9c2_openstack(260717ec-f898-4ccd-8a4f-c7d6b14ade6e): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 03 14:30:36 crc kubenswrapper[4805]: E1203 14:30:36.829507 4805 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"cinder-db-sync\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/cinder-db-sync-fh9c2" podUID="260717ec-f898-4ccd-8a4f-c7d6b14ade6e" Dec 03 14:30:36 crc kubenswrapper[4805]: I1203 14:30:36.848133 4805 scope.go:117] "RemoveContainer" containerID="f9c31ba4a3601f19c3ce752a2e09ba3a90a39dfe55c8ca130f6564acd0157752" Dec 03 14:30:36 crc kubenswrapper[4805]: I1203 14:30:36.944341 4805 scope.go:117] "RemoveContainer" containerID="392489bc555f119580c8997d1ea4e8089d6644ef798f7247ca27ec78f95e9855" Dec 03 14:30:37 crc kubenswrapper[4805]: I1203 14:30:37.339628 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 03 14:30:37 crc kubenswrapper[4805]: I1203 14:30:37.384691 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-bootstrap-74pdd"] Dec 03 14:30:37 crc kubenswrapper[4805]: W1203 14:30:37.418213 4805 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod10877f19_dd2a_4eb2_99e0_2d0cf82d5ad5.slice/crio-53cfaf6021f91b400709e52ab56b25b8a1e47a90e0bf4947947db07aec4309ef WatchSource:0}: Error finding container 53cfaf6021f91b400709e52ab56b25b8a1e47a90e0bf4947947db07aec4309ef: Status 404 returned error can't find the container with id 53cfaf6021f91b400709e52ab56b25b8a1e47a90e0bf4947947db07aec4309ef Dec 03 14:30:37 crc kubenswrapper[4805]: I1203 14:30:37.452856 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"osp-secret" Dec 03 14:30:37 crc kubenswrapper[4805]: I1203 14:30:37.732498 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"10877f19-dd2a-4eb2-99e0-2d0cf82d5ad5","Type":"ContainerStarted","Data":"53cfaf6021f91b400709e52ab56b25b8a1e47a90e0bf4947947db07aec4309ef"} Dec 03 14:30:37 crc kubenswrapper[4805]: I1203 14:30:37.738989 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-sync-sj4wt" event={"ID":"7c1b5525-de7b-4a8c-9062-bb9a68336989","Type":"ContainerStarted","Data":"d44bffcdc75d0c2fe9722cdbe85e33a78e3ee76b7cbfda1d8e0f176918810006"} Dec 03 14:30:37 crc kubenswrapper[4805]: I1203 14:30:37.742590 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"8533d470-22f2-401a-a39d-c37391920d81","Type":"ContainerStarted","Data":"dd9cf491b38857ec975e532024385fa6530bdcbb0987559861725061e1a31b29"} Dec 03 14:30:37 crc kubenswrapper[4805]: I1203 14:30:37.751272 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-db-sync-jvqmk" event={"ID":"14c327e3-e880-4b49-b366-364a0395cc55","Type":"ContainerStarted","Data":"baebfb4ff35a0b981da87af49115db6766414aa8463f0e39b209086ad47d1d10"} Dec 03 14:30:37 crc kubenswrapper[4805]: I1203 14:30:37.758811 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-74pdd" event={"ID":"b4e050ed-47f5-4d9f-a288-b6d2a7ab294a","Type":"ContainerStarted","Data":"c74bd6f2e58840be38059668ab74c768c407bb15b81d002cdfe1258aa66a081b"} Dec 03 14:30:37 crc kubenswrapper[4805]: I1203 14:30:37.758887 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-74pdd" event={"ID":"b4e050ed-47f5-4d9f-a288-b6d2a7ab294a","Type":"ContainerStarted","Data":"99f0a8d0649c06d7eb41305312ae7f07fcada9db1e232d6c809324c8bcdb5a75"} Dec 03 14:30:37 crc kubenswrapper[4805]: I1203 14:30:37.759613 4805 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/placement-db-sync-sj4wt" podStartSLOduration=2.381669256 podStartE2EDuration="29.759598407s" podCreationTimestamp="2025-12-03 14:30:08 +0000 UTC" firstStartedPulling="2025-12-03 14:30:09.415544653 +0000 UTC m=+1239.078461576" lastFinishedPulling="2025-12-03 14:30:36.793473794 +0000 UTC m=+1266.456390727" observedRunningTime="2025-12-03 14:30:37.758318173 +0000 UTC m=+1267.421235096" watchObservedRunningTime="2025-12-03 14:30:37.759598407 +0000 UTC m=+1267.422515350" Dec 03 14:30:37 crc kubenswrapper[4805]: E1203 14:30:37.763474 4805 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"cinder-db-sync\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-cinder-api:current-podified\\\"\"" pod="openstack/cinder-db-sync-fh9c2" podUID="260717ec-f898-4ccd-8a4f-c7d6b14ade6e" Dec 03 14:30:37 crc kubenswrapper[4805]: I1203 14:30:37.786692 4805 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-bootstrap-74pdd" podStartSLOduration=9.7866734 podStartE2EDuration="9.7866734s" podCreationTimestamp="2025-12-03 14:30:28 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 14:30:37.77929054 +0000 UTC m=+1267.442207473" watchObservedRunningTime="2025-12-03 14:30:37.7866734 +0000 UTC m=+1267.449590333" Dec 03 14:30:38 crc kubenswrapper[4805]: I1203 14:30:38.035382 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Dec 03 14:30:38 crc kubenswrapper[4805]: W1203 14:30:38.049462 4805 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod9ed6dfc0_82db_4d86_ad92_cfe94fa370a2.slice/crio-fb9cf095ff0fd5118500bc41d0e009beeadf259bbfe428a14abc1cdf0703dbfa WatchSource:0}: Error finding container fb9cf095ff0fd5118500bc41d0e009beeadf259bbfe428a14abc1cdf0703dbfa: Status 404 returned error can't find the container with id fb9cf095ff0fd5118500bc41d0e009beeadf259bbfe428a14abc1cdf0703dbfa Dec 03 14:30:38 crc kubenswrapper[4805]: I1203 14:30:38.780559 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"10877f19-dd2a-4eb2-99e0-2d0cf82d5ad5","Type":"ContainerStarted","Data":"24356197f69f0bfd999f1654211fef34606d3c50328026f4b3c6d2eda195db85"} Dec 03 14:30:38 crc kubenswrapper[4805]: I1203 14:30:38.783022 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"9ed6dfc0-82db-4d86-ad92-cfe94fa370a2","Type":"ContainerStarted","Data":"18b272180201caa258fce70a3153b9e82a5103c2066cac10db4b226952843d2f"} Dec 03 14:30:38 crc kubenswrapper[4805]: I1203 14:30:38.783077 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"9ed6dfc0-82db-4d86-ad92-cfe94fa370a2","Type":"ContainerStarted","Data":"fb9cf095ff0fd5118500bc41d0e009beeadf259bbfe428a14abc1cdf0703dbfa"} Dec 03 14:30:39 crc kubenswrapper[4805]: I1203 14:30:39.511662 4805 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-7ff5475cc9-mbdbt" podUID="1ca67122-920a-43a6-a434-82ec573d0160" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.125:5353: i/o timeout" Dec 03 14:30:39 crc kubenswrapper[4805]: I1203 14:30:39.794304 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"10877f19-dd2a-4eb2-99e0-2d0cf82d5ad5","Type":"ContainerStarted","Data":"61aa5396071fa72fd8a539334f4fa1f4982e07e997aee839e400e0b116fff32e"} Dec 03 14:30:39 crc kubenswrapper[4805]: I1203 14:30:39.796894 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"9ed6dfc0-82db-4d86-ad92-cfe94fa370a2","Type":"ContainerStarted","Data":"b40b68276f1683e9c153b03843673b0fb798f28865161b6b4f97250e1dcc86f9"} Dec 03 14:30:40 crc kubenswrapper[4805]: I1203 14:30:40.731592 4805 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/heat-db-sync-jvqmk" podStartSLOduration=4.943444942 podStartE2EDuration="32.731575258s" podCreationTimestamp="2025-12-03 14:30:08 +0000 UTC" firstStartedPulling="2025-12-03 14:30:09.664030069 +0000 UTC m=+1239.326946992" lastFinishedPulling="2025-12-03 14:30:37.452160385 +0000 UTC m=+1267.115077308" observedRunningTime="2025-12-03 14:30:38.80956155 +0000 UTC m=+1268.472478473" watchObservedRunningTime="2025-12-03 14:30:40.731575258 +0000 UTC m=+1270.394492181" Dec 03 14:30:40 crc kubenswrapper[4805]: I1203 14:30:40.835152 4805 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-internal-api-0" podStartSLOduration=24.835135591 podStartE2EDuration="24.835135591s" podCreationTimestamp="2025-12-03 14:30:16 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 14:30:40.833830206 +0000 UTC m=+1270.496747169" watchObservedRunningTime="2025-12-03 14:30:40.835135591 +0000 UTC m=+1270.498052514" Dec 03 14:30:40 crc kubenswrapper[4805]: I1203 14:30:40.871120 4805 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-external-api-0" podStartSLOduration=27.871100975 podStartE2EDuration="27.871100975s" podCreationTimestamp="2025-12-03 14:30:13 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 14:30:40.863404976 +0000 UTC m=+1270.526321899" watchObservedRunningTime="2025-12-03 14:30:40.871100975 +0000 UTC m=+1270.534017898" Dec 03 14:30:42 crc kubenswrapper[4805]: I1203 14:30:42.825424 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"8533d470-22f2-401a-a39d-c37391920d81","Type":"ContainerStarted","Data":"0e63c808604f7e7fe89fba38cf60f74044c3c5721825fa4640ddeadceee1ca05"} Dec 03 14:30:42 crc kubenswrapper[4805]: I1203 14:30:42.826815 4805 generic.go:334] "Generic (PLEG): container finished" podID="b4e050ed-47f5-4d9f-a288-b6d2a7ab294a" containerID="c74bd6f2e58840be38059668ab74c768c407bb15b81d002cdfe1258aa66a081b" exitCode=0 Dec 03 14:30:42 crc kubenswrapper[4805]: I1203 14:30:42.826874 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-74pdd" event={"ID":"b4e050ed-47f5-4d9f-a288-b6d2a7ab294a","Type":"ContainerDied","Data":"c74bd6f2e58840be38059668ab74c768c407bb15b81d002cdfe1258aa66a081b"} Dec 03 14:30:43 crc kubenswrapper[4805]: I1203 14:30:43.845573 4805 generic.go:334] "Generic (PLEG): container finished" podID="7c1b5525-de7b-4a8c-9062-bb9a68336989" containerID="d44bffcdc75d0c2fe9722cdbe85e33a78e3ee76b7cbfda1d8e0f176918810006" exitCode=0 Dec 03 14:30:43 crc kubenswrapper[4805]: I1203 14:30:43.845685 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-sync-sj4wt" event={"ID":"7c1b5525-de7b-4a8c-9062-bb9a68336989","Type":"ContainerDied","Data":"d44bffcdc75d0c2fe9722cdbe85e33a78e3ee76b7cbfda1d8e0f176918810006"} Dec 03 14:30:43 crc kubenswrapper[4805]: I1203 14:30:43.898396 4805 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-external-api-0" Dec 03 14:30:43 crc kubenswrapper[4805]: I1203 14:30:43.898724 4805 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-external-api-0" Dec 03 14:30:43 crc kubenswrapper[4805]: I1203 14:30:43.898740 4805 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-external-api-0" Dec 03 14:30:43 crc kubenswrapper[4805]: I1203 14:30:43.898752 4805 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-external-api-0" Dec 03 14:30:43 crc kubenswrapper[4805]: I1203 14:30:43.918020 4805 patch_prober.go:28] interesting pod/machine-config-daemon-gskh4 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 03 14:30:43 crc kubenswrapper[4805]: I1203 14:30:43.918086 4805 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-gskh4" podUID="6098937f-e3f6-45e8-a647-4994a79cd711" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 03 14:30:43 crc kubenswrapper[4805]: I1203 14:30:43.918146 4805 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-gskh4" Dec 03 14:30:43 crc kubenswrapper[4805]: I1203 14:30:43.918929 4805 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"b84d108663e877514e58fb0f23591399246f6b24b642394275299320cb0e52d9"} pod="openshift-machine-config-operator/machine-config-daemon-gskh4" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 03 14:30:43 crc kubenswrapper[4805]: I1203 14:30:43.918990 4805 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-gskh4" podUID="6098937f-e3f6-45e8-a647-4994a79cd711" containerName="machine-config-daemon" containerID="cri-o://b84d108663e877514e58fb0f23591399246f6b24b642394275299320cb0e52d9" gracePeriod=600 Dec 03 14:30:43 crc kubenswrapper[4805]: I1203 14:30:43.939940 4805 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-external-api-0" Dec 03 14:30:43 crc kubenswrapper[4805]: I1203 14:30:43.951591 4805 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-external-api-0" Dec 03 14:30:45 crc kubenswrapper[4805]: I1203 14:30:44.856387 4805 generic.go:334] "Generic (PLEG): container finished" podID="6098937f-e3f6-45e8-a647-4994a79cd711" containerID="b84d108663e877514e58fb0f23591399246f6b24b642394275299320cb0e52d9" exitCode=0 Dec 03 14:30:45 crc kubenswrapper[4805]: I1203 14:30:44.856792 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-gskh4" event={"ID":"6098937f-e3f6-45e8-a647-4994a79cd711","Type":"ContainerDied","Data":"b84d108663e877514e58fb0f23591399246f6b24b642394275299320cb0e52d9"} Dec 03 14:30:45 crc kubenswrapper[4805]: I1203 14:30:44.856821 4805 scope.go:117] "RemoveContainer" containerID="5079b541103d89d1ef42b4f7093c19ac19ef0baacd14c63b2eb7fb16b384fba4" Dec 03 14:30:45 crc kubenswrapper[4805]: I1203 14:30:45.867147 4805 generic.go:334] "Generic (PLEG): container finished" podID="14c327e3-e880-4b49-b366-364a0395cc55" containerID="baebfb4ff35a0b981da87af49115db6766414aa8463f0e39b209086ad47d1d10" exitCode=0 Dec 03 14:30:45 crc kubenswrapper[4805]: I1203 14:30:45.867201 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-db-sync-jvqmk" event={"ID":"14c327e3-e880-4b49-b366-364a0395cc55","Type":"ContainerDied","Data":"baebfb4ff35a0b981da87af49115db6766414aa8463f0e39b209086ad47d1d10"} Dec 03 14:30:45 crc kubenswrapper[4805]: I1203 14:30:45.869947 4805 generic.go:334] "Generic (PLEG): container finished" podID="32920f09-a728-4b35-a5d7-65dae2d70834" containerID="c8040c823d6c387af893c10d68844e6a3f47022e19a18567b34538fe04cfa697" exitCode=0 Dec 03 14:30:45 crc kubenswrapper[4805]: I1203 14:30:45.870358 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-sync-b6k7w" event={"ID":"32920f09-a728-4b35-a5d7-65dae2d70834","Type":"ContainerDied","Data":"c8040c823d6c387af893c10d68844e6a3f47022e19a18567b34538fe04cfa697"} Dec 03 14:30:46 crc kubenswrapper[4805]: I1203 14:30:46.708485 4805 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-sync-sj4wt" Dec 03 14:30:46 crc kubenswrapper[4805]: I1203 14:30:46.711139 4805 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-74pdd" Dec 03 14:30:46 crc kubenswrapper[4805]: I1203 14:30:46.778258 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7c1b5525-de7b-4a8c-9062-bb9a68336989-combined-ca-bundle\") pod \"7c1b5525-de7b-4a8c-9062-bb9a68336989\" (UID: \"7c1b5525-de7b-4a8c-9062-bb9a68336989\") " Dec 03 14:30:46 crc kubenswrapper[4805]: I1203 14:30:46.778329 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b4e050ed-47f5-4d9f-a288-b6d2a7ab294a-combined-ca-bundle\") pod \"b4e050ed-47f5-4d9f-a288-b6d2a7ab294a\" (UID: \"b4e050ed-47f5-4d9f-a288-b6d2a7ab294a\") " Dec 03 14:30:46 crc kubenswrapper[4805]: I1203 14:30:46.778373 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/7c1b5525-de7b-4a8c-9062-bb9a68336989-logs\") pod \"7c1b5525-de7b-4a8c-9062-bb9a68336989\" (UID: \"7c1b5525-de7b-4a8c-9062-bb9a68336989\") " Dec 03 14:30:46 crc kubenswrapper[4805]: I1203 14:30:46.778459 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-svwdk\" (UniqueName: \"kubernetes.io/projected/7c1b5525-de7b-4a8c-9062-bb9a68336989-kube-api-access-svwdk\") pod \"7c1b5525-de7b-4a8c-9062-bb9a68336989\" (UID: \"7c1b5525-de7b-4a8c-9062-bb9a68336989\") " Dec 03 14:30:46 crc kubenswrapper[4805]: I1203 14:30:46.778541 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7c1b5525-de7b-4a8c-9062-bb9a68336989-config-data\") pod \"7c1b5525-de7b-4a8c-9062-bb9a68336989\" (UID: \"7c1b5525-de7b-4a8c-9062-bb9a68336989\") " Dec 03 14:30:46 crc kubenswrapper[4805]: I1203 14:30:46.778625 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b4e050ed-47f5-4d9f-a288-b6d2a7ab294a-scripts\") pod \"b4e050ed-47f5-4d9f-a288-b6d2a7ab294a\" (UID: \"b4e050ed-47f5-4d9f-a288-b6d2a7ab294a\") " Dec 03 14:30:46 crc kubenswrapper[4805]: I1203 14:30:46.779014 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7c1b5525-de7b-4a8c-9062-bb9a68336989-logs" (OuterVolumeSpecName: "logs") pod "7c1b5525-de7b-4a8c-9062-bb9a68336989" (UID: "7c1b5525-de7b-4a8c-9062-bb9a68336989"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 14:30:46 crc kubenswrapper[4805]: I1203 14:30:46.779200 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/b4e050ed-47f5-4d9f-a288-b6d2a7ab294a-credential-keys\") pod \"b4e050ed-47f5-4d9f-a288-b6d2a7ab294a\" (UID: \"b4e050ed-47f5-4d9f-a288-b6d2a7ab294a\") " Dec 03 14:30:46 crc kubenswrapper[4805]: I1203 14:30:46.779273 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w7ppx\" (UniqueName: \"kubernetes.io/projected/b4e050ed-47f5-4d9f-a288-b6d2a7ab294a-kube-api-access-w7ppx\") pod \"b4e050ed-47f5-4d9f-a288-b6d2a7ab294a\" (UID: \"b4e050ed-47f5-4d9f-a288-b6d2a7ab294a\") " Dec 03 14:30:46 crc kubenswrapper[4805]: I1203 14:30:46.779308 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7c1b5525-de7b-4a8c-9062-bb9a68336989-scripts\") pod \"7c1b5525-de7b-4a8c-9062-bb9a68336989\" (UID: \"7c1b5525-de7b-4a8c-9062-bb9a68336989\") " Dec 03 14:30:46 crc kubenswrapper[4805]: I1203 14:30:46.779360 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b4e050ed-47f5-4d9f-a288-b6d2a7ab294a-config-data\") pod \"b4e050ed-47f5-4d9f-a288-b6d2a7ab294a\" (UID: \"b4e050ed-47f5-4d9f-a288-b6d2a7ab294a\") " Dec 03 14:30:46 crc kubenswrapper[4805]: I1203 14:30:46.779409 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/b4e050ed-47f5-4d9f-a288-b6d2a7ab294a-fernet-keys\") pod \"b4e050ed-47f5-4d9f-a288-b6d2a7ab294a\" (UID: \"b4e050ed-47f5-4d9f-a288-b6d2a7ab294a\") " Dec 03 14:30:46 crc kubenswrapper[4805]: I1203 14:30:46.780132 4805 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/7c1b5525-de7b-4a8c-9062-bb9a68336989-logs\") on node \"crc\" DevicePath \"\"" Dec 03 14:30:46 crc kubenswrapper[4805]: I1203 14:30:46.784484 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7c1b5525-de7b-4a8c-9062-bb9a68336989-kube-api-access-svwdk" (OuterVolumeSpecName: "kube-api-access-svwdk") pod "7c1b5525-de7b-4a8c-9062-bb9a68336989" (UID: "7c1b5525-de7b-4a8c-9062-bb9a68336989"). InnerVolumeSpecName "kube-api-access-svwdk". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 14:30:46 crc kubenswrapper[4805]: I1203 14:30:46.785008 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b4e050ed-47f5-4d9f-a288-b6d2a7ab294a-fernet-keys" (OuterVolumeSpecName: "fernet-keys") pod "b4e050ed-47f5-4d9f-a288-b6d2a7ab294a" (UID: "b4e050ed-47f5-4d9f-a288-b6d2a7ab294a"). InnerVolumeSpecName "fernet-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 14:30:46 crc kubenswrapper[4805]: I1203 14:30:46.790264 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b4e050ed-47f5-4d9f-a288-b6d2a7ab294a-kube-api-access-w7ppx" (OuterVolumeSpecName: "kube-api-access-w7ppx") pod "b4e050ed-47f5-4d9f-a288-b6d2a7ab294a" (UID: "b4e050ed-47f5-4d9f-a288-b6d2a7ab294a"). InnerVolumeSpecName "kube-api-access-w7ppx". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 14:30:46 crc kubenswrapper[4805]: I1203 14:30:46.790628 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7c1b5525-de7b-4a8c-9062-bb9a68336989-scripts" (OuterVolumeSpecName: "scripts") pod "7c1b5525-de7b-4a8c-9062-bb9a68336989" (UID: "7c1b5525-de7b-4a8c-9062-bb9a68336989"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 14:30:46 crc kubenswrapper[4805]: I1203 14:30:46.792804 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b4e050ed-47f5-4d9f-a288-b6d2a7ab294a-credential-keys" (OuterVolumeSpecName: "credential-keys") pod "b4e050ed-47f5-4d9f-a288-b6d2a7ab294a" (UID: "b4e050ed-47f5-4d9f-a288-b6d2a7ab294a"). InnerVolumeSpecName "credential-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 14:30:46 crc kubenswrapper[4805]: I1203 14:30:46.800123 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b4e050ed-47f5-4d9f-a288-b6d2a7ab294a-scripts" (OuterVolumeSpecName: "scripts") pod "b4e050ed-47f5-4d9f-a288-b6d2a7ab294a" (UID: "b4e050ed-47f5-4d9f-a288-b6d2a7ab294a"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 14:30:46 crc kubenswrapper[4805]: I1203 14:30:46.816826 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7c1b5525-de7b-4a8c-9062-bb9a68336989-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "7c1b5525-de7b-4a8c-9062-bb9a68336989" (UID: "7c1b5525-de7b-4a8c-9062-bb9a68336989"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 14:30:46 crc kubenswrapper[4805]: I1203 14:30:46.825161 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b4e050ed-47f5-4d9f-a288-b6d2a7ab294a-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "b4e050ed-47f5-4d9f-a288-b6d2a7ab294a" (UID: "b4e050ed-47f5-4d9f-a288-b6d2a7ab294a"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 14:30:46 crc kubenswrapper[4805]: I1203 14:30:46.833652 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7c1b5525-de7b-4a8c-9062-bb9a68336989-config-data" (OuterVolumeSpecName: "config-data") pod "7c1b5525-de7b-4a8c-9062-bb9a68336989" (UID: "7c1b5525-de7b-4a8c-9062-bb9a68336989"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 14:30:46 crc kubenswrapper[4805]: I1203 14:30:46.841321 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b4e050ed-47f5-4d9f-a288-b6d2a7ab294a-config-data" (OuterVolumeSpecName: "config-data") pod "b4e050ed-47f5-4d9f-a288-b6d2a7ab294a" (UID: "b4e050ed-47f5-4d9f-a288-b6d2a7ab294a"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 14:30:46 crc kubenswrapper[4805]: I1203 14:30:46.879248 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-sync-sj4wt" event={"ID":"7c1b5525-de7b-4a8c-9062-bb9a68336989","Type":"ContainerDied","Data":"438e761f1425123da258ad8b5ac54c3ef29d62c8d78ab6103c5ecee84a795ff2"} Dec 03 14:30:46 crc kubenswrapper[4805]: I1203 14:30:46.879289 4805 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="438e761f1425123da258ad8b5ac54c3ef29d62c8d78ab6103c5ecee84a795ff2" Dec 03 14:30:46 crc kubenswrapper[4805]: I1203 14:30:46.879268 4805 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-sync-sj4wt" Dec 03 14:30:46 crc kubenswrapper[4805]: I1203 14:30:46.881220 4805 reconciler_common.go:293] "Volume detached for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/b4e050ed-47f5-4d9f-a288-b6d2a7ab294a-fernet-keys\") on node \"crc\" DevicePath \"\"" Dec 03 14:30:46 crc kubenswrapper[4805]: I1203 14:30:46.881301 4805 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7c1b5525-de7b-4a8c-9062-bb9a68336989-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 14:30:46 crc kubenswrapper[4805]: I1203 14:30:46.881355 4805 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b4e050ed-47f5-4d9f-a288-b6d2a7ab294a-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 14:30:46 crc kubenswrapper[4805]: I1203 14:30:46.881412 4805 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-svwdk\" (UniqueName: \"kubernetes.io/projected/7c1b5525-de7b-4a8c-9062-bb9a68336989-kube-api-access-svwdk\") on node \"crc\" DevicePath \"\"" Dec 03 14:30:46 crc kubenswrapper[4805]: I1203 14:30:46.881461 4805 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7c1b5525-de7b-4a8c-9062-bb9a68336989-config-data\") on node \"crc\" DevicePath \"\"" Dec 03 14:30:46 crc kubenswrapper[4805]: I1203 14:30:46.881515 4805 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b4e050ed-47f5-4d9f-a288-b6d2a7ab294a-scripts\") on node \"crc\" DevicePath \"\"" Dec 03 14:30:46 crc kubenswrapper[4805]: I1203 14:30:46.881568 4805 reconciler_common.go:293] "Volume detached for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/b4e050ed-47f5-4d9f-a288-b6d2a7ab294a-credential-keys\") on node \"crc\" DevicePath \"\"" Dec 03 14:30:46 crc kubenswrapper[4805]: I1203 14:30:46.881617 4805 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w7ppx\" (UniqueName: \"kubernetes.io/projected/b4e050ed-47f5-4d9f-a288-b6d2a7ab294a-kube-api-access-w7ppx\") on node \"crc\" DevicePath \"\"" Dec 03 14:30:46 crc kubenswrapper[4805]: I1203 14:30:46.881663 4805 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7c1b5525-de7b-4a8c-9062-bb9a68336989-scripts\") on node \"crc\" DevicePath \"\"" Dec 03 14:30:46 crc kubenswrapper[4805]: I1203 14:30:46.881710 4805 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b4e050ed-47f5-4d9f-a288-b6d2a7ab294a-config-data\") on node \"crc\" DevicePath \"\"" Dec 03 14:30:46 crc kubenswrapper[4805]: I1203 14:30:46.883165 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-74pdd" event={"ID":"b4e050ed-47f5-4d9f-a288-b6d2a7ab294a","Type":"ContainerDied","Data":"99f0a8d0649c06d7eb41305312ae7f07fcada9db1e232d6c809324c8bcdb5a75"} Dec 03 14:30:46 crc kubenswrapper[4805]: I1203 14:30:46.883267 4805 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="99f0a8d0649c06d7eb41305312ae7f07fcada9db1e232d6c809324c8bcdb5a75" Dec 03 14:30:46 crc kubenswrapper[4805]: I1203 14:30:46.883183 4805 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-74pdd" Dec 03 14:30:46 crc kubenswrapper[4805]: I1203 14:30:46.950568 4805 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-internal-api-0" Dec 03 14:30:46 crc kubenswrapper[4805]: I1203 14:30:46.950792 4805 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-internal-api-0" Dec 03 14:30:46 crc kubenswrapper[4805]: I1203 14:30:46.950898 4805 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-internal-api-0" Dec 03 14:30:46 crc kubenswrapper[4805]: I1203 14:30:46.951121 4805 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-internal-api-0" Dec 03 14:30:46 crc kubenswrapper[4805]: I1203 14:30:46.974591 4805 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-external-api-0" Dec 03 14:30:46 crc kubenswrapper[4805]: I1203 14:30:46.974980 4805 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Dec 03 14:30:46 crc kubenswrapper[4805]: I1203 14:30:46.981602 4805 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-external-api-0" Dec 03 14:30:46 crc kubenswrapper[4805]: I1203 14:30:46.985688 4805 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-internal-api-0" Dec 03 14:30:47 crc kubenswrapper[4805]: I1203 14:30:47.009329 4805 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-internal-api-0" Dec 03 14:30:47 crc kubenswrapper[4805]: I1203 14:30:47.303346 4805 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/heat-db-sync-jvqmk" Dec 03 14:30:47 crc kubenswrapper[4805]: I1203 14:30:47.319512 4805 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-sync-b6k7w" Dec 03 14:30:47 crc kubenswrapper[4805]: I1203 14:30:47.490922 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-v5gd6\" (UniqueName: \"kubernetes.io/projected/32920f09-a728-4b35-a5d7-65dae2d70834-kube-api-access-v5gd6\") pod \"32920f09-a728-4b35-a5d7-65dae2d70834\" (UID: \"32920f09-a728-4b35-a5d7-65dae2d70834\") " Dec 03 14:30:47 crc kubenswrapper[4805]: I1203 14:30:47.490985 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/14c327e3-e880-4b49-b366-364a0395cc55-config-data\") pod \"14c327e3-e880-4b49-b366-364a0395cc55\" (UID: \"14c327e3-e880-4b49-b366-364a0395cc55\") " Dec 03 14:30:47 crc kubenswrapper[4805]: I1203 14:30:47.491064 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/14c327e3-e880-4b49-b366-364a0395cc55-combined-ca-bundle\") pod \"14c327e3-e880-4b49-b366-364a0395cc55\" (UID: \"14c327e3-e880-4b49-b366-364a0395cc55\") " Dec 03 14:30:47 crc kubenswrapper[4805]: I1203 14:30:47.491200 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zh8k9\" (UniqueName: \"kubernetes.io/projected/14c327e3-e880-4b49-b366-364a0395cc55-kube-api-access-zh8k9\") pod \"14c327e3-e880-4b49-b366-364a0395cc55\" (UID: \"14c327e3-e880-4b49-b366-364a0395cc55\") " Dec 03 14:30:47 crc kubenswrapper[4805]: I1203 14:30:47.491661 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/32920f09-a728-4b35-a5d7-65dae2d70834-combined-ca-bundle\") pod \"32920f09-a728-4b35-a5d7-65dae2d70834\" (UID: \"32920f09-a728-4b35-a5d7-65dae2d70834\") " Dec 03 14:30:47 crc kubenswrapper[4805]: I1203 14:30:47.491705 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/32920f09-a728-4b35-a5d7-65dae2d70834-config\") pod \"32920f09-a728-4b35-a5d7-65dae2d70834\" (UID: \"32920f09-a728-4b35-a5d7-65dae2d70834\") " Dec 03 14:30:47 crc kubenswrapper[4805]: I1203 14:30:47.496057 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/32920f09-a728-4b35-a5d7-65dae2d70834-kube-api-access-v5gd6" (OuterVolumeSpecName: "kube-api-access-v5gd6") pod "32920f09-a728-4b35-a5d7-65dae2d70834" (UID: "32920f09-a728-4b35-a5d7-65dae2d70834"). InnerVolumeSpecName "kube-api-access-v5gd6". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 14:30:47 crc kubenswrapper[4805]: I1203 14:30:47.496806 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/14c327e3-e880-4b49-b366-364a0395cc55-kube-api-access-zh8k9" (OuterVolumeSpecName: "kube-api-access-zh8k9") pod "14c327e3-e880-4b49-b366-364a0395cc55" (UID: "14c327e3-e880-4b49-b366-364a0395cc55"). InnerVolumeSpecName "kube-api-access-zh8k9". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 14:30:47 crc kubenswrapper[4805]: I1203 14:30:47.528091 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/32920f09-a728-4b35-a5d7-65dae2d70834-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "32920f09-a728-4b35-a5d7-65dae2d70834" (UID: "32920f09-a728-4b35-a5d7-65dae2d70834"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 14:30:47 crc kubenswrapper[4805]: I1203 14:30:47.530887 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/32920f09-a728-4b35-a5d7-65dae2d70834-config" (OuterVolumeSpecName: "config") pod "32920f09-a728-4b35-a5d7-65dae2d70834" (UID: "32920f09-a728-4b35-a5d7-65dae2d70834"). InnerVolumeSpecName "config". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 14:30:47 crc kubenswrapper[4805]: I1203 14:30:47.531993 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/14c327e3-e880-4b49-b366-364a0395cc55-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "14c327e3-e880-4b49-b366-364a0395cc55" (UID: "14c327e3-e880-4b49-b366-364a0395cc55"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 14:30:47 crc kubenswrapper[4805]: I1203 14:30:47.578374 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/14c327e3-e880-4b49-b366-364a0395cc55-config-data" (OuterVolumeSpecName: "config-data") pod "14c327e3-e880-4b49-b366-364a0395cc55" (UID: "14c327e3-e880-4b49-b366-364a0395cc55"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 14:30:47 crc kubenswrapper[4805]: I1203 14:30:47.593623 4805 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zh8k9\" (UniqueName: \"kubernetes.io/projected/14c327e3-e880-4b49-b366-364a0395cc55-kube-api-access-zh8k9\") on node \"crc\" DevicePath \"\"" Dec 03 14:30:47 crc kubenswrapper[4805]: I1203 14:30:47.593666 4805 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/32920f09-a728-4b35-a5d7-65dae2d70834-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 14:30:47 crc kubenswrapper[4805]: I1203 14:30:47.593676 4805 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/secret/32920f09-a728-4b35-a5d7-65dae2d70834-config\") on node \"crc\" DevicePath \"\"" Dec 03 14:30:47 crc kubenswrapper[4805]: I1203 14:30:47.593685 4805 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-v5gd6\" (UniqueName: \"kubernetes.io/projected/32920f09-a728-4b35-a5d7-65dae2d70834-kube-api-access-v5gd6\") on node \"crc\" DevicePath \"\"" Dec 03 14:30:47 crc kubenswrapper[4805]: I1203 14:30:47.593693 4805 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/14c327e3-e880-4b49-b366-364a0395cc55-config-data\") on node \"crc\" DevicePath \"\"" Dec 03 14:30:47 crc kubenswrapper[4805]: I1203 14:30:47.593740 4805 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/14c327e3-e880-4b49-b366-364a0395cc55-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 14:30:47 crc kubenswrapper[4805]: I1203 14:30:47.893416 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"8533d470-22f2-401a-a39d-c37391920d81","Type":"ContainerStarted","Data":"8cdd783ebc4c2832d701981f18a14850bb3c14ceb07034817106c55661990ccc"} Dec 03 14:30:47 crc kubenswrapper[4805]: I1203 14:30:47.895353 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-db-sync-jvqmk" event={"ID":"14c327e3-e880-4b49-b366-364a0395cc55","Type":"ContainerDied","Data":"1a8457670ac140f0c542b92f1e3130ca93dd7a523ed000d606773dd4046ddee3"} Dec 03 14:30:47 crc kubenswrapper[4805]: I1203 14:30:47.895401 4805 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="1a8457670ac140f0c542b92f1e3130ca93dd7a523ed000d606773dd4046ddee3" Dec 03 14:30:47 crc kubenswrapper[4805]: I1203 14:30:47.895478 4805 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/heat-db-sync-jvqmk" Dec 03 14:30:47 crc kubenswrapper[4805]: I1203 14:30:47.906171 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-gskh4" event={"ID":"6098937f-e3f6-45e8-a647-4994a79cd711","Type":"ContainerStarted","Data":"a8c52725eecea0a7c49a5c7b85e82a2cd4d97ad1c78cf047a931df09ab5530e8"} Dec 03 14:30:47 crc kubenswrapper[4805]: I1203 14:30:47.908733 4805 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-sync-b6k7w" Dec 03 14:30:47 crc kubenswrapper[4805]: I1203 14:30:47.909036 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-sync-b6k7w" event={"ID":"32920f09-a728-4b35-a5d7-65dae2d70834","Type":"ContainerDied","Data":"e197e76c91b87d530bb2b8c17e87b85b444572fd3c9af48e537d98f418a87313"} Dec 03 14:30:47 crc kubenswrapper[4805]: I1203 14:30:47.909082 4805 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="e197e76c91b87d530bb2b8c17e87b85b444572fd3c9af48e537d98f418a87313" Dec 03 14:30:47 crc kubenswrapper[4805]: I1203 14:30:47.926439 4805 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-585fb5cffc-n26gq"] Dec 03 14:30:47 crc kubenswrapper[4805]: E1203 14:30:47.927717 4805 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b4e050ed-47f5-4d9f-a288-b6d2a7ab294a" containerName="keystone-bootstrap" Dec 03 14:30:47 crc kubenswrapper[4805]: I1203 14:30:47.927739 4805 state_mem.go:107] "Deleted CPUSet assignment" podUID="b4e050ed-47f5-4d9f-a288-b6d2a7ab294a" containerName="keystone-bootstrap" Dec 03 14:30:47 crc kubenswrapper[4805]: E1203 14:30:47.927772 4805 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1ca67122-920a-43a6-a434-82ec573d0160" containerName="dnsmasq-dns" Dec 03 14:30:47 crc kubenswrapper[4805]: I1203 14:30:47.927779 4805 state_mem.go:107] "Deleted CPUSet assignment" podUID="1ca67122-920a-43a6-a434-82ec573d0160" containerName="dnsmasq-dns" Dec 03 14:30:47 crc kubenswrapper[4805]: E1203 14:30:47.927830 4805 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="32920f09-a728-4b35-a5d7-65dae2d70834" containerName="neutron-db-sync" Dec 03 14:30:47 crc kubenswrapper[4805]: I1203 14:30:47.927860 4805 state_mem.go:107] "Deleted CPUSet assignment" podUID="32920f09-a728-4b35-a5d7-65dae2d70834" containerName="neutron-db-sync" Dec 03 14:30:47 crc kubenswrapper[4805]: E1203 14:30:47.927874 4805 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1ca67122-920a-43a6-a434-82ec573d0160" containerName="init" Dec 03 14:30:47 crc kubenswrapper[4805]: I1203 14:30:47.927882 4805 state_mem.go:107] "Deleted CPUSet assignment" podUID="1ca67122-920a-43a6-a434-82ec573d0160" containerName="init" Dec 03 14:30:47 crc kubenswrapper[4805]: E1203 14:30:47.927914 4805 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="14c327e3-e880-4b49-b366-364a0395cc55" containerName="heat-db-sync" Dec 03 14:30:47 crc kubenswrapper[4805]: I1203 14:30:47.927923 4805 state_mem.go:107] "Deleted CPUSet assignment" podUID="14c327e3-e880-4b49-b366-364a0395cc55" containerName="heat-db-sync" Dec 03 14:30:47 crc kubenswrapper[4805]: E1203 14:30:47.927952 4805 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7c1b5525-de7b-4a8c-9062-bb9a68336989" containerName="placement-db-sync" Dec 03 14:30:47 crc kubenswrapper[4805]: I1203 14:30:47.927966 4805 state_mem.go:107] "Deleted CPUSet assignment" podUID="7c1b5525-de7b-4a8c-9062-bb9a68336989" containerName="placement-db-sync" Dec 03 14:30:47 crc kubenswrapper[4805]: I1203 14:30:47.928375 4805 memory_manager.go:354] "RemoveStaleState removing state" podUID="b4e050ed-47f5-4d9f-a288-b6d2a7ab294a" containerName="keystone-bootstrap" Dec 03 14:30:47 crc kubenswrapper[4805]: I1203 14:30:47.928410 4805 memory_manager.go:354] "RemoveStaleState removing state" podUID="7c1b5525-de7b-4a8c-9062-bb9a68336989" containerName="placement-db-sync" Dec 03 14:30:47 crc kubenswrapper[4805]: I1203 14:30:47.928418 4805 memory_manager.go:354] "RemoveStaleState removing state" podUID="14c327e3-e880-4b49-b366-364a0395cc55" containerName="heat-db-sync" Dec 03 14:30:47 crc kubenswrapper[4805]: I1203 14:30:47.928443 4805 memory_manager.go:354] "RemoveStaleState removing state" podUID="32920f09-a728-4b35-a5d7-65dae2d70834" containerName="neutron-db-sync" Dec 03 14:30:47 crc kubenswrapper[4805]: I1203 14:30:47.928468 4805 memory_manager.go:354] "RemoveStaleState removing state" podUID="1ca67122-920a-43a6-a434-82ec573d0160" containerName="dnsmasq-dns" Dec 03 14:30:47 crc kubenswrapper[4805]: I1203 14:30:47.929506 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-585fb5cffc-n26gq" Dec 03 14:30:47 crc kubenswrapper[4805]: I1203 14:30:47.942589 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-config-data" Dec 03 14:30:47 crc kubenswrapper[4805]: I1203 14:30:47.943407 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-keystone-internal-svc" Dec 03 14:30:47 crc kubenswrapper[4805]: I1203 14:30:47.950296 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone" Dec 03 14:30:47 crc kubenswrapper[4805]: I1203 14:30:47.951414 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-scripts" Dec 03 14:30:47 crc kubenswrapper[4805]: I1203 14:30:47.954513 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-keystone-dockercfg-5hcz5" Dec 03 14:30:47 crc kubenswrapper[4805]: I1203 14:30:47.971014 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-keystone-public-svc" Dec 03 14:30:47 crc kubenswrapper[4805]: I1203 14:30:47.973639 4805 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement-98cb4d7c6-cbjgt"] Dec 03 14:30:47 crc kubenswrapper[4805]: I1203 14:30:47.984715 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-98cb4d7c6-cbjgt" Dec 03 14:30:48 crc kubenswrapper[4805]: I1203 14:30:48.005820 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-placement-dockercfg-nmchz" Dec 03 14:30:48 crc kubenswrapper[4805]: I1203 14:30:48.005936 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-placement-internal-svc" Dec 03 14:30:48 crc kubenswrapper[4805]: I1203 14:30:48.006073 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-scripts" Dec 03 14:30:48 crc kubenswrapper[4805]: I1203 14:30:48.006290 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-placement-public-svc" Dec 03 14:30:48 crc kubenswrapper[4805]: I1203 14:30:48.006860 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-config-data" Dec 03 14:30:48 crc kubenswrapper[4805]: I1203 14:30:48.031914 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-585fb5cffc-n26gq"] Dec 03 14:30:48 crc kubenswrapper[4805]: I1203 14:30:48.054279 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-98cb4d7c6-cbjgt"] Dec 03 14:30:48 crc kubenswrapper[4805]: I1203 14:30:48.119186 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/d099a9d3-b941-499a-829f-7590e1fe7b0a-public-tls-certs\") pod \"placement-98cb4d7c6-cbjgt\" (UID: \"d099a9d3-b941-499a-829f-7590e1fe7b0a\") " pod="openstack/placement-98cb4d7c6-cbjgt" Dec 03 14:30:48 crc kubenswrapper[4805]: I1203 14:30:48.119413 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/3e723db7-3ea3-4360-b5b8-49f1e0524f8b-fernet-keys\") pod \"keystone-585fb5cffc-n26gq\" (UID: \"3e723db7-3ea3-4360-b5b8-49f1e0524f8b\") " pod="openstack/keystone-585fb5cffc-n26gq" Dec 03 14:30:48 crc kubenswrapper[4805]: I1203 14:30:48.119523 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/d099a9d3-b941-499a-829f-7590e1fe7b0a-internal-tls-certs\") pod \"placement-98cb4d7c6-cbjgt\" (UID: \"d099a9d3-b941-499a-829f-7590e1fe7b0a\") " pod="openstack/placement-98cb4d7c6-cbjgt" Dec 03 14:30:48 crc kubenswrapper[4805]: I1203 14:30:48.119600 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/3e723db7-3ea3-4360-b5b8-49f1e0524f8b-credential-keys\") pod \"keystone-585fb5cffc-n26gq\" (UID: \"3e723db7-3ea3-4360-b5b8-49f1e0524f8b\") " pod="openstack/keystone-585fb5cffc-n26gq" Dec 03 14:30:48 crc kubenswrapper[4805]: I1203 14:30:48.119922 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d099a9d3-b941-499a-829f-7590e1fe7b0a-scripts\") pod \"placement-98cb4d7c6-cbjgt\" (UID: \"d099a9d3-b941-499a-829f-7590e1fe7b0a\") " pod="openstack/placement-98cb4d7c6-cbjgt" Dec 03 14:30:48 crc kubenswrapper[4805]: I1203 14:30:48.120029 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/3e723db7-3ea3-4360-b5b8-49f1e0524f8b-internal-tls-certs\") pod \"keystone-585fb5cffc-n26gq\" (UID: \"3e723db7-3ea3-4360-b5b8-49f1e0524f8b\") " pod="openstack/keystone-585fb5cffc-n26gq" Dec 03 14:30:48 crc kubenswrapper[4805]: I1203 14:30:48.120144 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d099a9d3-b941-499a-829f-7590e1fe7b0a-combined-ca-bundle\") pod \"placement-98cb4d7c6-cbjgt\" (UID: \"d099a9d3-b941-499a-829f-7590e1fe7b0a\") " pod="openstack/placement-98cb4d7c6-cbjgt" Dec 03 14:30:48 crc kubenswrapper[4805]: I1203 14:30:48.122154 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-24vx5\" (UniqueName: \"kubernetes.io/projected/3e723db7-3ea3-4360-b5b8-49f1e0524f8b-kube-api-access-24vx5\") pod \"keystone-585fb5cffc-n26gq\" (UID: \"3e723db7-3ea3-4360-b5b8-49f1e0524f8b\") " pod="openstack/keystone-585fb5cffc-n26gq" Dec 03 14:30:48 crc kubenswrapper[4805]: I1203 14:30:48.122256 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-q8tnl\" (UniqueName: \"kubernetes.io/projected/d099a9d3-b941-499a-829f-7590e1fe7b0a-kube-api-access-q8tnl\") pod \"placement-98cb4d7c6-cbjgt\" (UID: \"d099a9d3-b941-499a-829f-7590e1fe7b0a\") " pod="openstack/placement-98cb4d7c6-cbjgt" Dec 03 14:30:48 crc kubenswrapper[4805]: I1203 14:30:48.122500 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3e723db7-3ea3-4360-b5b8-49f1e0524f8b-config-data\") pod \"keystone-585fb5cffc-n26gq\" (UID: \"3e723db7-3ea3-4360-b5b8-49f1e0524f8b\") " pod="openstack/keystone-585fb5cffc-n26gq" Dec 03 14:30:48 crc kubenswrapper[4805]: I1203 14:30:48.122635 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3e723db7-3ea3-4360-b5b8-49f1e0524f8b-combined-ca-bundle\") pod \"keystone-585fb5cffc-n26gq\" (UID: \"3e723db7-3ea3-4360-b5b8-49f1e0524f8b\") " pod="openstack/keystone-585fb5cffc-n26gq" Dec 03 14:30:48 crc kubenswrapper[4805]: I1203 14:30:48.122700 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/3e723db7-3ea3-4360-b5b8-49f1e0524f8b-public-tls-certs\") pod \"keystone-585fb5cffc-n26gq\" (UID: \"3e723db7-3ea3-4360-b5b8-49f1e0524f8b\") " pod="openstack/keystone-585fb5cffc-n26gq" Dec 03 14:30:48 crc kubenswrapper[4805]: I1203 14:30:48.122764 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3e723db7-3ea3-4360-b5b8-49f1e0524f8b-scripts\") pod \"keystone-585fb5cffc-n26gq\" (UID: \"3e723db7-3ea3-4360-b5b8-49f1e0524f8b\") " pod="openstack/keystone-585fb5cffc-n26gq" Dec 03 14:30:48 crc kubenswrapper[4805]: I1203 14:30:48.123015 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d099a9d3-b941-499a-829f-7590e1fe7b0a-config-data\") pod \"placement-98cb4d7c6-cbjgt\" (UID: \"d099a9d3-b941-499a-829f-7590e1fe7b0a\") " pod="openstack/placement-98cb4d7c6-cbjgt" Dec 03 14:30:48 crc kubenswrapper[4805]: I1203 14:30:48.123087 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d099a9d3-b941-499a-829f-7590e1fe7b0a-logs\") pod \"placement-98cb4d7c6-cbjgt\" (UID: \"d099a9d3-b941-499a-829f-7590e1fe7b0a\") " pod="openstack/placement-98cb4d7c6-cbjgt" Dec 03 14:30:48 crc kubenswrapper[4805]: I1203 14:30:48.146616 4805 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-84b966f6c9-zwsmd"] Dec 03 14:30:48 crc kubenswrapper[4805]: I1203 14:30:48.148426 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-84b966f6c9-zwsmd" Dec 03 14:30:48 crc kubenswrapper[4805]: I1203 14:30:48.154957 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-84b966f6c9-zwsmd"] Dec 03 14:30:48 crc kubenswrapper[4805]: I1203 14:30:48.224303 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d099a9d3-b941-499a-829f-7590e1fe7b0a-combined-ca-bundle\") pod \"placement-98cb4d7c6-cbjgt\" (UID: \"d099a9d3-b941-499a-829f-7590e1fe7b0a\") " pod="openstack/placement-98cb4d7c6-cbjgt" Dec 03 14:30:48 crc kubenswrapper[4805]: I1203 14:30:48.224349 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-24vx5\" (UniqueName: \"kubernetes.io/projected/3e723db7-3ea3-4360-b5b8-49f1e0524f8b-kube-api-access-24vx5\") pod \"keystone-585fb5cffc-n26gq\" (UID: \"3e723db7-3ea3-4360-b5b8-49f1e0524f8b\") " pod="openstack/keystone-585fb5cffc-n26gq" Dec 03 14:30:48 crc kubenswrapper[4805]: I1203 14:30:48.224374 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-q8tnl\" (UniqueName: \"kubernetes.io/projected/d099a9d3-b941-499a-829f-7590e1fe7b0a-kube-api-access-q8tnl\") pod \"placement-98cb4d7c6-cbjgt\" (UID: \"d099a9d3-b941-499a-829f-7590e1fe7b0a\") " pod="openstack/placement-98cb4d7c6-cbjgt" Dec 03 14:30:48 crc kubenswrapper[4805]: I1203 14:30:48.224393 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3e723db7-3ea3-4360-b5b8-49f1e0524f8b-config-data\") pod \"keystone-585fb5cffc-n26gq\" (UID: \"3e723db7-3ea3-4360-b5b8-49f1e0524f8b\") " pod="openstack/keystone-585fb5cffc-n26gq" Dec 03 14:30:48 crc kubenswrapper[4805]: I1203 14:30:48.224408 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3e723db7-3ea3-4360-b5b8-49f1e0524f8b-combined-ca-bundle\") pod \"keystone-585fb5cffc-n26gq\" (UID: \"3e723db7-3ea3-4360-b5b8-49f1e0524f8b\") " pod="openstack/keystone-585fb5cffc-n26gq" Dec 03 14:30:48 crc kubenswrapper[4805]: I1203 14:30:48.224422 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/3e723db7-3ea3-4360-b5b8-49f1e0524f8b-public-tls-certs\") pod \"keystone-585fb5cffc-n26gq\" (UID: \"3e723db7-3ea3-4360-b5b8-49f1e0524f8b\") " pod="openstack/keystone-585fb5cffc-n26gq" Dec 03 14:30:48 crc kubenswrapper[4805]: I1203 14:30:48.224439 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3e723db7-3ea3-4360-b5b8-49f1e0524f8b-scripts\") pod \"keystone-585fb5cffc-n26gq\" (UID: \"3e723db7-3ea3-4360-b5b8-49f1e0524f8b\") " pod="openstack/keystone-585fb5cffc-n26gq" Dec 03 14:30:48 crc kubenswrapper[4805]: I1203 14:30:48.224464 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d099a9d3-b941-499a-829f-7590e1fe7b0a-config-data\") pod \"placement-98cb4d7c6-cbjgt\" (UID: \"d099a9d3-b941-499a-829f-7590e1fe7b0a\") " pod="openstack/placement-98cb4d7c6-cbjgt" Dec 03 14:30:48 crc kubenswrapper[4805]: I1203 14:30:48.224497 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d099a9d3-b941-499a-829f-7590e1fe7b0a-logs\") pod \"placement-98cb4d7c6-cbjgt\" (UID: \"d099a9d3-b941-499a-829f-7590e1fe7b0a\") " pod="openstack/placement-98cb4d7c6-cbjgt" Dec 03 14:30:48 crc kubenswrapper[4805]: I1203 14:30:48.224577 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/d099a9d3-b941-499a-829f-7590e1fe7b0a-public-tls-certs\") pod \"placement-98cb4d7c6-cbjgt\" (UID: \"d099a9d3-b941-499a-829f-7590e1fe7b0a\") " pod="openstack/placement-98cb4d7c6-cbjgt" Dec 03 14:30:48 crc kubenswrapper[4805]: I1203 14:30:48.224593 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/3e723db7-3ea3-4360-b5b8-49f1e0524f8b-fernet-keys\") pod \"keystone-585fb5cffc-n26gq\" (UID: \"3e723db7-3ea3-4360-b5b8-49f1e0524f8b\") " pod="openstack/keystone-585fb5cffc-n26gq" Dec 03 14:30:48 crc kubenswrapper[4805]: I1203 14:30:48.224619 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/d099a9d3-b941-499a-829f-7590e1fe7b0a-internal-tls-certs\") pod \"placement-98cb4d7c6-cbjgt\" (UID: \"d099a9d3-b941-499a-829f-7590e1fe7b0a\") " pod="openstack/placement-98cb4d7c6-cbjgt" Dec 03 14:30:48 crc kubenswrapper[4805]: I1203 14:30:48.224638 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/3e723db7-3ea3-4360-b5b8-49f1e0524f8b-credential-keys\") pod \"keystone-585fb5cffc-n26gq\" (UID: \"3e723db7-3ea3-4360-b5b8-49f1e0524f8b\") " pod="openstack/keystone-585fb5cffc-n26gq" Dec 03 14:30:48 crc kubenswrapper[4805]: I1203 14:30:48.224661 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d099a9d3-b941-499a-829f-7590e1fe7b0a-scripts\") pod \"placement-98cb4d7c6-cbjgt\" (UID: \"d099a9d3-b941-499a-829f-7590e1fe7b0a\") " pod="openstack/placement-98cb4d7c6-cbjgt" Dec 03 14:30:48 crc kubenswrapper[4805]: I1203 14:30:48.224681 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/3e723db7-3ea3-4360-b5b8-49f1e0524f8b-internal-tls-certs\") pod \"keystone-585fb5cffc-n26gq\" (UID: \"3e723db7-3ea3-4360-b5b8-49f1e0524f8b\") " pod="openstack/keystone-585fb5cffc-n26gq" Dec 03 14:30:48 crc kubenswrapper[4805]: I1203 14:30:48.235275 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d099a9d3-b941-499a-829f-7590e1fe7b0a-combined-ca-bundle\") pod \"placement-98cb4d7c6-cbjgt\" (UID: \"d099a9d3-b941-499a-829f-7590e1fe7b0a\") " pod="openstack/placement-98cb4d7c6-cbjgt" Dec 03 14:30:48 crc kubenswrapper[4805]: I1203 14:30:48.235687 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d099a9d3-b941-499a-829f-7590e1fe7b0a-logs\") pod \"placement-98cb4d7c6-cbjgt\" (UID: \"d099a9d3-b941-499a-829f-7590e1fe7b0a\") " pod="openstack/placement-98cb4d7c6-cbjgt" Dec 03 14:30:48 crc kubenswrapper[4805]: I1203 14:30:48.245056 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3e723db7-3ea3-4360-b5b8-49f1e0524f8b-scripts\") pod \"keystone-585fb5cffc-n26gq\" (UID: \"3e723db7-3ea3-4360-b5b8-49f1e0524f8b\") " pod="openstack/keystone-585fb5cffc-n26gq" Dec 03 14:30:48 crc kubenswrapper[4805]: I1203 14:30:48.245818 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/3e723db7-3ea3-4360-b5b8-49f1e0524f8b-internal-tls-certs\") pod \"keystone-585fb5cffc-n26gq\" (UID: \"3e723db7-3ea3-4360-b5b8-49f1e0524f8b\") " pod="openstack/keystone-585fb5cffc-n26gq" Dec 03 14:30:48 crc kubenswrapper[4805]: I1203 14:30:48.253316 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/d099a9d3-b941-499a-829f-7590e1fe7b0a-public-tls-certs\") pod \"placement-98cb4d7c6-cbjgt\" (UID: \"d099a9d3-b941-499a-829f-7590e1fe7b0a\") " pod="openstack/placement-98cb4d7c6-cbjgt" Dec 03 14:30:48 crc kubenswrapper[4805]: I1203 14:30:48.262191 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d099a9d3-b941-499a-829f-7590e1fe7b0a-scripts\") pod \"placement-98cb4d7c6-cbjgt\" (UID: \"d099a9d3-b941-499a-829f-7590e1fe7b0a\") " pod="openstack/placement-98cb4d7c6-cbjgt" Dec 03 14:30:48 crc kubenswrapper[4805]: I1203 14:30:48.267469 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/3e723db7-3ea3-4360-b5b8-49f1e0524f8b-credential-keys\") pod \"keystone-585fb5cffc-n26gq\" (UID: \"3e723db7-3ea3-4360-b5b8-49f1e0524f8b\") " pod="openstack/keystone-585fb5cffc-n26gq" Dec 03 14:30:48 crc kubenswrapper[4805]: I1203 14:30:48.268263 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-q8tnl\" (UniqueName: \"kubernetes.io/projected/d099a9d3-b941-499a-829f-7590e1fe7b0a-kube-api-access-q8tnl\") pod \"placement-98cb4d7c6-cbjgt\" (UID: \"d099a9d3-b941-499a-829f-7590e1fe7b0a\") " pod="openstack/placement-98cb4d7c6-cbjgt" Dec 03 14:30:48 crc kubenswrapper[4805]: I1203 14:30:48.269298 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/3e723db7-3ea3-4360-b5b8-49f1e0524f8b-public-tls-certs\") pod \"keystone-585fb5cffc-n26gq\" (UID: \"3e723db7-3ea3-4360-b5b8-49f1e0524f8b\") " pod="openstack/keystone-585fb5cffc-n26gq" Dec 03 14:30:48 crc kubenswrapper[4805]: I1203 14:30:48.273655 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-24vx5\" (UniqueName: \"kubernetes.io/projected/3e723db7-3ea3-4360-b5b8-49f1e0524f8b-kube-api-access-24vx5\") pod \"keystone-585fb5cffc-n26gq\" (UID: \"3e723db7-3ea3-4360-b5b8-49f1e0524f8b\") " pod="openstack/keystone-585fb5cffc-n26gq" Dec 03 14:30:48 crc kubenswrapper[4805]: I1203 14:30:48.274435 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d099a9d3-b941-499a-829f-7590e1fe7b0a-config-data\") pod \"placement-98cb4d7c6-cbjgt\" (UID: \"d099a9d3-b941-499a-829f-7590e1fe7b0a\") " pod="openstack/placement-98cb4d7c6-cbjgt" Dec 03 14:30:48 crc kubenswrapper[4805]: I1203 14:30:48.274452 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/3e723db7-3ea3-4360-b5b8-49f1e0524f8b-fernet-keys\") pod \"keystone-585fb5cffc-n26gq\" (UID: \"3e723db7-3ea3-4360-b5b8-49f1e0524f8b\") " pod="openstack/keystone-585fb5cffc-n26gq" Dec 03 14:30:48 crc kubenswrapper[4805]: I1203 14:30:48.275359 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3e723db7-3ea3-4360-b5b8-49f1e0524f8b-config-data\") pod \"keystone-585fb5cffc-n26gq\" (UID: \"3e723db7-3ea3-4360-b5b8-49f1e0524f8b\") " pod="openstack/keystone-585fb5cffc-n26gq" Dec 03 14:30:48 crc kubenswrapper[4805]: I1203 14:30:48.276207 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/d099a9d3-b941-499a-829f-7590e1fe7b0a-internal-tls-certs\") pod \"placement-98cb4d7c6-cbjgt\" (UID: \"d099a9d3-b941-499a-829f-7590e1fe7b0a\") " pod="openstack/placement-98cb4d7c6-cbjgt" Dec 03 14:30:48 crc kubenswrapper[4805]: I1203 14:30:48.278199 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3e723db7-3ea3-4360-b5b8-49f1e0524f8b-combined-ca-bundle\") pod \"keystone-585fb5cffc-n26gq\" (UID: \"3e723db7-3ea3-4360-b5b8-49f1e0524f8b\") " pod="openstack/keystone-585fb5cffc-n26gq" Dec 03 14:30:48 crc kubenswrapper[4805]: I1203 14:30:48.295016 4805 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-75f57c664d-4tsg2"] Dec 03 14:30:48 crc kubenswrapper[4805]: I1203 14:30:48.296626 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-75f57c664d-4tsg2" Dec 03 14:30:48 crc kubenswrapper[4805]: I1203 14:30:48.300763 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-neutron-ovndbs" Dec 03 14:30:48 crc kubenswrapper[4805]: I1203 14:30:48.301033 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-httpd-config" Dec 03 14:30:48 crc kubenswrapper[4805]: I1203 14:30:48.301161 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-config" Dec 03 14:30:48 crc kubenswrapper[4805]: I1203 14:30:48.301359 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-neutron-dockercfg-68v4l" Dec 03 14:30:48 crc kubenswrapper[4805]: I1203 14:30:48.308169 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-585fb5cffc-n26gq" Dec 03 14:30:48 crc kubenswrapper[4805]: I1203 14:30:48.313444 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-75f57c664d-4tsg2"] Dec 03 14:30:48 crc kubenswrapper[4805]: I1203 14:30:48.326303 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b3ca6772-26a0-410c-9609-1f49d08b1864-config\") pod \"dnsmasq-dns-84b966f6c9-zwsmd\" (UID: \"b3ca6772-26a0-410c-9609-1f49d08b1864\") " pod="openstack/dnsmasq-dns-84b966f6c9-zwsmd" Dec 03 14:30:48 crc kubenswrapper[4805]: I1203 14:30:48.326364 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/b3ca6772-26a0-410c-9609-1f49d08b1864-ovsdbserver-sb\") pod \"dnsmasq-dns-84b966f6c9-zwsmd\" (UID: \"b3ca6772-26a0-410c-9609-1f49d08b1864\") " pod="openstack/dnsmasq-dns-84b966f6c9-zwsmd" Dec 03 14:30:48 crc kubenswrapper[4805]: I1203 14:30:48.326395 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7n7vx\" (UniqueName: \"kubernetes.io/projected/b3ca6772-26a0-410c-9609-1f49d08b1864-kube-api-access-7n7vx\") pod \"dnsmasq-dns-84b966f6c9-zwsmd\" (UID: \"b3ca6772-26a0-410c-9609-1f49d08b1864\") " pod="openstack/dnsmasq-dns-84b966f6c9-zwsmd" Dec 03 14:30:48 crc kubenswrapper[4805]: I1203 14:30:48.326432 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/b3ca6772-26a0-410c-9609-1f49d08b1864-dns-svc\") pod \"dnsmasq-dns-84b966f6c9-zwsmd\" (UID: \"b3ca6772-26a0-410c-9609-1f49d08b1864\") " pod="openstack/dnsmasq-dns-84b966f6c9-zwsmd" Dec 03 14:30:48 crc kubenswrapper[4805]: I1203 14:30:48.326482 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/b3ca6772-26a0-410c-9609-1f49d08b1864-ovsdbserver-nb\") pod \"dnsmasq-dns-84b966f6c9-zwsmd\" (UID: \"b3ca6772-26a0-410c-9609-1f49d08b1864\") " pod="openstack/dnsmasq-dns-84b966f6c9-zwsmd" Dec 03 14:30:48 crc kubenswrapper[4805]: I1203 14:30:48.326516 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/b3ca6772-26a0-410c-9609-1f49d08b1864-dns-swift-storage-0\") pod \"dnsmasq-dns-84b966f6c9-zwsmd\" (UID: \"b3ca6772-26a0-410c-9609-1f49d08b1864\") " pod="openstack/dnsmasq-dns-84b966f6c9-zwsmd" Dec 03 14:30:48 crc kubenswrapper[4805]: I1203 14:30:48.334069 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-98cb4d7c6-cbjgt" Dec 03 14:30:48 crc kubenswrapper[4805]: I1203 14:30:48.427763 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/318792ec-dc8b-4a37-92ab-d31773d66350-combined-ca-bundle\") pod \"neutron-75f57c664d-4tsg2\" (UID: \"318792ec-dc8b-4a37-92ab-d31773d66350\") " pod="openstack/neutron-75f57c664d-4tsg2" Dec 03 14:30:48 crc kubenswrapper[4805]: I1203 14:30:48.428024 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/b3ca6772-26a0-410c-9609-1f49d08b1864-dns-svc\") pod \"dnsmasq-dns-84b966f6c9-zwsmd\" (UID: \"b3ca6772-26a0-410c-9609-1f49d08b1864\") " pod="openstack/dnsmasq-dns-84b966f6c9-zwsmd" Dec 03 14:30:48 crc kubenswrapper[4805]: I1203 14:30:48.428061 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-28nqz\" (UniqueName: \"kubernetes.io/projected/318792ec-dc8b-4a37-92ab-d31773d66350-kube-api-access-28nqz\") pod \"neutron-75f57c664d-4tsg2\" (UID: \"318792ec-dc8b-4a37-92ab-d31773d66350\") " pod="openstack/neutron-75f57c664d-4tsg2" Dec 03 14:30:48 crc kubenswrapper[4805]: I1203 14:30:48.428094 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/b3ca6772-26a0-410c-9609-1f49d08b1864-ovsdbserver-nb\") pod \"dnsmasq-dns-84b966f6c9-zwsmd\" (UID: \"b3ca6772-26a0-410c-9609-1f49d08b1864\") " pod="openstack/dnsmasq-dns-84b966f6c9-zwsmd" Dec 03 14:30:48 crc kubenswrapper[4805]: I1203 14:30:48.428114 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/b3ca6772-26a0-410c-9609-1f49d08b1864-dns-swift-storage-0\") pod \"dnsmasq-dns-84b966f6c9-zwsmd\" (UID: \"b3ca6772-26a0-410c-9609-1f49d08b1864\") " pod="openstack/dnsmasq-dns-84b966f6c9-zwsmd" Dec 03 14:30:48 crc kubenswrapper[4805]: I1203 14:30:48.428161 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/318792ec-dc8b-4a37-92ab-d31773d66350-config\") pod \"neutron-75f57c664d-4tsg2\" (UID: \"318792ec-dc8b-4a37-92ab-d31773d66350\") " pod="openstack/neutron-75f57c664d-4tsg2" Dec 03 14:30:48 crc kubenswrapper[4805]: I1203 14:30:48.428181 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/318792ec-dc8b-4a37-92ab-d31773d66350-httpd-config\") pod \"neutron-75f57c664d-4tsg2\" (UID: \"318792ec-dc8b-4a37-92ab-d31773d66350\") " pod="openstack/neutron-75f57c664d-4tsg2" Dec 03 14:30:48 crc kubenswrapper[4805]: I1203 14:30:48.428410 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b3ca6772-26a0-410c-9609-1f49d08b1864-config\") pod \"dnsmasq-dns-84b966f6c9-zwsmd\" (UID: \"b3ca6772-26a0-410c-9609-1f49d08b1864\") " pod="openstack/dnsmasq-dns-84b966f6c9-zwsmd" Dec 03 14:30:48 crc kubenswrapper[4805]: I1203 14:30:48.428765 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/b3ca6772-26a0-410c-9609-1f49d08b1864-ovsdbserver-sb\") pod \"dnsmasq-dns-84b966f6c9-zwsmd\" (UID: \"b3ca6772-26a0-410c-9609-1f49d08b1864\") " pod="openstack/dnsmasq-dns-84b966f6c9-zwsmd" Dec 03 14:30:48 crc kubenswrapper[4805]: I1203 14:30:48.429052 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/b3ca6772-26a0-410c-9609-1f49d08b1864-dns-swift-storage-0\") pod \"dnsmasq-dns-84b966f6c9-zwsmd\" (UID: \"b3ca6772-26a0-410c-9609-1f49d08b1864\") " pod="openstack/dnsmasq-dns-84b966f6c9-zwsmd" Dec 03 14:30:48 crc kubenswrapper[4805]: I1203 14:30:48.429308 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7n7vx\" (UniqueName: \"kubernetes.io/projected/b3ca6772-26a0-410c-9609-1f49d08b1864-kube-api-access-7n7vx\") pod \"dnsmasq-dns-84b966f6c9-zwsmd\" (UID: \"b3ca6772-26a0-410c-9609-1f49d08b1864\") " pod="openstack/dnsmasq-dns-84b966f6c9-zwsmd" Dec 03 14:30:48 crc kubenswrapper[4805]: I1203 14:30:48.429348 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/318792ec-dc8b-4a37-92ab-d31773d66350-ovndb-tls-certs\") pod \"neutron-75f57c664d-4tsg2\" (UID: \"318792ec-dc8b-4a37-92ab-d31773d66350\") " pod="openstack/neutron-75f57c664d-4tsg2" Dec 03 14:30:48 crc kubenswrapper[4805]: I1203 14:30:48.429349 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/b3ca6772-26a0-410c-9609-1f49d08b1864-ovsdbserver-nb\") pod \"dnsmasq-dns-84b966f6c9-zwsmd\" (UID: \"b3ca6772-26a0-410c-9609-1f49d08b1864\") " pod="openstack/dnsmasq-dns-84b966f6c9-zwsmd" Dec 03 14:30:48 crc kubenswrapper[4805]: I1203 14:30:48.430051 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b3ca6772-26a0-410c-9609-1f49d08b1864-config\") pod \"dnsmasq-dns-84b966f6c9-zwsmd\" (UID: \"b3ca6772-26a0-410c-9609-1f49d08b1864\") " pod="openstack/dnsmasq-dns-84b966f6c9-zwsmd" Dec 03 14:30:48 crc kubenswrapper[4805]: I1203 14:30:48.430285 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/b3ca6772-26a0-410c-9609-1f49d08b1864-dns-svc\") pod \"dnsmasq-dns-84b966f6c9-zwsmd\" (UID: \"b3ca6772-26a0-410c-9609-1f49d08b1864\") " pod="openstack/dnsmasq-dns-84b966f6c9-zwsmd" Dec 03 14:30:48 crc kubenswrapper[4805]: I1203 14:30:48.442721 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/b3ca6772-26a0-410c-9609-1f49d08b1864-ovsdbserver-sb\") pod \"dnsmasq-dns-84b966f6c9-zwsmd\" (UID: \"b3ca6772-26a0-410c-9609-1f49d08b1864\") " pod="openstack/dnsmasq-dns-84b966f6c9-zwsmd" Dec 03 14:30:48 crc kubenswrapper[4805]: I1203 14:30:48.464669 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7n7vx\" (UniqueName: \"kubernetes.io/projected/b3ca6772-26a0-410c-9609-1f49d08b1864-kube-api-access-7n7vx\") pod \"dnsmasq-dns-84b966f6c9-zwsmd\" (UID: \"b3ca6772-26a0-410c-9609-1f49d08b1864\") " pod="openstack/dnsmasq-dns-84b966f6c9-zwsmd" Dec 03 14:30:48 crc kubenswrapper[4805]: I1203 14:30:48.531089 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-28nqz\" (UniqueName: \"kubernetes.io/projected/318792ec-dc8b-4a37-92ab-d31773d66350-kube-api-access-28nqz\") pod \"neutron-75f57c664d-4tsg2\" (UID: \"318792ec-dc8b-4a37-92ab-d31773d66350\") " pod="openstack/neutron-75f57c664d-4tsg2" Dec 03 14:30:48 crc kubenswrapper[4805]: I1203 14:30:48.531180 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/318792ec-dc8b-4a37-92ab-d31773d66350-config\") pod \"neutron-75f57c664d-4tsg2\" (UID: \"318792ec-dc8b-4a37-92ab-d31773d66350\") " pod="openstack/neutron-75f57c664d-4tsg2" Dec 03 14:30:48 crc kubenswrapper[4805]: I1203 14:30:48.531202 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/318792ec-dc8b-4a37-92ab-d31773d66350-httpd-config\") pod \"neutron-75f57c664d-4tsg2\" (UID: \"318792ec-dc8b-4a37-92ab-d31773d66350\") " pod="openstack/neutron-75f57c664d-4tsg2" Dec 03 14:30:48 crc kubenswrapper[4805]: I1203 14:30:48.531280 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/318792ec-dc8b-4a37-92ab-d31773d66350-ovndb-tls-certs\") pod \"neutron-75f57c664d-4tsg2\" (UID: \"318792ec-dc8b-4a37-92ab-d31773d66350\") " pod="openstack/neutron-75f57c664d-4tsg2" Dec 03 14:30:48 crc kubenswrapper[4805]: I1203 14:30:48.531310 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/318792ec-dc8b-4a37-92ab-d31773d66350-combined-ca-bundle\") pod \"neutron-75f57c664d-4tsg2\" (UID: \"318792ec-dc8b-4a37-92ab-d31773d66350\") " pod="openstack/neutron-75f57c664d-4tsg2" Dec 03 14:30:48 crc kubenswrapper[4805]: I1203 14:30:48.538411 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/secret/318792ec-dc8b-4a37-92ab-d31773d66350-config\") pod \"neutron-75f57c664d-4tsg2\" (UID: \"318792ec-dc8b-4a37-92ab-d31773d66350\") " pod="openstack/neutron-75f57c664d-4tsg2" Dec 03 14:30:48 crc kubenswrapper[4805]: I1203 14:30:48.546104 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/318792ec-dc8b-4a37-92ab-d31773d66350-httpd-config\") pod \"neutron-75f57c664d-4tsg2\" (UID: \"318792ec-dc8b-4a37-92ab-d31773d66350\") " pod="openstack/neutron-75f57c664d-4tsg2" Dec 03 14:30:48 crc kubenswrapper[4805]: I1203 14:30:48.552450 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-28nqz\" (UniqueName: \"kubernetes.io/projected/318792ec-dc8b-4a37-92ab-d31773d66350-kube-api-access-28nqz\") pod \"neutron-75f57c664d-4tsg2\" (UID: \"318792ec-dc8b-4a37-92ab-d31773d66350\") " pod="openstack/neutron-75f57c664d-4tsg2" Dec 03 14:30:48 crc kubenswrapper[4805]: I1203 14:30:48.556943 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/318792ec-dc8b-4a37-92ab-d31773d66350-ovndb-tls-certs\") pod \"neutron-75f57c664d-4tsg2\" (UID: \"318792ec-dc8b-4a37-92ab-d31773d66350\") " pod="openstack/neutron-75f57c664d-4tsg2" Dec 03 14:30:48 crc kubenswrapper[4805]: I1203 14:30:48.564568 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/318792ec-dc8b-4a37-92ab-d31773d66350-combined-ca-bundle\") pod \"neutron-75f57c664d-4tsg2\" (UID: \"318792ec-dc8b-4a37-92ab-d31773d66350\") " pod="openstack/neutron-75f57c664d-4tsg2" Dec 03 14:30:48 crc kubenswrapper[4805]: I1203 14:30:48.689651 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-84b966f6c9-zwsmd" Dec 03 14:30:48 crc kubenswrapper[4805]: I1203 14:30:48.698447 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-75f57c664d-4tsg2" Dec 03 14:30:48 crc kubenswrapper[4805]: I1203 14:30:48.911111 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-98cb4d7c6-cbjgt"] Dec 03 14:30:48 crc kubenswrapper[4805]: I1203 14:30:48.977268 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-sync-74tz4" event={"ID":"13ba03f0-acc2-4fb8-90a9-583b4f8694cb","Type":"ContainerStarted","Data":"76a9c6e35bfc792baf01e7282b5b7319e4e528a8abaa7f19a90eb5f470bdeab4"} Dec 03 14:30:49 crc kubenswrapper[4805]: I1203 14:30:49.012909 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-585fb5cffc-n26gq"] Dec 03 14:30:49 crc kubenswrapper[4805]: I1203 14:30:49.015940 4805 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-db-sync-74tz4" podStartSLOduration=1.914708845 podStartE2EDuration="41.015918524s" podCreationTimestamp="2025-12-03 14:30:08 +0000 UTC" firstStartedPulling="2025-12-03 14:30:09.051796446 +0000 UTC m=+1238.714713369" lastFinishedPulling="2025-12-03 14:30:48.153006125 +0000 UTC m=+1277.815923048" observedRunningTime="2025-12-03 14:30:48.994602827 +0000 UTC m=+1278.657519760" watchObservedRunningTime="2025-12-03 14:30:49.015918524 +0000 UTC m=+1278.678835447" Dec 03 14:30:49 crc kubenswrapper[4805]: I1203 14:30:49.453570 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-75f57c664d-4tsg2"] Dec 03 14:30:49 crc kubenswrapper[4805]: I1203 14:30:49.468460 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-84b966f6c9-zwsmd"] Dec 03 14:30:50 crc kubenswrapper[4805]: I1203 14:30:50.027161 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-84b966f6c9-zwsmd" event={"ID":"b3ca6772-26a0-410c-9609-1f49d08b1864","Type":"ContainerStarted","Data":"96e6ebbb257a1732acc3947ee1476daee89061ce93c32795c04d4108a648f662"} Dec 03 14:30:50 crc kubenswrapper[4805]: I1203 14:30:50.036045 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-75f57c664d-4tsg2" event={"ID":"318792ec-dc8b-4a37-92ab-d31773d66350","Type":"ContainerStarted","Data":"3c8b8134864ece92d0ee18d4cf46deec6ae2c8b4b4887927a2459f0ddac12a07"} Dec 03 14:30:50 crc kubenswrapper[4805]: I1203 14:30:50.075704 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-98cb4d7c6-cbjgt" event={"ID":"d099a9d3-b941-499a-829f-7590e1fe7b0a","Type":"ContainerStarted","Data":"ca39f232c8f42441c998f7db00c4b917e9d9eaf4ce299bbff39e7386aea7bb9e"} Dec 03 14:30:50 crc kubenswrapper[4805]: I1203 14:30:50.075747 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-98cb4d7c6-cbjgt" event={"ID":"d099a9d3-b941-499a-829f-7590e1fe7b0a","Type":"ContainerStarted","Data":"763a3c94defaa90df136a071d306b8fefd647c09cb22e6ff5208799080de65ff"} Dec 03 14:30:50 crc kubenswrapper[4805]: I1203 14:30:50.075756 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-98cb4d7c6-cbjgt" event={"ID":"d099a9d3-b941-499a-829f-7590e1fe7b0a","Type":"ContainerStarted","Data":"68ee7705c58b5ad088e44397c3865400b3b7d94469e6ce2b814ef1dd6e11e332"} Dec 03 14:30:50 crc kubenswrapper[4805]: I1203 14:30:50.076303 4805 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/placement-98cb4d7c6-cbjgt" Dec 03 14:30:50 crc kubenswrapper[4805]: I1203 14:30:50.076372 4805 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/placement-98cb4d7c6-cbjgt" Dec 03 14:30:50 crc kubenswrapper[4805]: I1203 14:30:50.089054 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-585fb5cffc-n26gq" event={"ID":"3e723db7-3ea3-4360-b5b8-49f1e0524f8b","Type":"ContainerStarted","Data":"a72d8091b22550ebb8d1576a4ea07d22bbd6483b7c8c3b2cbc30bdcca551b9a6"} Dec 03 14:30:50 crc kubenswrapper[4805]: I1203 14:30:50.089119 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-585fb5cffc-n26gq" event={"ID":"3e723db7-3ea3-4360-b5b8-49f1e0524f8b","Type":"ContainerStarted","Data":"8a87fa6197873d755c774ca37a8375cb8e52271f52c6fca068cefd34e64dd785"} Dec 03 14:30:50 crc kubenswrapper[4805]: I1203 14:30:50.090016 4805 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/keystone-585fb5cffc-n26gq" Dec 03 14:30:50 crc kubenswrapper[4805]: I1203 14:30:50.119906 4805 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/placement-98cb4d7c6-cbjgt" podStartSLOduration=3.119887607 podStartE2EDuration="3.119887607s" podCreationTimestamp="2025-12-03 14:30:47 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 14:30:50.114974934 +0000 UTC m=+1279.777891857" watchObservedRunningTime="2025-12-03 14:30:50.119887607 +0000 UTC m=+1279.782804530" Dec 03 14:30:50 crc kubenswrapper[4805]: I1203 14:30:50.136777 4805 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-585fb5cffc-n26gq" podStartSLOduration=3.136761944 podStartE2EDuration="3.136761944s" podCreationTimestamp="2025-12-03 14:30:47 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 14:30:50.135341966 +0000 UTC m=+1279.798258879" watchObservedRunningTime="2025-12-03 14:30:50.136761944 +0000 UTC m=+1279.799678867" Dec 03 14:30:50 crc kubenswrapper[4805]: I1203 14:30:50.539542 4805 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-internal-api-0" Dec 03 14:30:50 crc kubenswrapper[4805]: I1203 14:30:50.539933 4805 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Dec 03 14:30:50 crc kubenswrapper[4805]: I1203 14:30:50.542109 4805 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-internal-api-0" Dec 03 14:30:50 crc kubenswrapper[4805]: I1203 14:30:50.708012 4805 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-9764b468f-7sr7n"] Dec 03 14:30:50 crc kubenswrapper[4805]: I1203 14:30:50.709494 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-9764b468f-7sr7n" Dec 03 14:30:50 crc kubenswrapper[4805]: I1203 14:30:50.715164 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-neutron-internal-svc" Dec 03 14:30:50 crc kubenswrapper[4805]: I1203 14:30:50.715230 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-neutron-public-svc" Dec 03 14:30:50 crc kubenswrapper[4805]: I1203 14:30:50.739248 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-9764b468f-7sr7n"] Dec 03 14:30:50 crc kubenswrapper[4805]: I1203 14:30:50.787412 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/31537e49-7f76-4c8e-8c14-3fe55f9615b9-combined-ca-bundle\") pod \"neutron-9764b468f-7sr7n\" (UID: \"31537e49-7f76-4c8e-8c14-3fe55f9615b9\") " pod="openstack/neutron-9764b468f-7sr7n" Dec 03 14:30:50 crc kubenswrapper[4805]: I1203 14:30:50.787532 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/31537e49-7f76-4c8e-8c14-3fe55f9615b9-config\") pod \"neutron-9764b468f-7sr7n\" (UID: \"31537e49-7f76-4c8e-8c14-3fe55f9615b9\") " pod="openstack/neutron-9764b468f-7sr7n" Dec 03 14:30:50 crc kubenswrapper[4805]: I1203 14:30:50.787554 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/31537e49-7f76-4c8e-8c14-3fe55f9615b9-ovndb-tls-certs\") pod \"neutron-9764b468f-7sr7n\" (UID: \"31537e49-7f76-4c8e-8c14-3fe55f9615b9\") " pod="openstack/neutron-9764b468f-7sr7n" Dec 03 14:30:50 crc kubenswrapper[4805]: I1203 14:30:50.787576 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/31537e49-7f76-4c8e-8c14-3fe55f9615b9-httpd-config\") pod \"neutron-9764b468f-7sr7n\" (UID: \"31537e49-7f76-4c8e-8c14-3fe55f9615b9\") " pod="openstack/neutron-9764b468f-7sr7n" Dec 03 14:30:50 crc kubenswrapper[4805]: I1203 14:30:50.787601 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/31537e49-7f76-4c8e-8c14-3fe55f9615b9-public-tls-certs\") pod \"neutron-9764b468f-7sr7n\" (UID: \"31537e49-7f76-4c8e-8c14-3fe55f9615b9\") " pod="openstack/neutron-9764b468f-7sr7n" Dec 03 14:30:50 crc kubenswrapper[4805]: I1203 14:30:50.787619 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5sz66\" (UniqueName: \"kubernetes.io/projected/31537e49-7f76-4c8e-8c14-3fe55f9615b9-kube-api-access-5sz66\") pod \"neutron-9764b468f-7sr7n\" (UID: \"31537e49-7f76-4c8e-8c14-3fe55f9615b9\") " pod="openstack/neutron-9764b468f-7sr7n" Dec 03 14:30:50 crc kubenswrapper[4805]: I1203 14:30:50.787947 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/31537e49-7f76-4c8e-8c14-3fe55f9615b9-internal-tls-certs\") pod \"neutron-9764b468f-7sr7n\" (UID: \"31537e49-7f76-4c8e-8c14-3fe55f9615b9\") " pod="openstack/neutron-9764b468f-7sr7n" Dec 03 14:30:50 crc kubenswrapper[4805]: I1203 14:30:50.889810 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/31537e49-7f76-4c8e-8c14-3fe55f9615b9-public-tls-certs\") pod \"neutron-9764b468f-7sr7n\" (UID: \"31537e49-7f76-4c8e-8c14-3fe55f9615b9\") " pod="openstack/neutron-9764b468f-7sr7n" Dec 03 14:30:50 crc kubenswrapper[4805]: I1203 14:30:50.889866 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5sz66\" (UniqueName: \"kubernetes.io/projected/31537e49-7f76-4c8e-8c14-3fe55f9615b9-kube-api-access-5sz66\") pod \"neutron-9764b468f-7sr7n\" (UID: \"31537e49-7f76-4c8e-8c14-3fe55f9615b9\") " pod="openstack/neutron-9764b468f-7sr7n" Dec 03 14:30:50 crc kubenswrapper[4805]: I1203 14:30:50.889933 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/31537e49-7f76-4c8e-8c14-3fe55f9615b9-internal-tls-certs\") pod \"neutron-9764b468f-7sr7n\" (UID: \"31537e49-7f76-4c8e-8c14-3fe55f9615b9\") " pod="openstack/neutron-9764b468f-7sr7n" Dec 03 14:30:50 crc kubenswrapper[4805]: I1203 14:30:50.889996 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/31537e49-7f76-4c8e-8c14-3fe55f9615b9-combined-ca-bundle\") pod \"neutron-9764b468f-7sr7n\" (UID: \"31537e49-7f76-4c8e-8c14-3fe55f9615b9\") " pod="openstack/neutron-9764b468f-7sr7n" Dec 03 14:30:50 crc kubenswrapper[4805]: I1203 14:30:50.890036 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/31537e49-7f76-4c8e-8c14-3fe55f9615b9-config\") pod \"neutron-9764b468f-7sr7n\" (UID: \"31537e49-7f76-4c8e-8c14-3fe55f9615b9\") " pod="openstack/neutron-9764b468f-7sr7n" Dec 03 14:30:50 crc kubenswrapper[4805]: I1203 14:30:50.890053 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/31537e49-7f76-4c8e-8c14-3fe55f9615b9-ovndb-tls-certs\") pod \"neutron-9764b468f-7sr7n\" (UID: \"31537e49-7f76-4c8e-8c14-3fe55f9615b9\") " pod="openstack/neutron-9764b468f-7sr7n" Dec 03 14:30:50 crc kubenswrapper[4805]: I1203 14:30:50.890069 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/31537e49-7f76-4c8e-8c14-3fe55f9615b9-httpd-config\") pod \"neutron-9764b468f-7sr7n\" (UID: \"31537e49-7f76-4c8e-8c14-3fe55f9615b9\") " pod="openstack/neutron-9764b468f-7sr7n" Dec 03 14:30:50 crc kubenswrapper[4805]: I1203 14:30:50.895805 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/31537e49-7f76-4c8e-8c14-3fe55f9615b9-internal-tls-certs\") pod \"neutron-9764b468f-7sr7n\" (UID: \"31537e49-7f76-4c8e-8c14-3fe55f9615b9\") " pod="openstack/neutron-9764b468f-7sr7n" Dec 03 14:30:50 crc kubenswrapper[4805]: I1203 14:30:50.896258 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/31537e49-7f76-4c8e-8c14-3fe55f9615b9-ovndb-tls-certs\") pod \"neutron-9764b468f-7sr7n\" (UID: \"31537e49-7f76-4c8e-8c14-3fe55f9615b9\") " pod="openstack/neutron-9764b468f-7sr7n" Dec 03 14:30:50 crc kubenswrapper[4805]: I1203 14:30:50.897071 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/31537e49-7f76-4c8e-8c14-3fe55f9615b9-httpd-config\") pod \"neutron-9764b468f-7sr7n\" (UID: \"31537e49-7f76-4c8e-8c14-3fe55f9615b9\") " pod="openstack/neutron-9764b468f-7sr7n" Dec 03 14:30:50 crc kubenswrapper[4805]: I1203 14:30:50.910724 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/secret/31537e49-7f76-4c8e-8c14-3fe55f9615b9-config\") pod \"neutron-9764b468f-7sr7n\" (UID: \"31537e49-7f76-4c8e-8c14-3fe55f9615b9\") " pod="openstack/neutron-9764b468f-7sr7n" Dec 03 14:30:50 crc kubenswrapper[4805]: I1203 14:30:50.914060 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/31537e49-7f76-4c8e-8c14-3fe55f9615b9-combined-ca-bundle\") pod \"neutron-9764b468f-7sr7n\" (UID: \"31537e49-7f76-4c8e-8c14-3fe55f9615b9\") " pod="openstack/neutron-9764b468f-7sr7n" Dec 03 14:30:50 crc kubenswrapper[4805]: I1203 14:30:50.914165 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/31537e49-7f76-4c8e-8c14-3fe55f9615b9-public-tls-certs\") pod \"neutron-9764b468f-7sr7n\" (UID: \"31537e49-7f76-4c8e-8c14-3fe55f9615b9\") " pod="openstack/neutron-9764b468f-7sr7n" Dec 03 14:30:50 crc kubenswrapper[4805]: I1203 14:30:50.914756 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5sz66\" (UniqueName: \"kubernetes.io/projected/31537e49-7f76-4c8e-8c14-3fe55f9615b9-kube-api-access-5sz66\") pod \"neutron-9764b468f-7sr7n\" (UID: \"31537e49-7f76-4c8e-8c14-3fe55f9615b9\") " pod="openstack/neutron-9764b468f-7sr7n" Dec 03 14:30:51 crc kubenswrapper[4805]: I1203 14:30:51.031189 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-9764b468f-7sr7n" Dec 03 14:30:52 crc kubenswrapper[4805]: I1203 14:30:52.666709 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-9764b468f-7sr7n"] Dec 03 14:30:52 crc kubenswrapper[4805]: W1203 14:30:52.671092 4805 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod31537e49_7f76_4c8e_8c14_3fe55f9615b9.slice/crio-637a50106d2e077eadce218490d0d2b7f95723db34cbec01c4fa88cd4f8e1524 WatchSource:0}: Error finding container 637a50106d2e077eadce218490d0d2b7f95723db34cbec01c4fa88cd4f8e1524: Status 404 returned error can't find the container with id 637a50106d2e077eadce218490d0d2b7f95723db34cbec01c4fa88cd4f8e1524 Dec 03 14:30:53 crc kubenswrapper[4805]: I1203 14:30:53.115110 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-75f57c664d-4tsg2" event={"ID":"318792ec-dc8b-4a37-92ab-d31773d66350","Type":"ContainerStarted","Data":"48672ddefb5f29199ffd9090e43f5c584cf661e8cf1e4313d963fb96b28729b6"} Dec 03 14:30:53 crc kubenswrapper[4805]: I1203 14:30:53.117454 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-9764b468f-7sr7n" event={"ID":"31537e49-7f76-4c8e-8c14-3fe55f9615b9","Type":"ContainerStarted","Data":"637a50106d2e077eadce218490d0d2b7f95723db34cbec01c4fa88cd4f8e1524"} Dec 03 14:30:53 crc kubenswrapper[4805]: I1203 14:30:53.119129 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-84b966f6c9-zwsmd" event={"ID":"b3ca6772-26a0-410c-9609-1f49d08b1864","Type":"ContainerStarted","Data":"c3ec7bccf3f72b9e0eadf2e3235f7e5b3ecfefa7d6e5cabfb4409cfb4ee310b0"} Dec 03 14:30:54 crc kubenswrapper[4805]: I1203 14:30:54.131134 4805 generic.go:334] "Generic (PLEG): container finished" podID="b3ca6772-26a0-410c-9609-1f49d08b1864" containerID="c3ec7bccf3f72b9e0eadf2e3235f7e5b3ecfefa7d6e5cabfb4409cfb4ee310b0" exitCode=0 Dec 03 14:30:54 crc kubenswrapper[4805]: I1203 14:30:54.131543 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-84b966f6c9-zwsmd" event={"ID":"b3ca6772-26a0-410c-9609-1f49d08b1864","Type":"ContainerDied","Data":"c3ec7bccf3f72b9e0eadf2e3235f7e5b3ecfefa7d6e5cabfb4409cfb4ee310b0"} Dec 03 14:30:54 crc kubenswrapper[4805]: I1203 14:30:54.136007 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-75f57c664d-4tsg2" event={"ID":"318792ec-dc8b-4a37-92ab-d31773d66350","Type":"ContainerStarted","Data":"76ff7768f1c3210b2f1ffab05aefd717027127581dc36328eff11ac6503fdb3f"} Dec 03 14:30:54 crc kubenswrapper[4805]: I1203 14:30:54.136097 4805 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/neutron-75f57c664d-4tsg2" Dec 03 14:30:54 crc kubenswrapper[4805]: I1203 14:30:54.139239 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-9764b468f-7sr7n" event={"ID":"31537e49-7f76-4c8e-8c14-3fe55f9615b9","Type":"ContainerStarted","Data":"9b50bccb260ef9de08b827c72e8e6683996003d3c8982dbef3aaa1baf34b22ae"} Dec 03 14:30:54 crc kubenswrapper[4805]: I1203 14:30:54.139267 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-9764b468f-7sr7n" event={"ID":"31537e49-7f76-4c8e-8c14-3fe55f9615b9","Type":"ContainerStarted","Data":"1e89b1a54369086b939d15e59ad306d21201233b9a46975f2679cfb00d8749a2"} Dec 03 14:30:54 crc kubenswrapper[4805]: I1203 14:30:54.139436 4805 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/neutron-9764b468f-7sr7n" Dec 03 14:30:54 crc kubenswrapper[4805]: I1203 14:30:54.174117 4805 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-9764b468f-7sr7n" podStartSLOduration=4.174103984 podStartE2EDuration="4.174103984s" podCreationTimestamp="2025-12-03 14:30:50 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 14:30:54.170307341 +0000 UTC m=+1283.833224264" watchObservedRunningTime="2025-12-03 14:30:54.174103984 +0000 UTC m=+1283.837020907" Dec 03 14:30:54 crc kubenswrapper[4805]: I1203 14:30:54.200183 4805 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-75f57c664d-4tsg2" podStartSLOduration=6.200161769 podStartE2EDuration="6.200161769s" podCreationTimestamp="2025-12-03 14:30:48 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 14:30:54.193881799 +0000 UTC m=+1283.856798722" watchObservedRunningTime="2025-12-03 14:30:54.200161769 +0000 UTC m=+1283.863078692" Dec 03 14:31:03 crc kubenswrapper[4805]: I1203 14:31:03.232885 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-sync-fh9c2" event={"ID":"260717ec-f898-4ccd-8a4f-c7d6b14ade6e","Type":"ContainerStarted","Data":"3ae18d4a53cc1df81f75833ab27219aaab27ca80a4b5ccbb30481f900bcecd6f"} Dec 03 14:31:03 crc kubenswrapper[4805]: I1203 14:31:03.234825 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-84b966f6c9-zwsmd" event={"ID":"b3ca6772-26a0-410c-9609-1f49d08b1864","Type":"ContainerStarted","Data":"31628e51d65354f72d4dd433558b5b70b666da48c488d34fa61d3e72e666c631"} Dec 03 14:31:03 crc kubenswrapper[4805]: I1203 14:31:03.235026 4805 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-84b966f6c9-zwsmd" Dec 03 14:31:03 crc kubenswrapper[4805]: I1203 14:31:03.259710 4805 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-db-sync-fh9c2" podStartSLOduration=11.129583519 podStartE2EDuration="55.259689278s" podCreationTimestamp="2025-12-03 14:30:08 +0000 UTC" firstStartedPulling="2025-12-03 14:30:09.414216597 +0000 UTC m=+1239.077133510" lastFinishedPulling="2025-12-03 14:30:53.544322346 +0000 UTC m=+1283.207239269" observedRunningTime="2025-12-03 14:31:03.252161504 +0000 UTC m=+1292.915078427" watchObservedRunningTime="2025-12-03 14:31:03.259689278 +0000 UTC m=+1292.922606211" Dec 03 14:31:03 crc kubenswrapper[4805]: I1203 14:31:03.276822 4805 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-84b966f6c9-zwsmd" podStartSLOduration=15.276804622 podStartE2EDuration="15.276804622s" podCreationTimestamp="2025-12-03 14:30:48 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 14:31:03.269632087 +0000 UTC m=+1292.932549020" watchObservedRunningTime="2025-12-03 14:31:03.276804622 +0000 UTC m=+1292.939721545" Dec 03 14:31:06 crc kubenswrapper[4805]: I1203 14:31:06.263272 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"8533d470-22f2-401a-a39d-c37391920d81","Type":"ContainerStarted","Data":"b0453b9559e8e6be56e9ed69e25cfc3288b8cebccb3b351019de220c5feab03f"} Dec 03 14:31:06 crc kubenswrapper[4805]: I1203 14:31:06.263831 4805 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Dec 03 14:31:06 crc kubenswrapper[4805]: I1203 14:31:06.263470 4805 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="8533d470-22f2-401a-a39d-c37391920d81" containerName="proxy-httpd" containerID="cri-o://b0453b9559e8e6be56e9ed69e25cfc3288b8cebccb3b351019de220c5feab03f" gracePeriod=30 Dec 03 14:31:06 crc kubenswrapper[4805]: I1203 14:31:06.263426 4805 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="8533d470-22f2-401a-a39d-c37391920d81" containerName="ceilometer-central-agent" containerID="cri-o://dd9cf491b38857ec975e532024385fa6530bdcbb0987559861725061e1a31b29" gracePeriod=30 Dec 03 14:31:06 crc kubenswrapper[4805]: I1203 14:31:06.263483 4805 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="8533d470-22f2-401a-a39d-c37391920d81" containerName="sg-core" containerID="cri-o://8cdd783ebc4c2832d701981f18a14850bb3c14ceb07034817106c55661990ccc" gracePeriod=30 Dec 03 14:31:06 crc kubenswrapper[4805]: I1203 14:31:06.263514 4805 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="8533d470-22f2-401a-a39d-c37391920d81" containerName="ceilometer-notification-agent" containerID="cri-o://0e63c808604f7e7fe89fba38cf60f74044c3c5721825fa4640ddeadceee1ca05" gracePeriod=30 Dec 03 14:31:06 crc kubenswrapper[4805]: I1203 14:31:06.297518 4805 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=3.5945350080000003 podStartE2EDuration="59.297499721s" podCreationTimestamp="2025-12-03 14:30:07 +0000 UTC" firstStartedPulling="2025-12-03 14:30:09.083016141 +0000 UTC m=+1238.745933054" lastFinishedPulling="2025-12-03 14:31:04.785980824 +0000 UTC m=+1294.448897767" observedRunningTime="2025-12-03 14:31:06.296079592 +0000 UTC m=+1295.958996515" watchObservedRunningTime="2025-12-03 14:31:06.297499721 +0000 UTC m=+1295.960416654" Dec 03 14:31:07 crc kubenswrapper[4805]: I1203 14:31:07.275920 4805 generic.go:334] "Generic (PLEG): container finished" podID="8533d470-22f2-401a-a39d-c37391920d81" containerID="b0453b9559e8e6be56e9ed69e25cfc3288b8cebccb3b351019de220c5feab03f" exitCode=0 Dec 03 14:31:07 crc kubenswrapper[4805]: I1203 14:31:07.277179 4805 generic.go:334] "Generic (PLEG): container finished" podID="8533d470-22f2-401a-a39d-c37391920d81" containerID="8cdd783ebc4c2832d701981f18a14850bb3c14ceb07034817106c55661990ccc" exitCode=2 Dec 03 14:31:07 crc kubenswrapper[4805]: I1203 14:31:07.277307 4805 generic.go:334] "Generic (PLEG): container finished" podID="8533d470-22f2-401a-a39d-c37391920d81" containerID="dd9cf491b38857ec975e532024385fa6530bdcbb0987559861725061e1a31b29" exitCode=0 Dec 03 14:31:07 crc kubenswrapper[4805]: I1203 14:31:07.275981 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"8533d470-22f2-401a-a39d-c37391920d81","Type":"ContainerDied","Data":"b0453b9559e8e6be56e9ed69e25cfc3288b8cebccb3b351019de220c5feab03f"} Dec 03 14:31:07 crc kubenswrapper[4805]: I1203 14:31:07.277480 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"8533d470-22f2-401a-a39d-c37391920d81","Type":"ContainerDied","Data":"8cdd783ebc4c2832d701981f18a14850bb3c14ceb07034817106c55661990ccc"} Dec 03 14:31:07 crc kubenswrapper[4805]: I1203 14:31:07.277568 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"8533d470-22f2-401a-a39d-c37391920d81","Type":"ContainerDied","Data":"dd9cf491b38857ec975e532024385fa6530bdcbb0987559861725061e1a31b29"} Dec 03 14:31:08 crc kubenswrapper[4805]: I1203 14:31:08.692161 4805 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-84b966f6c9-zwsmd" Dec 03 14:31:08 crc kubenswrapper[4805]: I1203 14:31:08.771255 4805 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-8b5c85b87-bvb8w"] Dec 03 14:31:08 crc kubenswrapper[4805]: I1203 14:31:08.772080 4805 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-8b5c85b87-bvb8w" podUID="a8299a7f-082a-4a51-becd-7c393a6f4b51" containerName="dnsmasq-dns" containerID="cri-o://58d07340f3e9637400fe3855db1f26dfcf2ec01f9006fa34bbfe5a88846c9ff2" gracePeriod=10 Dec 03 14:31:09 crc kubenswrapper[4805]: I1203 14:31:09.321069 4805 generic.go:334] "Generic (PLEG): container finished" podID="a8299a7f-082a-4a51-becd-7c393a6f4b51" containerID="58d07340f3e9637400fe3855db1f26dfcf2ec01f9006fa34bbfe5a88846c9ff2" exitCode=0 Dec 03 14:31:09 crc kubenswrapper[4805]: I1203 14:31:09.321162 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-8b5c85b87-bvb8w" event={"ID":"a8299a7f-082a-4a51-becd-7c393a6f4b51","Type":"ContainerDied","Data":"58d07340f3e9637400fe3855db1f26dfcf2ec01f9006fa34bbfe5a88846c9ff2"} Dec 03 14:31:09 crc kubenswrapper[4805]: I1203 14:31:09.321657 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-8b5c85b87-bvb8w" event={"ID":"a8299a7f-082a-4a51-becd-7c393a6f4b51","Type":"ContainerDied","Data":"ee5acaf567b19c4061d7d5cc37b1adc10bd8c89de090f9b575df0e9c5c942ea9"} Dec 03 14:31:09 crc kubenswrapper[4805]: I1203 14:31:09.321674 4805 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="ee5acaf567b19c4061d7d5cc37b1adc10bd8c89de090f9b575df0e9c5c942ea9" Dec 03 14:31:09 crc kubenswrapper[4805]: I1203 14:31:09.336693 4805 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-8b5c85b87-bvb8w" Dec 03 14:31:09 crc kubenswrapper[4805]: I1203 14:31:09.475275 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/a8299a7f-082a-4a51-becd-7c393a6f4b51-dns-swift-storage-0\") pod \"a8299a7f-082a-4a51-becd-7c393a6f4b51\" (UID: \"a8299a7f-082a-4a51-becd-7c393a6f4b51\") " Dec 03 14:31:09 crc kubenswrapper[4805]: I1203 14:31:09.475486 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/a8299a7f-082a-4a51-becd-7c393a6f4b51-ovsdbserver-sb\") pod \"a8299a7f-082a-4a51-becd-7c393a6f4b51\" (UID: \"a8299a7f-082a-4a51-becd-7c393a6f4b51\") " Dec 03 14:31:09 crc kubenswrapper[4805]: I1203 14:31:09.475538 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a8299a7f-082a-4a51-becd-7c393a6f4b51-config\") pod \"a8299a7f-082a-4a51-becd-7c393a6f4b51\" (UID: \"a8299a7f-082a-4a51-becd-7c393a6f4b51\") " Dec 03 14:31:09 crc kubenswrapper[4805]: I1203 14:31:09.475579 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/a8299a7f-082a-4a51-becd-7c393a6f4b51-ovsdbserver-nb\") pod \"a8299a7f-082a-4a51-becd-7c393a6f4b51\" (UID: \"a8299a7f-082a-4a51-becd-7c393a6f4b51\") " Dec 03 14:31:09 crc kubenswrapper[4805]: I1203 14:31:09.475608 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wkmfm\" (UniqueName: \"kubernetes.io/projected/a8299a7f-082a-4a51-becd-7c393a6f4b51-kube-api-access-wkmfm\") pod \"a8299a7f-082a-4a51-becd-7c393a6f4b51\" (UID: \"a8299a7f-082a-4a51-becd-7c393a6f4b51\") " Dec 03 14:31:09 crc kubenswrapper[4805]: I1203 14:31:09.475644 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/a8299a7f-082a-4a51-becd-7c393a6f4b51-dns-svc\") pod \"a8299a7f-082a-4a51-becd-7c393a6f4b51\" (UID: \"a8299a7f-082a-4a51-becd-7c393a6f4b51\") " Dec 03 14:31:09 crc kubenswrapper[4805]: I1203 14:31:09.483069 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a8299a7f-082a-4a51-becd-7c393a6f4b51-kube-api-access-wkmfm" (OuterVolumeSpecName: "kube-api-access-wkmfm") pod "a8299a7f-082a-4a51-becd-7c393a6f4b51" (UID: "a8299a7f-082a-4a51-becd-7c393a6f4b51"). InnerVolumeSpecName "kube-api-access-wkmfm". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 14:31:09 crc kubenswrapper[4805]: I1203 14:31:09.529439 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a8299a7f-082a-4a51-becd-7c393a6f4b51-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "a8299a7f-082a-4a51-becd-7c393a6f4b51" (UID: "a8299a7f-082a-4a51-becd-7c393a6f4b51"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 14:31:09 crc kubenswrapper[4805]: I1203 14:31:09.530325 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a8299a7f-082a-4a51-becd-7c393a6f4b51-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "a8299a7f-082a-4a51-becd-7c393a6f4b51" (UID: "a8299a7f-082a-4a51-becd-7c393a6f4b51"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 14:31:09 crc kubenswrapper[4805]: I1203 14:31:09.535546 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a8299a7f-082a-4a51-becd-7c393a6f4b51-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "a8299a7f-082a-4a51-becd-7c393a6f4b51" (UID: "a8299a7f-082a-4a51-becd-7c393a6f4b51"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 14:31:09 crc kubenswrapper[4805]: I1203 14:31:09.537652 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a8299a7f-082a-4a51-becd-7c393a6f4b51-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "a8299a7f-082a-4a51-becd-7c393a6f4b51" (UID: "a8299a7f-082a-4a51-becd-7c393a6f4b51"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 14:31:09 crc kubenswrapper[4805]: I1203 14:31:09.538670 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a8299a7f-082a-4a51-becd-7c393a6f4b51-config" (OuterVolumeSpecName: "config") pod "a8299a7f-082a-4a51-becd-7c393a6f4b51" (UID: "a8299a7f-082a-4a51-becd-7c393a6f4b51"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 14:31:09 crc kubenswrapper[4805]: I1203 14:31:09.577158 4805 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/a8299a7f-082a-4a51-becd-7c393a6f4b51-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 03 14:31:09 crc kubenswrapper[4805]: I1203 14:31:09.577191 4805 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/a8299a7f-082a-4a51-becd-7c393a6f4b51-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Dec 03 14:31:09 crc kubenswrapper[4805]: I1203 14:31:09.577202 4805 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/a8299a7f-082a-4a51-becd-7c393a6f4b51-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Dec 03 14:31:09 crc kubenswrapper[4805]: I1203 14:31:09.577212 4805 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a8299a7f-082a-4a51-becd-7c393a6f4b51-config\") on node \"crc\" DevicePath \"\"" Dec 03 14:31:09 crc kubenswrapper[4805]: I1203 14:31:09.577220 4805 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/a8299a7f-082a-4a51-becd-7c393a6f4b51-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Dec 03 14:31:09 crc kubenswrapper[4805]: I1203 14:31:09.577228 4805 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wkmfm\" (UniqueName: \"kubernetes.io/projected/a8299a7f-082a-4a51-becd-7c393a6f4b51-kube-api-access-wkmfm\") on node \"crc\" DevicePath \"\"" Dec 03 14:31:10 crc kubenswrapper[4805]: I1203 14:31:10.262092 4805 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 03 14:31:10 crc kubenswrapper[4805]: I1203 14:31:10.341413 4805 generic.go:334] "Generic (PLEG): container finished" podID="8533d470-22f2-401a-a39d-c37391920d81" containerID="0e63c808604f7e7fe89fba38cf60f74044c3c5721825fa4640ddeadceee1ca05" exitCode=0 Dec 03 14:31:10 crc kubenswrapper[4805]: I1203 14:31:10.341473 4805 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 03 14:31:10 crc kubenswrapper[4805]: I1203 14:31:10.341494 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"8533d470-22f2-401a-a39d-c37391920d81","Type":"ContainerDied","Data":"0e63c808604f7e7fe89fba38cf60f74044c3c5721825fa4640ddeadceee1ca05"} Dec 03 14:31:10 crc kubenswrapper[4805]: I1203 14:31:10.341552 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"8533d470-22f2-401a-a39d-c37391920d81","Type":"ContainerDied","Data":"bc0c877e9184355cf11dd77131fca5e9b83afff5f85a897efcbb6fe87c45a6c1"} Dec 03 14:31:10 crc kubenswrapper[4805]: I1203 14:31:10.341577 4805 scope.go:117] "RemoveContainer" containerID="b0453b9559e8e6be56e9ed69e25cfc3288b8cebccb3b351019de220c5feab03f" Dec 03 14:31:10 crc kubenswrapper[4805]: I1203 14:31:10.342771 4805 generic.go:334] "Generic (PLEG): container finished" podID="13ba03f0-acc2-4fb8-90a9-583b4f8694cb" containerID="76a9c6e35bfc792baf01e7282b5b7319e4e528a8abaa7f19a90eb5f470bdeab4" exitCode=0 Dec 03 14:31:10 crc kubenswrapper[4805]: I1203 14:31:10.342879 4805 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-8b5c85b87-bvb8w" Dec 03 14:31:10 crc kubenswrapper[4805]: I1203 14:31:10.342879 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-sync-74tz4" event={"ID":"13ba03f0-acc2-4fb8-90a9-583b4f8694cb","Type":"ContainerDied","Data":"76a9c6e35bfc792baf01e7282b5b7319e4e528a8abaa7f19a90eb5f470bdeab4"} Dec 03 14:31:10 crc kubenswrapper[4805]: I1203 14:31:10.366456 4805 scope.go:117] "RemoveContainer" containerID="8cdd783ebc4c2832d701981f18a14850bb3c14ceb07034817106c55661990ccc" Dec 03 14:31:10 crc kubenswrapper[4805]: I1203 14:31:10.388230 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/8533d470-22f2-401a-a39d-c37391920d81-sg-core-conf-yaml\") pod \"8533d470-22f2-401a-a39d-c37391920d81\" (UID: \"8533d470-22f2-401a-a39d-c37391920d81\") " Dec 03 14:31:10 crc kubenswrapper[4805]: I1203 14:31:10.388320 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/8533d470-22f2-401a-a39d-c37391920d81-log-httpd\") pod \"8533d470-22f2-401a-a39d-c37391920d81\" (UID: \"8533d470-22f2-401a-a39d-c37391920d81\") " Dec 03 14:31:10 crc kubenswrapper[4805]: I1203 14:31:10.388338 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/8533d470-22f2-401a-a39d-c37391920d81-run-httpd\") pod \"8533d470-22f2-401a-a39d-c37391920d81\" (UID: \"8533d470-22f2-401a-a39d-c37391920d81\") " Dec 03 14:31:10 crc kubenswrapper[4805]: I1203 14:31:10.388368 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kvkpr\" (UniqueName: \"kubernetes.io/projected/8533d470-22f2-401a-a39d-c37391920d81-kube-api-access-kvkpr\") pod \"8533d470-22f2-401a-a39d-c37391920d81\" (UID: \"8533d470-22f2-401a-a39d-c37391920d81\") " Dec 03 14:31:10 crc kubenswrapper[4805]: I1203 14:31:10.388498 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8533d470-22f2-401a-a39d-c37391920d81-combined-ca-bundle\") pod \"8533d470-22f2-401a-a39d-c37391920d81\" (UID: \"8533d470-22f2-401a-a39d-c37391920d81\") " Dec 03 14:31:10 crc kubenswrapper[4805]: I1203 14:31:10.388531 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8533d470-22f2-401a-a39d-c37391920d81-scripts\") pod \"8533d470-22f2-401a-a39d-c37391920d81\" (UID: \"8533d470-22f2-401a-a39d-c37391920d81\") " Dec 03 14:31:10 crc kubenswrapper[4805]: I1203 14:31:10.388558 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8533d470-22f2-401a-a39d-c37391920d81-config-data\") pod \"8533d470-22f2-401a-a39d-c37391920d81\" (UID: \"8533d470-22f2-401a-a39d-c37391920d81\") " Dec 03 14:31:10 crc kubenswrapper[4805]: I1203 14:31:10.389098 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8533d470-22f2-401a-a39d-c37391920d81-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "8533d470-22f2-401a-a39d-c37391920d81" (UID: "8533d470-22f2-401a-a39d-c37391920d81"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 14:31:10 crc kubenswrapper[4805]: I1203 14:31:10.389505 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8533d470-22f2-401a-a39d-c37391920d81-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "8533d470-22f2-401a-a39d-c37391920d81" (UID: "8533d470-22f2-401a-a39d-c37391920d81"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 14:31:10 crc kubenswrapper[4805]: I1203 14:31:10.392562 4805 scope.go:117] "RemoveContainer" containerID="0e63c808604f7e7fe89fba38cf60f74044c3c5721825fa4640ddeadceee1ca05" Dec 03 14:31:10 crc kubenswrapper[4805]: I1203 14:31:10.394692 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8533d470-22f2-401a-a39d-c37391920d81-scripts" (OuterVolumeSpecName: "scripts") pod "8533d470-22f2-401a-a39d-c37391920d81" (UID: "8533d470-22f2-401a-a39d-c37391920d81"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 14:31:10 crc kubenswrapper[4805]: I1203 14:31:10.395357 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8533d470-22f2-401a-a39d-c37391920d81-kube-api-access-kvkpr" (OuterVolumeSpecName: "kube-api-access-kvkpr") pod "8533d470-22f2-401a-a39d-c37391920d81" (UID: "8533d470-22f2-401a-a39d-c37391920d81"). InnerVolumeSpecName "kube-api-access-kvkpr". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 14:31:10 crc kubenswrapper[4805]: I1203 14:31:10.399719 4805 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-8b5c85b87-bvb8w"] Dec 03 14:31:10 crc kubenswrapper[4805]: I1203 14:31:10.408736 4805 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-8b5c85b87-bvb8w"] Dec 03 14:31:10 crc kubenswrapper[4805]: I1203 14:31:10.412991 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8533d470-22f2-401a-a39d-c37391920d81-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "8533d470-22f2-401a-a39d-c37391920d81" (UID: "8533d470-22f2-401a-a39d-c37391920d81"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 14:31:10 crc kubenswrapper[4805]: I1203 14:31:10.446768 4805 scope.go:117] "RemoveContainer" containerID="dd9cf491b38857ec975e532024385fa6530bdcbb0987559861725061e1a31b29" Dec 03 14:31:10 crc kubenswrapper[4805]: I1203 14:31:10.463318 4805 scope.go:117] "RemoveContainer" containerID="b0453b9559e8e6be56e9ed69e25cfc3288b8cebccb3b351019de220c5feab03f" Dec 03 14:31:10 crc kubenswrapper[4805]: E1203 14:31:10.463637 4805 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b0453b9559e8e6be56e9ed69e25cfc3288b8cebccb3b351019de220c5feab03f\": container with ID starting with b0453b9559e8e6be56e9ed69e25cfc3288b8cebccb3b351019de220c5feab03f not found: ID does not exist" containerID="b0453b9559e8e6be56e9ed69e25cfc3288b8cebccb3b351019de220c5feab03f" Dec 03 14:31:10 crc kubenswrapper[4805]: I1203 14:31:10.463664 4805 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b0453b9559e8e6be56e9ed69e25cfc3288b8cebccb3b351019de220c5feab03f"} err="failed to get container status \"b0453b9559e8e6be56e9ed69e25cfc3288b8cebccb3b351019de220c5feab03f\": rpc error: code = NotFound desc = could not find container \"b0453b9559e8e6be56e9ed69e25cfc3288b8cebccb3b351019de220c5feab03f\": container with ID starting with b0453b9559e8e6be56e9ed69e25cfc3288b8cebccb3b351019de220c5feab03f not found: ID does not exist" Dec 03 14:31:10 crc kubenswrapper[4805]: I1203 14:31:10.463683 4805 scope.go:117] "RemoveContainer" containerID="8cdd783ebc4c2832d701981f18a14850bb3c14ceb07034817106c55661990ccc" Dec 03 14:31:10 crc kubenswrapper[4805]: E1203 14:31:10.463934 4805 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8cdd783ebc4c2832d701981f18a14850bb3c14ceb07034817106c55661990ccc\": container with ID starting with 8cdd783ebc4c2832d701981f18a14850bb3c14ceb07034817106c55661990ccc not found: ID does not exist" containerID="8cdd783ebc4c2832d701981f18a14850bb3c14ceb07034817106c55661990ccc" Dec 03 14:31:10 crc kubenswrapper[4805]: I1203 14:31:10.463978 4805 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8cdd783ebc4c2832d701981f18a14850bb3c14ceb07034817106c55661990ccc"} err="failed to get container status \"8cdd783ebc4c2832d701981f18a14850bb3c14ceb07034817106c55661990ccc\": rpc error: code = NotFound desc = could not find container \"8cdd783ebc4c2832d701981f18a14850bb3c14ceb07034817106c55661990ccc\": container with ID starting with 8cdd783ebc4c2832d701981f18a14850bb3c14ceb07034817106c55661990ccc not found: ID does not exist" Dec 03 14:31:10 crc kubenswrapper[4805]: I1203 14:31:10.464005 4805 scope.go:117] "RemoveContainer" containerID="0e63c808604f7e7fe89fba38cf60f74044c3c5721825fa4640ddeadceee1ca05" Dec 03 14:31:10 crc kubenswrapper[4805]: E1203 14:31:10.464226 4805 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0e63c808604f7e7fe89fba38cf60f74044c3c5721825fa4640ddeadceee1ca05\": container with ID starting with 0e63c808604f7e7fe89fba38cf60f74044c3c5721825fa4640ddeadceee1ca05 not found: ID does not exist" containerID="0e63c808604f7e7fe89fba38cf60f74044c3c5721825fa4640ddeadceee1ca05" Dec 03 14:31:10 crc kubenswrapper[4805]: I1203 14:31:10.464251 4805 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0e63c808604f7e7fe89fba38cf60f74044c3c5721825fa4640ddeadceee1ca05"} err="failed to get container status \"0e63c808604f7e7fe89fba38cf60f74044c3c5721825fa4640ddeadceee1ca05\": rpc error: code = NotFound desc = could not find container \"0e63c808604f7e7fe89fba38cf60f74044c3c5721825fa4640ddeadceee1ca05\": container with ID starting with 0e63c808604f7e7fe89fba38cf60f74044c3c5721825fa4640ddeadceee1ca05 not found: ID does not exist" Dec 03 14:31:10 crc kubenswrapper[4805]: I1203 14:31:10.464265 4805 scope.go:117] "RemoveContainer" containerID="dd9cf491b38857ec975e532024385fa6530bdcbb0987559861725061e1a31b29" Dec 03 14:31:10 crc kubenswrapper[4805]: E1203 14:31:10.464513 4805 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"dd9cf491b38857ec975e532024385fa6530bdcbb0987559861725061e1a31b29\": container with ID starting with dd9cf491b38857ec975e532024385fa6530bdcbb0987559861725061e1a31b29 not found: ID does not exist" containerID="dd9cf491b38857ec975e532024385fa6530bdcbb0987559861725061e1a31b29" Dec 03 14:31:10 crc kubenswrapper[4805]: I1203 14:31:10.464561 4805 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"dd9cf491b38857ec975e532024385fa6530bdcbb0987559861725061e1a31b29"} err="failed to get container status \"dd9cf491b38857ec975e532024385fa6530bdcbb0987559861725061e1a31b29\": rpc error: code = NotFound desc = could not find container \"dd9cf491b38857ec975e532024385fa6530bdcbb0987559861725061e1a31b29\": container with ID starting with dd9cf491b38857ec975e532024385fa6530bdcbb0987559861725061e1a31b29 not found: ID does not exist" Dec 03 14:31:10 crc kubenswrapper[4805]: I1203 14:31:10.470919 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8533d470-22f2-401a-a39d-c37391920d81-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "8533d470-22f2-401a-a39d-c37391920d81" (UID: "8533d470-22f2-401a-a39d-c37391920d81"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 14:31:10 crc kubenswrapper[4805]: I1203 14:31:10.480115 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8533d470-22f2-401a-a39d-c37391920d81-config-data" (OuterVolumeSpecName: "config-data") pod "8533d470-22f2-401a-a39d-c37391920d81" (UID: "8533d470-22f2-401a-a39d-c37391920d81"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 14:31:10 crc kubenswrapper[4805]: I1203 14:31:10.490169 4805 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8533d470-22f2-401a-a39d-c37391920d81-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 14:31:10 crc kubenswrapper[4805]: I1203 14:31:10.490191 4805 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8533d470-22f2-401a-a39d-c37391920d81-scripts\") on node \"crc\" DevicePath \"\"" Dec 03 14:31:10 crc kubenswrapper[4805]: I1203 14:31:10.490202 4805 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8533d470-22f2-401a-a39d-c37391920d81-config-data\") on node \"crc\" DevicePath \"\"" Dec 03 14:31:10 crc kubenswrapper[4805]: I1203 14:31:10.490210 4805 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/8533d470-22f2-401a-a39d-c37391920d81-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Dec 03 14:31:10 crc kubenswrapper[4805]: I1203 14:31:10.490217 4805 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/8533d470-22f2-401a-a39d-c37391920d81-log-httpd\") on node \"crc\" DevicePath \"\"" Dec 03 14:31:10 crc kubenswrapper[4805]: I1203 14:31:10.490225 4805 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/8533d470-22f2-401a-a39d-c37391920d81-run-httpd\") on node \"crc\" DevicePath \"\"" Dec 03 14:31:10 crc kubenswrapper[4805]: I1203 14:31:10.490233 4805 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kvkpr\" (UniqueName: \"kubernetes.io/projected/8533d470-22f2-401a-a39d-c37391920d81-kube-api-access-kvkpr\") on node \"crc\" DevicePath \"\"" Dec 03 14:31:10 crc kubenswrapper[4805]: I1203 14:31:10.676287 4805 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 03 14:31:10 crc kubenswrapper[4805]: I1203 14:31:10.682905 4805 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Dec 03 14:31:10 crc kubenswrapper[4805]: I1203 14:31:10.716432 4805 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8533d470-22f2-401a-a39d-c37391920d81" path="/var/lib/kubelet/pods/8533d470-22f2-401a-a39d-c37391920d81/volumes" Dec 03 14:31:10 crc kubenswrapper[4805]: I1203 14:31:10.717363 4805 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a8299a7f-082a-4a51-becd-7c393a6f4b51" path="/var/lib/kubelet/pods/a8299a7f-082a-4a51-becd-7c393a6f4b51/volumes" Dec 03 14:31:10 crc kubenswrapper[4805]: I1203 14:31:10.718002 4805 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Dec 03 14:31:10 crc kubenswrapper[4805]: E1203 14:31:10.718577 4805 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a8299a7f-082a-4a51-becd-7c393a6f4b51" containerName="init" Dec 03 14:31:10 crc kubenswrapper[4805]: I1203 14:31:10.718590 4805 state_mem.go:107] "Deleted CPUSet assignment" podUID="a8299a7f-082a-4a51-becd-7c393a6f4b51" containerName="init" Dec 03 14:31:10 crc kubenswrapper[4805]: E1203 14:31:10.718604 4805 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8533d470-22f2-401a-a39d-c37391920d81" containerName="ceilometer-notification-agent" Dec 03 14:31:10 crc kubenswrapper[4805]: I1203 14:31:10.718610 4805 state_mem.go:107] "Deleted CPUSet assignment" podUID="8533d470-22f2-401a-a39d-c37391920d81" containerName="ceilometer-notification-agent" Dec 03 14:31:10 crc kubenswrapper[4805]: E1203 14:31:10.718628 4805 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8533d470-22f2-401a-a39d-c37391920d81" containerName="proxy-httpd" Dec 03 14:31:10 crc kubenswrapper[4805]: I1203 14:31:10.718634 4805 state_mem.go:107] "Deleted CPUSet assignment" podUID="8533d470-22f2-401a-a39d-c37391920d81" containerName="proxy-httpd" Dec 03 14:31:10 crc kubenswrapper[4805]: E1203 14:31:10.718647 4805 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8533d470-22f2-401a-a39d-c37391920d81" containerName="ceilometer-central-agent" Dec 03 14:31:10 crc kubenswrapper[4805]: I1203 14:31:10.718653 4805 state_mem.go:107] "Deleted CPUSet assignment" podUID="8533d470-22f2-401a-a39d-c37391920d81" containerName="ceilometer-central-agent" Dec 03 14:31:10 crc kubenswrapper[4805]: E1203 14:31:10.718671 4805 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8533d470-22f2-401a-a39d-c37391920d81" containerName="sg-core" Dec 03 14:31:10 crc kubenswrapper[4805]: I1203 14:31:10.718678 4805 state_mem.go:107] "Deleted CPUSet assignment" podUID="8533d470-22f2-401a-a39d-c37391920d81" containerName="sg-core" Dec 03 14:31:10 crc kubenswrapper[4805]: E1203 14:31:10.718710 4805 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a8299a7f-082a-4a51-becd-7c393a6f4b51" containerName="dnsmasq-dns" Dec 03 14:31:10 crc kubenswrapper[4805]: I1203 14:31:10.718716 4805 state_mem.go:107] "Deleted CPUSet assignment" podUID="a8299a7f-082a-4a51-becd-7c393a6f4b51" containerName="dnsmasq-dns" Dec 03 14:31:10 crc kubenswrapper[4805]: I1203 14:31:10.719423 4805 memory_manager.go:354] "RemoveStaleState removing state" podUID="8533d470-22f2-401a-a39d-c37391920d81" containerName="sg-core" Dec 03 14:31:10 crc kubenswrapper[4805]: I1203 14:31:10.719485 4805 memory_manager.go:354] "RemoveStaleState removing state" podUID="8533d470-22f2-401a-a39d-c37391920d81" containerName="ceilometer-central-agent" Dec 03 14:31:10 crc kubenswrapper[4805]: I1203 14:31:10.719501 4805 memory_manager.go:354] "RemoveStaleState removing state" podUID="8533d470-22f2-401a-a39d-c37391920d81" containerName="ceilometer-notification-agent" Dec 03 14:31:10 crc kubenswrapper[4805]: I1203 14:31:10.719540 4805 memory_manager.go:354] "RemoveStaleState removing state" podUID="8533d470-22f2-401a-a39d-c37391920d81" containerName="proxy-httpd" Dec 03 14:31:10 crc kubenswrapper[4805]: I1203 14:31:10.719547 4805 memory_manager.go:354] "RemoveStaleState removing state" podUID="a8299a7f-082a-4a51-becd-7c393a6f4b51" containerName="dnsmasq-dns" Dec 03 14:31:10 crc kubenswrapper[4805]: I1203 14:31:10.724288 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 03 14:31:10 crc kubenswrapper[4805]: I1203 14:31:10.727237 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Dec 03 14:31:10 crc kubenswrapper[4805]: I1203 14:31:10.727652 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Dec 03 14:31:10 crc kubenswrapper[4805]: I1203 14:31:10.734615 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 03 14:31:10 crc kubenswrapper[4805]: I1203 14:31:10.897497 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/58c98434-0b4d-4c2d-ac35-109e38f96803-log-httpd\") pod \"ceilometer-0\" (UID: \"58c98434-0b4d-4c2d-ac35-109e38f96803\") " pod="openstack/ceilometer-0" Dec 03 14:31:10 crc kubenswrapper[4805]: I1203 14:31:10.897669 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/58c98434-0b4d-4c2d-ac35-109e38f96803-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"58c98434-0b4d-4c2d-ac35-109e38f96803\") " pod="openstack/ceilometer-0" Dec 03 14:31:10 crc kubenswrapper[4805]: I1203 14:31:10.897803 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/58c98434-0b4d-4c2d-ac35-109e38f96803-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"58c98434-0b4d-4c2d-ac35-109e38f96803\") " pod="openstack/ceilometer-0" Dec 03 14:31:10 crc kubenswrapper[4805]: I1203 14:31:10.897941 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/58c98434-0b4d-4c2d-ac35-109e38f96803-config-data\") pod \"ceilometer-0\" (UID: \"58c98434-0b4d-4c2d-ac35-109e38f96803\") " pod="openstack/ceilometer-0" Dec 03 14:31:10 crc kubenswrapper[4805]: I1203 14:31:10.898031 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8d5p7\" (UniqueName: \"kubernetes.io/projected/58c98434-0b4d-4c2d-ac35-109e38f96803-kube-api-access-8d5p7\") pod \"ceilometer-0\" (UID: \"58c98434-0b4d-4c2d-ac35-109e38f96803\") " pod="openstack/ceilometer-0" Dec 03 14:31:10 crc kubenswrapper[4805]: I1203 14:31:10.898203 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/58c98434-0b4d-4c2d-ac35-109e38f96803-run-httpd\") pod \"ceilometer-0\" (UID: \"58c98434-0b4d-4c2d-ac35-109e38f96803\") " pod="openstack/ceilometer-0" Dec 03 14:31:10 crc kubenswrapper[4805]: I1203 14:31:10.898377 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/58c98434-0b4d-4c2d-ac35-109e38f96803-scripts\") pod \"ceilometer-0\" (UID: \"58c98434-0b4d-4c2d-ac35-109e38f96803\") " pod="openstack/ceilometer-0" Dec 03 14:31:11 crc kubenswrapper[4805]: I1203 14:31:10.999690 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/58c98434-0b4d-4c2d-ac35-109e38f96803-log-httpd\") pod \"ceilometer-0\" (UID: \"58c98434-0b4d-4c2d-ac35-109e38f96803\") " pod="openstack/ceilometer-0" Dec 03 14:31:11 crc kubenswrapper[4805]: I1203 14:31:10.999748 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/58c98434-0b4d-4c2d-ac35-109e38f96803-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"58c98434-0b4d-4c2d-ac35-109e38f96803\") " pod="openstack/ceilometer-0" Dec 03 14:31:11 crc kubenswrapper[4805]: I1203 14:31:10.999784 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/58c98434-0b4d-4c2d-ac35-109e38f96803-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"58c98434-0b4d-4c2d-ac35-109e38f96803\") " pod="openstack/ceilometer-0" Dec 03 14:31:11 crc kubenswrapper[4805]: I1203 14:31:10.999809 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/58c98434-0b4d-4c2d-ac35-109e38f96803-config-data\") pod \"ceilometer-0\" (UID: \"58c98434-0b4d-4c2d-ac35-109e38f96803\") " pod="openstack/ceilometer-0" Dec 03 14:31:11 crc kubenswrapper[4805]: I1203 14:31:10.999857 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8d5p7\" (UniqueName: \"kubernetes.io/projected/58c98434-0b4d-4c2d-ac35-109e38f96803-kube-api-access-8d5p7\") pod \"ceilometer-0\" (UID: \"58c98434-0b4d-4c2d-ac35-109e38f96803\") " pod="openstack/ceilometer-0" Dec 03 14:31:11 crc kubenswrapper[4805]: I1203 14:31:10.999886 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/58c98434-0b4d-4c2d-ac35-109e38f96803-run-httpd\") pod \"ceilometer-0\" (UID: \"58c98434-0b4d-4c2d-ac35-109e38f96803\") " pod="openstack/ceilometer-0" Dec 03 14:31:11 crc kubenswrapper[4805]: I1203 14:31:10.999925 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/58c98434-0b4d-4c2d-ac35-109e38f96803-scripts\") pod \"ceilometer-0\" (UID: \"58c98434-0b4d-4c2d-ac35-109e38f96803\") " pod="openstack/ceilometer-0" Dec 03 14:31:11 crc kubenswrapper[4805]: I1203 14:31:11.000956 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/58c98434-0b4d-4c2d-ac35-109e38f96803-run-httpd\") pod \"ceilometer-0\" (UID: \"58c98434-0b4d-4c2d-ac35-109e38f96803\") " pod="openstack/ceilometer-0" Dec 03 14:31:11 crc kubenswrapper[4805]: I1203 14:31:11.001058 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/58c98434-0b4d-4c2d-ac35-109e38f96803-log-httpd\") pod \"ceilometer-0\" (UID: \"58c98434-0b4d-4c2d-ac35-109e38f96803\") " pod="openstack/ceilometer-0" Dec 03 14:31:11 crc kubenswrapper[4805]: I1203 14:31:11.005933 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/58c98434-0b4d-4c2d-ac35-109e38f96803-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"58c98434-0b4d-4c2d-ac35-109e38f96803\") " pod="openstack/ceilometer-0" Dec 03 14:31:11 crc kubenswrapper[4805]: I1203 14:31:11.006568 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/58c98434-0b4d-4c2d-ac35-109e38f96803-scripts\") pod \"ceilometer-0\" (UID: \"58c98434-0b4d-4c2d-ac35-109e38f96803\") " pod="openstack/ceilometer-0" Dec 03 14:31:11 crc kubenswrapper[4805]: I1203 14:31:11.006709 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/58c98434-0b4d-4c2d-ac35-109e38f96803-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"58c98434-0b4d-4c2d-ac35-109e38f96803\") " pod="openstack/ceilometer-0" Dec 03 14:31:11 crc kubenswrapper[4805]: I1203 14:31:11.008165 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/58c98434-0b4d-4c2d-ac35-109e38f96803-config-data\") pod \"ceilometer-0\" (UID: \"58c98434-0b4d-4c2d-ac35-109e38f96803\") " pod="openstack/ceilometer-0" Dec 03 14:31:11 crc kubenswrapper[4805]: I1203 14:31:11.026912 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8d5p7\" (UniqueName: \"kubernetes.io/projected/58c98434-0b4d-4c2d-ac35-109e38f96803-kube-api-access-8d5p7\") pod \"ceilometer-0\" (UID: \"58c98434-0b4d-4c2d-ac35-109e38f96803\") " pod="openstack/ceilometer-0" Dec 03 14:31:11 crc kubenswrapper[4805]: I1203 14:31:11.046588 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 03 14:31:11 crc kubenswrapper[4805]: I1203 14:31:11.541663 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 03 14:31:11 crc kubenswrapper[4805]: W1203 14:31:11.546205 4805 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod58c98434_0b4d_4c2d_ac35_109e38f96803.slice/crio-b98faa6bdfacd7a3dbe424082fb5dc2be3f9f63287fa6fd24e34505f98826444 WatchSource:0}: Error finding container b98faa6bdfacd7a3dbe424082fb5dc2be3f9f63287fa6fd24e34505f98826444: Status 404 returned error can't find the container with id b98faa6bdfacd7a3dbe424082fb5dc2be3f9f63287fa6fd24e34505f98826444 Dec 03 14:31:11 crc kubenswrapper[4805]: I1203 14:31:11.641820 4805 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-sync-74tz4" Dec 03 14:31:11 crc kubenswrapper[4805]: I1203 14:31:11.816599 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-c5jlg\" (UniqueName: \"kubernetes.io/projected/13ba03f0-acc2-4fb8-90a9-583b4f8694cb-kube-api-access-c5jlg\") pod \"13ba03f0-acc2-4fb8-90a9-583b4f8694cb\" (UID: \"13ba03f0-acc2-4fb8-90a9-583b4f8694cb\") " Dec 03 14:31:11 crc kubenswrapper[4805]: I1203 14:31:11.816654 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/13ba03f0-acc2-4fb8-90a9-583b4f8694cb-combined-ca-bundle\") pod \"13ba03f0-acc2-4fb8-90a9-583b4f8694cb\" (UID: \"13ba03f0-acc2-4fb8-90a9-583b4f8694cb\") " Dec 03 14:31:11 crc kubenswrapper[4805]: I1203 14:31:11.816680 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/13ba03f0-acc2-4fb8-90a9-583b4f8694cb-db-sync-config-data\") pod \"13ba03f0-acc2-4fb8-90a9-583b4f8694cb\" (UID: \"13ba03f0-acc2-4fb8-90a9-583b4f8694cb\") " Dec 03 14:31:11 crc kubenswrapper[4805]: I1203 14:31:11.821693 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/13ba03f0-acc2-4fb8-90a9-583b4f8694cb-db-sync-config-data" (OuterVolumeSpecName: "db-sync-config-data") pod "13ba03f0-acc2-4fb8-90a9-583b4f8694cb" (UID: "13ba03f0-acc2-4fb8-90a9-583b4f8694cb"). InnerVolumeSpecName "db-sync-config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 14:31:11 crc kubenswrapper[4805]: I1203 14:31:11.822078 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/13ba03f0-acc2-4fb8-90a9-583b4f8694cb-kube-api-access-c5jlg" (OuterVolumeSpecName: "kube-api-access-c5jlg") pod "13ba03f0-acc2-4fb8-90a9-583b4f8694cb" (UID: "13ba03f0-acc2-4fb8-90a9-583b4f8694cb"). InnerVolumeSpecName "kube-api-access-c5jlg". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 14:31:11 crc kubenswrapper[4805]: I1203 14:31:11.847675 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/13ba03f0-acc2-4fb8-90a9-583b4f8694cb-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "13ba03f0-acc2-4fb8-90a9-583b4f8694cb" (UID: "13ba03f0-acc2-4fb8-90a9-583b4f8694cb"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 14:31:11 crc kubenswrapper[4805]: I1203 14:31:11.920470 4805 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-c5jlg\" (UniqueName: \"kubernetes.io/projected/13ba03f0-acc2-4fb8-90a9-583b4f8694cb-kube-api-access-c5jlg\") on node \"crc\" DevicePath \"\"" Dec 03 14:31:11 crc kubenswrapper[4805]: I1203 14:31:11.920513 4805 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/13ba03f0-acc2-4fb8-90a9-583b4f8694cb-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 14:31:11 crc kubenswrapper[4805]: I1203 14:31:11.920526 4805 reconciler_common.go:293] "Volume detached for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/13ba03f0-acc2-4fb8-90a9-583b4f8694cb-db-sync-config-data\") on node \"crc\" DevicePath \"\"" Dec 03 14:31:12 crc kubenswrapper[4805]: I1203 14:31:12.364616 4805 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-sync-74tz4" Dec 03 14:31:12 crc kubenswrapper[4805]: I1203 14:31:12.365025 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-sync-74tz4" event={"ID":"13ba03f0-acc2-4fb8-90a9-583b4f8694cb","Type":"ContainerDied","Data":"6c6724ca7aa2fbf1c806e79de571dbe6e1f2da3605e0f56f04d421afcbd569f6"} Dec 03 14:31:12 crc kubenswrapper[4805]: I1203 14:31:12.365152 4805 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="6c6724ca7aa2fbf1c806e79de571dbe6e1f2da3605e0f56f04d421afcbd569f6" Dec 03 14:31:12 crc kubenswrapper[4805]: I1203 14:31:12.366177 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"58c98434-0b4d-4c2d-ac35-109e38f96803","Type":"ContainerStarted","Data":"b98faa6bdfacd7a3dbe424082fb5dc2be3f9f63287fa6fd24e34505f98826444"} Dec 03 14:31:12 crc kubenswrapper[4805]: I1203 14:31:12.660100 4805 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-worker-8559568497-vwxwh"] Dec 03 14:31:12 crc kubenswrapper[4805]: E1203 14:31:12.660476 4805 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="13ba03f0-acc2-4fb8-90a9-583b4f8694cb" containerName="barbican-db-sync" Dec 03 14:31:12 crc kubenswrapper[4805]: I1203 14:31:12.660490 4805 state_mem.go:107] "Deleted CPUSet assignment" podUID="13ba03f0-acc2-4fb8-90a9-583b4f8694cb" containerName="barbican-db-sync" Dec 03 14:31:12 crc kubenswrapper[4805]: I1203 14:31:12.660645 4805 memory_manager.go:354] "RemoveStaleState removing state" podUID="13ba03f0-acc2-4fb8-90a9-583b4f8694cb" containerName="barbican-db-sync" Dec 03 14:31:12 crc kubenswrapper[4805]: I1203 14:31:12.662945 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-worker-8559568497-vwxwh" Dec 03 14:31:12 crc kubenswrapper[4805]: I1203 14:31:12.666125 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-barbican-dockercfg-kkm6j" Dec 03 14:31:12 crc kubenswrapper[4805]: I1203 14:31:12.666365 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-worker-config-data" Dec 03 14:31:12 crc kubenswrapper[4805]: I1203 14:31:12.666611 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-config-data" Dec 03 14:31:12 crc kubenswrapper[4805]: I1203 14:31:12.678817 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-worker-8559568497-vwxwh"] Dec 03 14:31:12 crc kubenswrapper[4805]: I1203 14:31:12.735332 4805 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-75c8ddd69c-s5r44"] Dec 03 14:31:12 crc kubenswrapper[4805]: I1203 14:31:12.737142 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-75c8ddd69c-s5r44" Dec 03 14:31:12 crc kubenswrapper[4805]: I1203 14:31:12.744634 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-75c8ddd69c-s5r44"] Dec 03 14:31:12 crc kubenswrapper[4805]: I1203 14:31:12.799051 4805 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-keystone-listener-8484c694-6frjp"] Dec 03 14:31:12 crc kubenswrapper[4805]: I1203 14:31:12.800582 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-keystone-listener-8484c694-6frjp" Dec 03 14:31:12 crc kubenswrapper[4805]: I1203 14:31:12.805565 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-keystone-listener-config-data" Dec 03 14:31:12 crc kubenswrapper[4805]: I1203 14:31:12.808685 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-keystone-listener-8484c694-6frjp"] Dec 03 14:31:12 crc kubenswrapper[4805]: I1203 14:31:12.836516 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wjs7b\" (UniqueName: \"kubernetes.io/projected/9f381959-280e-4eb1-9fb2-2365db2f9c43-kube-api-access-wjs7b\") pod \"dnsmasq-dns-75c8ddd69c-s5r44\" (UID: \"9f381959-280e-4eb1-9fb2-2365db2f9c43\") " pod="openstack/dnsmasq-dns-75c8ddd69c-s5r44" Dec 03 14:31:12 crc kubenswrapper[4805]: I1203 14:31:12.836576 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7fmrb\" (UniqueName: \"kubernetes.io/projected/93ad1063-a321-44a3-bd91-c00c7ba1ef96-kube-api-access-7fmrb\") pod \"barbican-worker-8559568497-vwxwh\" (UID: \"93ad1063-a321-44a3-bd91-c00c7ba1ef96\") " pod="openstack/barbican-worker-8559568497-vwxwh" Dec 03 14:31:12 crc kubenswrapper[4805]: I1203 14:31:12.836621 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/93ad1063-a321-44a3-bd91-c00c7ba1ef96-config-data\") pod \"barbican-worker-8559568497-vwxwh\" (UID: \"93ad1063-a321-44a3-bd91-c00c7ba1ef96\") " pod="openstack/barbican-worker-8559568497-vwxwh" Dec 03 14:31:12 crc kubenswrapper[4805]: I1203 14:31:12.836652 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9f381959-280e-4eb1-9fb2-2365db2f9c43-config\") pod \"dnsmasq-dns-75c8ddd69c-s5r44\" (UID: \"9f381959-280e-4eb1-9fb2-2365db2f9c43\") " pod="openstack/dnsmasq-dns-75c8ddd69c-s5r44" Dec 03 14:31:12 crc kubenswrapper[4805]: I1203 14:31:12.836680 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/93ad1063-a321-44a3-bd91-c00c7ba1ef96-config-data-custom\") pod \"barbican-worker-8559568497-vwxwh\" (UID: \"93ad1063-a321-44a3-bd91-c00c7ba1ef96\") " pod="openstack/barbican-worker-8559568497-vwxwh" Dec 03 14:31:12 crc kubenswrapper[4805]: I1203 14:31:12.836746 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/9f381959-280e-4eb1-9fb2-2365db2f9c43-ovsdbserver-nb\") pod \"dnsmasq-dns-75c8ddd69c-s5r44\" (UID: \"9f381959-280e-4eb1-9fb2-2365db2f9c43\") " pod="openstack/dnsmasq-dns-75c8ddd69c-s5r44" Dec 03 14:31:12 crc kubenswrapper[4805]: I1203 14:31:12.836797 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/93ad1063-a321-44a3-bd91-c00c7ba1ef96-logs\") pod \"barbican-worker-8559568497-vwxwh\" (UID: \"93ad1063-a321-44a3-bd91-c00c7ba1ef96\") " pod="openstack/barbican-worker-8559568497-vwxwh" Dec 03 14:31:12 crc kubenswrapper[4805]: I1203 14:31:12.836823 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/9f381959-280e-4eb1-9fb2-2365db2f9c43-dns-swift-storage-0\") pod \"dnsmasq-dns-75c8ddd69c-s5r44\" (UID: \"9f381959-280e-4eb1-9fb2-2365db2f9c43\") " pod="openstack/dnsmasq-dns-75c8ddd69c-s5r44" Dec 03 14:31:12 crc kubenswrapper[4805]: I1203 14:31:12.836865 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/9f381959-280e-4eb1-9fb2-2365db2f9c43-dns-svc\") pod \"dnsmasq-dns-75c8ddd69c-s5r44\" (UID: \"9f381959-280e-4eb1-9fb2-2365db2f9c43\") " pod="openstack/dnsmasq-dns-75c8ddd69c-s5r44" Dec 03 14:31:12 crc kubenswrapper[4805]: I1203 14:31:12.836888 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/93ad1063-a321-44a3-bd91-c00c7ba1ef96-combined-ca-bundle\") pod \"barbican-worker-8559568497-vwxwh\" (UID: \"93ad1063-a321-44a3-bd91-c00c7ba1ef96\") " pod="openstack/barbican-worker-8559568497-vwxwh" Dec 03 14:31:12 crc kubenswrapper[4805]: I1203 14:31:12.836936 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/9f381959-280e-4eb1-9fb2-2365db2f9c43-ovsdbserver-sb\") pod \"dnsmasq-dns-75c8ddd69c-s5r44\" (UID: \"9f381959-280e-4eb1-9fb2-2365db2f9c43\") " pod="openstack/dnsmasq-dns-75c8ddd69c-s5r44" Dec 03 14:31:12 crc kubenswrapper[4805]: I1203 14:31:12.898772 4805 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-api-845f47d448-thqcr"] Dec 03 14:31:12 crc kubenswrapper[4805]: I1203 14:31:12.900161 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-845f47d448-thqcr" Dec 03 14:31:12 crc kubenswrapper[4805]: I1203 14:31:12.908385 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-api-config-data" Dec 03 14:31:12 crc kubenswrapper[4805]: I1203 14:31:12.930029 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-api-845f47d448-thqcr"] Dec 03 14:31:12 crc kubenswrapper[4805]: I1203 14:31:12.942588 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/93ad1063-a321-44a3-bd91-c00c7ba1ef96-config-data\") pod \"barbican-worker-8559568497-vwxwh\" (UID: \"93ad1063-a321-44a3-bd91-c00c7ba1ef96\") " pod="openstack/barbican-worker-8559568497-vwxwh" Dec 03 14:31:12 crc kubenswrapper[4805]: I1203 14:31:12.942654 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9f381959-280e-4eb1-9fb2-2365db2f9c43-config\") pod \"dnsmasq-dns-75c8ddd69c-s5r44\" (UID: \"9f381959-280e-4eb1-9fb2-2365db2f9c43\") " pod="openstack/dnsmasq-dns-75c8ddd69c-s5r44" Dec 03 14:31:12 crc kubenswrapper[4805]: I1203 14:31:12.942681 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/93ad1063-a321-44a3-bd91-c00c7ba1ef96-config-data-custom\") pod \"barbican-worker-8559568497-vwxwh\" (UID: \"93ad1063-a321-44a3-bd91-c00c7ba1ef96\") " pod="openstack/barbican-worker-8559568497-vwxwh" Dec 03 14:31:12 crc kubenswrapper[4805]: I1203 14:31:12.942708 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/18349f43-5c89-4c11-8ff2-fee30d6cceb7-config-data-custom\") pod \"barbican-keystone-listener-8484c694-6frjp\" (UID: \"18349f43-5c89-4c11-8ff2-fee30d6cceb7\") " pod="openstack/barbican-keystone-listener-8484c694-6frjp" Dec 03 14:31:12 crc kubenswrapper[4805]: I1203 14:31:12.942825 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/9f381959-280e-4eb1-9fb2-2365db2f9c43-ovsdbserver-nb\") pod \"dnsmasq-dns-75c8ddd69c-s5r44\" (UID: \"9f381959-280e-4eb1-9fb2-2365db2f9c43\") " pod="openstack/dnsmasq-dns-75c8ddd69c-s5r44" Dec 03 14:31:12 crc kubenswrapper[4805]: I1203 14:31:12.942872 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/93ad1063-a321-44a3-bd91-c00c7ba1ef96-logs\") pod \"barbican-worker-8559568497-vwxwh\" (UID: \"93ad1063-a321-44a3-bd91-c00c7ba1ef96\") " pod="openstack/barbican-worker-8559568497-vwxwh" Dec 03 14:31:12 crc kubenswrapper[4805]: I1203 14:31:12.942896 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/9f381959-280e-4eb1-9fb2-2365db2f9c43-dns-swift-storage-0\") pod \"dnsmasq-dns-75c8ddd69c-s5r44\" (UID: \"9f381959-280e-4eb1-9fb2-2365db2f9c43\") " pod="openstack/dnsmasq-dns-75c8ddd69c-s5r44" Dec 03 14:31:12 crc kubenswrapper[4805]: I1203 14:31:12.942943 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/18349f43-5c89-4c11-8ff2-fee30d6cceb7-config-data\") pod \"barbican-keystone-listener-8484c694-6frjp\" (UID: \"18349f43-5c89-4c11-8ff2-fee30d6cceb7\") " pod="openstack/barbican-keystone-listener-8484c694-6frjp" Dec 03 14:31:12 crc kubenswrapper[4805]: I1203 14:31:12.942972 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/9f381959-280e-4eb1-9fb2-2365db2f9c43-dns-svc\") pod \"dnsmasq-dns-75c8ddd69c-s5r44\" (UID: \"9f381959-280e-4eb1-9fb2-2365db2f9c43\") " pod="openstack/dnsmasq-dns-75c8ddd69c-s5r44" Dec 03 14:31:12 crc kubenswrapper[4805]: I1203 14:31:12.942995 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/93ad1063-a321-44a3-bd91-c00c7ba1ef96-combined-ca-bundle\") pod \"barbican-worker-8559568497-vwxwh\" (UID: \"93ad1063-a321-44a3-bd91-c00c7ba1ef96\") " pod="openstack/barbican-worker-8559568497-vwxwh" Dec 03 14:31:12 crc kubenswrapper[4805]: I1203 14:31:12.943104 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4wvh4\" (UniqueName: \"kubernetes.io/projected/18349f43-5c89-4c11-8ff2-fee30d6cceb7-kube-api-access-4wvh4\") pod \"barbican-keystone-listener-8484c694-6frjp\" (UID: \"18349f43-5c89-4c11-8ff2-fee30d6cceb7\") " pod="openstack/barbican-keystone-listener-8484c694-6frjp" Dec 03 14:31:12 crc kubenswrapper[4805]: I1203 14:31:12.943147 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/18349f43-5c89-4c11-8ff2-fee30d6cceb7-logs\") pod \"barbican-keystone-listener-8484c694-6frjp\" (UID: \"18349f43-5c89-4c11-8ff2-fee30d6cceb7\") " pod="openstack/barbican-keystone-listener-8484c694-6frjp" Dec 03 14:31:12 crc kubenswrapper[4805]: I1203 14:31:12.943206 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/9f381959-280e-4eb1-9fb2-2365db2f9c43-ovsdbserver-sb\") pod \"dnsmasq-dns-75c8ddd69c-s5r44\" (UID: \"9f381959-280e-4eb1-9fb2-2365db2f9c43\") " pod="openstack/dnsmasq-dns-75c8ddd69c-s5r44" Dec 03 14:31:12 crc kubenswrapper[4805]: I1203 14:31:12.943258 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/18349f43-5c89-4c11-8ff2-fee30d6cceb7-combined-ca-bundle\") pod \"barbican-keystone-listener-8484c694-6frjp\" (UID: \"18349f43-5c89-4c11-8ff2-fee30d6cceb7\") " pod="openstack/barbican-keystone-listener-8484c694-6frjp" Dec 03 14:31:12 crc kubenswrapper[4805]: I1203 14:31:12.943285 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wjs7b\" (UniqueName: \"kubernetes.io/projected/9f381959-280e-4eb1-9fb2-2365db2f9c43-kube-api-access-wjs7b\") pod \"dnsmasq-dns-75c8ddd69c-s5r44\" (UID: \"9f381959-280e-4eb1-9fb2-2365db2f9c43\") " pod="openstack/dnsmasq-dns-75c8ddd69c-s5r44" Dec 03 14:31:12 crc kubenswrapper[4805]: I1203 14:31:12.943309 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7fmrb\" (UniqueName: \"kubernetes.io/projected/93ad1063-a321-44a3-bd91-c00c7ba1ef96-kube-api-access-7fmrb\") pod \"barbican-worker-8559568497-vwxwh\" (UID: \"93ad1063-a321-44a3-bd91-c00c7ba1ef96\") " pod="openstack/barbican-worker-8559568497-vwxwh" Dec 03 14:31:12 crc kubenswrapper[4805]: I1203 14:31:12.944359 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/9f381959-280e-4eb1-9fb2-2365db2f9c43-dns-svc\") pod \"dnsmasq-dns-75c8ddd69c-s5r44\" (UID: \"9f381959-280e-4eb1-9fb2-2365db2f9c43\") " pod="openstack/dnsmasq-dns-75c8ddd69c-s5r44" Dec 03 14:31:12 crc kubenswrapper[4805]: I1203 14:31:12.944933 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/9f381959-280e-4eb1-9fb2-2365db2f9c43-ovsdbserver-nb\") pod \"dnsmasq-dns-75c8ddd69c-s5r44\" (UID: \"9f381959-280e-4eb1-9fb2-2365db2f9c43\") " pod="openstack/dnsmasq-dns-75c8ddd69c-s5r44" Dec 03 14:31:12 crc kubenswrapper[4805]: I1203 14:31:12.945177 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/93ad1063-a321-44a3-bd91-c00c7ba1ef96-logs\") pod \"barbican-worker-8559568497-vwxwh\" (UID: \"93ad1063-a321-44a3-bd91-c00c7ba1ef96\") " pod="openstack/barbican-worker-8559568497-vwxwh" Dec 03 14:31:12 crc kubenswrapper[4805]: I1203 14:31:12.945788 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/9f381959-280e-4eb1-9fb2-2365db2f9c43-ovsdbserver-sb\") pod \"dnsmasq-dns-75c8ddd69c-s5r44\" (UID: \"9f381959-280e-4eb1-9fb2-2365db2f9c43\") " pod="openstack/dnsmasq-dns-75c8ddd69c-s5r44" Dec 03 14:31:12 crc kubenswrapper[4805]: I1203 14:31:12.946277 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9f381959-280e-4eb1-9fb2-2365db2f9c43-config\") pod \"dnsmasq-dns-75c8ddd69c-s5r44\" (UID: \"9f381959-280e-4eb1-9fb2-2365db2f9c43\") " pod="openstack/dnsmasq-dns-75c8ddd69c-s5r44" Dec 03 14:31:12 crc kubenswrapper[4805]: I1203 14:31:12.947280 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/9f381959-280e-4eb1-9fb2-2365db2f9c43-dns-swift-storage-0\") pod \"dnsmasq-dns-75c8ddd69c-s5r44\" (UID: \"9f381959-280e-4eb1-9fb2-2365db2f9c43\") " pod="openstack/dnsmasq-dns-75c8ddd69c-s5r44" Dec 03 14:31:12 crc kubenswrapper[4805]: I1203 14:31:12.948286 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/93ad1063-a321-44a3-bd91-c00c7ba1ef96-config-data\") pod \"barbican-worker-8559568497-vwxwh\" (UID: \"93ad1063-a321-44a3-bd91-c00c7ba1ef96\") " pod="openstack/barbican-worker-8559568497-vwxwh" Dec 03 14:31:12 crc kubenswrapper[4805]: I1203 14:31:12.952178 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/93ad1063-a321-44a3-bd91-c00c7ba1ef96-config-data-custom\") pod \"barbican-worker-8559568497-vwxwh\" (UID: \"93ad1063-a321-44a3-bd91-c00c7ba1ef96\") " pod="openstack/barbican-worker-8559568497-vwxwh" Dec 03 14:31:12 crc kubenswrapper[4805]: I1203 14:31:12.952972 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/93ad1063-a321-44a3-bd91-c00c7ba1ef96-combined-ca-bundle\") pod \"barbican-worker-8559568497-vwxwh\" (UID: \"93ad1063-a321-44a3-bd91-c00c7ba1ef96\") " pod="openstack/barbican-worker-8559568497-vwxwh" Dec 03 14:31:12 crc kubenswrapper[4805]: I1203 14:31:12.968991 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wjs7b\" (UniqueName: \"kubernetes.io/projected/9f381959-280e-4eb1-9fb2-2365db2f9c43-kube-api-access-wjs7b\") pod \"dnsmasq-dns-75c8ddd69c-s5r44\" (UID: \"9f381959-280e-4eb1-9fb2-2365db2f9c43\") " pod="openstack/dnsmasq-dns-75c8ddd69c-s5r44" Dec 03 14:31:12 crc kubenswrapper[4805]: I1203 14:31:12.976299 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7fmrb\" (UniqueName: \"kubernetes.io/projected/93ad1063-a321-44a3-bd91-c00c7ba1ef96-kube-api-access-7fmrb\") pod \"barbican-worker-8559568497-vwxwh\" (UID: \"93ad1063-a321-44a3-bd91-c00c7ba1ef96\") " pod="openstack/barbican-worker-8559568497-vwxwh" Dec 03 14:31:12 crc kubenswrapper[4805]: I1203 14:31:12.987453 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-worker-8559568497-vwxwh" Dec 03 14:31:13 crc kubenswrapper[4805]: I1203 14:31:13.044908 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/37658956-d75a-43ce-81d8-38f9bca42957-config-data-custom\") pod \"barbican-api-845f47d448-thqcr\" (UID: \"37658956-d75a-43ce-81d8-38f9bca42957\") " pod="openstack/barbican-api-845f47d448-thqcr" Dec 03 14:31:13 crc kubenswrapper[4805]: I1203 14:31:13.044974 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/37658956-d75a-43ce-81d8-38f9bca42957-combined-ca-bundle\") pod \"barbican-api-845f47d448-thqcr\" (UID: \"37658956-d75a-43ce-81d8-38f9bca42957\") " pod="openstack/barbican-api-845f47d448-thqcr" Dec 03 14:31:13 crc kubenswrapper[4805]: I1203 14:31:13.045006 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/18349f43-5c89-4c11-8ff2-fee30d6cceb7-config-data\") pod \"barbican-keystone-listener-8484c694-6frjp\" (UID: \"18349f43-5c89-4c11-8ff2-fee30d6cceb7\") " pod="openstack/barbican-keystone-listener-8484c694-6frjp" Dec 03 14:31:13 crc kubenswrapper[4805]: I1203 14:31:13.045032 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4wvh4\" (UniqueName: \"kubernetes.io/projected/18349f43-5c89-4c11-8ff2-fee30d6cceb7-kube-api-access-4wvh4\") pod \"barbican-keystone-listener-8484c694-6frjp\" (UID: \"18349f43-5c89-4c11-8ff2-fee30d6cceb7\") " pod="openstack/barbican-keystone-listener-8484c694-6frjp" Dec 03 14:31:13 crc kubenswrapper[4805]: I1203 14:31:13.045062 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/37658956-d75a-43ce-81d8-38f9bca42957-logs\") pod \"barbican-api-845f47d448-thqcr\" (UID: \"37658956-d75a-43ce-81d8-38f9bca42957\") " pod="openstack/barbican-api-845f47d448-thqcr" Dec 03 14:31:13 crc kubenswrapper[4805]: I1203 14:31:13.045083 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/18349f43-5c89-4c11-8ff2-fee30d6cceb7-logs\") pod \"barbican-keystone-listener-8484c694-6frjp\" (UID: \"18349f43-5c89-4c11-8ff2-fee30d6cceb7\") " pod="openstack/barbican-keystone-listener-8484c694-6frjp" Dec 03 14:31:13 crc kubenswrapper[4805]: I1203 14:31:13.045107 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/37658956-d75a-43ce-81d8-38f9bca42957-config-data\") pod \"barbican-api-845f47d448-thqcr\" (UID: \"37658956-d75a-43ce-81d8-38f9bca42957\") " pod="openstack/barbican-api-845f47d448-thqcr" Dec 03 14:31:13 crc kubenswrapper[4805]: I1203 14:31:13.045142 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/18349f43-5c89-4c11-8ff2-fee30d6cceb7-combined-ca-bundle\") pod \"barbican-keystone-listener-8484c694-6frjp\" (UID: \"18349f43-5c89-4c11-8ff2-fee30d6cceb7\") " pod="openstack/barbican-keystone-listener-8484c694-6frjp" Dec 03 14:31:13 crc kubenswrapper[4805]: I1203 14:31:13.045163 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wdmrf\" (UniqueName: \"kubernetes.io/projected/37658956-d75a-43ce-81d8-38f9bca42957-kube-api-access-wdmrf\") pod \"barbican-api-845f47d448-thqcr\" (UID: \"37658956-d75a-43ce-81d8-38f9bca42957\") " pod="openstack/barbican-api-845f47d448-thqcr" Dec 03 14:31:13 crc kubenswrapper[4805]: I1203 14:31:13.045199 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/18349f43-5c89-4c11-8ff2-fee30d6cceb7-config-data-custom\") pod \"barbican-keystone-listener-8484c694-6frjp\" (UID: \"18349f43-5c89-4c11-8ff2-fee30d6cceb7\") " pod="openstack/barbican-keystone-listener-8484c694-6frjp" Dec 03 14:31:13 crc kubenswrapper[4805]: I1203 14:31:13.050436 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/18349f43-5c89-4c11-8ff2-fee30d6cceb7-logs\") pod \"barbican-keystone-listener-8484c694-6frjp\" (UID: \"18349f43-5c89-4c11-8ff2-fee30d6cceb7\") " pod="openstack/barbican-keystone-listener-8484c694-6frjp" Dec 03 14:31:13 crc kubenswrapper[4805]: I1203 14:31:13.054630 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/18349f43-5c89-4c11-8ff2-fee30d6cceb7-config-data\") pod \"barbican-keystone-listener-8484c694-6frjp\" (UID: \"18349f43-5c89-4c11-8ff2-fee30d6cceb7\") " pod="openstack/barbican-keystone-listener-8484c694-6frjp" Dec 03 14:31:13 crc kubenswrapper[4805]: I1203 14:31:13.056826 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/18349f43-5c89-4c11-8ff2-fee30d6cceb7-combined-ca-bundle\") pod \"barbican-keystone-listener-8484c694-6frjp\" (UID: \"18349f43-5c89-4c11-8ff2-fee30d6cceb7\") " pod="openstack/barbican-keystone-listener-8484c694-6frjp" Dec 03 14:31:13 crc kubenswrapper[4805]: I1203 14:31:13.059279 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-75c8ddd69c-s5r44" Dec 03 14:31:13 crc kubenswrapper[4805]: I1203 14:31:13.064327 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/18349f43-5c89-4c11-8ff2-fee30d6cceb7-config-data-custom\") pod \"barbican-keystone-listener-8484c694-6frjp\" (UID: \"18349f43-5c89-4c11-8ff2-fee30d6cceb7\") " pod="openstack/barbican-keystone-listener-8484c694-6frjp" Dec 03 14:31:13 crc kubenswrapper[4805]: I1203 14:31:13.067636 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4wvh4\" (UniqueName: \"kubernetes.io/projected/18349f43-5c89-4c11-8ff2-fee30d6cceb7-kube-api-access-4wvh4\") pod \"barbican-keystone-listener-8484c694-6frjp\" (UID: \"18349f43-5c89-4c11-8ff2-fee30d6cceb7\") " pod="openstack/barbican-keystone-listener-8484c694-6frjp" Dec 03 14:31:13 crc kubenswrapper[4805]: I1203 14:31:13.115938 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-keystone-listener-8484c694-6frjp" Dec 03 14:31:13 crc kubenswrapper[4805]: I1203 14:31:13.147517 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/37658956-d75a-43ce-81d8-38f9bca42957-logs\") pod \"barbican-api-845f47d448-thqcr\" (UID: \"37658956-d75a-43ce-81d8-38f9bca42957\") " pod="openstack/barbican-api-845f47d448-thqcr" Dec 03 14:31:13 crc kubenswrapper[4805]: I1203 14:31:13.147585 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/37658956-d75a-43ce-81d8-38f9bca42957-config-data\") pod \"barbican-api-845f47d448-thqcr\" (UID: \"37658956-d75a-43ce-81d8-38f9bca42957\") " pod="openstack/barbican-api-845f47d448-thqcr" Dec 03 14:31:13 crc kubenswrapper[4805]: I1203 14:31:13.147645 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wdmrf\" (UniqueName: \"kubernetes.io/projected/37658956-d75a-43ce-81d8-38f9bca42957-kube-api-access-wdmrf\") pod \"barbican-api-845f47d448-thqcr\" (UID: \"37658956-d75a-43ce-81d8-38f9bca42957\") " pod="openstack/barbican-api-845f47d448-thqcr" Dec 03 14:31:13 crc kubenswrapper[4805]: I1203 14:31:13.147725 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/37658956-d75a-43ce-81d8-38f9bca42957-config-data-custom\") pod \"barbican-api-845f47d448-thqcr\" (UID: \"37658956-d75a-43ce-81d8-38f9bca42957\") " pod="openstack/barbican-api-845f47d448-thqcr" Dec 03 14:31:13 crc kubenswrapper[4805]: I1203 14:31:13.147768 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/37658956-d75a-43ce-81d8-38f9bca42957-combined-ca-bundle\") pod \"barbican-api-845f47d448-thqcr\" (UID: \"37658956-d75a-43ce-81d8-38f9bca42957\") " pod="openstack/barbican-api-845f47d448-thqcr" Dec 03 14:31:13 crc kubenswrapper[4805]: I1203 14:31:13.148053 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/37658956-d75a-43ce-81d8-38f9bca42957-logs\") pod \"barbican-api-845f47d448-thqcr\" (UID: \"37658956-d75a-43ce-81d8-38f9bca42957\") " pod="openstack/barbican-api-845f47d448-thqcr" Dec 03 14:31:13 crc kubenswrapper[4805]: I1203 14:31:13.153353 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/37658956-d75a-43ce-81d8-38f9bca42957-config-data-custom\") pod \"barbican-api-845f47d448-thqcr\" (UID: \"37658956-d75a-43ce-81d8-38f9bca42957\") " pod="openstack/barbican-api-845f47d448-thqcr" Dec 03 14:31:13 crc kubenswrapper[4805]: I1203 14:31:13.153374 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/37658956-d75a-43ce-81d8-38f9bca42957-combined-ca-bundle\") pod \"barbican-api-845f47d448-thqcr\" (UID: \"37658956-d75a-43ce-81d8-38f9bca42957\") " pod="openstack/barbican-api-845f47d448-thqcr" Dec 03 14:31:13 crc kubenswrapper[4805]: I1203 14:31:13.155866 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/37658956-d75a-43ce-81d8-38f9bca42957-config-data\") pod \"barbican-api-845f47d448-thqcr\" (UID: \"37658956-d75a-43ce-81d8-38f9bca42957\") " pod="openstack/barbican-api-845f47d448-thqcr" Dec 03 14:31:13 crc kubenswrapper[4805]: I1203 14:31:13.167138 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wdmrf\" (UniqueName: \"kubernetes.io/projected/37658956-d75a-43ce-81d8-38f9bca42957-kube-api-access-wdmrf\") pod \"barbican-api-845f47d448-thqcr\" (UID: \"37658956-d75a-43ce-81d8-38f9bca42957\") " pod="openstack/barbican-api-845f47d448-thqcr" Dec 03 14:31:13 crc kubenswrapper[4805]: I1203 14:31:13.230427 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-845f47d448-thqcr" Dec 03 14:31:13 crc kubenswrapper[4805]: I1203 14:31:13.399382 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"58c98434-0b4d-4c2d-ac35-109e38f96803","Type":"ContainerStarted","Data":"f98e610e128d27a8cf1f7ca39fe6a3f759d8784ae234f8e1a4c4d5e228950639"} Dec 03 14:31:13 crc kubenswrapper[4805]: I1203 14:31:13.477865 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-worker-8559568497-vwxwh"] Dec 03 14:31:13 crc kubenswrapper[4805]: I1203 14:31:13.626439 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-75c8ddd69c-s5r44"] Dec 03 14:31:13 crc kubenswrapper[4805]: W1203 14:31:13.626666 4805 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod9f381959_280e_4eb1_9fb2_2365db2f9c43.slice/crio-792a5a7dd26d0a43f6c48455682e32c7fb8958acbc12cea35384caede78c4f97 WatchSource:0}: Error finding container 792a5a7dd26d0a43f6c48455682e32c7fb8958acbc12cea35384caede78c4f97: Status 404 returned error can't find the container with id 792a5a7dd26d0a43f6c48455682e32c7fb8958acbc12cea35384caede78c4f97 Dec 03 14:31:13 crc kubenswrapper[4805]: I1203 14:31:13.699228 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-keystone-listener-8484c694-6frjp"] Dec 03 14:31:13 crc kubenswrapper[4805]: I1203 14:31:13.762486 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-api-845f47d448-thqcr"] Dec 03 14:31:13 crc kubenswrapper[4805]: W1203 14:31:13.772291 4805 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod37658956_d75a_43ce_81d8_38f9bca42957.slice/crio-49d130dec5d88f2be0467ebb81763437af671eee08cac357241332cdcd156538 WatchSource:0}: Error finding container 49d130dec5d88f2be0467ebb81763437af671eee08cac357241332cdcd156538: Status 404 returned error can't find the container with id 49d130dec5d88f2be0467ebb81763437af671eee08cac357241332cdcd156538 Dec 03 14:31:14 crc kubenswrapper[4805]: I1203 14:31:14.418674 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-keystone-listener-8484c694-6frjp" event={"ID":"18349f43-5c89-4c11-8ff2-fee30d6cceb7","Type":"ContainerStarted","Data":"c451c3da83954b1770839d1182dabd6b5a45794185254c625bdb2463d77a9812"} Dec 03 14:31:14 crc kubenswrapper[4805]: I1203 14:31:14.421663 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-845f47d448-thqcr" event={"ID":"37658956-d75a-43ce-81d8-38f9bca42957","Type":"ContainerStarted","Data":"49d130dec5d88f2be0467ebb81763437af671eee08cac357241332cdcd156538"} Dec 03 14:31:14 crc kubenswrapper[4805]: I1203 14:31:14.423928 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-75c8ddd69c-s5r44" event={"ID":"9f381959-280e-4eb1-9fb2-2365db2f9c43","Type":"ContainerStarted","Data":"dfafb0c8fd4a424b2f580f9c4f5ed8901168d0cee724096075c3a6574fb656a5"} Dec 03 14:31:14 crc kubenswrapper[4805]: I1203 14:31:14.423966 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-75c8ddd69c-s5r44" event={"ID":"9f381959-280e-4eb1-9fb2-2365db2f9c43","Type":"ContainerStarted","Data":"792a5a7dd26d0a43f6c48455682e32c7fb8958acbc12cea35384caede78c4f97"} Dec 03 14:31:14 crc kubenswrapper[4805]: I1203 14:31:14.433000 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-worker-8559568497-vwxwh" event={"ID":"93ad1063-a321-44a3-bd91-c00c7ba1ef96","Type":"ContainerStarted","Data":"f97fc8494a87ec7151250f95cfcaa848d86a7a36df58a3edead94d4e0dc09bb8"} Dec 03 14:31:15 crc kubenswrapper[4805]: I1203 14:31:15.444121 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-845f47d448-thqcr" event={"ID":"37658956-d75a-43ce-81d8-38f9bca42957","Type":"ContainerStarted","Data":"a54729ca2abce4df438fff733d8adb4838b1bcfb1e74dd34f90719edf58bd259"} Dec 03 14:31:15 crc kubenswrapper[4805]: I1203 14:31:15.445513 4805 generic.go:334] "Generic (PLEG): container finished" podID="9f381959-280e-4eb1-9fb2-2365db2f9c43" containerID="dfafb0c8fd4a424b2f580f9c4f5ed8901168d0cee724096075c3a6574fb656a5" exitCode=0 Dec 03 14:31:15 crc kubenswrapper[4805]: I1203 14:31:15.445537 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-75c8ddd69c-s5r44" event={"ID":"9f381959-280e-4eb1-9fb2-2365db2f9c43","Type":"ContainerDied","Data":"dfafb0c8fd4a424b2f580f9c4f5ed8901168d0cee724096075c3a6574fb656a5"} Dec 03 14:31:15 crc kubenswrapper[4805]: I1203 14:31:15.556896 4805 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-api-58794647f4-rtjlq"] Dec 03 14:31:15 crc kubenswrapper[4805]: I1203 14:31:15.559432 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-58794647f4-rtjlq" Dec 03 14:31:15 crc kubenswrapper[4805]: I1203 14:31:15.564258 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-api-58794647f4-rtjlq"] Dec 03 14:31:15 crc kubenswrapper[4805]: I1203 14:31:15.564904 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-barbican-internal-svc" Dec 03 14:31:15 crc kubenswrapper[4805]: I1203 14:31:15.565112 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-barbican-public-svc" Dec 03 14:31:15 crc kubenswrapper[4805]: I1203 14:31:15.712899 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/5a55a37b-c678-481e-bf36-a8961f5b26ed-public-tls-certs\") pod \"barbican-api-58794647f4-rtjlq\" (UID: \"5a55a37b-c678-481e-bf36-a8961f5b26ed\") " pod="openstack/barbican-api-58794647f4-rtjlq" Dec 03 14:31:15 crc kubenswrapper[4805]: I1203 14:31:15.712986 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/5a55a37b-c678-481e-bf36-a8961f5b26ed-config-data-custom\") pod \"barbican-api-58794647f4-rtjlq\" (UID: \"5a55a37b-c678-481e-bf36-a8961f5b26ed\") " pod="openstack/barbican-api-58794647f4-rtjlq" Dec 03 14:31:15 crc kubenswrapper[4805]: I1203 14:31:15.713062 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/5a55a37b-c678-481e-bf36-a8961f5b26ed-logs\") pod \"barbican-api-58794647f4-rtjlq\" (UID: \"5a55a37b-c678-481e-bf36-a8961f5b26ed\") " pod="openstack/barbican-api-58794647f4-rtjlq" Dec 03 14:31:15 crc kubenswrapper[4805]: I1203 14:31:15.713137 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5a55a37b-c678-481e-bf36-a8961f5b26ed-config-data\") pod \"barbican-api-58794647f4-rtjlq\" (UID: \"5a55a37b-c678-481e-bf36-a8961f5b26ed\") " pod="openstack/barbican-api-58794647f4-rtjlq" Dec 03 14:31:15 crc kubenswrapper[4805]: I1203 14:31:15.713278 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bqn2c\" (UniqueName: \"kubernetes.io/projected/5a55a37b-c678-481e-bf36-a8961f5b26ed-kube-api-access-bqn2c\") pod \"barbican-api-58794647f4-rtjlq\" (UID: \"5a55a37b-c678-481e-bf36-a8961f5b26ed\") " pod="openstack/barbican-api-58794647f4-rtjlq" Dec 03 14:31:15 crc kubenswrapper[4805]: I1203 14:31:15.713306 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5a55a37b-c678-481e-bf36-a8961f5b26ed-combined-ca-bundle\") pod \"barbican-api-58794647f4-rtjlq\" (UID: \"5a55a37b-c678-481e-bf36-a8961f5b26ed\") " pod="openstack/barbican-api-58794647f4-rtjlq" Dec 03 14:31:15 crc kubenswrapper[4805]: I1203 14:31:15.713474 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/5a55a37b-c678-481e-bf36-a8961f5b26ed-internal-tls-certs\") pod \"barbican-api-58794647f4-rtjlq\" (UID: \"5a55a37b-c678-481e-bf36-a8961f5b26ed\") " pod="openstack/barbican-api-58794647f4-rtjlq" Dec 03 14:31:15 crc kubenswrapper[4805]: I1203 14:31:15.814697 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/5a55a37b-c678-481e-bf36-a8961f5b26ed-internal-tls-certs\") pod \"barbican-api-58794647f4-rtjlq\" (UID: \"5a55a37b-c678-481e-bf36-a8961f5b26ed\") " pod="openstack/barbican-api-58794647f4-rtjlq" Dec 03 14:31:15 crc kubenswrapper[4805]: I1203 14:31:15.814859 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/5a55a37b-c678-481e-bf36-a8961f5b26ed-public-tls-certs\") pod \"barbican-api-58794647f4-rtjlq\" (UID: \"5a55a37b-c678-481e-bf36-a8961f5b26ed\") " pod="openstack/barbican-api-58794647f4-rtjlq" Dec 03 14:31:15 crc kubenswrapper[4805]: I1203 14:31:15.814885 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/5a55a37b-c678-481e-bf36-a8961f5b26ed-config-data-custom\") pod \"barbican-api-58794647f4-rtjlq\" (UID: \"5a55a37b-c678-481e-bf36-a8961f5b26ed\") " pod="openstack/barbican-api-58794647f4-rtjlq" Dec 03 14:31:15 crc kubenswrapper[4805]: I1203 14:31:15.814919 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/5a55a37b-c678-481e-bf36-a8961f5b26ed-logs\") pod \"barbican-api-58794647f4-rtjlq\" (UID: \"5a55a37b-c678-481e-bf36-a8961f5b26ed\") " pod="openstack/barbican-api-58794647f4-rtjlq" Dec 03 14:31:15 crc kubenswrapper[4805]: I1203 14:31:15.814957 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5a55a37b-c678-481e-bf36-a8961f5b26ed-config-data\") pod \"barbican-api-58794647f4-rtjlq\" (UID: \"5a55a37b-c678-481e-bf36-a8961f5b26ed\") " pod="openstack/barbican-api-58794647f4-rtjlq" Dec 03 14:31:15 crc kubenswrapper[4805]: I1203 14:31:15.815015 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bqn2c\" (UniqueName: \"kubernetes.io/projected/5a55a37b-c678-481e-bf36-a8961f5b26ed-kube-api-access-bqn2c\") pod \"barbican-api-58794647f4-rtjlq\" (UID: \"5a55a37b-c678-481e-bf36-a8961f5b26ed\") " pod="openstack/barbican-api-58794647f4-rtjlq" Dec 03 14:31:15 crc kubenswrapper[4805]: I1203 14:31:15.815037 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5a55a37b-c678-481e-bf36-a8961f5b26ed-combined-ca-bundle\") pod \"barbican-api-58794647f4-rtjlq\" (UID: \"5a55a37b-c678-481e-bf36-a8961f5b26ed\") " pod="openstack/barbican-api-58794647f4-rtjlq" Dec 03 14:31:15 crc kubenswrapper[4805]: I1203 14:31:15.816395 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/5a55a37b-c678-481e-bf36-a8961f5b26ed-logs\") pod \"barbican-api-58794647f4-rtjlq\" (UID: \"5a55a37b-c678-481e-bf36-a8961f5b26ed\") " pod="openstack/barbican-api-58794647f4-rtjlq" Dec 03 14:31:15 crc kubenswrapper[4805]: I1203 14:31:15.819215 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5a55a37b-c678-481e-bf36-a8961f5b26ed-combined-ca-bundle\") pod \"barbican-api-58794647f4-rtjlq\" (UID: \"5a55a37b-c678-481e-bf36-a8961f5b26ed\") " pod="openstack/barbican-api-58794647f4-rtjlq" Dec 03 14:31:15 crc kubenswrapper[4805]: I1203 14:31:15.819729 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/5a55a37b-c678-481e-bf36-a8961f5b26ed-public-tls-certs\") pod \"barbican-api-58794647f4-rtjlq\" (UID: \"5a55a37b-c678-481e-bf36-a8961f5b26ed\") " pod="openstack/barbican-api-58794647f4-rtjlq" Dec 03 14:31:15 crc kubenswrapper[4805]: I1203 14:31:15.821777 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/5a55a37b-c678-481e-bf36-a8961f5b26ed-config-data-custom\") pod \"barbican-api-58794647f4-rtjlq\" (UID: \"5a55a37b-c678-481e-bf36-a8961f5b26ed\") " pod="openstack/barbican-api-58794647f4-rtjlq" Dec 03 14:31:15 crc kubenswrapper[4805]: I1203 14:31:15.823185 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5a55a37b-c678-481e-bf36-a8961f5b26ed-config-data\") pod \"barbican-api-58794647f4-rtjlq\" (UID: \"5a55a37b-c678-481e-bf36-a8961f5b26ed\") " pod="openstack/barbican-api-58794647f4-rtjlq" Dec 03 14:31:15 crc kubenswrapper[4805]: I1203 14:31:15.830820 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/5a55a37b-c678-481e-bf36-a8961f5b26ed-internal-tls-certs\") pod \"barbican-api-58794647f4-rtjlq\" (UID: \"5a55a37b-c678-481e-bf36-a8961f5b26ed\") " pod="openstack/barbican-api-58794647f4-rtjlq" Dec 03 14:31:15 crc kubenswrapper[4805]: I1203 14:31:15.832674 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bqn2c\" (UniqueName: \"kubernetes.io/projected/5a55a37b-c678-481e-bf36-a8961f5b26ed-kube-api-access-bqn2c\") pod \"barbican-api-58794647f4-rtjlq\" (UID: \"5a55a37b-c678-481e-bf36-a8961f5b26ed\") " pod="openstack/barbican-api-58794647f4-rtjlq" Dec 03 14:31:15 crc kubenswrapper[4805]: I1203 14:31:15.898712 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-58794647f4-rtjlq" Dec 03 14:31:16 crc kubenswrapper[4805]: I1203 14:31:16.387212 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-api-58794647f4-rtjlq"] Dec 03 14:31:16 crc kubenswrapper[4805]: I1203 14:31:16.456562 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-75c8ddd69c-s5r44" event={"ID":"9f381959-280e-4eb1-9fb2-2365db2f9c43","Type":"ContainerStarted","Data":"0f465542ed3b0873cf31f730f8258e4061dbb10f0ebce0003303d33cd5b14544"} Dec 03 14:31:16 crc kubenswrapper[4805]: I1203 14:31:16.457452 4805 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-75c8ddd69c-s5r44" Dec 03 14:31:16 crc kubenswrapper[4805]: I1203 14:31:16.470777 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"58c98434-0b4d-4c2d-ac35-109e38f96803","Type":"ContainerStarted","Data":"9cf635a2329565cc994a3a2337944c2eca2f7047534f27d7eeb5f3ceaa96458b"} Dec 03 14:31:16 crc kubenswrapper[4805]: I1203 14:31:16.483416 4805 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-75c8ddd69c-s5r44" podStartSLOduration=4.483389351 podStartE2EDuration="4.483389351s" podCreationTimestamp="2025-12-03 14:31:12 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 14:31:16.47497697 +0000 UTC m=+1306.137893903" watchObservedRunningTime="2025-12-03 14:31:16.483389351 +0000 UTC m=+1306.146306274" Dec 03 14:31:16 crc kubenswrapper[4805]: I1203 14:31:16.485694 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-845f47d448-thqcr" event={"ID":"37658956-d75a-43ce-81d8-38f9bca42957","Type":"ContainerStarted","Data":"7adb8b3c5e068f3ee12e9581aecb5bdfa43066e3e1f0991bc28977ea2af801da"} Dec 03 14:31:16 crc kubenswrapper[4805]: I1203 14:31:16.489371 4805 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/barbican-api-845f47d448-thqcr" Dec 03 14:31:16 crc kubenswrapper[4805]: I1203 14:31:16.489423 4805 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/barbican-api-845f47d448-thqcr" Dec 03 14:31:16 crc kubenswrapper[4805]: I1203 14:31:16.518200 4805 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-api-845f47d448-thqcr" podStartSLOduration=4.5181777929999996 podStartE2EDuration="4.518177793s" podCreationTimestamp="2025-12-03 14:31:12 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 14:31:16.505509676 +0000 UTC m=+1306.168426599" watchObservedRunningTime="2025-12-03 14:31:16.518177793 +0000 UTC m=+1306.181094716" Dec 03 14:31:17 crc kubenswrapper[4805]: W1203 14:31:17.089272 4805 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod5a55a37b_c678_481e_bf36_a8961f5b26ed.slice/crio-ac72c3aff3061c9e7f744a4087f0126d38161da3c7cb8ad20de706d90c322aee WatchSource:0}: Error finding container ac72c3aff3061c9e7f744a4087f0126d38161da3c7cb8ad20de706d90c322aee: Status 404 returned error can't find the container with id ac72c3aff3061c9e7f744a4087f0126d38161da3c7cb8ad20de706d90c322aee Dec 03 14:31:17 crc kubenswrapper[4805]: I1203 14:31:17.500091 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-58794647f4-rtjlq" event={"ID":"5a55a37b-c678-481e-bf36-a8961f5b26ed","Type":"ContainerStarted","Data":"ac72c3aff3061c9e7f744a4087f0126d38161da3c7cb8ad20de706d90c322aee"} Dec 03 14:31:18 crc kubenswrapper[4805]: I1203 14:31:18.511508 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-worker-8559568497-vwxwh" event={"ID":"93ad1063-a321-44a3-bd91-c00c7ba1ef96","Type":"ContainerStarted","Data":"2ea01f1471d28b2fde8295c78e07b79f0d374adfba3365977ecf2c5ba1280f60"} Dec 03 14:31:18 crc kubenswrapper[4805]: I1203 14:31:18.511908 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-worker-8559568497-vwxwh" event={"ID":"93ad1063-a321-44a3-bd91-c00c7ba1ef96","Type":"ContainerStarted","Data":"10bbe262c054c273168621d713196f8eef589dc8208f62d01fefc0fbd04d7623"} Dec 03 14:31:18 crc kubenswrapper[4805]: I1203 14:31:18.516468 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-58794647f4-rtjlq" event={"ID":"5a55a37b-c678-481e-bf36-a8961f5b26ed","Type":"ContainerStarted","Data":"130390e39a794742c18e93ca620ab12c666d164ad65581ce01399677cd7712a6"} Dec 03 14:31:18 crc kubenswrapper[4805]: I1203 14:31:18.516934 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-58794647f4-rtjlq" event={"ID":"5a55a37b-c678-481e-bf36-a8961f5b26ed","Type":"ContainerStarted","Data":"67daca659d85a1b0f2324ff297f7092b7fabc98f14b3226d92fbff0867064cf9"} Dec 03 14:31:18 crc kubenswrapper[4805]: I1203 14:31:18.516960 4805 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/barbican-api-58794647f4-rtjlq" Dec 03 14:31:18 crc kubenswrapper[4805]: I1203 14:31:18.519310 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-keystone-listener-8484c694-6frjp" event={"ID":"18349f43-5c89-4c11-8ff2-fee30d6cceb7","Type":"ContainerStarted","Data":"7fa58f6b323accfa8e7c157acb9b68e30711dc2de6d3109eb4986c5e0ff86c10"} Dec 03 14:31:18 crc kubenswrapper[4805]: I1203 14:31:18.519366 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-keystone-listener-8484c694-6frjp" event={"ID":"18349f43-5c89-4c11-8ff2-fee30d6cceb7","Type":"ContainerStarted","Data":"030be3dcb7271a031b714db68aa59fd3cd24d4837273a0c05c5e6ed425957ffc"} Dec 03 14:31:18 crc kubenswrapper[4805]: I1203 14:31:18.522108 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"58c98434-0b4d-4c2d-ac35-109e38f96803","Type":"ContainerStarted","Data":"20bba63c0e20fac3641cd24a4df07f94b5cb20bb86b1e7fd6adcb5e5071e51b0"} Dec 03 14:31:18 crc kubenswrapper[4805]: I1203 14:31:18.533257 4805 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-worker-8559568497-vwxwh" podStartSLOduration=2.737514896 podStartE2EDuration="6.533236536s" podCreationTimestamp="2025-12-03 14:31:12 +0000 UTC" firstStartedPulling="2025-12-03 14:31:13.48799757 +0000 UTC m=+1303.150914493" lastFinishedPulling="2025-12-03 14:31:17.28371921 +0000 UTC m=+1306.946636133" observedRunningTime="2025-12-03 14:31:18.528445285 +0000 UTC m=+1308.191362218" watchObservedRunningTime="2025-12-03 14:31:18.533236536 +0000 UTC m=+1308.196153459" Dec 03 14:31:18 crc kubenswrapper[4805]: I1203 14:31:18.555067 4805 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-keystone-listener-8484c694-6frjp" podStartSLOduration=2.949450498 podStartE2EDuration="6.555044823s" podCreationTimestamp="2025-12-03 14:31:12 +0000 UTC" firstStartedPulling="2025-12-03 14:31:13.699709296 +0000 UTC m=+1303.362626219" lastFinishedPulling="2025-12-03 14:31:17.305303621 +0000 UTC m=+1306.968220544" observedRunningTime="2025-12-03 14:31:18.548549785 +0000 UTC m=+1308.211466708" watchObservedRunningTime="2025-12-03 14:31:18.555044823 +0000 UTC m=+1308.217961746" Dec 03 14:31:18 crc kubenswrapper[4805]: I1203 14:31:18.576100 4805 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-api-58794647f4-rtjlq" podStartSLOduration=3.576082089 podStartE2EDuration="3.576082089s" podCreationTimestamp="2025-12-03 14:31:15 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 14:31:18.564874712 +0000 UTC m=+1308.227791635" watchObservedRunningTime="2025-12-03 14:31:18.576082089 +0000 UTC m=+1308.238999012" Dec 03 14:31:18 crc kubenswrapper[4805]: I1203 14:31:18.725826 4805 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/neutron-75f57c664d-4tsg2" Dec 03 14:31:19 crc kubenswrapper[4805]: I1203 14:31:19.530291 4805 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/barbican-api-58794647f4-rtjlq" Dec 03 14:31:20 crc kubenswrapper[4805]: I1203 14:31:20.050717 4805 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/placement-98cb4d7c6-cbjgt" Dec 03 14:31:20 crc kubenswrapper[4805]: I1203 14:31:20.151700 4805 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/placement-98cb4d7c6-cbjgt" Dec 03 14:31:20 crc kubenswrapper[4805]: I1203 14:31:20.557004 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"58c98434-0b4d-4c2d-ac35-109e38f96803","Type":"ContainerStarted","Data":"126ffcd2cdaa24e89e915efb3598e7f49248d24b6884776c5b6848a1429ab5c0"} Dec 03 14:31:20 crc kubenswrapper[4805]: I1203 14:31:20.557681 4805 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Dec 03 14:31:20 crc kubenswrapper[4805]: I1203 14:31:20.587124 4805 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=1.907288756 podStartE2EDuration="10.587101052s" podCreationTimestamp="2025-12-03 14:31:10 +0000 UTC" firstStartedPulling="2025-12-03 14:31:11.558106147 +0000 UTC m=+1301.221023070" lastFinishedPulling="2025-12-03 14:31:20.237918423 +0000 UTC m=+1309.900835366" observedRunningTime="2025-12-03 14:31:20.576343158 +0000 UTC m=+1310.239260101" watchObservedRunningTime="2025-12-03 14:31:20.587101052 +0000 UTC m=+1310.250017975" Dec 03 14:31:20 crc kubenswrapper[4805]: I1203 14:31:20.898088 4805 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/keystone-585fb5cffc-n26gq" Dec 03 14:31:21 crc kubenswrapper[4805]: I1203 14:31:21.073587 4805 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/neutron-9764b468f-7sr7n" Dec 03 14:31:21 crc kubenswrapper[4805]: I1203 14:31:21.250141 4805 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-75f57c664d-4tsg2"] Dec 03 14:31:21 crc kubenswrapper[4805]: I1203 14:31:21.250372 4805 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/neutron-75f57c664d-4tsg2" podUID="318792ec-dc8b-4a37-92ab-d31773d66350" containerName="neutron-api" containerID="cri-o://48672ddefb5f29199ffd9090e43f5c584cf661e8cf1e4313d963fb96b28729b6" gracePeriod=30 Dec 03 14:31:21 crc kubenswrapper[4805]: I1203 14:31:21.250533 4805 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/neutron-75f57c664d-4tsg2" podUID="318792ec-dc8b-4a37-92ab-d31773d66350" containerName="neutron-httpd" containerID="cri-o://76ff7768f1c3210b2f1ffab05aefd717027127581dc36328eff11ac6503fdb3f" gracePeriod=30 Dec 03 14:31:21 crc kubenswrapper[4805]: I1203 14:31:21.577537 4805 generic.go:334] "Generic (PLEG): container finished" podID="260717ec-f898-4ccd-8a4f-c7d6b14ade6e" containerID="3ae18d4a53cc1df81f75833ab27219aaab27ca80a4b5ccbb30481f900bcecd6f" exitCode=0 Dec 03 14:31:21 crc kubenswrapper[4805]: I1203 14:31:21.577604 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-sync-fh9c2" event={"ID":"260717ec-f898-4ccd-8a4f-c7d6b14ade6e","Type":"ContainerDied","Data":"3ae18d4a53cc1df81f75833ab27219aaab27ca80a4b5ccbb30481f900bcecd6f"} Dec 03 14:31:21 crc kubenswrapper[4805]: I1203 14:31:21.841375 4805 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/openstackclient"] Dec 03 14:31:21 crc kubenswrapper[4805]: I1203 14:31:21.842544 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Dec 03 14:31:21 crc kubenswrapper[4805]: I1203 14:31:21.849243 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-config-secret" Dec 03 14:31:21 crc kubenswrapper[4805]: I1203 14:31:21.849292 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstackclient-openstackclient-dockercfg-tkrm8" Dec 03 14:31:21 crc kubenswrapper[4805]: I1203 14:31:21.851933 4805 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-config" Dec 03 14:31:21 crc kubenswrapper[4805]: I1203 14:31:21.853685 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstackclient"] Dec 03 14:31:21 crc kubenswrapper[4805]: I1203 14:31:21.914788 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/29c53624-ef4b-43ab-8922-181a78908d24-openstack-config-secret\") pod \"openstackclient\" (UID: \"29c53624-ef4b-43ab-8922-181a78908d24\") " pod="openstack/openstackclient" Dec 03 14:31:21 crc kubenswrapper[4805]: I1203 14:31:21.914959 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/29c53624-ef4b-43ab-8922-181a78908d24-combined-ca-bundle\") pod \"openstackclient\" (UID: \"29c53624-ef4b-43ab-8922-181a78908d24\") " pod="openstack/openstackclient" Dec 03 14:31:21 crc kubenswrapper[4805]: I1203 14:31:21.914992 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/29c53624-ef4b-43ab-8922-181a78908d24-openstack-config\") pod \"openstackclient\" (UID: \"29c53624-ef4b-43ab-8922-181a78908d24\") " pod="openstack/openstackclient" Dec 03 14:31:21 crc kubenswrapper[4805]: I1203 14:31:21.915027 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9pll7\" (UniqueName: \"kubernetes.io/projected/29c53624-ef4b-43ab-8922-181a78908d24-kube-api-access-9pll7\") pod \"openstackclient\" (UID: \"29c53624-ef4b-43ab-8922-181a78908d24\") " pod="openstack/openstackclient" Dec 03 14:31:22 crc kubenswrapper[4805]: I1203 14:31:22.017321 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/29c53624-ef4b-43ab-8922-181a78908d24-openstack-config-secret\") pod \"openstackclient\" (UID: \"29c53624-ef4b-43ab-8922-181a78908d24\") " pod="openstack/openstackclient" Dec 03 14:31:22 crc kubenswrapper[4805]: I1203 14:31:22.017422 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/29c53624-ef4b-43ab-8922-181a78908d24-combined-ca-bundle\") pod \"openstackclient\" (UID: \"29c53624-ef4b-43ab-8922-181a78908d24\") " pod="openstack/openstackclient" Dec 03 14:31:22 crc kubenswrapper[4805]: I1203 14:31:22.017450 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/29c53624-ef4b-43ab-8922-181a78908d24-openstack-config\") pod \"openstackclient\" (UID: \"29c53624-ef4b-43ab-8922-181a78908d24\") " pod="openstack/openstackclient" Dec 03 14:31:22 crc kubenswrapper[4805]: I1203 14:31:22.017486 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9pll7\" (UniqueName: \"kubernetes.io/projected/29c53624-ef4b-43ab-8922-181a78908d24-kube-api-access-9pll7\") pod \"openstackclient\" (UID: \"29c53624-ef4b-43ab-8922-181a78908d24\") " pod="openstack/openstackclient" Dec 03 14:31:22 crc kubenswrapper[4805]: I1203 14:31:22.019600 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/29c53624-ef4b-43ab-8922-181a78908d24-openstack-config\") pod \"openstackclient\" (UID: \"29c53624-ef4b-43ab-8922-181a78908d24\") " pod="openstack/openstackclient" Dec 03 14:31:22 crc kubenswrapper[4805]: I1203 14:31:22.024310 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/29c53624-ef4b-43ab-8922-181a78908d24-openstack-config-secret\") pod \"openstackclient\" (UID: \"29c53624-ef4b-43ab-8922-181a78908d24\") " pod="openstack/openstackclient" Dec 03 14:31:22 crc kubenswrapper[4805]: I1203 14:31:22.024272 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/29c53624-ef4b-43ab-8922-181a78908d24-combined-ca-bundle\") pod \"openstackclient\" (UID: \"29c53624-ef4b-43ab-8922-181a78908d24\") " pod="openstack/openstackclient" Dec 03 14:31:22 crc kubenswrapper[4805]: I1203 14:31:22.037313 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9pll7\" (UniqueName: \"kubernetes.io/projected/29c53624-ef4b-43ab-8922-181a78908d24-kube-api-access-9pll7\") pod \"openstackclient\" (UID: \"29c53624-ef4b-43ab-8922-181a78908d24\") " pod="openstack/openstackclient" Dec 03 14:31:22 crc kubenswrapper[4805]: I1203 14:31:22.162968 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Dec 03 14:31:22 crc kubenswrapper[4805]: I1203 14:31:22.588016 4805 generic.go:334] "Generic (PLEG): container finished" podID="318792ec-dc8b-4a37-92ab-d31773d66350" containerID="76ff7768f1c3210b2f1ffab05aefd717027127581dc36328eff11ac6503fdb3f" exitCode=0 Dec 03 14:31:22 crc kubenswrapper[4805]: I1203 14:31:22.588226 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-75f57c664d-4tsg2" event={"ID":"318792ec-dc8b-4a37-92ab-d31773d66350","Type":"ContainerDied","Data":"76ff7768f1c3210b2f1ffab05aefd717027127581dc36328eff11ac6503fdb3f"} Dec 03 14:31:22 crc kubenswrapper[4805]: I1203 14:31:22.670397 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstackclient"] Dec 03 14:31:23 crc kubenswrapper[4805]: I1203 14:31:23.026288 4805 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-sync-fh9c2" Dec 03 14:31:23 crc kubenswrapper[4805]: I1203 14:31:23.065051 4805 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-75c8ddd69c-s5r44" Dec 03 14:31:23 crc kubenswrapper[4805]: I1203 14:31:23.141158 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/260717ec-f898-4ccd-8a4f-c7d6b14ade6e-db-sync-config-data\") pod \"260717ec-f898-4ccd-8a4f-c7d6b14ade6e\" (UID: \"260717ec-f898-4ccd-8a4f-c7d6b14ade6e\") " Dec 03 14:31:23 crc kubenswrapper[4805]: I1203 14:31:23.141220 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/260717ec-f898-4ccd-8a4f-c7d6b14ade6e-config-data\") pod \"260717ec-f898-4ccd-8a4f-c7d6b14ade6e\" (UID: \"260717ec-f898-4ccd-8a4f-c7d6b14ade6e\") " Dec 03 14:31:23 crc kubenswrapper[4805]: I1203 14:31:23.141260 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/260717ec-f898-4ccd-8a4f-c7d6b14ade6e-combined-ca-bundle\") pod \"260717ec-f898-4ccd-8a4f-c7d6b14ade6e\" (UID: \"260717ec-f898-4ccd-8a4f-c7d6b14ade6e\") " Dec 03 14:31:23 crc kubenswrapper[4805]: I1203 14:31:23.141303 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/260717ec-f898-4ccd-8a4f-c7d6b14ade6e-scripts\") pod \"260717ec-f898-4ccd-8a4f-c7d6b14ade6e\" (UID: \"260717ec-f898-4ccd-8a4f-c7d6b14ade6e\") " Dec 03 14:31:23 crc kubenswrapper[4805]: I1203 14:31:23.141358 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/260717ec-f898-4ccd-8a4f-c7d6b14ade6e-etc-machine-id\") pod \"260717ec-f898-4ccd-8a4f-c7d6b14ade6e\" (UID: \"260717ec-f898-4ccd-8a4f-c7d6b14ade6e\") " Dec 03 14:31:23 crc kubenswrapper[4805]: I1203 14:31:23.141385 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xxm8t\" (UniqueName: \"kubernetes.io/projected/260717ec-f898-4ccd-8a4f-c7d6b14ade6e-kube-api-access-xxm8t\") pod \"260717ec-f898-4ccd-8a4f-c7d6b14ade6e\" (UID: \"260717ec-f898-4ccd-8a4f-c7d6b14ade6e\") " Dec 03 14:31:23 crc kubenswrapper[4805]: I1203 14:31:23.143557 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/260717ec-f898-4ccd-8a4f-c7d6b14ade6e-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "260717ec-f898-4ccd-8a4f-c7d6b14ade6e" (UID: "260717ec-f898-4ccd-8a4f-c7d6b14ade6e"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 03 14:31:23 crc kubenswrapper[4805]: I1203 14:31:23.149696 4805 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-84b966f6c9-zwsmd"] Dec 03 14:31:23 crc kubenswrapper[4805]: I1203 14:31:23.149991 4805 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-84b966f6c9-zwsmd" podUID="b3ca6772-26a0-410c-9609-1f49d08b1864" containerName="dnsmasq-dns" containerID="cri-o://31628e51d65354f72d4dd433558b5b70b666da48c488d34fa61d3e72e666c631" gracePeriod=10 Dec 03 14:31:23 crc kubenswrapper[4805]: I1203 14:31:23.155813 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/260717ec-f898-4ccd-8a4f-c7d6b14ade6e-db-sync-config-data" (OuterVolumeSpecName: "db-sync-config-data") pod "260717ec-f898-4ccd-8a4f-c7d6b14ade6e" (UID: "260717ec-f898-4ccd-8a4f-c7d6b14ade6e"). InnerVolumeSpecName "db-sync-config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 14:31:23 crc kubenswrapper[4805]: I1203 14:31:23.161055 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/260717ec-f898-4ccd-8a4f-c7d6b14ade6e-scripts" (OuterVolumeSpecName: "scripts") pod "260717ec-f898-4ccd-8a4f-c7d6b14ade6e" (UID: "260717ec-f898-4ccd-8a4f-c7d6b14ade6e"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 14:31:23 crc kubenswrapper[4805]: I1203 14:31:23.177213 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/260717ec-f898-4ccd-8a4f-c7d6b14ade6e-kube-api-access-xxm8t" (OuterVolumeSpecName: "kube-api-access-xxm8t") pod "260717ec-f898-4ccd-8a4f-c7d6b14ade6e" (UID: "260717ec-f898-4ccd-8a4f-c7d6b14ade6e"). InnerVolumeSpecName "kube-api-access-xxm8t". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 14:31:23 crc kubenswrapper[4805]: I1203 14:31:23.203751 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/260717ec-f898-4ccd-8a4f-c7d6b14ade6e-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "260717ec-f898-4ccd-8a4f-c7d6b14ade6e" (UID: "260717ec-f898-4ccd-8a4f-c7d6b14ade6e"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 14:31:23 crc kubenswrapper[4805]: I1203 14:31:23.243573 4805 reconciler_common.go:293] "Volume detached for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/260717ec-f898-4ccd-8a4f-c7d6b14ade6e-db-sync-config-data\") on node \"crc\" DevicePath \"\"" Dec 03 14:31:23 crc kubenswrapper[4805]: I1203 14:31:23.243616 4805 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/260717ec-f898-4ccd-8a4f-c7d6b14ade6e-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 14:31:23 crc kubenswrapper[4805]: I1203 14:31:23.243629 4805 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/260717ec-f898-4ccd-8a4f-c7d6b14ade6e-scripts\") on node \"crc\" DevicePath \"\"" Dec 03 14:31:23 crc kubenswrapper[4805]: I1203 14:31:23.243640 4805 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/260717ec-f898-4ccd-8a4f-c7d6b14ade6e-etc-machine-id\") on node \"crc\" DevicePath \"\"" Dec 03 14:31:23 crc kubenswrapper[4805]: I1203 14:31:23.243652 4805 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xxm8t\" (UniqueName: \"kubernetes.io/projected/260717ec-f898-4ccd-8a4f-c7d6b14ade6e-kube-api-access-xxm8t\") on node \"crc\" DevicePath \"\"" Dec 03 14:31:23 crc kubenswrapper[4805]: I1203 14:31:23.274113 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/260717ec-f898-4ccd-8a4f-c7d6b14ade6e-config-data" (OuterVolumeSpecName: "config-data") pod "260717ec-f898-4ccd-8a4f-c7d6b14ade6e" (UID: "260717ec-f898-4ccd-8a4f-c7d6b14ade6e"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 14:31:23 crc kubenswrapper[4805]: I1203 14:31:23.345863 4805 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/260717ec-f898-4ccd-8a4f-c7d6b14ade6e-config-data\") on node \"crc\" DevicePath \"\"" Dec 03 14:31:23 crc kubenswrapper[4805]: I1203 14:31:23.597869 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstackclient" event={"ID":"29c53624-ef4b-43ab-8922-181a78908d24","Type":"ContainerStarted","Data":"c14fd86145d691ba729cdb9d5192c54b2b0614f2ce374e5fcf17fab5e9f82db3"} Dec 03 14:31:23 crc kubenswrapper[4805]: I1203 14:31:23.599410 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-sync-fh9c2" event={"ID":"260717ec-f898-4ccd-8a4f-c7d6b14ade6e","Type":"ContainerDied","Data":"d577be3bf13b8fce168f412977177621057fb7ad146b5d216cd41ffc22be9462"} Dec 03 14:31:23 crc kubenswrapper[4805]: I1203 14:31:23.599436 4805 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="d577be3bf13b8fce168f412977177621057fb7ad146b5d216cd41ffc22be9462" Dec 03 14:31:23 crc kubenswrapper[4805]: I1203 14:31:23.599489 4805 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-sync-fh9c2" Dec 03 14:31:23 crc kubenswrapper[4805]: I1203 14:31:23.691418 4805 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-84b966f6c9-zwsmd" podUID="b3ca6772-26a0-410c-9609-1f49d08b1864" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.152:5353: connect: connection refused" Dec 03 14:31:23 crc kubenswrapper[4805]: I1203 14:31:23.896001 4805 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-scheduler-0"] Dec 03 14:31:23 crc kubenswrapper[4805]: E1203 14:31:23.896931 4805 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="260717ec-f898-4ccd-8a4f-c7d6b14ade6e" containerName="cinder-db-sync" Dec 03 14:31:23 crc kubenswrapper[4805]: I1203 14:31:23.896954 4805 state_mem.go:107] "Deleted CPUSet assignment" podUID="260717ec-f898-4ccd-8a4f-c7d6b14ade6e" containerName="cinder-db-sync" Dec 03 14:31:23 crc kubenswrapper[4805]: I1203 14:31:23.897326 4805 memory_manager.go:354] "RemoveStaleState removing state" podUID="260717ec-f898-4ccd-8a4f-c7d6b14ade6e" containerName="cinder-db-sync" Dec 03 14:31:23 crc kubenswrapper[4805]: I1203 14:31:23.899015 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Dec 03 14:31:23 crc kubenswrapper[4805]: I1203 14:31:23.902568 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-cinder-dockercfg-r5964" Dec 03 14:31:23 crc kubenswrapper[4805]: I1203 14:31:23.902800 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-scripts" Dec 03 14:31:23 crc kubenswrapper[4805]: I1203 14:31:23.902938 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-config-data" Dec 03 14:31:23 crc kubenswrapper[4805]: I1203 14:31:23.903064 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-scheduler-config-data" Dec 03 14:31:23 crc kubenswrapper[4805]: I1203 14:31:23.953818 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-scheduler-0"] Dec 03 14:31:23 crc kubenswrapper[4805]: I1203 14:31:23.964884 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/29710b09-a1ad-468f-af26-b5d5eb516898-scripts\") pod \"cinder-scheduler-0\" (UID: \"29710b09-a1ad-468f-af26-b5d5eb516898\") " pod="openstack/cinder-scheduler-0" Dec 03 14:31:23 crc kubenswrapper[4805]: I1203 14:31:23.965122 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/29710b09-a1ad-468f-af26-b5d5eb516898-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"29710b09-a1ad-468f-af26-b5d5eb516898\") " pod="openstack/cinder-scheduler-0" Dec 03 14:31:23 crc kubenswrapper[4805]: I1203 14:31:23.965208 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/29710b09-a1ad-468f-af26-b5d5eb516898-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"29710b09-a1ad-468f-af26-b5d5eb516898\") " pod="openstack/cinder-scheduler-0" Dec 03 14:31:23 crc kubenswrapper[4805]: I1203 14:31:23.965243 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/29710b09-a1ad-468f-af26-b5d5eb516898-config-data\") pod \"cinder-scheduler-0\" (UID: \"29710b09-a1ad-468f-af26-b5d5eb516898\") " pod="openstack/cinder-scheduler-0" Dec 03 14:31:23 crc kubenswrapper[4805]: I1203 14:31:23.965361 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qfmb5\" (UniqueName: \"kubernetes.io/projected/29710b09-a1ad-468f-af26-b5d5eb516898-kube-api-access-qfmb5\") pod \"cinder-scheduler-0\" (UID: \"29710b09-a1ad-468f-af26-b5d5eb516898\") " pod="openstack/cinder-scheduler-0" Dec 03 14:31:23 crc kubenswrapper[4805]: I1203 14:31:23.965454 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/29710b09-a1ad-468f-af26-b5d5eb516898-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"29710b09-a1ad-468f-af26-b5d5eb516898\") " pod="openstack/cinder-scheduler-0" Dec 03 14:31:23 crc kubenswrapper[4805]: I1203 14:31:23.988993 4805 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-5784cf869f-l4bt9"] Dec 03 14:31:23 crc kubenswrapper[4805]: I1203 14:31:23.993735 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5784cf869f-l4bt9" Dec 03 14:31:23 crc kubenswrapper[4805]: I1203 14:31:23.994717 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5784cf869f-l4bt9"] Dec 03 14:31:24 crc kubenswrapper[4805]: I1203 14:31:24.078066 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/bc7fd7a7-02bc-48e1-9bde-d1680876da37-config\") pod \"dnsmasq-dns-5784cf869f-l4bt9\" (UID: \"bc7fd7a7-02bc-48e1-9bde-d1680876da37\") " pod="openstack/dnsmasq-dns-5784cf869f-l4bt9" Dec 03 14:31:24 crc kubenswrapper[4805]: I1203 14:31:24.078136 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/bc7fd7a7-02bc-48e1-9bde-d1680876da37-dns-svc\") pod \"dnsmasq-dns-5784cf869f-l4bt9\" (UID: \"bc7fd7a7-02bc-48e1-9bde-d1680876da37\") " pod="openstack/dnsmasq-dns-5784cf869f-l4bt9" Dec 03 14:31:24 crc kubenswrapper[4805]: I1203 14:31:24.078204 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/29710b09-a1ad-468f-af26-b5d5eb516898-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"29710b09-a1ad-468f-af26-b5d5eb516898\") " pod="openstack/cinder-scheduler-0" Dec 03 14:31:24 crc kubenswrapper[4805]: I1203 14:31:24.078251 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/29710b09-a1ad-468f-af26-b5d5eb516898-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"29710b09-a1ad-468f-af26-b5d5eb516898\") " pod="openstack/cinder-scheduler-0" Dec 03 14:31:24 crc kubenswrapper[4805]: I1203 14:31:24.078279 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/29710b09-a1ad-468f-af26-b5d5eb516898-config-data\") pod \"cinder-scheduler-0\" (UID: \"29710b09-a1ad-468f-af26-b5d5eb516898\") " pod="openstack/cinder-scheduler-0" Dec 03 14:31:24 crc kubenswrapper[4805]: I1203 14:31:24.078325 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dz2wl\" (UniqueName: \"kubernetes.io/projected/bc7fd7a7-02bc-48e1-9bde-d1680876da37-kube-api-access-dz2wl\") pod \"dnsmasq-dns-5784cf869f-l4bt9\" (UID: \"bc7fd7a7-02bc-48e1-9bde-d1680876da37\") " pod="openstack/dnsmasq-dns-5784cf869f-l4bt9" Dec 03 14:31:24 crc kubenswrapper[4805]: I1203 14:31:24.078374 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qfmb5\" (UniqueName: \"kubernetes.io/projected/29710b09-a1ad-468f-af26-b5d5eb516898-kube-api-access-qfmb5\") pod \"cinder-scheduler-0\" (UID: \"29710b09-a1ad-468f-af26-b5d5eb516898\") " pod="openstack/cinder-scheduler-0" Dec 03 14:31:24 crc kubenswrapper[4805]: I1203 14:31:24.078421 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/29710b09-a1ad-468f-af26-b5d5eb516898-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"29710b09-a1ad-468f-af26-b5d5eb516898\") " pod="openstack/cinder-scheduler-0" Dec 03 14:31:24 crc kubenswrapper[4805]: I1203 14:31:24.078449 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/bc7fd7a7-02bc-48e1-9bde-d1680876da37-ovsdbserver-sb\") pod \"dnsmasq-dns-5784cf869f-l4bt9\" (UID: \"bc7fd7a7-02bc-48e1-9bde-d1680876da37\") " pod="openstack/dnsmasq-dns-5784cf869f-l4bt9" Dec 03 14:31:24 crc kubenswrapper[4805]: I1203 14:31:24.078480 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/29710b09-a1ad-468f-af26-b5d5eb516898-scripts\") pod \"cinder-scheduler-0\" (UID: \"29710b09-a1ad-468f-af26-b5d5eb516898\") " pod="openstack/cinder-scheduler-0" Dec 03 14:31:24 crc kubenswrapper[4805]: I1203 14:31:24.078526 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/bc7fd7a7-02bc-48e1-9bde-d1680876da37-dns-swift-storage-0\") pod \"dnsmasq-dns-5784cf869f-l4bt9\" (UID: \"bc7fd7a7-02bc-48e1-9bde-d1680876da37\") " pod="openstack/dnsmasq-dns-5784cf869f-l4bt9" Dec 03 14:31:24 crc kubenswrapper[4805]: I1203 14:31:24.078570 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/bc7fd7a7-02bc-48e1-9bde-d1680876da37-ovsdbserver-nb\") pod \"dnsmasq-dns-5784cf869f-l4bt9\" (UID: \"bc7fd7a7-02bc-48e1-9bde-d1680876da37\") " pod="openstack/dnsmasq-dns-5784cf869f-l4bt9" Dec 03 14:31:24 crc kubenswrapper[4805]: I1203 14:31:24.078688 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/29710b09-a1ad-468f-af26-b5d5eb516898-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"29710b09-a1ad-468f-af26-b5d5eb516898\") " pod="openstack/cinder-scheduler-0" Dec 03 14:31:24 crc kubenswrapper[4805]: I1203 14:31:24.115631 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/29710b09-a1ad-468f-af26-b5d5eb516898-config-data\") pod \"cinder-scheduler-0\" (UID: \"29710b09-a1ad-468f-af26-b5d5eb516898\") " pod="openstack/cinder-scheduler-0" Dec 03 14:31:24 crc kubenswrapper[4805]: I1203 14:31:24.119357 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qfmb5\" (UniqueName: \"kubernetes.io/projected/29710b09-a1ad-468f-af26-b5d5eb516898-kube-api-access-qfmb5\") pod \"cinder-scheduler-0\" (UID: \"29710b09-a1ad-468f-af26-b5d5eb516898\") " pod="openstack/cinder-scheduler-0" Dec 03 14:31:24 crc kubenswrapper[4805]: I1203 14:31:24.119500 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/29710b09-a1ad-468f-af26-b5d5eb516898-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"29710b09-a1ad-468f-af26-b5d5eb516898\") " pod="openstack/cinder-scheduler-0" Dec 03 14:31:24 crc kubenswrapper[4805]: I1203 14:31:24.124481 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/29710b09-a1ad-468f-af26-b5d5eb516898-scripts\") pod \"cinder-scheduler-0\" (UID: \"29710b09-a1ad-468f-af26-b5d5eb516898\") " pod="openstack/cinder-scheduler-0" Dec 03 14:31:24 crc kubenswrapper[4805]: I1203 14:31:24.124820 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/29710b09-a1ad-468f-af26-b5d5eb516898-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"29710b09-a1ad-468f-af26-b5d5eb516898\") " pod="openstack/cinder-scheduler-0" Dec 03 14:31:24 crc kubenswrapper[4805]: I1203 14:31:24.182734 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/bc7fd7a7-02bc-48e1-9bde-d1680876da37-ovsdbserver-sb\") pod \"dnsmasq-dns-5784cf869f-l4bt9\" (UID: \"bc7fd7a7-02bc-48e1-9bde-d1680876da37\") " pod="openstack/dnsmasq-dns-5784cf869f-l4bt9" Dec 03 14:31:24 crc kubenswrapper[4805]: I1203 14:31:24.182904 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/bc7fd7a7-02bc-48e1-9bde-d1680876da37-dns-swift-storage-0\") pod \"dnsmasq-dns-5784cf869f-l4bt9\" (UID: \"bc7fd7a7-02bc-48e1-9bde-d1680876da37\") " pod="openstack/dnsmasq-dns-5784cf869f-l4bt9" Dec 03 14:31:24 crc kubenswrapper[4805]: I1203 14:31:24.182968 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/bc7fd7a7-02bc-48e1-9bde-d1680876da37-ovsdbserver-nb\") pod \"dnsmasq-dns-5784cf869f-l4bt9\" (UID: \"bc7fd7a7-02bc-48e1-9bde-d1680876da37\") " pod="openstack/dnsmasq-dns-5784cf869f-l4bt9" Dec 03 14:31:24 crc kubenswrapper[4805]: I1203 14:31:24.183009 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/bc7fd7a7-02bc-48e1-9bde-d1680876da37-config\") pod \"dnsmasq-dns-5784cf869f-l4bt9\" (UID: \"bc7fd7a7-02bc-48e1-9bde-d1680876da37\") " pod="openstack/dnsmasq-dns-5784cf869f-l4bt9" Dec 03 14:31:24 crc kubenswrapper[4805]: I1203 14:31:24.183064 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/bc7fd7a7-02bc-48e1-9bde-d1680876da37-dns-svc\") pod \"dnsmasq-dns-5784cf869f-l4bt9\" (UID: \"bc7fd7a7-02bc-48e1-9bde-d1680876da37\") " pod="openstack/dnsmasq-dns-5784cf869f-l4bt9" Dec 03 14:31:24 crc kubenswrapper[4805]: I1203 14:31:24.183269 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dz2wl\" (UniqueName: \"kubernetes.io/projected/bc7fd7a7-02bc-48e1-9bde-d1680876da37-kube-api-access-dz2wl\") pod \"dnsmasq-dns-5784cf869f-l4bt9\" (UID: \"bc7fd7a7-02bc-48e1-9bde-d1680876da37\") " pod="openstack/dnsmasq-dns-5784cf869f-l4bt9" Dec 03 14:31:24 crc kubenswrapper[4805]: I1203 14:31:24.184055 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/bc7fd7a7-02bc-48e1-9bde-d1680876da37-ovsdbserver-sb\") pod \"dnsmasq-dns-5784cf869f-l4bt9\" (UID: \"bc7fd7a7-02bc-48e1-9bde-d1680876da37\") " pod="openstack/dnsmasq-dns-5784cf869f-l4bt9" Dec 03 14:31:24 crc kubenswrapper[4805]: I1203 14:31:24.184513 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/bc7fd7a7-02bc-48e1-9bde-d1680876da37-dns-swift-storage-0\") pod \"dnsmasq-dns-5784cf869f-l4bt9\" (UID: \"bc7fd7a7-02bc-48e1-9bde-d1680876da37\") " pod="openstack/dnsmasq-dns-5784cf869f-l4bt9" Dec 03 14:31:24 crc kubenswrapper[4805]: I1203 14:31:24.184609 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/bc7fd7a7-02bc-48e1-9bde-d1680876da37-config\") pod \"dnsmasq-dns-5784cf869f-l4bt9\" (UID: \"bc7fd7a7-02bc-48e1-9bde-d1680876da37\") " pod="openstack/dnsmasq-dns-5784cf869f-l4bt9" Dec 03 14:31:24 crc kubenswrapper[4805]: I1203 14:31:24.185206 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/bc7fd7a7-02bc-48e1-9bde-d1680876da37-ovsdbserver-nb\") pod \"dnsmasq-dns-5784cf869f-l4bt9\" (UID: \"bc7fd7a7-02bc-48e1-9bde-d1680876da37\") " pod="openstack/dnsmasq-dns-5784cf869f-l4bt9" Dec 03 14:31:24 crc kubenswrapper[4805]: I1203 14:31:24.185258 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/bc7fd7a7-02bc-48e1-9bde-d1680876da37-dns-svc\") pod \"dnsmasq-dns-5784cf869f-l4bt9\" (UID: \"bc7fd7a7-02bc-48e1-9bde-d1680876da37\") " pod="openstack/dnsmasq-dns-5784cf869f-l4bt9" Dec 03 14:31:24 crc kubenswrapper[4805]: I1203 14:31:24.226660 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dz2wl\" (UniqueName: \"kubernetes.io/projected/bc7fd7a7-02bc-48e1-9bde-d1680876da37-kube-api-access-dz2wl\") pod \"dnsmasq-dns-5784cf869f-l4bt9\" (UID: \"bc7fd7a7-02bc-48e1-9bde-d1680876da37\") " pod="openstack/dnsmasq-dns-5784cf869f-l4bt9" Dec 03 14:31:24 crc kubenswrapper[4805]: I1203 14:31:24.237486 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Dec 03 14:31:24 crc kubenswrapper[4805]: I1203 14:31:24.306028 4805 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-api-0"] Dec 03 14:31:24 crc kubenswrapper[4805]: I1203 14:31:24.307493 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Dec 03 14:31:24 crc kubenswrapper[4805]: I1203 14:31:24.317211 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-api-config-data" Dec 03 14:31:24 crc kubenswrapper[4805]: I1203 14:31:24.344351 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5784cf869f-l4bt9" Dec 03 14:31:24 crc kubenswrapper[4805]: I1203 14:31:24.415216 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-api-0"] Dec 03 14:31:24 crc kubenswrapper[4805]: I1203 14:31:24.415770 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/afc5cad3-750c-429e-be48-6355bc07ac6a-logs\") pod \"cinder-api-0\" (UID: \"afc5cad3-750c-429e-be48-6355bc07ac6a\") " pod="openstack/cinder-api-0" Dec 03 14:31:24 crc kubenswrapper[4805]: I1203 14:31:24.436055 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/afc5cad3-750c-429e-be48-6355bc07ac6a-etc-machine-id\") pod \"cinder-api-0\" (UID: \"afc5cad3-750c-429e-be48-6355bc07ac6a\") " pod="openstack/cinder-api-0" Dec 03 14:31:24 crc kubenswrapper[4805]: I1203 14:31:24.436093 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9z95h\" (UniqueName: \"kubernetes.io/projected/afc5cad3-750c-429e-be48-6355bc07ac6a-kube-api-access-9z95h\") pod \"cinder-api-0\" (UID: \"afc5cad3-750c-429e-be48-6355bc07ac6a\") " pod="openstack/cinder-api-0" Dec 03 14:31:24 crc kubenswrapper[4805]: I1203 14:31:24.436258 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/afc5cad3-750c-429e-be48-6355bc07ac6a-scripts\") pod \"cinder-api-0\" (UID: \"afc5cad3-750c-429e-be48-6355bc07ac6a\") " pod="openstack/cinder-api-0" Dec 03 14:31:24 crc kubenswrapper[4805]: I1203 14:31:24.436289 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/afc5cad3-750c-429e-be48-6355bc07ac6a-config-data-custom\") pod \"cinder-api-0\" (UID: \"afc5cad3-750c-429e-be48-6355bc07ac6a\") " pod="openstack/cinder-api-0" Dec 03 14:31:24 crc kubenswrapper[4805]: I1203 14:31:24.436307 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/afc5cad3-750c-429e-be48-6355bc07ac6a-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"afc5cad3-750c-429e-be48-6355bc07ac6a\") " pod="openstack/cinder-api-0" Dec 03 14:31:24 crc kubenswrapper[4805]: I1203 14:31:24.436343 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/afc5cad3-750c-429e-be48-6355bc07ac6a-config-data\") pod \"cinder-api-0\" (UID: \"afc5cad3-750c-429e-be48-6355bc07ac6a\") " pod="openstack/cinder-api-0" Dec 03 14:31:24 crc kubenswrapper[4805]: I1203 14:31:24.556240 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/afc5cad3-750c-429e-be48-6355bc07ac6a-logs\") pod \"cinder-api-0\" (UID: \"afc5cad3-750c-429e-be48-6355bc07ac6a\") " pod="openstack/cinder-api-0" Dec 03 14:31:24 crc kubenswrapper[4805]: I1203 14:31:24.556505 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/afc5cad3-750c-429e-be48-6355bc07ac6a-etc-machine-id\") pod \"cinder-api-0\" (UID: \"afc5cad3-750c-429e-be48-6355bc07ac6a\") " pod="openstack/cinder-api-0" Dec 03 14:31:24 crc kubenswrapper[4805]: I1203 14:31:24.556525 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9z95h\" (UniqueName: \"kubernetes.io/projected/afc5cad3-750c-429e-be48-6355bc07ac6a-kube-api-access-9z95h\") pod \"cinder-api-0\" (UID: \"afc5cad3-750c-429e-be48-6355bc07ac6a\") " pod="openstack/cinder-api-0" Dec 03 14:31:24 crc kubenswrapper[4805]: I1203 14:31:24.556590 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/afc5cad3-750c-429e-be48-6355bc07ac6a-scripts\") pod \"cinder-api-0\" (UID: \"afc5cad3-750c-429e-be48-6355bc07ac6a\") " pod="openstack/cinder-api-0" Dec 03 14:31:24 crc kubenswrapper[4805]: I1203 14:31:24.556612 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/afc5cad3-750c-429e-be48-6355bc07ac6a-config-data-custom\") pod \"cinder-api-0\" (UID: \"afc5cad3-750c-429e-be48-6355bc07ac6a\") " pod="openstack/cinder-api-0" Dec 03 14:31:24 crc kubenswrapper[4805]: I1203 14:31:24.556629 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/afc5cad3-750c-429e-be48-6355bc07ac6a-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"afc5cad3-750c-429e-be48-6355bc07ac6a\") " pod="openstack/cinder-api-0" Dec 03 14:31:24 crc kubenswrapper[4805]: I1203 14:31:24.556651 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/afc5cad3-750c-429e-be48-6355bc07ac6a-config-data\") pod \"cinder-api-0\" (UID: \"afc5cad3-750c-429e-be48-6355bc07ac6a\") " pod="openstack/cinder-api-0" Dec 03 14:31:24 crc kubenswrapper[4805]: I1203 14:31:24.558571 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/afc5cad3-750c-429e-be48-6355bc07ac6a-logs\") pod \"cinder-api-0\" (UID: \"afc5cad3-750c-429e-be48-6355bc07ac6a\") " pod="openstack/cinder-api-0" Dec 03 14:31:24 crc kubenswrapper[4805]: I1203 14:31:24.558625 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/afc5cad3-750c-429e-be48-6355bc07ac6a-etc-machine-id\") pod \"cinder-api-0\" (UID: \"afc5cad3-750c-429e-be48-6355bc07ac6a\") " pod="openstack/cinder-api-0" Dec 03 14:31:24 crc kubenswrapper[4805]: I1203 14:31:24.566272 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/afc5cad3-750c-429e-be48-6355bc07ac6a-scripts\") pod \"cinder-api-0\" (UID: \"afc5cad3-750c-429e-be48-6355bc07ac6a\") " pod="openstack/cinder-api-0" Dec 03 14:31:24 crc kubenswrapper[4805]: I1203 14:31:24.569923 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/afc5cad3-750c-429e-be48-6355bc07ac6a-config-data\") pod \"cinder-api-0\" (UID: \"afc5cad3-750c-429e-be48-6355bc07ac6a\") " pod="openstack/cinder-api-0" Dec 03 14:31:24 crc kubenswrapper[4805]: I1203 14:31:24.574522 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/afc5cad3-750c-429e-be48-6355bc07ac6a-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"afc5cad3-750c-429e-be48-6355bc07ac6a\") " pod="openstack/cinder-api-0" Dec 03 14:31:24 crc kubenswrapper[4805]: I1203 14:31:24.581421 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/afc5cad3-750c-429e-be48-6355bc07ac6a-config-data-custom\") pod \"cinder-api-0\" (UID: \"afc5cad3-750c-429e-be48-6355bc07ac6a\") " pod="openstack/cinder-api-0" Dec 03 14:31:24 crc kubenswrapper[4805]: I1203 14:31:24.587496 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9z95h\" (UniqueName: \"kubernetes.io/projected/afc5cad3-750c-429e-be48-6355bc07ac6a-kube-api-access-9z95h\") pod \"cinder-api-0\" (UID: \"afc5cad3-750c-429e-be48-6355bc07ac6a\") " pod="openstack/cinder-api-0" Dec 03 14:31:24 crc kubenswrapper[4805]: I1203 14:31:24.591237 4805 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-84b966f6c9-zwsmd" Dec 03 14:31:24 crc kubenswrapper[4805]: I1203 14:31:24.635578 4805 generic.go:334] "Generic (PLEG): container finished" podID="b3ca6772-26a0-410c-9609-1f49d08b1864" containerID="31628e51d65354f72d4dd433558b5b70b666da48c488d34fa61d3e72e666c631" exitCode=0 Dec 03 14:31:24 crc kubenswrapper[4805]: I1203 14:31:24.635619 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-84b966f6c9-zwsmd" event={"ID":"b3ca6772-26a0-410c-9609-1f49d08b1864","Type":"ContainerDied","Data":"31628e51d65354f72d4dd433558b5b70b666da48c488d34fa61d3e72e666c631"} Dec 03 14:31:24 crc kubenswrapper[4805]: I1203 14:31:24.635644 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-84b966f6c9-zwsmd" event={"ID":"b3ca6772-26a0-410c-9609-1f49d08b1864","Type":"ContainerDied","Data":"96e6ebbb257a1732acc3947ee1476daee89061ce93c32795c04d4108a648f662"} Dec 03 14:31:24 crc kubenswrapper[4805]: I1203 14:31:24.635660 4805 scope.go:117] "RemoveContainer" containerID="31628e51d65354f72d4dd433558b5b70b666da48c488d34fa61d3e72e666c631" Dec 03 14:31:24 crc kubenswrapper[4805]: I1203 14:31:24.635780 4805 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-84b966f6c9-zwsmd" Dec 03 14:31:24 crc kubenswrapper[4805]: I1203 14:31:24.659923 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/b3ca6772-26a0-410c-9609-1f49d08b1864-ovsdbserver-sb\") pod \"b3ca6772-26a0-410c-9609-1f49d08b1864\" (UID: \"b3ca6772-26a0-410c-9609-1f49d08b1864\") " Dec 03 14:31:24 crc kubenswrapper[4805]: I1203 14:31:24.659979 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/b3ca6772-26a0-410c-9609-1f49d08b1864-dns-svc\") pod \"b3ca6772-26a0-410c-9609-1f49d08b1864\" (UID: \"b3ca6772-26a0-410c-9609-1f49d08b1864\") " Dec 03 14:31:24 crc kubenswrapper[4805]: I1203 14:31:24.660037 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/b3ca6772-26a0-410c-9609-1f49d08b1864-ovsdbserver-nb\") pod \"b3ca6772-26a0-410c-9609-1f49d08b1864\" (UID: \"b3ca6772-26a0-410c-9609-1f49d08b1864\") " Dec 03 14:31:24 crc kubenswrapper[4805]: I1203 14:31:24.660117 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7n7vx\" (UniqueName: \"kubernetes.io/projected/b3ca6772-26a0-410c-9609-1f49d08b1864-kube-api-access-7n7vx\") pod \"b3ca6772-26a0-410c-9609-1f49d08b1864\" (UID: \"b3ca6772-26a0-410c-9609-1f49d08b1864\") " Dec 03 14:31:24 crc kubenswrapper[4805]: I1203 14:31:24.660234 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b3ca6772-26a0-410c-9609-1f49d08b1864-config\") pod \"b3ca6772-26a0-410c-9609-1f49d08b1864\" (UID: \"b3ca6772-26a0-410c-9609-1f49d08b1864\") " Dec 03 14:31:24 crc kubenswrapper[4805]: I1203 14:31:24.660260 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/b3ca6772-26a0-410c-9609-1f49d08b1864-dns-swift-storage-0\") pod \"b3ca6772-26a0-410c-9609-1f49d08b1864\" (UID: \"b3ca6772-26a0-410c-9609-1f49d08b1864\") " Dec 03 14:31:24 crc kubenswrapper[4805]: I1203 14:31:24.690316 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b3ca6772-26a0-410c-9609-1f49d08b1864-kube-api-access-7n7vx" (OuterVolumeSpecName: "kube-api-access-7n7vx") pod "b3ca6772-26a0-410c-9609-1f49d08b1864" (UID: "b3ca6772-26a0-410c-9609-1f49d08b1864"). InnerVolumeSpecName "kube-api-access-7n7vx". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 14:31:24 crc kubenswrapper[4805]: I1203 14:31:24.747997 4805 scope.go:117] "RemoveContainer" containerID="c3ec7bccf3f72b9e0eadf2e3235f7e5b3ecfefa7d6e5cabfb4409cfb4ee310b0" Dec 03 14:31:24 crc kubenswrapper[4805]: I1203 14:31:24.763826 4805 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7n7vx\" (UniqueName: \"kubernetes.io/projected/b3ca6772-26a0-410c-9609-1f49d08b1864-kube-api-access-7n7vx\") on node \"crc\" DevicePath \"\"" Dec 03 14:31:24 crc kubenswrapper[4805]: I1203 14:31:24.766579 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b3ca6772-26a0-410c-9609-1f49d08b1864-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "b3ca6772-26a0-410c-9609-1f49d08b1864" (UID: "b3ca6772-26a0-410c-9609-1f49d08b1864"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 14:31:24 crc kubenswrapper[4805]: I1203 14:31:24.813069 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Dec 03 14:31:24 crc kubenswrapper[4805]: I1203 14:31:24.831444 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b3ca6772-26a0-410c-9609-1f49d08b1864-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "b3ca6772-26a0-410c-9609-1f49d08b1864" (UID: "b3ca6772-26a0-410c-9609-1f49d08b1864"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 14:31:24 crc kubenswrapper[4805]: I1203 14:31:24.831809 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b3ca6772-26a0-410c-9609-1f49d08b1864-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "b3ca6772-26a0-410c-9609-1f49d08b1864" (UID: "b3ca6772-26a0-410c-9609-1f49d08b1864"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 14:31:24 crc kubenswrapper[4805]: I1203 14:31:24.856470 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b3ca6772-26a0-410c-9609-1f49d08b1864-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "b3ca6772-26a0-410c-9609-1f49d08b1864" (UID: "b3ca6772-26a0-410c-9609-1f49d08b1864"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 14:31:24 crc kubenswrapper[4805]: I1203 14:31:24.861484 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b3ca6772-26a0-410c-9609-1f49d08b1864-config" (OuterVolumeSpecName: "config") pod "b3ca6772-26a0-410c-9609-1f49d08b1864" (UID: "b3ca6772-26a0-410c-9609-1f49d08b1864"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 14:31:24 crc kubenswrapper[4805]: I1203 14:31:24.875009 4805 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b3ca6772-26a0-410c-9609-1f49d08b1864-config\") on node \"crc\" DevicePath \"\"" Dec 03 14:31:24 crc kubenswrapper[4805]: I1203 14:31:24.875057 4805 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/b3ca6772-26a0-410c-9609-1f49d08b1864-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Dec 03 14:31:24 crc kubenswrapper[4805]: I1203 14:31:24.875071 4805 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/b3ca6772-26a0-410c-9609-1f49d08b1864-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Dec 03 14:31:24 crc kubenswrapper[4805]: I1203 14:31:24.875085 4805 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/b3ca6772-26a0-410c-9609-1f49d08b1864-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 03 14:31:24 crc kubenswrapper[4805]: I1203 14:31:24.875096 4805 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/b3ca6772-26a0-410c-9609-1f49d08b1864-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Dec 03 14:31:24 crc kubenswrapper[4805]: I1203 14:31:24.974380 4805 scope.go:117] "RemoveContainer" containerID="31628e51d65354f72d4dd433558b5b70b666da48c488d34fa61d3e72e666c631" Dec 03 14:31:24 crc kubenswrapper[4805]: E1203 14:31:24.994575 4805 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"31628e51d65354f72d4dd433558b5b70b666da48c488d34fa61d3e72e666c631\": container with ID starting with 31628e51d65354f72d4dd433558b5b70b666da48c488d34fa61d3e72e666c631 not found: ID does not exist" containerID="31628e51d65354f72d4dd433558b5b70b666da48c488d34fa61d3e72e666c631" Dec 03 14:31:24 crc kubenswrapper[4805]: I1203 14:31:24.994720 4805 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"31628e51d65354f72d4dd433558b5b70b666da48c488d34fa61d3e72e666c631"} err="failed to get container status \"31628e51d65354f72d4dd433558b5b70b666da48c488d34fa61d3e72e666c631\": rpc error: code = NotFound desc = could not find container \"31628e51d65354f72d4dd433558b5b70b666da48c488d34fa61d3e72e666c631\": container with ID starting with 31628e51d65354f72d4dd433558b5b70b666da48c488d34fa61d3e72e666c631 not found: ID does not exist" Dec 03 14:31:24 crc kubenswrapper[4805]: I1203 14:31:24.994751 4805 scope.go:117] "RemoveContainer" containerID="c3ec7bccf3f72b9e0eadf2e3235f7e5b3ecfefa7d6e5cabfb4409cfb4ee310b0" Dec 03 14:31:25 crc kubenswrapper[4805]: E1203 14:31:25.001193 4805 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c3ec7bccf3f72b9e0eadf2e3235f7e5b3ecfefa7d6e5cabfb4409cfb4ee310b0\": container with ID starting with c3ec7bccf3f72b9e0eadf2e3235f7e5b3ecfefa7d6e5cabfb4409cfb4ee310b0 not found: ID does not exist" containerID="c3ec7bccf3f72b9e0eadf2e3235f7e5b3ecfefa7d6e5cabfb4409cfb4ee310b0" Dec 03 14:31:25 crc kubenswrapper[4805]: I1203 14:31:25.001238 4805 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c3ec7bccf3f72b9e0eadf2e3235f7e5b3ecfefa7d6e5cabfb4409cfb4ee310b0"} err="failed to get container status \"c3ec7bccf3f72b9e0eadf2e3235f7e5b3ecfefa7d6e5cabfb4409cfb4ee310b0\": rpc error: code = NotFound desc = could not find container \"c3ec7bccf3f72b9e0eadf2e3235f7e5b3ecfefa7d6e5cabfb4409cfb4ee310b0\": container with ID starting with c3ec7bccf3f72b9e0eadf2e3235f7e5b3ecfefa7d6e5cabfb4409cfb4ee310b0 not found: ID does not exist" Dec 03 14:31:25 crc kubenswrapper[4805]: I1203 14:31:25.032957 4805 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-84b966f6c9-zwsmd"] Dec 03 14:31:25 crc kubenswrapper[4805]: I1203 14:31:25.041688 4805 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-84b966f6c9-zwsmd"] Dec 03 14:31:25 crc kubenswrapper[4805]: I1203 14:31:25.063958 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5784cf869f-l4bt9"] Dec 03 14:31:25 crc kubenswrapper[4805]: I1203 14:31:25.229283 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-scheduler-0"] Dec 03 14:31:25 crc kubenswrapper[4805]: I1203 14:31:25.449794 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-api-0"] Dec 03 14:31:25 crc kubenswrapper[4805]: I1203 14:31:25.671383 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"afc5cad3-750c-429e-be48-6355bc07ac6a","Type":"ContainerStarted","Data":"15abd9a67de7f1e48b9365a46a01d9a55f31ceb6300e45d8fcb784327b5d18a9"} Dec 03 14:31:25 crc kubenswrapper[4805]: I1203 14:31:25.676892 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5784cf869f-l4bt9" event={"ID":"bc7fd7a7-02bc-48e1-9bde-d1680876da37","Type":"ContainerStarted","Data":"f31912c14d4c8f23d67c323b6a6401e02aee26fb20bd559099f3c3aada4dffa3"} Dec 03 14:31:25 crc kubenswrapper[4805]: I1203 14:31:25.676947 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5784cf869f-l4bt9" event={"ID":"bc7fd7a7-02bc-48e1-9bde-d1680876da37","Type":"ContainerStarted","Data":"09a0e37d64b74a252a57f21a3a4a49eec5dc40794afbee941035747c1661f72e"} Dec 03 14:31:25 crc kubenswrapper[4805]: I1203 14:31:25.682048 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"29710b09-a1ad-468f-af26-b5d5eb516898","Type":"ContainerStarted","Data":"c6b5100cdc92b5421f6d6c7f7bc48f88da1ec74c305b02d496f9ff09c6110c39"} Dec 03 14:31:26 crc kubenswrapper[4805]: I1203 14:31:26.047267 4805 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/barbican-api-845f47d448-thqcr" Dec 03 14:31:26 crc kubenswrapper[4805]: I1203 14:31:26.245698 4805 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-api-0"] Dec 03 14:31:26 crc kubenswrapper[4805]: I1203 14:31:26.577610 4805 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/barbican-api-845f47d448-thqcr" Dec 03 14:31:26 crc kubenswrapper[4805]: I1203 14:31:26.741154 4805 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b3ca6772-26a0-410c-9609-1f49d08b1864" path="/var/lib/kubelet/pods/b3ca6772-26a0-410c-9609-1f49d08b1864/volumes" Dec 03 14:31:26 crc kubenswrapper[4805]: I1203 14:31:26.773889 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"afc5cad3-750c-429e-be48-6355bc07ac6a","Type":"ContainerStarted","Data":"da8cd59d46d4d0271f0708f93fc1e61436d3d96636caa2268abbe253d568f068"} Dec 03 14:31:26 crc kubenswrapper[4805]: I1203 14:31:26.803008 4805 generic.go:334] "Generic (PLEG): container finished" podID="bc7fd7a7-02bc-48e1-9bde-d1680876da37" containerID="f31912c14d4c8f23d67c323b6a6401e02aee26fb20bd559099f3c3aada4dffa3" exitCode=0 Dec 03 14:31:26 crc kubenswrapper[4805]: I1203 14:31:26.803911 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5784cf869f-l4bt9" event={"ID":"bc7fd7a7-02bc-48e1-9bde-d1680876da37","Type":"ContainerDied","Data":"f31912c14d4c8f23d67c323b6a6401e02aee26fb20bd559099f3c3aada4dffa3"} Dec 03 14:31:27 crc kubenswrapper[4805]: I1203 14:31:27.814964 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"29710b09-a1ad-468f-af26-b5d5eb516898","Type":"ContainerStarted","Data":"cb75dea47e869ccbdcc4f2f309b4ef5bddded1bea1f31601b29cb63b5646b269"} Dec 03 14:31:27 crc kubenswrapper[4805]: I1203 14:31:27.817698 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5784cf869f-l4bt9" event={"ID":"bc7fd7a7-02bc-48e1-9bde-d1680876da37","Type":"ContainerStarted","Data":"d957799f184adf285454f33a4f7aacfb2a9823b898bb6d66acdb0d27983681d6"} Dec 03 14:31:27 crc kubenswrapper[4805]: I1203 14:31:27.817873 4805 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-5784cf869f-l4bt9" Dec 03 14:31:27 crc kubenswrapper[4805]: I1203 14:31:27.835928 4805 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-5784cf869f-l4bt9" podStartSLOduration=4.835913393 podStartE2EDuration="4.835913393s" podCreationTimestamp="2025-12-03 14:31:23 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 14:31:27.835111581 +0000 UTC m=+1317.498028504" watchObservedRunningTime="2025-12-03 14:31:27.835913393 +0000 UTC m=+1317.498830316" Dec 03 14:31:28 crc kubenswrapper[4805]: I1203 14:31:28.509279 4805 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/barbican-api-58794647f4-rtjlq" Dec 03 14:31:28 crc kubenswrapper[4805]: I1203 14:31:28.532577 4805 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/barbican-api-58794647f4-rtjlq" Dec 03 14:31:28 crc kubenswrapper[4805]: I1203 14:31:28.591829 4805 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-api-845f47d448-thqcr"] Dec 03 14:31:28 crc kubenswrapper[4805]: I1203 14:31:28.593591 4805 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/barbican-api-845f47d448-thqcr" podUID="37658956-d75a-43ce-81d8-38f9bca42957" containerName="barbican-api-log" containerID="cri-o://a54729ca2abce4df438fff733d8adb4838b1bcfb1e74dd34f90719edf58bd259" gracePeriod=30 Dec 03 14:31:28 crc kubenswrapper[4805]: I1203 14:31:28.593978 4805 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/barbican-api-845f47d448-thqcr" podUID="37658956-d75a-43ce-81d8-38f9bca42957" containerName="barbican-api" containerID="cri-o://7adb8b3c5e068f3ee12e9581aecb5bdfa43066e3e1f0991bc28977ea2af801da" gracePeriod=30 Dec 03 14:31:28 crc kubenswrapper[4805]: I1203 14:31:28.606336 4805 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/barbican-api-845f47d448-thqcr" podUID="37658956-d75a-43ce-81d8-38f9bca42957" containerName="barbican-api" probeResult="failure" output="Get \"http://10.217.0.159:9311/healthcheck\": EOF" Dec 03 14:31:29 crc kubenswrapper[4805]: I1203 14:31:29.840633 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"afc5cad3-750c-429e-be48-6355bc07ac6a","Type":"ContainerStarted","Data":"a7a77d1494bbadc672dbcab26f3b73cc3725f5cbea40bc88d21a85a1779e81a6"} Dec 03 14:31:29 crc kubenswrapper[4805]: I1203 14:31:29.841480 4805 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/cinder-api-0" Dec 03 14:31:29 crc kubenswrapper[4805]: I1203 14:31:29.840857 4805 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-api-0" podUID="afc5cad3-750c-429e-be48-6355bc07ac6a" containerName="cinder-api" containerID="cri-o://a7a77d1494bbadc672dbcab26f3b73cc3725f5cbea40bc88d21a85a1779e81a6" gracePeriod=30 Dec 03 14:31:29 crc kubenswrapper[4805]: I1203 14:31:29.840759 4805 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-api-0" podUID="afc5cad3-750c-429e-be48-6355bc07ac6a" containerName="cinder-api-log" containerID="cri-o://da8cd59d46d4d0271f0708f93fc1e61436d3d96636caa2268abbe253d568f068" gracePeriod=30 Dec 03 14:31:29 crc kubenswrapper[4805]: I1203 14:31:29.851512 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"29710b09-a1ad-468f-af26-b5d5eb516898","Type":"ContainerStarted","Data":"95b1ef4794618699f48a7350bb18cecf536b9ae5725dda4c6c96e5147bb28baf"} Dec 03 14:31:29 crc kubenswrapper[4805]: I1203 14:31:29.856373 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-845f47d448-thqcr" event={"ID":"37658956-d75a-43ce-81d8-38f9bca42957","Type":"ContainerDied","Data":"a54729ca2abce4df438fff733d8adb4838b1bcfb1e74dd34f90719edf58bd259"} Dec 03 14:31:29 crc kubenswrapper[4805]: I1203 14:31:29.856330 4805 generic.go:334] "Generic (PLEG): container finished" podID="37658956-d75a-43ce-81d8-38f9bca42957" containerID="a54729ca2abce4df438fff733d8adb4838b1bcfb1e74dd34f90719edf58bd259" exitCode=143 Dec 03 14:31:29 crc kubenswrapper[4805]: I1203 14:31:29.874785 4805 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-api-0" podStartSLOduration=5.874767228 podStartE2EDuration="5.874767228s" podCreationTimestamp="2025-12-03 14:31:24 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 14:31:29.865195696 +0000 UTC m=+1319.528112649" watchObservedRunningTime="2025-12-03 14:31:29.874767228 +0000 UTC m=+1319.537684151" Dec 03 14:31:29 crc kubenswrapper[4805]: I1203 14:31:29.905139 4805 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-scheduler-0" podStartSLOduration=6.02881757 podStartE2EDuration="6.905113578s" podCreationTimestamp="2025-12-03 14:31:23 +0000 UTC" firstStartedPulling="2025-12-03 14:31:25.307103446 +0000 UTC m=+1314.970020369" lastFinishedPulling="2025-12-03 14:31:26.183399454 +0000 UTC m=+1315.846316377" observedRunningTime="2025-12-03 14:31:29.896152663 +0000 UTC m=+1319.559069576" watchObservedRunningTime="2025-12-03 14:31:29.905113578 +0000 UTC m=+1319.568030501" Dec 03 14:31:30 crc kubenswrapper[4805]: I1203 14:31:30.861075 4805 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Dec 03 14:31:30 crc kubenswrapper[4805]: I1203 14:31:30.877771 4805 generic.go:334] "Generic (PLEG): container finished" podID="afc5cad3-750c-429e-be48-6355bc07ac6a" containerID="a7a77d1494bbadc672dbcab26f3b73cc3725f5cbea40bc88d21a85a1779e81a6" exitCode=0 Dec 03 14:31:30 crc kubenswrapper[4805]: I1203 14:31:30.877822 4805 generic.go:334] "Generic (PLEG): container finished" podID="afc5cad3-750c-429e-be48-6355bc07ac6a" containerID="da8cd59d46d4d0271f0708f93fc1e61436d3d96636caa2268abbe253d568f068" exitCode=143 Dec 03 14:31:30 crc kubenswrapper[4805]: I1203 14:31:30.877902 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"afc5cad3-750c-429e-be48-6355bc07ac6a","Type":"ContainerDied","Data":"a7a77d1494bbadc672dbcab26f3b73cc3725f5cbea40bc88d21a85a1779e81a6"} Dec 03 14:31:30 crc kubenswrapper[4805]: I1203 14:31:30.877952 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"afc5cad3-750c-429e-be48-6355bc07ac6a","Type":"ContainerDied","Data":"da8cd59d46d4d0271f0708f93fc1e61436d3d96636caa2268abbe253d568f068"} Dec 03 14:31:30 crc kubenswrapper[4805]: I1203 14:31:30.877963 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"afc5cad3-750c-429e-be48-6355bc07ac6a","Type":"ContainerDied","Data":"15abd9a67de7f1e48b9365a46a01d9a55f31ceb6300e45d8fcb784327b5d18a9"} Dec 03 14:31:30 crc kubenswrapper[4805]: I1203 14:31:30.877979 4805 scope.go:117] "RemoveContainer" containerID="a7a77d1494bbadc672dbcab26f3b73cc3725f5cbea40bc88d21a85a1779e81a6" Dec 03 14:31:30 crc kubenswrapper[4805]: I1203 14:31:30.878151 4805 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Dec 03 14:31:30 crc kubenswrapper[4805]: I1203 14:31:30.908414 4805 generic.go:334] "Generic (PLEG): container finished" podID="318792ec-dc8b-4a37-92ab-d31773d66350" containerID="48672ddefb5f29199ffd9090e43f5c584cf661e8cf1e4313d963fb96b28729b6" exitCode=0 Dec 03 14:31:30 crc kubenswrapper[4805]: I1203 14:31:30.911673 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-75f57c664d-4tsg2" event={"ID":"318792ec-dc8b-4a37-92ab-d31773d66350","Type":"ContainerDied","Data":"48672ddefb5f29199ffd9090e43f5c584cf661e8cf1e4313d963fb96b28729b6"} Dec 03 14:31:30 crc kubenswrapper[4805]: I1203 14:31:30.965111 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/afc5cad3-750c-429e-be48-6355bc07ac6a-etc-machine-id\") pod \"afc5cad3-750c-429e-be48-6355bc07ac6a\" (UID: \"afc5cad3-750c-429e-be48-6355bc07ac6a\") " Dec 03 14:31:30 crc kubenswrapper[4805]: I1203 14:31:30.965231 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/afc5cad3-750c-429e-be48-6355bc07ac6a-combined-ca-bundle\") pod \"afc5cad3-750c-429e-be48-6355bc07ac6a\" (UID: \"afc5cad3-750c-429e-be48-6355bc07ac6a\") " Dec 03 14:31:30 crc kubenswrapper[4805]: I1203 14:31:30.965238 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/afc5cad3-750c-429e-be48-6355bc07ac6a-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "afc5cad3-750c-429e-be48-6355bc07ac6a" (UID: "afc5cad3-750c-429e-be48-6355bc07ac6a"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 03 14:31:30 crc kubenswrapper[4805]: I1203 14:31:30.965262 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/afc5cad3-750c-429e-be48-6355bc07ac6a-scripts\") pod \"afc5cad3-750c-429e-be48-6355bc07ac6a\" (UID: \"afc5cad3-750c-429e-be48-6355bc07ac6a\") " Dec 03 14:31:30 crc kubenswrapper[4805]: I1203 14:31:30.965338 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/afc5cad3-750c-429e-be48-6355bc07ac6a-config-data\") pod \"afc5cad3-750c-429e-be48-6355bc07ac6a\" (UID: \"afc5cad3-750c-429e-be48-6355bc07ac6a\") " Dec 03 14:31:30 crc kubenswrapper[4805]: I1203 14:31:30.965436 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9z95h\" (UniqueName: \"kubernetes.io/projected/afc5cad3-750c-429e-be48-6355bc07ac6a-kube-api-access-9z95h\") pod \"afc5cad3-750c-429e-be48-6355bc07ac6a\" (UID: \"afc5cad3-750c-429e-be48-6355bc07ac6a\") " Dec 03 14:31:30 crc kubenswrapper[4805]: I1203 14:31:30.965471 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/afc5cad3-750c-429e-be48-6355bc07ac6a-config-data-custom\") pod \"afc5cad3-750c-429e-be48-6355bc07ac6a\" (UID: \"afc5cad3-750c-429e-be48-6355bc07ac6a\") " Dec 03 14:31:30 crc kubenswrapper[4805]: I1203 14:31:30.965534 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/afc5cad3-750c-429e-be48-6355bc07ac6a-logs\") pod \"afc5cad3-750c-429e-be48-6355bc07ac6a\" (UID: \"afc5cad3-750c-429e-be48-6355bc07ac6a\") " Dec 03 14:31:30 crc kubenswrapper[4805]: I1203 14:31:30.966088 4805 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/afc5cad3-750c-429e-be48-6355bc07ac6a-etc-machine-id\") on node \"crc\" DevicePath \"\"" Dec 03 14:31:30 crc kubenswrapper[4805]: I1203 14:31:30.967027 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/afc5cad3-750c-429e-be48-6355bc07ac6a-logs" (OuterVolumeSpecName: "logs") pod "afc5cad3-750c-429e-be48-6355bc07ac6a" (UID: "afc5cad3-750c-429e-be48-6355bc07ac6a"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 14:31:30 crc kubenswrapper[4805]: I1203 14:31:30.973983 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/afc5cad3-750c-429e-be48-6355bc07ac6a-kube-api-access-9z95h" (OuterVolumeSpecName: "kube-api-access-9z95h") pod "afc5cad3-750c-429e-be48-6355bc07ac6a" (UID: "afc5cad3-750c-429e-be48-6355bc07ac6a"). InnerVolumeSpecName "kube-api-access-9z95h". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 14:31:30 crc kubenswrapper[4805]: I1203 14:31:30.985911 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/afc5cad3-750c-429e-be48-6355bc07ac6a-scripts" (OuterVolumeSpecName: "scripts") pod "afc5cad3-750c-429e-be48-6355bc07ac6a" (UID: "afc5cad3-750c-429e-be48-6355bc07ac6a"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 14:31:30 crc kubenswrapper[4805]: I1203 14:31:30.985978 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/afc5cad3-750c-429e-be48-6355bc07ac6a-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "afc5cad3-750c-429e-be48-6355bc07ac6a" (UID: "afc5cad3-750c-429e-be48-6355bc07ac6a"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 14:31:31 crc kubenswrapper[4805]: I1203 14:31:31.013966 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/afc5cad3-750c-429e-be48-6355bc07ac6a-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "afc5cad3-750c-429e-be48-6355bc07ac6a" (UID: "afc5cad3-750c-429e-be48-6355bc07ac6a"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 14:31:31 crc kubenswrapper[4805]: I1203 14:31:31.067936 4805 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/afc5cad3-750c-429e-be48-6355bc07ac6a-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 14:31:31 crc kubenswrapper[4805]: I1203 14:31:31.067973 4805 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/afc5cad3-750c-429e-be48-6355bc07ac6a-scripts\") on node \"crc\" DevicePath \"\"" Dec 03 14:31:31 crc kubenswrapper[4805]: I1203 14:31:31.067987 4805 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9z95h\" (UniqueName: \"kubernetes.io/projected/afc5cad3-750c-429e-be48-6355bc07ac6a-kube-api-access-9z95h\") on node \"crc\" DevicePath \"\"" Dec 03 14:31:31 crc kubenswrapper[4805]: I1203 14:31:31.068002 4805 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/afc5cad3-750c-429e-be48-6355bc07ac6a-config-data-custom\") on node \"crc\" DevicePath \"\"" Dec 03 14:31:31 crc kubenswrapper[4805]: I1203 14:31:31.068004 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/afc5cad3-750c-429e-be48-6355bc07ac6a-config-data" (OuterVolumeSpecName: "config-data") pod "afc5cad3-750c-429e-be48-6355bc07ac6a" (UID: "afc5cad3-750c-429e-be48-6355bc07ac6a"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 14:31:31 crc kubenswrapper[4805]: I1203 14:31:31.068017 4805 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/afc5cad3-750c-429e-be48-6355bc07ac6a-logs\") on node \"crc\" DevicePath \"\"" Dec 03 14:31:31 crc kubenswrapper[4805]: I1203 14:31:31.148595 4805 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-75f57c664d-4tsg2" Dec 03 14:31:31 crc kubenswrapper[4805]: I1203 14:31:31.151017 4805 scope.go:117] "RemoveContainer" containerID="da8cd59d46d4d0271f0708f93fc1e61436d3d96636caa2268abbe253d568f068" Dec 03 14:31:31 crc kubenswrapper[4805]: I1203 14:31:31.169502 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/318792ec-dc8b-4a37-92ab-d31773d66350-config\") pod \"318792ec-dc8b-4a37-92ab-d31773d66350\" (UID: \"318792ec-dc8b-4a37-92ab-d31773d66350\") " Dec 03 14:31:31 crc kubenswrapper[4805]: I1203 14:31:31.169544 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/318792ec-dc8b-4a37-92ab-d31773d66350-ovndb-tls-certs\") pod \"318792ec-dc8b-4a37-92ab-d31773d66350\" (UID: \"318792ec-dc8b-4a37-92ab-d31773d66350\") " Dec 03 14:31:31 crc kubenswrapper[4805]: I1203 14:31:31.169609 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/318792ec-dc8b-4a37-92ab-d31773d66350-httpd-config\") pod \"318792ec-dc8b-4a37-92ab-d31773d66350\" (UID: \"318792ec-dc8b-4a37-92ab-d31773d66350\") " Dec 03 14:31:31 crc kubenswrapper[4805]: I1203 14:31:31.169663 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/318792ec-dc8b-4a37-92ab-d31773d66350-combined-ca-bundle\") pod \"318792ec-dc8b-4a37-92ab-d31773d66350\" (UID: \"318792ec-dc8b-4a37-92ab-d31773d66350\") " Dec 03 14:31:31 crc kubenswrapper[4805]: I1203 14:31:31.169779 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-28nqz\" (UniqueName: \"kubernetes.io/projected/318792ec-dc8b-4a37-92ab-d31773d66350-kube-api-access-28nqz\") pod \"318792ec-dc8b-4a37-92ab-d31773d66350\" (UID: \"318792ec-dc8b-4a37-92ab-d31773d66350\") " Dec 03 14:31:31 crc kubenswrapper[4805]: I1203 14:31:31.170161 4805 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/afc5cad3-750c-429e-be48-6355bc07ac6a-config-data\") on node \"crc\" DevicePath \"\"" Dec 03 14:31:31 crc kubenswrapper[4805]: I1203 14:31:31.198878 4805 scope.go:117] "RemoveContainer" containerID="a7a77d1494bbadc672dbcab26f3b73cc3725f5cbea40bc88d21a85a1779e81a6" Dec 03 14:31:31 crc kubenswrapper[4805]: E1203 14:31:31.203166 4805 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a7a77d1494bbadc672dbcab26f3b73cc3725f5cbea40bc88d21a85a1779e81a6\": container with ID starting with a7a77d1494bbadc672dbcab26f3b73cc3725f5cbea40bc88d21a85a1779e81a6 not found: ID does not exist" containerID="a7a77d1494bbadc672dbcab26f3b73cc3725f5cbea40bc88d21a85a1779e81a6" Dec 03 14:31:31 crc kubenswrapper[4805]: I1203 14:31:31.203219 4805 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a7a77d1494bbadc672dbcab26f3b73cc3725f5cbea40bc88d21a85a1779e81a6"} err="failed to get container status \"a7a77d1494bbadc672dbcab26f3b73cc3725f5cbea40bc88d21a85a1779e81a6\": rpc error: code = NotFound desc = could not find container \"a7a77d1494bbadc672dbcab26f3b73cc3725f5cbea40bc88d21a85a1779e81a6\": container with ID starting with a7a77d1494bbadc672dbcab26f3b73cc3725f5cbea40bc88d21a85a1779e81a6 not found: ID does not exist" Dec 03 14:31:31 crc kubenswrapper[4805]: I1203 14:31:31.203253 4805 scope.go:117] "RemoveContainer" containerID="da8cd59d46d4d0271f0708f93fc1e61436d3d96636caa2268abbe253d568f068" Dec 03 14:31:31 crc kubenswrapper[4805]: E1203 14:31:31.207273 4805 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"da8cd59d46d4d0271f0708f93fc1e61436d3d96636caa2268abbe253d568f068\": container with ID starting with da8cd59d46d4d0271f0708f93fc1e61436d3d96636caa2268abbe253d568f068 not found: ID does not exist" containerID="da8cd59d46d4d0271f0708f93fc1e61436d3d96636caa2268abbe253d568f068" Dec 03 14:31:31 crc kubenswrapper[4805]: I1203 14:31:31.207321 4805 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"da8cd59d46d4d0271f0708f93fc1e61436d3d96636caa2268abbe253d568f068"} err="failed to get container status \"da8cd59d46d4d0271f0708f93fc1e61436d3d96636caa2268abbe253d568f068\": rpc error: code = NotFound desc = could not find container \"da8cd59d46d4d0271f0708f93fc1e61436d3d96636caa2268abbe253d568f068\": container with ID starting with da8cd59d46d4d0271f0708f93fc1e61436d3d96636caa2268abbe253d568f068 not found: ID does not exist" Dec 03 14:31:31 crc kubenswrapper[4805]: I1203 14:31:31.207353 4805 scope.go:117] "RemoveContainer" containerID="a7a77d1494bbadc672dbcab26f3b73cc3725f5cbea40bc88d21a85a1779e81a6" Dec 03 14:31:31 crc kubenswrapper[4805]: I1203 14:31:31.208293 4805 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a7a77d1494bbadc672dbcab26f3b73cc3725f5cbea40bc88d21a85a1779e81a6"} err="failed to get container status \"a7a77d1494bbadc672dbcab26f3b73cc3725f5cbea40bc88d21a85a1779e81a6\": rpc error: code = NotFound desc = could not find container \"a7a77d1494bbadc672dbcab26f3b73cc3725f5cbea40bc88d21a85a1779e81a6\": container with ID starting with a7a77d1494bbadc672dbcab26f3b73cc3725f5cbea40bc88d21a85a1779e81a6 not found: ID does not exist" Dec 03 14:31:31 crc kubenswrapper[4805]: I1203 14:31:31.208323 4805 scope.go:117] "RemoveContainer" containerID="da8cd59d46d4d0271f0708f93fc1e61436d3d96636caa2268abbe253d568f068" Dec 03 14:31:31 crc kubenswrapper[4805]: I1203 14:31:31.210209 4805 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"da8cd59d46d4d0271f0708f93fc1e61436d3d96636caa2268abbe253d568f068"} err="failed to get container status \"da8cd59d46d4d0271f0708f93fc1e61436d3d96636caa2268abbe253d568f068\": rpc error: code = NotFound desc = could not find container \"da8cd59d46d4d0271f0708f93fc1e61436d3d96636caa2268abbe253d568f068\": container with ID starting with da8cd59d46d4d0271f0708f93fc1e61436d3d96636caa2268abbe253d568f068 not found: ID does not exist" Dec 03 14:31:31 crc kubenswrapper[4805]: I1203 14:31:31.220301 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/318792ec-dc8b-4a37-92ab-d31773d66350-httpd-config" (OuterVolumeSpecName: "httpd-config") pod "318792ec-dc8b-4a37-92ab-d31773d66350" (UID: "318792ec-dc8b-4a37-92ab-d31773d66350"). InnerVolumeSpecName "httpd-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 14:31:31 crc kubenswrapper[4805]: I1203 14:31:31.241118 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/318792ec-dc8b-4a37-92ab-d31773d66350-kube-api-access-28nqz" (OuterVolumeSpecName: "kube-api-access-28nqz") pod "318792ec-dc8b-4a37-92ab-d31773d66350" (UID: "318792ec-dc8b-4a37-92ab-d31773d66350"). InnerVolumeSpecName "kube-api-access-28nqz". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 14:31:31 crc kubenswrapper[4805]: I1203 14:31:31.250105 4805 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-api-0"] Dec 03 14:31:31 crc kubenswrapper[4805]: I1203 14:31:31.258140 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/318792ec-dc8b-4a37-92ab-d31773d66350-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "318792ec-dc8b-4a37-92ab-d31773d66350" (UID: "318792ec-dc8b-4a37-92ab-d31773d66350"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 14:31:31 crc kubenswrapper[4805]: I1203 14:31:31.266797 4805 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-api-0"] Dec 03 14:31:31 crc kubenswrapper[4805]: I1203 14:31:31.271490 4805 reconciler_common.go:293] "Volume detached for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/318792ec-dc8b-4a37-92ab-d31773d66350-httpd-config\") on node \"crc\" DevicePath \"\"" Dec 03 14:31:31 crc kubenswrapper[4805]: I1203 14:31:31.271531 4805 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/318792ec-dc8b-4a37-92ab-d31773d66350-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 14:31:31 crc kubenswrapper[4805]: I1203 14:31:31.271548 4805 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-28nqz\" (UniqueName: \"kubernetes.io/projected/318792ec-dc8b-4a37-92ab-d31773d66350-kube-api-access-28nqz\") on node \"crc\" DevicePath \"\"" Dec 03 14:31:31 crc kubenswrapper[4805]: I1203 14:31:31.280243 4805 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-api-0"] Dec 03 14:31:31 crc kubenswrapper[4805]: E1203 14:31:31.280713 4805 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b3ca6772-26a0-410c-9609-1f49d08b1864" containerName="init" Dec 03 14:31:31 crc kubenswrapper[4805]: I1203 14:31:31.280728 4805 state_mem.go:107] "Deleted CPUSet assignment" podUID="b3ca6772-26a0-410c-9609-1f49d08b1864" containerName="init" Dec 03 14:31:31 crc kubenswrapper[4805]: E1203 14:31:31.280744 4805 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="afc5cad3-750c-429e-be48-6355bc07ac6a" containerName="cinder-api-log" Dec 03 14:31:31 crc kubenswrapper[4805]: I1203 14:31:31.280751 4805 state_mem.go:107] "Deleted CPUSet assignment" podUID="afc5cad3-750c-429e-be48-6355bc07ac6a" containerName="cinder-api-log" Dec 03 14:31:31 crc kubenswrapper[4805]: E1203 14:31:31.280786 4805 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="afc5cad3-750c-429e-be48-6355bc07ac6a" containerName="cinder-api" Dec 03 14:31:31 crc kubenswrapper[4805]: I1203 14:31:31.280795 4805 state_mem.go:107] "Deleted CPUSet assignment" podUID="afc5cad3-750c-429e-be48-6355bc07ac6a" containerName="cinder-api" Dec 03 14:31:31 crc kubenswrapper[4805]: E1203 14:31:31.280807 4805 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="318792ec-dc8b-4a37-92ab-d31773d66350" containerName="neutron-api" Dec 03 14:31:31 crc kubenswrapper[4805]: I1203 14:31:31.280831 4805 state_mem.go:107] "Deleted CPUSet assignment" podUID="318792ec-dc8b-4a37-92ab-d31773d66350" containerName="neutron-api" Dec 03 14:31:31 crc kubenswrapper[4805]: E1203 14:31:31.280859 4805 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="318792ec-dc8b-4a37-92ab-d31773d66350" containerName="neutron-httpd" Dec 03 14:31:31 crc kubenswrapper[4805]: I1203 14:31:31.280869 4805 state_mem.go:107] "Deleted CPUSet assignment" podUID="318792ec-dc8b-4a37-92ab-d31773d66350" containerName="neutron-httpd" Dec 03 14:31:31 crc kubenswrapper[4805]: E1203 14:31:31.280887 4805 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b3ca6772-26a0-410c-9609-1f49d08b1864" containerName="dnsmasq-dns" Dec 03 14:31:31 crc kubenswrapper[4805]: I1203 14:31:31.280895 4805 state_mem.go:107] "Deleted CPUSet assignment" podUID="b3ca6772-26a0-410c-9609-1f49d08b1864" containerName="dnsmasq-dns" Dec 03 14:31:31 crc kubenswrapper[4805]: I1203 14:31:31.281120 4805 memory_manager.go:354] "RemoveStaleState removing state" podUID="318792ec-dc8b-4a37-92ab-d31773d66350" containerName="neutron-httpd" Dec 03 14:31:31 crc kubenswrapper[4805]: I1203 14:31:31.281146 4805 memory_manager.go:354] "RemoveStaleState removing state" podUID="afc5cad3-750c-429e-be48-6355bc07ac6a" containerName="cinder-api" Dec 03 14:31:31 crc kubenswrapper[4805]: I1203 14:31:31.281167 4805 memory_manager.go:354] "RemoveStaleState removing state" podUID="318792ec-dc8b-4a37-92ab-d31773d66350" containerName="neutron-api" Dec 03 14:31:31 crc kubenswrapper[4805]: I1203 14:31:31.281182 4805 memory_manager.go:354] "RemoveStaleState removing state" podUID="afc5cad3-750c-429e-be48-6355bc07ac6a" containerName="cinder-api-log" Dec 03 14:31:31 crc kubenswrapper[4805]: I1203 14:31:31.281191 4805 memory_manager.go:354] "RemoveStaleState removing state" podUID="b3ca6772-26a0-410c-9609-1f49d08b1864" containerName="dnsmasq-dns" Dec 03 14:31:31 crc kubenswrapper[4805]: I1203 14:31:31.282332 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Dec 03 14:31:31 crc kubenswrapper[4805]: I1203 14:31:31.287437 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-cinder-public-svc" Dec 03 14:31:31 crc kubenswrapper[4805]: I1203 14:31:31.288024 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-cinder-internal-svc" Dec 03 14:31:31 crc kubenswrapper[4805]: I1203 14:31:31.293644 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-api-0"] Dec 03 14:31:31 crc kubenswrapper[4805]: I1203 14:31:31.297909 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-api-config-data" Dec 03 14:31:31 crc kubenswrapper[4805]: I1203 14:31:31.331949 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/318792ec-dc8b-4a37-92ab-d31773d66350-config" (OuterVolumeSpecName: "config") pod "318792ec-dc8b-4a37-92ab-d31773d66350" (UID: "318792ec-dc8b-4a37-92ab-d31773d66350"). InnerVolumeSpecName "config". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 14:31:31 crc kubenswrapper[4805]: I1203 14:31:31.333563 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/318792ec-dc8b-4a37-92ab-d31773d66350-ovndb-tls-certs" (OuterVolumeSpecName: "ovndb-tls-certs") pod "318792ec-dc8b-4a37-92ab-d31773d66350" (UID: "318792ec-dc8b-4a37-92ab-d31773d66350"). InnerVolumeSpecName "ovndb-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 14:31:31 crc kubenswrapper[4805]: I1203 14:31:31.372854 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7b5b9f5a-5af6-41d7-be26-3f4c091098b7-scripts\") pod \"cinder-api-0\" (UID: \"7b5b9f5a-5af6-41d7-be26-3f4c091098b7\") " pod="openstack/cinder-api-0" Dec 03 14:31:31 crc kubenswrapper[4805]: I1203 14:31:31.372894 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/7b5b9f5a-5af6-41d7-be26-3f4c091098b7-config-data-custom\") pod \"cinder-api-0\" (UID: \"7b5b9f5a-5af6-41d7-be26-3f4c091098b7\") " pod="openstack/cinder-api-0" Dec 03 14:31:31 crc kubenswrapper[4805]: I1203 14:31:31.372933 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7b5b9f5a-5af6-41d7-be26-3f4c091098b7-config-data\") pod \"cinder-api-0\" (UID: \"7b5b9f5a-5af6-41d7-be26-3f4c091098b7\") " pod="openstack/cinder-api-0" Dec 03 14:31:31 crc kubenswrapper[4805]: I1203 14:31:31.372955 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/7b5b9f5a-5af6-41d7-be26-3f4c091098b7-internal-tls-certs\") pod \"cinder-api-0\" (UID: \"7b5b9f5a-5af6-41d7-be26-3f4c091098b7\") " pod="openstack/cinder-api-0" Dec 03 14:31:31 crc kubenswrapper[4805]: I1203 14:31:31.372980 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/7b5b9f5a-5af6-41d7-be26-3f4c091098b7-public-tls-certs\") pod \"cinder-api-0\" (UID: \"7b5b9f5a-5af6-41d7-be26-3f4c091098b7\") " pod="openstack/cinder-api-0" Dec 03 14:31:31 crc kubenswrapper[4805]: I1203 14:31:31.372998 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-96rmv\" (UniqueName: \"kubernetes.io/projected/7b5b9f5a-5af6-41d7-be26-3f4c091098b7-kube-api-access-96rmv\") pod \"cinder-api-0\" (UID: \"7b5b9f5a-5af6-41d7-be26-3f4c091098b7\") " pod="openstack/cinder-api-0" Dec 03 14:31:31 crc kubenswrapper[4805]: I1203 14:31:31.373053 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/7b5b9f5a-5af6-41d7-be26-3f4c091098b7-logs\") pod \"cinder-api-0\" (UID: \"7b5b9f5a-5af6-41d7-be26-3f4c091098b7\") " pod="openstack/cinder-api-0" Dec 03 14:31:31 crc kubenswrapper[4805]: I1203 14:31:31.373077 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/7b5b9f5a-5af6-41d7-be26-3f4c091098b7-etc-machine-id\") pod \"cinder-api-0\" (UID: \"7b5b9f5a-5af6-41d7-be26-3f4c091098b7\") " pod="openstack/cinder-api-0" Dec 03 14:31:31 crc kubenswrapper[4805]: I1203 14:31:31.373094 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7b5b9f5a-5af6-41d7-be26-3f4c091098b7-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"7b5b9f5a-5af6-41d7-be26-3f4c091098b7\") " pod="openstack/cinder-api-0" Dec 03 14:31:31 crc kubenswrapper[4805]: I1203 14:31:31.373172 4805 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/secret/318792ec-dc8b-4a37-92ab-d31773d66350-config\") on node \"crc\" DevicePath \"\"" Dec 03 14:31:31 crc kubenswrapper[4805]: I1203 14:31:31.373182 4805 reconciler_common.go:293] "Volume detached for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/318792ec-dc8b-4a37-92ab-d31773d66350-ovndb-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 03 14:31:31 crc kubenswrapper[4805]: I1203 14:31:31.474495 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7b5b9f5a-5af6-41d7-be26-3f4c091098b7-scripts\") pod \"cinder-api-0\" (UID: \"7b5b9f5a-5af6-41d7-be26-3f4c091098b7\") " pod="openstack/cinder-api-0" Dec 03 14:31:31 crc kubenswrapper[4805]: I1203 14:31:31.474554 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/7b5b9f5a-5af6-41d7-be26-3f4c091098b7-config-data-custom\") pod \"cinder-api-0\" (UID: \"7b5b9f5a-5af6-41d7-be26-3f4c091098b7\") " pod="openstack/cinder-api-0" Dec 03 14:31:31 crc kubenswrapper[4805]: I1203 14:31:31.474601 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7b5b9f5a-5af6-41d7-be26-3f4c091098b7-config-data\") pod \"cinder-api-0\" (UID: \"7b5b9f5a-5af6-41d7-be26-3f4c091098b7\") " pod="openstack/cinder-api-0" Dec 03 14:31:31 crc kubenswrapper[4805]: I1203 14:31:31.474633 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/7b5b9f5a-5af6-41d7-be26-3f4c091098b7-internal-tls-certs\") pod \"cinder-api-0\" (UID: \"7b5b9f5a-5af6-41d7-be26-3f4c091098b7\") " pod="openstack/cinder-api-0" Dec 03 14:31:31 crc kubenswrapper[4805]: I1203 14:31:31.474668 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/7b5b9f5a-5af6-41d7-be26-3f4c091098b7-public-tls-certs\") pod \"cinder-api-0\" (UID: \"7b5b9f5a-5af6-41d7-be26-3f4c091098b7\") " pod="openstack/cinder-api-0" Dec 03 14:31:31 crc kubenswrapper[4805]: I1203 14:31:31.474695 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-96rmv\" (UniqueName: \"kubernetes.io/projected/7b5b9f5a-5af6-41d7-be26-3f4c091098b7-kube-api-access-96rmv\") pod \"cinder-api-0\" (UID: \"7b5b9f5a-5af6-41d7-be26-3f4c091098b7\") " pod="openstack/cinder-api-0" Dec 03 14:31:31 crc kubenswrapper[4805]: I1203 14:31:31.474763 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/7b5b9f5a-5af6-41d7-be26-3f4c091098b7-logs\") pod \"cinder-api-0\" (UID: \"7b5b9f5a-5af6-41d7-be26-3f4c091098b7\") " pod="openstack/cinder-api-0" Dec 03 14:31:31 crc kubenswrapper[4805]: I1203 14:31:31.474790 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/7b5b9f5a-5af6-41d7-be26-3f4c091098b7-etc-machine-id\") pod \"cinder-api-0\" (UID: \"7b5b9f5a-5af6-41d7-be26-3f4c091098b7\") " pod="openstack/cinder-api-0" Dec 03 14:31:31 crc kubenswrapper[4805]: I1203 14:31:31.474819 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7b5b9f5a-5af6-41d7-be26-3f4c091098b7-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"7b5b9f5a-5af6-41d7-be26-3f4c091098b7\") " pod="openstack/cinder-api-0" Dec 03 14:31:31 crc kubenswrapper[4805]: I1203 14:31:31.475764 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/7b5b9f5a-5af6-41d7-be26-3f4c091098b7-logs\") pod \"cinder-api-0\" (UID: \"7b5b9f5a-5af6-41d7-be26-3f4c091098b7\") " pod="openstack/cinder-api-0" Dec 03 14:31:31 crc kubenswrapper[4805]: I1203 14:31:31.475863 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/7b5b9f5a-5af6-41d7-be26-3f4c091098b7-etc-machine-id\") pod \"cinder-api-0\" (UID: \"7b5b9f5a-5af6-41d7-be26-3f4c091098b7\") " pod="openstack/cinder-api-0" Dec 03 14:31:31 crc kubenswrapper[4805]: I1203 14:31:31.478978 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/7b5b9f5a-5af6-41d7-be26-3f4c091098b7-internal-tls-certs\") pod \"cinder-api-0\" (UID: \"7b5b9f5a-5af6-41d7-be26-3f4c091098b7\") " pod="openstack/cinder-api-0" Dec 03 14:31:31 crc kubenswrapper[4805]: I1203 14:31:31.479100 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/7b5b9f5a-5af6-41d7-be26-3f4c091098b7-config-data-custom\") pod \"cinder-api-0\" (UID: \"7b5b9f5a-5af6-41d7-be26-3f4c091098b7\") " pod="openstack/cinder-api-0" Dec 03 14:31:31 crc kubenswrapper[4805]: I1203 14:31:31.481610 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7b5b9f5a-5af6-41d7-be26-3f4c091098b7-config-data\") pod \"cinder-api-0\" (UID: \"7b5b9f5a-5af6-41d7-be26-3f4c091098b7\") " pod="openstack/cinder-api-0" Dec 03 14:31:31 crc kubenswrapper[4805]: I1203 14:31:31.482134 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7b5b9f5a-5af6-41d7-be26-3f4c091098b7-scripts\") pod \"cinder-api-0\" (UID: \"7b5b9f5a-5af6-41d7-be26-3f4c091098b7\") " pod="openstack/cinder-api-0" Dec 03 14:31:31 crc kubenswrapper[4805]: I1203 14:31:31.482533 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7b5b9f5a-5af6-41d7-be26-3f4c091098b7-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"7b5b9f5a-5af6-41d7-be26-3f4c091098b7\") " pod="openstack/cinder-api-0" Dec 03 14:31:31 crc kubenswrapper[4805]: I1203 14:31:31.483436 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/7b5b9f5a-5af6-41d7-be26-3f4c091098b7-public-tls-certs\") pod \"cinder-api-0\" (UID: \"7b5b9f5a-5af6-41d7-be26-3f4c091098b7\") " pod="openstack/cinder-api-0" Dec 03 14:31:31 crc kubenswrapper[4805]: I1203 14:31:31.492471 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-96rmv\" (UniqueName: \"kubernetes.io/projected/7b5b9f5a-5af6-41d7-be26-3f4c091098b7-kube-api-access-96rmv\") pod \"cinder-api-0\" (UID: \"7b5b9f5a-5af6-41d7-be26-3f4c091098b7\") " pod="openstack/cinder-api-0" Dec 03 14:31:31 crc kubenswrapper[4805]: I1203 14:31:31.599374 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Dec 03 14:31:31 crc kubenswrapper[4805]: I1203 14:31:31.805688 4805 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/barbican-api-845f47d448-thqcr" podUID="37658956-d75a-43ce-81d8-38f9bca42957" containerName="barbican-api" probeResult="failure" output="Get \"http://10.217.0.159:9311/healthcheck\": read tcp 10.217.0.2:58870->10.217.0.159:9311: read: connection reset by peer" Dec 03 14:31:31 crc kubenswrapper[4805]: I1203 14:31:31.914073 4805 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/swift-proxy-78bdd85b87-5tf8n"] Dec 03 14:31:31 crc kubenswrapper[4805]: I1203 14:31:31.940636 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-proxy-78bdd85b87-5tf8n" Dec 03 14:31:31 crc kubenswrapper[4805]: I1203 14:31:31.944919 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"swift-proxy-config-data" Dec 03 14:31:31 crc kubenswrapper[4805]: I1203 14:31:31.945829 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-swift-internal-svc" Dec 03 14:31:31 crc kubenswrapper[4805]: I1203 14:31:31.946109 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-swift-public-svc" Dec 03 14:31:31 crc kubenswrapper[4805]: I1203 14:31:31.976575 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-proxy-78bdd85b87-5tf8n"] Dec 03 14:31:31 crc kubenswrapper[4805]: I1203 14:31:31.985512 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gr5zs\" (UniqueName: \"kubernetes.io/projected/0fdb9aee-e509-4e04-88b7-61fc216cfc34-kube-api-access-gr5zs\") pod \"swift-proxy-78bdd85b87-5tf8n\" (UID: \"0fdb9aee-e509-4e04-88b7-61fc216cfc34\") " pod="openstack/swift-proxy-78bdd85b87-5tf8n" Dec 03 14:31:31 crc kubenswrapper[4805]: I1203 14:31:31.986343 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/0fdb9aee-e509-4e04-88b7-61fc216cfc34-etc-swift\") pod \"swift-proxy-78bdd85b87-5tf8n\" (UID: \"0fdb9aee-e509-4e04-88b7-61fc216cfc34\") " pod="openstack/swift-proxy-78bdd85b87-5tf8n" Dec 03 14:31:31 crc kubenswrapper[4805]: I1203 14:31:31.986780 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/0fdb9aee-e509-4e04-88b7-61fc216cfc34-public-tls-certs\") pod \"swift-proxy-78bdd85b87-5tf8n\" (UID: \"0fdb9aee-e509-4e04-88b7-61fc216cfc34\") " pod="openstack/swift-proxy-78bdd85b87-5tf8n" Dec 03 14:31:31 crc kubenswrapper[4805]: I1203 14:31:31.986972 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/0fdb9aee-e509-4e04-88b7-61fc216cfc34-log-httpd\") pod \"swift-proxy-78bdd85b87-5tf8n\" (UID: \"0fdb9aee-e509-4e04-88b7-61fc216cfc34\") " pod="openstack/swift-proxy-78bdd85b87-5tf8n" Dec 03 14:31:31 crc kubenswrapper[4805]: I1203 14:31:31.987187 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/0fdb9aee-e509-4e04-88b7-61fc216cfc34-run-httpd\") pod \"swift-proxy-78bdd85b87-5tf8n\" (UID: \"0fdb9aee-e509-4e04-88b7-61fc216cfc34\") " pod="openstack/swift-proxy-78bdd85b87-5tf8n" Dec 03 14:31:31 crc kubenswrapper[4805]: I1203 14:31:31.987305 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0fdb9aee-e509-4e04-88b7-61fc216cfc34-combined-ca-bundle\") pod \"swift-proxy-78bdd85b87-5tf8n\" (UID: \"0fdb9aee-e509-4e04-88b7-61fc216cfc34\") " pod="openstack/swift-proxy-78bdd85b87-5tf8n" Dec 03 14:31:31 crc kubenswrapper[4805]: I1203 14:31:31.987736 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/0fdb9aee-e509-4e04-88b7-61fc216cfc34-internal-tls-certs\") pod \"swift-proxy-78bdd85b87-5tf8n\" (UID: \"0fdb9aee-e509-4e04-88b7-61fc216cfc34\") " pod="openstack/swift-proxy-78bdd85b87-5tf8n" Dec 03 14:31:31 crc kubenswrapper[4805]: I1203 14:31:31.988146 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0fdb9aee-e509-4e04-88b7-61fc216cfc34-config-data\") pod \"swift-proxy-78bdd85b87-5tf8n\" (UID: \"0fdb9aee-e509-4e04-88b7-61fc216cfc34\") " pod="openstack/swift-proxy-78bdd85b87-5tf8n" Dec 03 14:31:31 crc kubenswrapper[4805]: I1203 14:31:31.990638 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-75f57c664d-4tsg2" event={"ID":"318792ec-dc8b-4a37-92ab-d31773d66350","Type":"ContainerDied","Data":"3c8b8134864ece92d0ee18d4cf46deec6ae2c8b4b4887927a2459f0ddac12a07"} Dec 03 14:31:31 crc kubenswrapper[4805]: I1203 14:31:31.990693 4805 scope.go:117] "RemoveContainer" containerID="76ff7768f1c3210b2f1ffab05aefd717027127581dc36328eff11ac6503fdb3f" Dec 03 14:31:31 crc kubenswrapper[4805]: I1203 14:31:31.990831 4805 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-75f57c664d-4tsg2" Dec 03 14:31:32 crc kubenswrapper[4805]: I1203 14:31:32.000095 4805 generic.go:334] "Generic (PLEG): container finished" podID="37658956-d75a-43ce-81d8-38f9bca42957" containerID="7adb8b3c5e068f3ee12e9581aecb5bdfa43066e3e1f0991bc28977ea2af801da" exitCode=0 Dec 03 14:31:32 crc kubenswrapper[4805]: I1203 14:31:32.000177 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-845f47d448-thqcr" event={"ID":"37658956-d75a-43ce-81d8-38f9bca42957","Type":"ContainerDied","Data":"7adb8b3c5e068f3ee12e9581aecb5bdfa43066e3e1f0991bc28977ea2af801da"} Dec 03 14:31:32 crc kubenswrapper[4805]: I1203 14:31:32.038499 4805 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-75f57c664d-4tsg2"] Dec 03 14:31:32 crc kubenswrapper[4805]: I1203 14:31:32.049670 4805 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-75f57c664d-4tsg2"] Dec 03 14:31:32 crc kubenswrapper[4805]: I1203 14:31:32.081304 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-api-0"] Dec 03 14:31:32 crc kubenswrapper[4805]: I1203 14:31:32.091804 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0fdb9aee-e509-4e04-88b7-61fc216cfc34-combined-ca-bundle\") pod \"swift-proxy-78bdd85b87-5tf8n\" (UID: \"0fdb9aee-e509-4e04-88b7-61fc216cfc34\") " pod="openstack/swift-proxy-78bdd85b87-5tf8n" Dec 03 14:31:32 crc kubenswrapper[4805]: I1203 14:31:32.091915 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/0fdb9aee-e509-4e04-88b7-61fc216cfc34-internal-tls-certs\") pod \"swift-proxy-78bdd85b87-5tf8n\" (UID: \"0fdb9aee-e509-4e04-88b7-61fc216cfc34\") " pod="openstack/swift-proxy-78bdd85b87-5tf8n" Dec 03 14:31:32 crc kubenswrapper[4805]: I1203 14:31:32.091954 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0fdb9aee-e509-4e04-88b7-61fc216cfc34-config-data\") pod \"swift-proxy-78bdd85b87-5tf8n\" (UID: \"0fdb9aee-e509-4e04-88b7-61fc216cfc34\") " pod="openstack/swift-proxy-78bdd85b87-5tf8n" Dec 03 14:31:32 crc kubenswrapper[4805]: I1203 14:31:32.092022 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gr5zs\" (UniqueName: \"kubernetes.io/projected/0fdb9aee-e509-4e04-88b7-61fc216cfc34-kube-api-access-gr5zs\") pod \"swift-proxy-78bdd85b87-5tf8n\" (UID: \"0fdb9aee-e509-4e04-88b7-61fc216cfc34\") " pod="openstack/swift-proxy-78bdd85b87-5tf8n" Dec 03 14:31:32 crc kubenswrapper[4805]: I1203 14:31:32.092066 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/0fdb9aee-e509-4e04-88b7-61fc216cfc34-etc-swift\") pod \"swift-proxy-78bdd85b87-5tf8n\" (UID: \"0fdb9aee-e509-4e04-88b7-61fc216cfc34\") " pod="openstack/swift-proxy-78bdd85b87-5tf8n" Dec 03 14:31:32 crc kubenswrapper[4805]: I1203 14:31:32.092089 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/0fdb9aee-e509-4e04-88b7-61fc216cfc34-public-tls-certs\") pod \"swift-proxy-78bdd85b87-5tf8n\" (UID: \"0fdb9aee-e509-4e04-88b7-61fc216cfc34\") " pod="openstack/swift-proxy-78bdd85b87-5tf8n" Dec 03 14:31:32 crc kubenswrapper[4805]: I1203 14:31:32.092114 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/0fdb9aee-e509-4e04-88b7-61fc216cfc34-log-httpd\") pod \"swift-proxy-78bdd85b87-5tf8n\" (UID: \"0fdb9aee-e509-4e04-88b7-61fc216cfc34\") " pod="openstack/swift-proxy-78bdd85b87-5tf8n" Dec 03 14:31:32 crc kubenswrapper[4805]: I1203 14:31:32.092231 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/0fdb9aee-e509-4e04-88b7-61fc216cfc34-run-httpd\") pod \"swift-proxy-78bdd85b87-5tf8n\" (UID: \"0fdb9aee-e509-4e04-88b7-61fc216cfc34\") " pod="openstack/swift-proxy-78bdd85b87-5tf8n" Dec 03 14:31:32 crc kubenswrapper[4805]: I1203 14:31:32.092800 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/0fdb9aee-e509-4e04-88b7-61fc216cfc34-run-httpd\") pod \"swift-proxy-78bdd85b87-5tf8n\" (UID: \"0fdb9aee-e509-4e04-88b7-61fc216cfc34\") " pod="openstack/swift-proxy-78bdd85b87-5tf8n" Dec 03 14:31:32 crc kubenswrapper[4805]: I1203 14:31:32.096186 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/0fdb9aee-e509-4e04-88b7-61fc216cfc34-log-httpd\") pod \"swift-proxy-78bdd85b87-5tf8n\" (UID: \"0fdb9aee-e509-4e04-88b7-61fc216cfc34\") " pod="openstack/swift-proxy-78bdd85b87-5tf8n" Dec 03 14:31:32 crc kubenswrapper[4805]: I1203 14:31:32.107415 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/0fdb9aee-e509-4e04-88b7-61fc216cfc34-internal-tls-certs\") pod \"swift-proxy-78bdd85b87-5tf8n\" (UID: \"0fdb9aee-e509-4e04-88b7-61fc216cfc34\") " pod="openstack/swift-proxy-78bdd85b87-5tf8n" Dec 03 14:31:32 crc kubenswrapper[4805]: I1203 14:31:32.107927 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/0fdb9aee-e509-4e04-88b7-61fc216cfc34-public-tls-certs\") pod \"swift-proxy-78bdd85b87-5tf8n\" (UID: \"0fdb9aee-e509-4e04-88b7-61fc216cfc34\") " pod="openstack/swift-proxy-78bdd85b87-5tf8n" Dec 03 14:31:32 crc kubenswrapper[4805]: I1203 14:31:32.108423 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0fdb9aee-e509-4e04-88b7-61fc216cfc34-config-data\") pod \"swift-proxy-78bdd85b87-5tf8n\" (UID: \"0fdb9aee-e509-4e04-88b7-61fc216cfc34\") " pod="openstack/swift-proxy-78bdd85b87-5tf8n" Dec 03 14:31:32 crc kubenswrapper[4805]: I1203 14:31:32.109061 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/0fdb9aee-e509-4e04-88b7-61fc216cfc34-etc-swift\") pod \"swift-proxy-78bdd85b87-5tf8n\" (UID: \"0fdb9aee-e509-4e04-88b7-61fc216cfc34\") " pod="openstack/swift-proxy-78bdd85b87-5tf8n" Dec 03 14:31:32 crc kubenswrapper[4805]: I1203 14:31:32.109222 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0fdb9aee-e509-4e04-88b7-61fc216cfc34-combined-ca-bundle\") pod \"swift-proxy-78bdd85b87-5tf8n\" (UID: \"0fdb9aee-e509-4e04-88b7-61fc216cfc34\") " pod="openstack/swift-proxy-78bdd85b87-5tf8n" Dec 03 14:31:32 crc kubenswrapper[4805]: I1203 14:31:32.113370 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gr5zs\" (UniqueName: \"kubernetes.io/projected/0fdb9aee-e509-4e04-88b7-61fc216cfc34-kube-api-access-gr5zs\") pod \"swift-proxy-78bdd85b87-5tf8n\" (UID: \"0fdb9aee-e509-4e04-88b7-61fc216cfc34\") " pod="openstack/swift-proxy-78bdd85b87-5tf8n" Dec 03 14:31:32 crc kubenswrapper[4805]: E1203 14:31:32.266507 4805 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod318792ec_dc8b_4a37_92ab_d31773d66350.slice/crio-3c8b8134864ece92d0ee18d4cf46deec6ae2c8b4b4887927a2459f0ddac12a07\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod318792ec_dc8b_4a37_92ab_d31773d66350.slice\": RecentStats: unable to find data in memory cache]" Dec 03 14:31:32 crc kubenswrapper[4805]: I1203 14:31:32.295010 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-proxy-78bdd85b87-5tf8n" Dec 03 14:31:32 crc kubenswrapper[4805]: I1203 14:31:32.710184 4805 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="318792ec-dc8b-4a37-92ab-d31773d66350" path="/var/lib/kubelet/pods/318792ec-dc8b-4a37-92ab-d31773d66350/volumes" Dec 03 14:31:32 crc kubenswrapper[4805]: I1203 14:31:32.711158 4805 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="afc5cad3-750c-429e-be48-6355bc07ac6a" path="/var/lib/kubelet/pods/afc5cad3-750c-429e-be48-6355bc07ac6a/volumes" Dec 03 14:31:33 crc kubenswrapper[4805]: I1203 14:31:33.176350 4805 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/heat-engine-5fdfc8886c-v7jx4"] Dec 03 14:31:33 crc kubenswrapper[4805]: I1203 14:31:33.177489 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-engine-5fdfc8886c-v7jx4" Dec 03 14:31:33 crc kubenswrapper[4805]: I1203 14:31:33.188201 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"heat-engine-config-data" Dec 03 14:31:33 crc kubenswrapper[4805]: I1203 14:31:33.188600 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"heat-config-data" Dec 03 14:31:33 crc kubenswrapper[4805]: I1203 14:31:33.188765 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"heat-heat-dockercfg-2wxds" Dec 03 14:31:33 crc kubenswrapper[4805]: I1203 14:31:33.210895 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-engine-5fdfc8886c-v7jx4"] Dec 03 14:31:33 crc kubenswrapper[4805]: I1203 14:31:33.211901 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/36af56b0-4335-48cf-82ba-bd95a50b14f3-combined-ca-bundle\") pod \"heat-engine-5fdfc8886c-v7jx4\" (UID: \"36af56b0-4335-48cf-82ba-bd95a50b14f3\") " pod="openstack/heat-engine-5fdfc8886c-v7jx4" Dec 03 14:31:33 crc kubenswrapper[4805]: I1203 14:31:33.211956 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/36af56b0-4335-48cf-82ba-bd95a50b14f3-config-data\") pod \"heat-engine-5fdfc8886c-v7jx4\" (UID: \"36af56b0-4335-48cf-82ba-bd95a50b14f3\") " pod="openstack/heat-engine-5fdfc8886c-v7jx4" Dec 03 14:31:33 crc kubenswrapper[4805]: I1203 14:31:33.211992 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sb4qs\" (UniqueName: \"kubernetes.io/projected/36af56b0-4335-48cf-82ba-bd95a50b14f3-kube-api-access-sb4qs\") pod \"heat-engine-5fdfc8886c-v7jx4\" (UID: \"36af56b0-4335-48cf-82ba-bd95a50b14f3\") " pod="openstack/heat-engine-5fdfc8886c-v7jx4" Dec 03 14:31:33 crc kubenswrapper[4805]: I1203 14:31:33.212026 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/36af56b0-4335-48cf-82ba-bd95a50b14f3-config-data-custom\") pod \"heat-engine-5fdfc8886c-v7jx4\" (UID: \"36af56b0-4335-48cf-82ba-bd95a50b14f3\") " pod="openstack/heat-engine-5fdfc8886c-v7jx4" Dec 03 14:31:33 crc kubenswrapper[4805]: I1203 14:31:33.231359 4805 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/barbican-api-845f47d448-thqcr" podUID="37658956-d75a-43ce-81d8-38f9bca42957" containerName="barbican-api-log" probeResult="failure" output="Get \"http://10.217.0.159:9311/healthcheck\": dial tcp 10.217.0.159:9311: connect: connection refused" Dec 03 14:31:33 crc kubenswrapper[4805]: I1203 14:31:33.231705 4805 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/barbican-api-845f47d448-thqcr" podUID="37658956-d75a-43ce-81d8-38f9bca42957" containerName="barbican-api" probeResult="failure" output="Get \"http://10.217.0.159:9311/healthcheck\": dial tcp 10.217.0.159:9311: connect: connection refused" Dec 03 14:31:33 crc kubenswrapper[4805]: I1203 14:31:33.231775 4805 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/barbican-api-845f47d448-thqcr" Dec 03 14:31:33 crc kubenswrapper[4805]: I1203 14:31:33.300184 4805 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/heat-cfnapi-795fdbdfbf-8zpm7"] Dec 03 14:31:33 crc kubenswrapper[4805]: I1203 14:31:33.301409 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-cfnapi-795fdbdfbf-8zpm7" Dec 03 14:31:33 crc kubenswrapper[4805]: I1203 14:31:33.310800 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"heat-cfnapi-config-data" Dec 03 14:31:33 crc kubenswrapper[4805]: I1203 14:31:33.313213 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/33d42113-07b3-4db3-95b8-f43810991ce4-config-data-custom\") pod \"heat-cfnapi-795fdbdfbf-8zpm7\" (UID: \"33d42113-07b3-4db3-95b8-f43810991ce4\") " pod="openstack/heat-cfnapi-795fdbdfbf-8zpm7" Dec 03 14:31:33 crc kubenswrapper[4805]: I1203 14:31:33.313269 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/36af56b0-4335-48cf-82ba-bd95a50b14f3-combined-ca-bundle\") pod \"heat-engine-5fdfc8886c-v7jx4\" (UID: \"36af56b0-4335-48cf-82ba-bd95a50b14f3\") " pod="openstack/heat-engine-5fdfc8886c-v7jx4" Dec 03 14:31:33 crc kubenswrapper[4805]: I1203 14:31:33.313303 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4k7rf\" (UniqueName: \"kubernetes.io/projected/33d42113-07b3-4db3-95b8-f43810991ce4-kube-api-access-4k7rf\") pod \"heat-cfnapi-795fdbdfbf-8zpm7\" (UID: \"33d42113-07b3-4db3-95b8-f43810991ce4\") " pod="openstack/heat-cfnapi-795fdbdfbf-8zpm7" Dec 03 14:31:33 crc kubenswrapper[4805]: I1203 14:31:33.313884 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/36af56b0-4335-48cf-82ba-bd95a50b14f3-config-data\") pod \"heat-engine-5fdfc8886c-v7jx4\" (UID: \"36af56b0-4335-48cf-82ba-bd95a50b14f3\") " pod="openstack/heat-engine-5fdfc8886c-v7jx4" Dec 03 14:31:33 crc kubenswrapper[4805]: I1203 14:31:33.313941 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sb4qs\" (UniqueName: \"kubernetes.io/projected/36af56b0-4335-48cf-82ba-bd95a50b14f3-kube-api-access-sb4qs\") pod \"heat-engine-5fdfc8886c-v7jx4\" (UID: \"36af56b0-4335-48cf-82ba-bd95a50b14f3\") " pod="openstack/heat-engine-5fdfc8886c-v7jx4" Dec 03 14:31:33 crc kubenswrapper[4805]: I1203 14:31:33.317897 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/36af56b0-4335-48cf-82ba-bd95a50b14f3-config-data-custom\") pod \"heat-engine-5fdfc8886c-v7jx4\" (UID: \"36af56b0-4335-48cf-82ba-bd95a50b14f3\") " pod="openstack/heat-engine-5fdfc8886c-v7jx4" Dec 03 14:31:33 crc kubenswrapper[4805]: I1203 14:31:33.318005 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/33d42113-07b3-4db3-95b8-f43810991ce4-combined-ca-bundle\") pod \"heat-cfnapi-795fdbdfbf-8zpm7\" (UID: \"33d42113-07b3-4db3-95b8-f43810991ce4\") " pod="openstack/heat-cfnapi-795fdbdfbf-8zpm7" Dec 03 14:31:33 crc kubenswrapper[4805]: I1203 14:31:33.318061 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/33d42113-07b3-4db3-95b8-f43810991ce4-config-data\") pod \"heat-cfnapi-795fdbdfbf-8zpm7\" (UID: \"33d42113-07b3-4db3-95b8-f43810991ce4\") " pod="openstack/heat-cfnapi-795fdbdfbf-8zpm7" Dec 03 14:31:33 crc kubenswrapper[4805]: I1203 14:31:33.324611 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/36af56b0-4335-48cf-82ba-bd95a50b14f3-combined-ca-bundle\") pod \"heat-engine-5fdfc8886c-v7jx4\" (UID: \"36af56b0-4335-48cf-82ba-bd95a50b14f3\") " pod="openstack/heat-engine-5fdfc8886c-v7jx4" Dec 03 14:31:33 crc kubenswrapper[4805]: I1203 14:31:33.326174 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/36af56b0-4335-48cf-82ba-bd95a50b14f3-config-data\") pod \"heat-engine-5fdfc8886c-v7jx4\" (UID: \"36af56b0-4335-48cf-82ba-bd95a50b14f3\") " pod="openstack/heat-engine-5fdfc8886c-v7jx4" Dec 03 14:31:33 crc kubenswrapper[4805]: I1203 14:31:33.332873 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-cfnapi-795fdbdfbf-8zpm7"] Dec 03 14:31:33 crc kubenswrapper[4805]: I1203 14:31:33.347899 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/36af56b0-4335-48cf-82ba-bd95a50b14f3-config-data-custom\") pod \"heat-engine-5fdfc8886c-v7jx4\" (UID: \"36af56b0-4335-48cf-82ba-bd95a50b14f3\") " pod="openstack/heat-engine-5fdfc8886c-v7jx4" Dec 03 14:31:33 crc kubenswrapper[4805]: I1203 14:31:33.361317 4805 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5784cf869f-l4bt9"] Dec 03 14:31:33 crc kubenswrapper[4805]: I1203 14:31:33.361594 4805 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-5784cf869f-l4bt9" podUID="bc7fd7a7-02bc-48e1-9bde-d1680876da37" containerName="dnsmasq-dns" containerID="cri-o://d957799f184adf285454f33a4f7aacfb2a9823b898bb6d66acdb0d27983681d6" gracePeriod=10 Dec 03 14:31:33 crc kubenswrapper[4805]: I1203 14:31:33.371043 4805 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-5784cf869f-l4bt9" Dec 03 14:31:33 crc kubenswrapper[4805]: I1203 14:31:33.385698 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sb4qs\" (UniqueName: \"kubernetes.io/projected/36af56b0-4335-48cf-82ba-bd95a50b14f3-kube-api-access-sb4qs\") pod \"heat-engine-5fdfc8886c-v7jx4\" (UID: \"36af56b0-4335-48cf-82ba-bd95a50b14f3\") " pod="openstack/heat-engine-5fdfc8886c-v7jx4" Dec 03 14:31:33 crc kubenswrapper[4805]: I1203 14:31:33.390537 4805 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/heat-api-6bd4dc884b-6qst5"] Dec 03 14:31:33 crc kubenswrapper[4805]: I1203 14:31:33.391878 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-api-6bd4dc884b-6qst5" Dec 03 14:31:33 crc kubenswrapper[4805]: I1203 14:31:33.400260 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"heat-api-config-data" Dec 03 14:31:33 crc kubenswrapper[4805]: I1203 14:31:33.423342 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/33d42113-07b3-4db3-95b8-f43810991ce4-combined-ca-bundle\") pod \"heat-cfnapi-795fdbdfbf-8zpm7\" (UID: \"33d42113-07b3-4db3-95b8-f43810991ce4\") " pod="openstack/heat-cfnapi-795fdbdfbf-8zpm7" Dec 03 14:31:33 crc kubenswrapper[4805]: I1203 14:31:33.423390 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/33d42113-07b3-4db3-95b8-f43810991ce4-config-data\") pod \"heat-cfnapi-795fdbdfbf-8zpm7\" (UID: \"33d42113-07b3-4db3-95b8-f43810991ce4\") " pod="openstack/heat-cfnapi-795fdbdfbf-8zpm7" Dec 03 14:31:33 crc kubenswrapper[4805]: I1203 14:31:33.423433 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rd575\" (UniqueName: \"kubernetes.io/projected/c4431968-a92e-4fc0-951f-f42a15942f33-kube-api-access-rd575\") pod \"heat-api-6bd4dc884b-6qst5\" (UID: \"c4431968-a92e-4fc0-951f-f42a15942f33\") " pod="openstack/heat-api-6bd4dc884b-6qst5" Dec 03 14:31:33 crc kubenswrapper[4805]: I1203 14:31:33.423486 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c4431968-a92e-4fc0-951f-f42a15942f33-combined-ca-bundle\") pod \"heat-api-6bd4dc884b-6qst5\" (UID: \"c4431968-a92e-4fc0-951f-f42a15942f33\") " pod="openstack/heat-api-6bd4dc884b-6qst5" Dec 03 14:31:33 crc kubenswrapper[4805]: I1203 14:31:33.423516 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/33d42113-07b3-4db3-95b8-f43810991ce4-config-data-custom\") pod \"heat-cfnapi-795fdbdfbf-8zpm7\" (UID: \"33d42113-07b3-4db3-95b8-f43810991ce4\") " pod="openstack/heat-cfnapi-795fdbdfbf-8zpm7" Dec 03 14:31:33 crc kubenswrapper[4805]: I1203 14:31:33.423535 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c4431968-a92e-4fc0-951f-f42a15942f33-config-data\") pod \"heat-api-6bd4dc884b-6qst5\" (UID: \"c4431968-a92e-4fc0-951f-f42a15942f33\") " pod="openstack/heat-api-6bd4dc884b-6qst5" Dec 03 14:31:33 crc kubenswrapper[4805]: I1203 14:31:33.423552 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4k7rf\" (UniqueName: \"kubernetes.io/projected/33d42113-07b3-4db3-95b8-f43810991ce4-kube-api-access-4k7rf\") pod \"heat-cfnapi-795fdbdfbf-8zpm7\" (UID: \"33d42113-07b3-4db3-95b8-f43810991ce4\") " pod="openstack/heat-cfnapi-795fdbdfbf-8zpm7" Dec 03 14:31:33 crc kubenswrapper[4805]: I1203 14:31:33.423590 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/c4431968-a92e-4fc0-951f-f42a15942f33-config-data-custom\") pod \"heat-api-6bd4dc884b-6qst5\" (UID: \"c4431968-a92e-4fc0-951f-f42a15942f33\") " pod="openstack/heat-api-6bd4dc884b-6qst5" Dec 03 14:31:33 crc kubenswrapper[4805]: I1203 14:31:33.428660 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/33d42113-07b3-4db3-95b8-f43810991ce4-config-data\") pod \"heat-cfnapi-795fdbdfbf-8zpm7\" (UID: \"33d42113-07b3-4db3-95b8-f43810991ce4\") " pod="openstack/heat-cfnapi-795fdbdfbf-8zpm7" Dec 03 14:31:33 crc kubenswrapper[4805]: I1203 14:31:33.431620 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/33d42113-07b3-4db3-95b8-f43810991ce4-combined-ca-bundle\") pod \"heat-cfnapi-795fdbdfbf-8zpm7\" (UID: \"33d42113-07b3-4db3-95b8-f43810991ce4\") " pod="openstack/heat-cfnapi-795fdbdfbf-8zpm7" Dec 03 14:31:33 crc kubenswrapper[4805]: I1203 14:31:33.434767 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-api-6bd4dc884b-6qst5"] Dec 03 14:31:33 crc kubenswrapper[4805]: I1203 14:31:33.449745 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/33d42113-07b3-4db3-95b8-f43810991ce4-config-data-custom\") pod \"heat-cfnapi-795fdbdfbf-8zpm7\" (UID: \"33d42113-07b3-4db3-95b8-f43810991ce4\") " pod="openstack/heat-cfnapi-795fdbdfbf-8zpm7" Dec 03 14:31:33 crc kubenswrapper[4805]: I1203 14:31:33.452651 4805 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-f6bc4c6c9-x5tdg"] Dec 03 14:31:33 crc kubenswrapper[4805]: I1203 14:31:33.460789 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4k7rf\" (UniqueName: \"kubernetes.io/projected/33d42113-07b3-4db3-95b8-f43810991ce4-kube-api-access-4k7rf\") pod \"heat-cfnapi-795fdbdfbf-8zpm7\" (UID: \"33d42113-07b3-4db3-95b8-f43810991ce4\") " pod="openstack/heat-cfnapi-795fdbdfbf-8zpm7" Dec 03 14:31:33 crc kubenswrapper[4805]: I1203 14:31:33.462236 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-f6bc4c6c9-x5tdg" Dec 03 14:31:33 crc kubenswrapper[4805]: I1203 14:31:33.473316 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-f6bc4c6c9-x5tdg"] Dec 03 14:31:33 crc kubenswrapper[4805]: I1203 14:31:33.520669 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-engine-5fdfc8886c-v7jx4" Dec 03 14:31:33 crc kubenswrapper[4805]: I1203 14:31:33.527062 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/4994372a-ac42-4661-aed6-734070bee008-ovsdbserver-nb\") pod \"dnsmasq-dns-f6bc4c6c9-x5tdg\" (UID: \"4994372a-ac42-4661-aed6-734070bee008\") " pod="openstack/dnsmasq-dns-f6bc4c6c9-x5tdg" Dec 03 14:31:33 crc kubenswrapper[4805]: I1203 14:31:33.527138 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4994372a-ac42-4661-aed6-734070bee008-config\") pod \"dnsmasq-dns-f6bc4c6c9-x5tdg\" (UID: \"4994372a-ac42-4661-aed6-734070bee008\") " pod="openstack/dnsmasq-dns-f6bc4c6c9-x5tdg" Dec 03 14:31:33 crc kubenswrapper[4805]: I1203 14:31:33.527225 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rd575\" (UniqueName: \"kubernetes.io/projected/c4431968-a92e-4fc0-951f-f42a15942f33-kube-api-access-rd575\") pod \"heat-api-6bd4dc884b-6qst5\" (UID: \"c4431968-a92e-4fc0-951f-f42a15942f33\") " pod="openstack/heat-api-6bd4dc884b-6qst5" Dec 03 14:31:33 crc kubenswrapper[4805]: I1203 14:31:33.527297 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/4994372a-ac42-4661-aed6-734070bee008-dns-svc\") pod \"dnsmasq-dns-f6bc4c6c9-x5tdg\" (UID: \"4994372a-ac42-4661-aed6-734070bee008\") " pod="openstack/dnsmasq-dns-f6bc4c6c9-x5tdg" Dec 03 14:31:33 crc kubenswrapper[4805]: I1203 14:31:33.527333 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/4994372a-ac42-4661-aed6-734070bee008-dns-swift-storage-0\") pod \"dnsmasq-dns-f6bc4c6c9-x5tdg\" (UID: \"4994372a-ac42-4661-aed6-734070bee008\") " pod="openstack/dnsmasq-dns-f6bc4c6c9-x5tdg" Dec 03 14:31:33 crc kubenswrapper[4805]: I1203 14:31:33.527373 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/4994372a-ac42-4661-aed6-734070bee008-ovsdbserver-sb\") pod \"dnsmasq-dns-f6bc4c6c9-x5tdg\" (UID: \"4994372a-ac42-4661-aed6-734070bee008\") " pod="openstack/dnsmasq-dns-f6bc4c6c9-x5tdg" Dec 03 14:31:33 crc kubenswrapper[4805]: I1203 14:31:33.527414 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c4431968-a92e-4fc0-951f-f42a15942f33-combined-ca-bundle\") pod \"heat-api-6bd4dc884b-6qst5\" (UID: \"c4431968-a92e-4fc0-951f-f42a15942f33\") " pod="openstack/heat-api-6bd4dc884b-6qst5" Dec 03 14:31:33 crc kubenswrapper[4805]: I1203 14:31:33.527513 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4dfjm\" (UniqueName: \"kubernetes.io/projected/4994372a-ac42-4661-aed6-734070bee008-kube-api-access-4dfjm\") pod \"dnsmasq-dns-f6bc4c6c9-x5tdg\" (UID: \"4994372a-ac42-4661-aed6-734070bee008\") " pod="openstack/dnsmasq-dns-f6bc4c6c9-x5tdg" Dec 03 14:31:33 crc kubenswrapper[4805]: I1203 14:31:33.527555 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c4431968-a92e-4fc0-951f-f42a15942f33-config-data\") pod \"heat-api-6bd4dc884b-6qst5\" (UID: \"c4431968-a92e-4fc0-951f-f42a15942f33\") " pod="openstack/heat-api-6bd4dc884b-6qst5" Dec 03 14:31:33 crc kubenswrapper[4805]: I1203 14:31:33.529409 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/c4431968-a92e-4fc0-951f-f42a15942f33-config-data-custom\") pod \"heat-api-6bd4dc884b-6qst5\" (UID: \"c4431968-a92e-4fc0-951f-f42a15942f33\") " pod="openstack/heat-api-6bd4dc884b-6qst5" Dec 03 14:31:33 crc kubenswrapper[4805]: I1203 14:31:33.538019 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/c4431968-a92e-4fc0-951f-f42a15942f33-config-data-custom\") pod \"heat-api-6bd4dc884b-6qst5\" (UID: \"c4431968-a92e-4fc0-951f-f42a15942f33\") " pod="openstack/heat-api-6bd4dc884b-6qst5" Dec 03 14:31:33 crc kubenswrapper[4805]: I1203 14:31:33.541591 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c4431968-a92e-4fc0-951f-f42a15942f33-config-data\") pod \"heat-api-6bd4dc884b-6qst5\" (UID: \"c4431968-a92e-4fc0-951f-f42a15942f33\") " pod="openstack/heat-api-6bd4dc884b-6qst5" Dec 03 14:31:33 crc kubenswrapper[4805]: I1203 14:31:33.545479 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c4431968-a92e-4fc0-951f-f42a15942f33-combined-ca-bundle\") pod \"heat-api-6bd4dc884b-6qst5\" (UID: \"c4431968-a92e-4fc0-951f-f42a15942f33\") " pod="openstack/heat-api-6bd4dc884b-6qst5" Dec 03 14:31:33 crc kubenswrapper[4805]: I1203 14:31:33.552870 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rd575\" (UniqueName: \"kubernetes.io/projected/c4431968-a92e-4fc0-951f-f42a15942f33-kube-api-access-rd575\") pod \"heat-api-6bd4dc884b-6qst5\" (UID: \"c4431968-a92e-4fc0-951f-f42a15942f33\") " pod="openstack/heat-api-6bd4dc884b-6qst5" Dec 03 14:31:33 crc kubenswrapper[4805]: I1203 14:31:33.602487 4805 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 03 14:31:33 crc kubenswrapper[4805]: I1203 14:31:33.603640 4805 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="58c98434-0b4d-4c2d-ac35-109e38f96803" containerName="ceilometer-central-agent" containerID="cri-o://f98e610e128d27a8cf1f7ca39fe6a3f759d8784ae234f8e1a4c4d5e228950639" gracePeriod=30 Dec 03 14:31:33 crc kubenswrapper[4805]: I1203 14:31:33.603782 4805 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="58c98434-0b4d-4c2d-ac35-109e38f96803" containerName="proxy-httpd" containerID="cri-o://126ffcd2cdaa24e89e915efb3598e7f49248d24b6884776c5b6848a1429ab5c0" gracePeriod=30 Dec 03 14:31:33 crc kubenswrapper[4805]: I1203 14:31:33.603828 4805 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="58c98434-0b4d-4c2d-ac35-109e38f96803" containerName="sg-core" containerID="cri-o://20bba63c0e20fac3641cd24a4df07f94b5cb20bb86b1e7fd6adcb5e5071e51b0" gracePeriod=30 Dec 03 14:31:33 crc kubenswrapper[4805]: I1203 14:31:33.603876 4805 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="58c98434-0b4d-4c2d-ac35-109e38f96803" containerName="ceilometer-notification-agent" containerID="cri-o://9cf635a2329565cc994a3a2337944c2eca2f7047534f27d7eeb5f3ceaa96458b" gracePeriod=30 Dec 03 14:31:33 crc kubenswrapper[4805]: I1203 14:31:33.631810 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/4994372a-ac42-4661-aed6-734070bee008-ovsdbserver-nb\") pod \"dnsmasq-dns-f6bc4c6c9-x5tdg\" (UID: \"4994372a-ac42-4661-aed6-734070bee008\") " pod="openstack/dnsmasq-dns-f6bc4c6c9-x5tdg" Dec 03 14:31:33 crc kubenswrapper[4805]: I1203 14:31:33.631966 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4994372a-ac42-4661-aed6-734070bee008-config\") pod \"dnsmasq-dns-f6bc4c6c9-x5tdg\" (UID: \"4994372a-ac42-4661-aed6-734070bee008\") " pod="openstack/dnsmasq-dns-f6bc4c6c9-x5tdg" Dec 03 14:31:33 crc kubenswrapper[4805]: I1203 14:31:33.632040 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/4994372a-ac42-4661-aed6-734070bee008-dns-svc\") pod \"dnsmasq-dns-f6bc4c6c9-x5tdg\" (UID: \"4994372a-ac42-4661-aed6-734070bee008\") " pod="openstack/dnsmasq-dns-f6bc4c6c9-x5tdg" Dec 03 14:31:33 crc kubenswrapper[4805]: I1203 14:31:33.632059 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/4994372a-ac42-4661-aed6-734070bee008-dns-swift-storage-0\") pod \"dnsmasq-dns-f6bc4c6c9-x5tdg\" (UID: \"4994372a-ac42-4661-aed6-734070bee008\") " pod="openstack/dnsmasq-dns-f6bc4c6c9-x5tdg" Dec 03 14:31:33 crc kubenswrapper[4805]: I1203 14:31:33.632155 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/4994372a-ac42-4661-aed6-734070bee008-ovsdbserver-sb\") pod \"dnsmasq-dns-f6bc4c6c9-x5tdg\" (UID: \"4994372a-ac42-4661-aed6-734070bee008\") " pod="openstack/dnsmasq-dns-f6bc4c6c9-x5tdg" Dec 03 14:31:33 crc kubenswrapper[4805]: I1203 14:31:33.632219 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4dfjm\" (UniqueName: \"kubernetes.io/projected/4994372a-ac42-4661-aed6-734070bee008-kube-api-access-4dfjm\") pod \"dnsmasq-dns-f6bc4c6c9-x5tdg\" (UID: \"4994372a-ac42-4661-aed6-734070bee008\") " pod="openstack/dnsmasq-dns-f6bc4c6c9-x5tdg" Dec 03 14:31:33 crc kubenswrapper[4805]: I1203 14:31:33.634477 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/4994372a-ac42-4661-aed6-734070bee008-dns-svc\") pod \"dnsmasq-dns-f6bc4c6c9-x5tdg\" (UID: \"4994372a-ac42-4661-aed6-734070bee008\") " pod="openstack/dnsmasq-dns-f6bc4c6c9-x5tdg" Dec 03 14:31:33 crc kubenswrapper[4805]: I1203 14:31:33.638043 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-cfnapi-795fdbdfbf-8zpm7" Dec 03 14:31:33 crc kubenswrapper[4805]: I1203 14:31:33.639380 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4994372a-ac42-4661-aed6-734070bee008-config\") pod \"dnsmasq-dns-f6bc4c6c9-x5tdg\" (UID: \"4994372a-ac42-4661-aed6-734070bee008\") " pod="openstack/dnsmasq-dns-f6bc4c6c9-x5tdg" Dec 03 14:31:33 crc kubenswrapper[4805]: I1203 14:31:33.639743 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/4994372a-ac42-4661-aed6-734070bee008-dns-swift-storage-0\") pod \"dnsmasq-dns-f6bc4c6c9-x5tdg\" (UID: \"4994372a-ac42-4661-aed6-734070bee008\") " pod="openstack/dnsmasq-dns-f6bc4c6c9-x5tdg" Dec 03 14:31:33 crc kubenswrapper[4805]: I1203 14:31:33.640044 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/4994372a-ac42-4661-aed6-734070bee008-ovsdbserver-nb\") pod \"dnsmasq-dns-f6bc4c6c9-x5tdg\" (UID: \"4994372a-ac42-4661-aed6-734070bee008\") " pod="openstack/dnsmasq-dns-f6bc4c6c9-x5tdg" Dec 03 14:31:33 crc kubenswrapper[4805]: I1203 14:31:33.640547 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/4994372a-ac42-4661-aed6-734070bee008-ovsdbserver-sb\") pod \"dnsmasq-dns-f6bc4c6c9-x5tdg\" (UID: \"4994372a-ac42-4661-aed6-734070bee008\") " pod="openstack/dnsmasq-dns-f6bc4c6c9-x5tdg" Dec 03 14:31:33 crc kubenswrapper[4805]: I1203 14:31:33.650245 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-api-6bd4dc884b-6qst5" Dec 03 14:31:33 crc kubenswrapper[4805]: I1203 14:31:33.656116 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4dfjm\" (UniqueName: \"kubernetes.io/projected/4994372a-ac42-4661-aed6-734070bee008-kube-api-access-4dfjm\") pod \"dnsmasq-dns-f6bc4c6c9-x5tdg\" (UID: \"4994372a-ac42-4661-aed6-734070bee008\") " pod="openstack/dnsmasq-dns-f6bc4c6c9-x5tdg" Dec 03 14:31:33 crc kubenswrapper[4805]: I1203 14:31:33.659258 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-f6bc4c6c9-x5tdg" Dec 03 14:31:33 crc kubenswrapper[4805]: I1203 14:31:33.720540 4805 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/ceilometer-0" podUID="58c98434-0b4d-4c2d-ac35-109e38f96803" containerName="proxy-httpd" probeResult="failure" output="Get \"http://10.217.0.155:3000/\": read tcp 10.217.0.2:49008->10.217.0.155:3000: read: connection reset by peer" Dec 03 14:31:34 crc kubenswrapper[4805]: I1203 14:31:34.028059 4805 generic.go:334] "Generic (PLEG): container finished" podID="58c98434-0b4d-4c2d-ac35-109e38f96803" containerID="126ffcd2cdaa24e89e915efb3598e7f49248d24b6884776c5b6848a1429ab5c0" exitCode=0 Dec 03 14:31:34 crc kubenswrapper[4805]: I1203 14:31:34.028090 4805 generic.go:334] "Generic (PLEG): container finished" podID="58c98434-0b4d-4c2d-ac35-109e38f96803" containerID="20bba63c0e20fac3641cd24a4df07f94b5cb20bb86b1e7fd6adcb5e5071e51b0" exitCode=2 Dec 03 14:31:34 crc kubenswrapper[4805]: I1203 14:31:34.028132 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"58c98434-0b4d-4c2d-ac35-109e38f96803","Type":"ContainerDied","Data":"126ffcd2cdaa24e89e915efb3598e7f49248d24b6884776c5b6848a1429ab5c0"} Dec 03 14:31:34 crc kubenswrapper[4805]: I1203 14:31:34.028157 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"58c98434-0b4d-4c2d-ac35-109e38f96803","Type":"ContainerDied","Data":"20bba63c0e20fac3641cd24a4df07f94b5cb20bb86b1e7fd6adcb5e5071e51b0"} Dec 03 14:31:34 crc kubenswrapper[4805]: I1203 14:31:34.029877 4805 generic.go:334] "Generic (PLEG): container finished" podID="bc7fd7a7-02bc-48e1-9bde-d1680876da37" containerID="d957799f184adf285454f33a4f7aacfb2a9823b898bb6d66acdb0d27983681d6" exitCode=0 Dec 03 14:31:34 crc kubenswrapper[4805]: I1203 14:31:34.029901 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5784cf869f-l4bt9" event={"ID":"bc7fd7a7-02bc-48e1-9bde-d1680876da37","Type":"ContainerDied","Data":"d957799f184adf285454f33a4f7aacfb2a9823b898bb6d66acdb0d27983681d6"} Dec 03 14:31:34 crc kubenswrapper[4805]: I1203 14:31:34.238653 4805 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/cinder-scheduler-0" Dec 03 14:31:34 crc kubenswrapper[4805]: I1203 14:31:34.345437 4805 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-5784cf869f-l4bt9" podUID="bc7fd7a7-02bc-48e1-9bde-d1680876da37" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.163:5353: connect: connection refused" Dec 03 14:31:34 crc kubenswrapper[4805]: I1203 14:31:34.566678 4805 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-db-create-z7rqx"] Dec 03 14:31:34 crc kubenswrapper[4805]: I1203 14:31:34.567831 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-db-create-z7rqx" Dec 03 14:31:34 crc kubenswrapper[4805]: I1203 14:31:34.592411 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-db-create-z7rqx"] Dec 03 14:31:34 crc kubenswrapper[4805]: I1203 14:31:34.599830 4805 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/cinder-scheduler-0" Dec 03 14:31:34 crc kubenswrapper[4805]: I1203 14:31:34.649364 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fqfx6\" (UniqueName: \"kubernetes.io/projected/649e0d3a-332c-43d7-90bd-8ad10015268f-kube-api-access-fqfx6\") pod \"nova-api-db-create-z7rqx\" (UID: \"649e0d3a-332c-43d7-90bd-8ad10015268f\") " pod="openstack/nova-api-db-create-z7rqx" Dec 03 14:31:34 crc kubenswrapper[4805]: I1203 14:31:34.649467 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/649e0d3a-332c-43d7-90bd-8ad10015268f-operator-scripts\") pod \"nova-api-db-create-z7rqx\" (UID: \"649e0d3a-332c-43d7-90bd-8ad10015268f\") " pod="openstack/nova-api-db-create-z7rqx" Dec 03 14:31:34 crc kubenswrapper[4805]: I1203 14:31:34.667051 4805 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-db-create-bc8gs"] Dec 03 14:31:34 crc kubenswrapper[4805]: I1203 14:31:34.668568 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-db-create-bc8gs" Dec 03 14:31:34 crc kubenswrapper[4805]: I1203 14:31:34.692316 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-db-create-bc8gs"] Dec 03 14:31:34 crc kubenswrapper[4805]: I1203 14:31:34.716585 4805 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-1be1-account-create-update-58nc2"] Dec 03 14:31:34 crc kubenswrapper[4805]: I1203 14:31:34.717908 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-1be1-account-create-update-58nc2" Dec 03 14:31:34 crc kubenswrapper[4805]: I1203 14:31:34.720317 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-db-secret" Dec 03 14:31:34 crc kubenswrapper[4805]: I1203 14:31:34.751288 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nmhpq\" (UniqueName: \"kubernetes.io/projected/1245f02d-543a-40d6-bc19-745a48249604-kube-api-access-nmhpq\") pod \"nova-cell0-db-create-bc8gs\" (UID: \"1245f02d-543a-40d6-bc19-745a48249604\") " pod="openstack/nova-cell0-db-create-bc8gs" Dec 03 14:31:34 crc kubenswrapper[4805]: I1203 14:31:34.751518 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/1082c862-47c9-4382-b3e8-09e58fb5d85d-operator-scripts\") pod \"nova-api-1be1-account-create-update-58nc2\" (UID: \"1082c862-47c9-4382-b3e8-09e58fb5d85d\") " pod="openstack/nova-api-1be1-account-create-update-58nc2" Dec 03 14:31:34 crc kubenswrapper[4805]: I1203 14:31:34.751658 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fqfx6\" (UniqueName: \"kubernetes.io/projected/649e0d3a-332c-43d7-90bd-8ad10015268f-kube-api-access-fqfx6\") pod \"nova-api-db-create-z7rqx\" (UID: \"649e0d3a-332c-43d7-90bd-8ad10015268f\") " pod="openstack/nova-api-db-create-z7rqx" Dec 03 14:31:34 crc kubenswrapper[4805]: I1203 14:31:34.751798 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/1245f02d-543a-40d6-bc19-745a48249604-operator-scripts\") pod \"nova-cell0-db-create-bc8gs\" (UID: \"1245f02d-543a-40d6-bc19-745a48249604\") " pod="openstack/nova-cell0-db-create-bc8gs" Dec 03 14:31:34 crc kubenswrapper[4805]: I1203 14:31:34.751913 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-n6mc6\" (UniqueName: \"kubernetes.io/projected/1082c862-47c9-4382-b3e8-09e58fb5d85d-kube-api-access-n6mc6\") pod \"nova-api-1be1-account-create-update-58nc2\" (UID: \"1082c862-47c9-4382-b3e8-09e58fb5d85d\") " pod="openstack/nova-api-1be1-account-create-update-58nc2" Dec 03 14:31:34 crc kubenswrapper[4805]: I1203 14:31:34.751997 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/649e0d3a-332c-43d7-90bd-8ad10015268f-operator-scripts\") pod \"nova-api-db-create-z7rqx\" (UID: \"649e0d3a-332c-43d7-90bd-8ad10015268f\") " pod="openstack/nova-api-db-create-z7rqx" Dec 03 14:31:34 crc kubenswrapper[4805]: I1203 14:31:34.752778 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/649e0d3a-332c-43d7-90bd-8ad10015268f-operator-scripts\") pod \"nova-api-db-create-z7rqx\" (UID: \"649e0d3a-332c-43d7-90bd-8ad10015268f\") " pod="openstack/nova-api-db-create-z7rqx" Dec 03 14:31:34 crc kubenswrapper[4805]: I1203 14:31:34.773475 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-1be1-account-create-update-58nc2"] Dec 03 14:31:34 crc kubenswrapper[4805]: I1203 14:31:34.782427 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fqfx6\" (UniqueName: \"kubernetes.io/projected/649e0d3a-332c-43d7-90bd-8ad10015268f-kube-api-access-fqfx6\") pod \"nova-api-db-create-z7rqx\" (UID: \"649e0d3a-332c-43d7-90bd-8ad10015268f\") " pod="openstack/nova-api-db-create-z7rqx" Dec 03 14:31:34 crc kubenswrapper[4805]: I1203 14:31:34.854782 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nmhpq\" (UniqueName: \"kubernetes.io/projected/1245f02d-543a-40d6-bc19-745a48249604-kube-api-access-nmhpq\") pod \"nova-cell0-db-create-bc8gs\" (UID: \"1245f02d-543a-40d6-bc19-745a48249604\") " pod="openstack/nova-cell0-db-create-bc8gs" Dec 03 14:31:34 crc kubenswrapper[4805]: I1203 14:31:34.854888 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/1082c862-47c9-4382-b3e8-09e58fb5d85d-operator-scripts\") pod \"nova-api-1be1-account-create-update-58nc2\" (UID: \"1082c862-47c9-4382-b3e8-09e58fb5d85d\") " pod="openstack/nova-api-1be1-account-create-update-58nc2" Dec 03 14:31:34 crc kubenswrapper[4805]: I1203 14:31:34.854988 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/1245f02d-543a-40d6-bc19-745a48249604-operator-scripts\") pod \"nova-cell0-db-create-bc8gs\" (UID: \"1245f02d-543a-40d6-bc19-745a48249604\") " pod="openstack/nova-cell0-db-create-bc8gs" Dec 03 14:31:34 crc kubenswrapper[4805]: I1203 14:31:34.855031 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-n6mc6\" (UniqueName: \"kubernetes.io/projected/1082c862-47c9-4382-b3e8-09e58fb5d85d-kube-api-access-n6mc6\") pod \"nova-api-1be1-account-create-update-58nc2\" (UID: \"1082c862-47c9-4382-b3e8-09e58fb5d85d\") " pod="openstack/nova-api-1be1-account-create-update-58nc2" Dec 03 14:31:34 crc kubenswrapper[4805]: I1203 14:31:34.856170 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/1082c862-47c9-4382-b3e8-09e58fb5d85d-operator-scripts\") pod \"nova-api-1be1-account-create-update-58nc2\" (UID: \"1082c862-47c9-4382-b3e8-09e58fb5d85d\") " pod="openstack/nova-api-1be1-account-create-update-58nc2" Dec 03 14:31:34 crc kubenswrapper[4805]: I1203 14:31:34.856635 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/1245f02d-543a-40d6-bc19-745a48249604-operator-scripts\") pod \"nova-cell0-db-create-bc8gs\" (UID: \"1245f02d-543a-40d6-bc19-745a48249604\") " pod="openstack/nova-cell0-db-create-bc8gs" Dec 03 14:31:34 crc kubenswrapper[4805]: I1203 14:31:34.872221 4805 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-db-create-b8n7x"] Dec 03 14:31:34 crc kubenswrapper[4805]: I1203 14:31:34.873767 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-db-create-b8n7x" Dec 03 14:31:34 crc kubenswrapper[4805]: I1203 14:31:34.883071 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-n6mc6\" (UniqueName: \"kubernetes.io/projected/1082c862-47c9-4382-b3e8-09e58fb5d85d-kube-api-access-n6mc6\") pod \"nova-api-1be1-account-create-update-58nc2\" (UID: \"1082c862-47c9-4382-b3e8-09e58fb5d85d\") " pod="openstack/nova-api-1be1-account-create-update-58nc2" Dec 03 14:31:34 crc kubenswrapper[4805]: I1203 14:31:34.887289 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nmhpq\" (UniqueName: \"kubernetes.io/projected/1245f02d-543a-40d6-bc19-745a48249604-kube-api-access-nmhpq\") pod \"nova-cell0-db-create-bc8gs\" (UID: \"1245f02d-543a-40d6-bc19-745a48249604\") " pod="openstack/nova-cell0-db-create-bc8gs" Dec 03 14:31:34 crc kubenswrapper[4805]: I1203 14:31:34.894656 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-db-create-z7rqx" Dec 03 14:31:34 crc kubenswrapper[4805]: I1203 14:31:34.914967 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-db-create-b8n7x"] Dec 03 14:31:34 crc kubenswrapper[4805]: I1203 14:31:34.926513 4805 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-7801-account-create-update-rjbpl"] Dec 03 14:31:34 crc kubenswrapper[4805]: I1203 14:31:34.928271 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-7801-account-create-update-rjbpl" Dec 03 14:31:34 crc kubenswrapper[4805]: I1203 14:31:34.932296 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-db-secret" Dec 03 14:31:34 crc kubenswrapper[4805]: I1203 14:31:34.957095 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-k5wks\" (UniqueName: \"kubernetes.io/projected/5c694d3e-75d4-40d4-924d-cfe9e9ef5ef7-kube-api-access-k5wks\") pod \"nova-cell0-7801-account-create-update-rjbpl\" (UID: \"5c694d3e-75d4-40d4-924d-cfe9e9ef5ef7\") " pod="openstack/nova-cell0-7801-account-create-update-rjbpl" Dec 03 14:31:34 crc kubenswrapper[4805]: I1203 14:31:34.957161 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/596937d0-4958-4d99-a677-698ca4b7d4eb-operator-scripts\") pod \"nova-cell1-db-create-b8n7x\" (UID: \"596937d0-4958-4d99-a677-698ca4b7d4eb\") " pod="openstack/nova-cell1-db-create-b8n7x" Dec 03 14:31:34 crc kubenswrapper[4805]: I1203 14:31:34.957799 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-7801-account-create-update-rjbpl"] Dec 03 14:31:34 crc kubenswrapper[4805]: I1203 14:31:34.957965 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2kg82\" (UniqueName: \"kubernetes.io/projected/596937d0-4958-4d99-a677-698ca4b7d4eb-kube-api-access-2kg82\") pod \"nova-cell1-db-create-b8n7x\" (UID: \"596937d0-4958-4d99-a677-698ca4b7d4eb\") " pod="openstack/nova-cell1-db-create-b8n7x" Dec 03 14:31:34 crc kubenswrapper[4805]: I1203 14:31:34.958142 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/5c694d3e-75d4-40d4-924d-cfe9e9ef5ef7-operator-scripts\") pod \"nova-cell0-7801-account-create-update-rjbpl\" (UID: \"5c694d3e-75d4-40d4-924d-cfe9e9ef5ef7\") " pod="openstack/nova-cell0-7801-account-create-update-rjbpl" Dec 03 14:31:35 crc kubenswrapper[4805]: I1203 14:31:35.006362 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-db-create-bc8gs" Dec 03 14:31:35 crc kubenswrapper[4805]: I1203 14:31:35.043794 4805 generic.go:334] "Generic (PLEG): container finished" podID="58c98434-0b4d-4c2d-ac35-109e38f96803" containerID="f98e610e128d27a8cf1f7ca39fe6a3f759d8784ae234f8e1a4c4d5e228950639" exitCode=0 Dec 03 14:31:35 crc kubenswrapper[4805]: I1203 14:31:35.044019 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"58c98434-0b4d-4c2d-ac35-109e38f96803","Type":"ContainerDied","Data":"f98e610e128d27a8cf1f7ca39fe6a3f759d8784ae234f8e1a4c4d5e228950639"} Dec 03 14:31:35 crc kubenswrapper[4805]: I1203 14:31:35.050563 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-1be1-account-create-update-58nc2" Dec 03 14:31:35 crc kubenswrapper[4805]: I1203 14:31:35.064490 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2kg82\" (UniqueName: \"kubernetes.io/projected/596937d0-4958-4d99-a677-698ca4b7d4eb-kube-api-access-2kg82\") pod \"nova-cell1-db-create-b8n7x\" (UID: \"596937d0-4958-4d99-a677-698ca4b7d4eb\") " pod="openstack/nova-cell1-db-create-b8n7x" Dec 03 14:31:35 crc kubenswrapper[4805]: I1203 14:31:35.064825 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/5c694d3e-75d4-40d4-924d-cfe9e9ef5ef7-operator-scripts\") pod \"nova-cell0-7801-account-create-update-rjbpl\" (UID: \"5c694d3e-75d4-40d4-924d-cfe9e9ef5ef7\") " pod="openstack/nova-cell0-7801-account-create-update-rjbpl" Dec 03 14:31:35 crc kubenswrapper[4805]: I1203 14:31:35.064976 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-k5wks\" (UniqueName: \"kubernetes.io/projected/5c694d3e-75d4-40d4-924d-cfe9e9ef5ef7-kube-api-access-k5wks\") pod \"nova-cell0-7801-account-create-update-rjbpl\" (UID: \"5c694d3e-75d4-40d4-924d-cfe9e9ef5ef7\") " pod="openstack/nova-cell0-7801-account-create-update-rjbpl" Dec 03 14:31:35 crc kubenswrapper[4805]: I1203 14:31:35.065098 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/596937d0-4958-4d99-a677-698ca4b7d4eb-operator-scripts\") pod \"nova-cell1-db-create-b8n7x\" (UID: \"596937d0-4958-4d99-a677-698ca4b7d4eb\") " pod="openstack/nova-cell1-db-create-b8n7x" Dec 03 14:31:35 crc kubenswrapper[4805]: I1203 14:31:35.065454 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/5c694d3e-75d4-40d4-924d-cfe9e9ef5ef7-operator-scripts\") pod \"nova-cell0-7801-account-create-update-rjbpl\" (UID: \"5c694d3e-75d4-40d4-924d-cfe9e9ef5ef7\") " pod="openstack/nova-cell0-7801-account-create-update-rjbpl" Dec 03 14:31:35 crc kubenswrapper[4805]: I1203 14:31:35.065880 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/596937d0-4958-4d99-a677-698ca4b7d4eb-operator-scripts\") pod \"nova-cell1-db-create-b8n7x\" (UID: \"596937d0-4958-4d99-a677-698ca4b7d4eb\") " pod="openstack/nova-cell1-db-create-b8n7x" Dec 03 14:31:35 crc kubenswrapper[4805]: I1203 14:31:35.078952 4805 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-f183-account-create-update-p7mnx"] Dec 03 14:31:35 crc kubenswrapper[4805]: I1203 14:31:35.080933 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-f183-account-create-update-p7mnx" Dec 03 14:31:35 crc kubenswrapper[4805]: I1203 14:31:35.084498 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-db-secret" Dec 03 14:31:35 crc kubenswrapper[4805]: I1203 14:31:35.086113 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2kg82\" (UniqueName: \"kubernetes.io/projected/596937d0-4958-4d99-a677-698ca4b7d4eb-kube-api-access-2kg82\") pod \"nova-cell1-db-create-b8n7x\" (UID: \"596937d0-4958-4d99-a677-698ca4b7d4eb\") " pod="openstack/nova-cell1-db-create-b8n7x" Dec 03 14:31:35 crc kubenswrapper[4805]: I1203 14:31:35.088763 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-k5wks\" (UniqueName: \"kubernetes.io/projected/5c694d3e-75d4-40d4-924d-cfe9e9ef5ef7-kube-api-access-k5wks\") pod \"nova-cell0-7801-account-create-update-rjbpl\" (UID: \"5c694d3e-75d4-40d4-924d-cfe9e9ef5ef7\") " pod="openstack/nova-cell0-7801-account-create-update-rjbpl" Dec 03 14:31:35 crc kubenswrapper[4805]: I1203 14:31:35.158903 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-f183-account-create-update-p7mnx"] Dec 03 14:31:35 crc kubenswrapper[4805]: I1203 14:31:35.166897 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-z82gm\" (UniqueName: \"kubernetes.io/projected/0c5c12d6-db52-4367-8898-e88fc8cdde41-kube-api-access-z82gm\") pod \"nova-cell1-f183-account-create-update-p7mnx\" (UID: \"0c5c12d6-db52-4367-8898-e88fc8cdde41\") " pod="openstack/nova-cell1-f183-account-create-update-p7mnx" Dec 03 14:31:35 crc kubenswrapper[4805]: I1203 14:31:35.167387 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/0c5c12d6-db52-4367-8898-e88fc8cdde41-operator-scripts\") pod \"nova-cell1-f183-account-create-update-p7mnx\" (UID: \"0c5c12d6-db52-4367-8898-e88fc8cdde41\") " pod="openstack/nova-cell1-f183-account-create-update-p7mnx" Dec 03 14:31:35 crc kubenswrapper[4805]: I1203 14:31:35.168764 4805 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-scheduler-0"] Dec 03 14:31:35 crc kubenswrapper[4805]: I1203 14:31:35.270558 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-z82gm\" (UniqueName: \"kubernetes.io/projected/0c5c12d6-db52-4367-8898-e88fc8cdde41-kube-api-access-z82gm\") pod \"nova-cell1-f183-account-create-update-p7mnx\" (UID: \"0c5c12d6-db52-4367-8898-e88fc8cdde41\") " pod="openstack/nova-cell1-f183-account-create-update-p7mnx" Dec 03 14:31:35 crc kubenswrapper[4805]: I1203 14:31:35.270742 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/0c5c12d6-db52-4367-8898-e88fc8cdde41-operator-scripts\") pod \"nova-cell1-f183-account-create-update-p7mnx\" (UID: \"0c5c12d6-db52-4367-8898-e88fc8cdde41\") " pod="openstack/nova-cell1-f183-account-create-update-p7mnx" Dec 03 14:31:35 crc kubenswrapper[4805]: I1203 14:31:35.271512 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/0c5c12d6-db52-4367-8898-e88fc8cdde41-operator-scripts\") pod \"nova-cell1-f183-account-create-update-p7mnx\" (UID: \"0c5c12d6-db52-4367-8898-e88fc8cdde41\") " pod="openstack/nova-cell1-f183-account-create-update-p7mnx" Dec 03 14:31:35 crc kubenswrapper[4805]: I1203 14:31:35.289998 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-db-create-b8n7x" Dec 03 14:31:35 crc kubenswrapper[4805]: I1203 14:31:35.293546 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-z82gm\" (UniqueName: \"kubernetes.io/projected/0c5c12d6-db52-4367-8898-e88fc8cdde41-kube-api-access-z82gm\") pod \"nova-cell1-f183-account-create-update-p7mnx\" (UID: \"0c5c12d6-db52-4367-8898-e88fc8cdde41\") " pod="openstack/nova-cell1-f183-account-create-update-p7mnx" Dec 03 14:31:35 crc kubenswrapper[4805]: I1203 14:31:35.298907 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-7801-account-create-update-rjbpl" Dec 03 14:31:35 crc kubenswrapper[4805]: I1203 14:31:35.465484 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-f183-account-create-update-p7mnx" Dec 03 14:31:36 crc kubenswrapper[4805]: I1203 14:31:36.056692 4805 generic.go:334] "Generic (PLEG): container finished" podID="58c98434-0b4d-4c2d-ac35-109e38f96803" containerID="9cf635a2329565cc994a3a2337944c2eca2f7047534f27d7eeb5f3ceaa96458b" exitCode=0 Dec 03 14:31:36 crc kubenswrapper[4805]: I1203 14:31:36.057230 4805 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-scheduler-0" podUID="29710b09-a1ad-468f-af26-b5d5eb516898" containerName="cinder-scheduler" containerID="cri-o://cb75dea47e869ccbdcc4f2f309b4ef5bddded1bea1f31601b29cb63b5646b269" gracePeriod=30 Dec 03 14:31:36 crc kubenswrapper[4805]: I1203 14:31:36.056802 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"58c98434-0b4d-4c2d-ac35-109e38f96803","Type":"ContainerDied","Data":"9cf635a2329565cc994a3a2337944c2eca2f7047534f27d7eeb5f3ceaa96458b"} Dec 03 14:31:36 crc kubenswrapper[4805]: I1203 14:31:36.057681 4805 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-scheduler-0" podUID="29710b09-a1ad-468f-af26-b5d5eb516898" containerName="probe" containerID="cri-o://95b1ef4794618699f48a7350bb18cecf536b9ae5725dda4c6c96e5147bb28baf" gracePeriod=30 Dec 03 14:31:37 crc kubenswrapper[4805]: I1203 14:31:37.069461 4805 generic.go:334] "Generic (PLEG): container finished" podID="29710b09-a1ad-468f-af26-b5d5eb516898" containerID="95b1ef4794618699f48a7350bb18cecf536b9ae5725dda4c6c96e5147bb28baf" exitCode=0 Dec 03 14:31:37 crc kubenswrapper[4805]: I1203 14:31:37.069504 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"29710b09-a1ad-468f-af26-b5d5eb516898","Type":"ContainerDied","Data":"95b1ef4794618699f48a7350bb18cecf536b9ae5725dda4c6c96e5147bb28baf"} Dec 03 14:31:38 crc kubenswrapper[4805]: I1203 14:31:38.231823 4805 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/barbican-api-845f47d448-thqcr" podUID="37658956-d75a-43ce-81d8-38f9bca42957" containerName="barbican-api-log" probeResult="failure" output="Get \"http://10.217.0.159:9311/healthcheck\": dial tcp 10.217.0.159:9311: connect: connection refused" Dec 03 14:31:38 crc kubenswrapper[4805]: I1203 14:31:38.232083 4805 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/barbican-api-845f47d448-thqcr" podUID="37658956-d75a-43ce-81d8-38f9bca42957" containerName="barbican-api" probeResult="failure" output="Get \"http://10.217.0.159:9311/healthcheck\": dial tcp 10.217.0.159:9311: connect: connection refused" Dec 03 14:31:39 crc kubenswrapper[4805]: W1203 14:31:39.094781 4805 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod7b5b9f5a_5af6_41d7_be26_3f4c091098b7.slice/crio-f6cc1a7c62128f0e7fb8cab5df7f53433c5a44d654fa753a7e9430e505bb0561 WatchSource:0}: Error finding container f6cc1a7c62128f0e7fb8cab5df7f53433c5a44d654fa753a7e9430e505bb0561: Status 404 returned error can't find the container with id f6cc1a7c62128f0e7fb8cab5df7f53433c5a44d654fa753a7e9430e505bb0561 Dec 03 14:31:39 crc kubenswrapper[4805]: I1203 14:31:39.465970 4805 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/heat-engine-76dcfc959c-swfs4"] Dec 03 14:31:39 crc kubenswrapper[4805]: I1203 14:31:39.467413 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-engine-76dcfc959c-swfs4" Dec 03 14:31:39 crc kubenswrapper[4805]: I1203 14:31:39.497768 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-engine-76dcfc959c-swfs4"] Dec 03 14:31:39 crc kubenswrapper[4805]: I1203 14:31:39.515883 4805 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/heat-cfnapi-6cd94c587b-nshpc"] Dec 03 14:31:39 crc kubenswrapper[4805]: I1203 14:31:39.517039 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-cfnapi-6cd94c587b-nshpc" Dec 03 14:31:39 crc kubenswrapper[4805]: I1203 14:31:39.532789 4805 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/heat-api-69cb5bcd7b-bwpl9"] Dec 03 14:31:39 crc kubenswrapper[4805]: I1203 14:31:39.536198 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-api-69cb5bcd7b-bwpl9" Dec 03 14:31:39 crc kubenswrapper[4805]: I1203 14:31:39.555556 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/d1d26c95-d05a-4551-8d00-655941ce122b-config-data-custom\") pod \"heat-cfnapi-6cd94c587b-nshpc\" (UID: \"d1d26c95-d05a-4551-8d00-655941ce122b\") " pod="openstack/heat-cfnapi-6cd94c587b-nshpc" Dec 03 14:31:39 crc kubenswrapper[4805]: I1203 14:31:39.555596 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-p2m4h\" (UniqueName: \"kubernetes.io/projected/854cde2d-c692-4ab4-9172-02ad6a2b2f6a-kube-api-access-p2m4h\") pod \"heat-engine-76dcfc959c-swfs4\" (UID: \"854cde2d-c692-4ab4-9172-02ad6a2b2f6a\") " pod="openstack/heat-engine-76dcfc959c-swfs4" Dec 03 14:31:39 crc kubenswrapper[4805]: I1203 14:31:39.555614 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d1d26c95-d05a-4551-8d00-655941ce122b-config-data\") pod \"heat-cfnapi-6cd94c587b-nshpc\" (UID: \"d1d26c95-d05a-4551-8d00-655941ce122b\") " pod="openstack/heat-cfnapi-6cd94c587b-nshpc" Dec 03 14:31:39 crc kubenswrapper[4805]: I1203 14:31:39.555631 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/854cde2d-c692-4ab4-9172-02ad6a2b2f6a-config-data-custom\") pod \"heat-engine-76dcfc959c-swfs4\" (UID: \"854cde2d-c692-4ab4-9172-02ad6a2b2f6a\") " pod="openstack/heat-engine-76dcfc959c-swfs4" Dec 03 14:31:39 crc kubenswrapper[4805]: I1203 14:31:39.555651 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/756ba517-3759-4ee7-8718-a5d8c6f1be72-combined-ca-bundle\") pod \"heat-api-69cb5bcd7b-bwpl9\" (UID: \"756ba517-3759-4ee7-8718-a5d8c6f1be72\") " pod="openstack/heat-api-69cb5bcd7b-bwpl9" Dec 03 14:31:39 crc kubenswrapper[4805]: I1203 14:31:39.555669 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d1d26c95-d05a-4551-8d00-655941ce122b-combined-ca-bundle\") pod \"heat-cfnapi-6cd94c587b-nshpc\" (UID: \"d1d26c95-d05a-4551-8d00-655941ce122b\") " pod="openstack/heat-cfnapi-6cd94c587b-nshpc" Dec 03 14:31:39 crc kubenswrapper[4805]: I1203 14:31:39.555689 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/756ba517-3759-4ee7-8718-a5d8c6f1be72-config-data-custom\") pod \"heat-api-69cb5bcd7b-bwpl9\" (UID: \"756ba517-3759-4ee7-8718-a5d8c6f1be72\") " pod="openstack/heat-api-69cb5bcd7b-bwpl9" Dec 03 14:31:39 crc kubenswrapper[4805]: I1203 14:31:39.555753 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-z7p7f\" (UniqueName: \"kubernetes.io/projected/d1d26c95-d05a-4551-8d00-655941ce122b-kube-api-access-z7p7f\") pod \"heat-cfnapi-6cd94c587b-nshpc\" (UID: \"d1d26c95-d05a-4551-8d00-655941ce122b\") " pod="openstack/heat-cfnapi-6cd94c587b-nshpc" Dec 03 14:31:39 crc kubenswrapper[4805]: I1203 14:31:39.555770 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ckx8m\" (UniqueName: \"kubernetes.io/projected/756ba517-3759-4ee7-8718-a5d8c6f1be72-kube-api-access-ckx8m\") pod \"heat-api-69cb5bcd7b-bwpl9\" (UID: \"756ba517-3759-4ee7-8718-a5d8c6f1be72\") " pod="openstack/heat-api-69cb5bcd7b-bwpl9" Dec 03 14:31:39 crc kubenswrapper[4805]: I1203 14:31:39.555812 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/854cde2d-c692-4ab4-9172-02ad6a2b2f6a-config-data\") pod \"heat-engine-76dcfc959c-swfs4\" (UID: \"854cde2d-c692-4ab4-9172-02ad6a2b2f6a\") " pod="openstack/heat-engine-76dcfc959c-swfs4" Dec 03 14:31:39 crc kubenswrapper[4805]: I1203 14:31:39.555852 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/756ba517-3759-4ee7-8718-a5d8c6f1be72-config-data\") pod \"heat-api-69cb5bcd7b-bwpl9\" (UID: \"756ba517-3759-4ee7-8718-a5d8c6f1be72\") " pod="openstack/heat-api-69cb5bcd7b-bwpl9" Dec 03 14:31:39 crc kubenswrapper[4805]: I1203 14:31:39.555869 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/854cde2d-c692-4ab4-9172-02ad6a2b2f6a-combined-ca-bundle\") pod \"heat-engine-76dcfc959c-swfs4\" (UID: \"854cde2d-c692-4ab4-9172-02ad6a2b2f6a\") " pod="openstack/heat-engine-76dcfc959c-swfs4" Dec 03 14:31:39 crc kubenswrapper[4805]: I1203 14:31:39.556880 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-cfnapi-6cd94c587b-nshpc"] Dec 03 14:31:39 crc kubenswrapper[4805]: I1203 14:31:39.568919 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-api-69cb5bcd7b-bwpl9"] Dec 03 14:31:39 crc kubenswrapper[4805]: I1203 14:31:39.657873 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/d1d26c95-d05a-4551-8d00-655941ce122b-config-data-custom\") pod \"heat-cfnapi-6cd94c587b-nshpc\" (UID: \"d1d26c95-d05a-4551-8d00-655941ce122b\") " pod="openstack/heat-cfnapi-6cd94c587b-nshpc" Dec 03 14:31:39 crc kubenswrapper[4805]: I1203 14:31:39.658071 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-p2m4h\" (UniqueName: \"kubernetes.io/projected/854cde2d-c692-4ab4-9172-02ad6a2b2f6a-kube-api-access-p2m4h\") pod \"heat-engine-76dcfc959c-swfs4\" (UID: \"854cde2d-c692-4ab4-9172-02ad6a2b2f6a\") " pod="openstack/heat-engine-76dcfc959c-swfs4" Dec 03 14:31:39 crc kubenswrapper[4805]: I1203 14:31:39.658110 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d1d26c95-d05a-4551-8d00-655941ce122b-config-data\") pod \"heat-cfnapi-6cd94c587b-nshpc\" (UID: \"d1d26c95-d05a-4551-8d00-655941ce122b\") " pod="openstack/heat-cfnapi-6cd94c587b-nshpc" Dec 03 14:31:39 crc kubenswrapper[4805]: I1203 14:31:39.658129 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/854cde2d-c692-4ab4-9172-02ad6a2b2f6a-config-data-custom\") pod \"heat-engine-76dcfc959c-swfs4\" (UID: \"854cde2d-c692-4ab4-9172-02ad6a2b2f6a\") " pod="openstack/heat-engine-76dcfc959c-swfs4" Dec 03 14:31:39 crc kubenswrapper[4805]: I1203 14:31:39.658152 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/756ba517-3759-4ee7-8718-a5d8c6f1be72-combined-ca-bundle\") pod \"heat-api-69cb5bcd7b-bwpl9\" (UID: \"756ba517-3759-4ee7-8718-a5d8c6f1be72\") " pod="openstack/heat-api-69cb5bcd7b-bwpl9" Dec 03 14:31:39 crc kubenswrapper[4805]: I1203 14:31:39.658168 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d1d26c95-d05a-4551-8d00-655941ce122b-combined-ca-bundle\") pod \"heat-cfnapi-6cd94c587b-nshpc\" (UID: \"d1d26c95-d05a-4551-8d00-655941ce122b\") " pod="openstack/heat-cfnapi-6cd94c587b-nshpc" Dec 03 14:31:39 crc kubenswrapper[4805]: I1203 14:31:39.658191 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/756ba517-3759-4ee7-8718-a5d8c6f1be72-config-data-custom\") pod \"heat-api-69cb5bcd7b-bwpl9\" (UID: \"756ba517-3759-4ee7-8718-a5d8c6f1be72\") " pod="openstack/heat-api-69cb5bcd7b-bwpl9" Dec 03 14:31:39 crc kubenswrapper[4805]: I1203 14:31:39.658256 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-z7p7f\" (UniqueName: \"kubernetes.io/projected/d1d26c95-d05a-4551-8d00-655941ce122b-kube-api-access-z7p7f\") pod \"heat-cfnapi-6cd94c587b-nshpc\" (UID: \"d1d26c95-d05a-4551-8d00-655941ce122b\") " pod="openstack/heat-cfnapi-6cd94c587b-nshpc" Dec 03 14:31:39 crc kubenswrapper[4805]: I1203 14:31:39.658274 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ckx8m\" (UniqueName: \"kubernetes.io/projected/756ba517-3759-4ee7-8718-a5d8c6f1be72-kube-api-access-ckx8m\") pod \"heat-api-69cb5bcd7b-bwpl9\" (UID: \"756ba517-3759-4ee7-8718-a5d8c6f1be72\") " pod="openstack/heat-api-69cb5bcd7b-bwpl9" Dec 03 14:31:39 crc kubenswrapper[4805]: I1203 14:31:39.658314 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/854cde2d-c692-4ab4-9172-02ad6a2b2f6a-config-data\") pod \"heat-engine-76dcfc959c-swfs4\" (UID: \"854cde2d-c692-4ab4-9172-02ad6a2b2f6a\") " pod="openstack/heat-engine-76dcfc959c-swfs4" Dec 03 14:31:39 crc kubenswrapper[4805]: I1203 14:31:39.658335 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/756ba517-3759-4ee7-8718-a5d8c6f1be72-config-data\") pod \"heat-api-69cb5bcd7b-bwpl9\" (UID: \"756ba517-3759-4ee7-8718-a5d8c6f1be72\") " pod="openstack/heat-api-69cb5bcd7b-bwpl9" Dec 03 14:31:39 crc kubenswrapper[4805]: I1203 14:31:39.658355 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/854cde2d-c692-4ab4-9172-02ad6a2b2f6a-combined-ca-bundle\") pod \"heat-engine-76dcfc959c-swfs4\" (UID: \"854cde2d-c692-4ab4-9172-02ad6a2b2f6a\") " pod="openstack/heat-engine-76dcfc959c-swfs4" Dec 03 14:31:39 crc kubenswrapper[4805]: I1203 14:31:39.681997 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/756ba517-3759-4ee7-8718-a5d8c6f1be72-combined-ca-bundle\") pod \"heat-api-69cb5bcd7b-bwpl9\" (UID: \"756ba517-3759-4ee7-8718-a5d8c6f1be72\") " pod="openstack/heat-api-69cb5bcd7b-bwpl9" Dec 03 14:31:39 crc kubenswrapper[4805]: I1203 14:31:39.682801 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d1d26c95-d05a-4551-8d00-655941ce122b-combined-ca-bundle\") pod \"heat-cfnapi-6cd94c587b-nshpc\" (UID: \"d1d26c95-d05a-4551-8d00-655941ce122b\") " pod="openstack/heat-cfnapi-6cd94c587b-nshpc" Dec 03 14:31:39 crc kubenswrapper[4805]: I1203 14:31:39.683637 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/756ba517-3759-4ee7-8718-a5d8c6f1be72-config-data-custom\") pod \"heat-api-69cb5bcd7b-bwpl9\" (UID: \"756ba517-3759-4ee7-8718-a5d8c6f1be72\") " pod="openstack/heat-api-69cb5bcd7b-bwpl9" Dec 03 14:31:39 crc kubenswrapper[4805]: I1203 14:31:39.683970 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/854cde2d-c692-4ab4-9172-02ad6a2b2f6a-config-data\") pod \"heat-engine-76dcfc959c-swfs4\" (UID: \"854cde2d-c692-4ab4-9172-02ad6a2b2f6a\") " pod="openstack/heat-engine-76dcfc959c-swfs4" Dec 03 14:31:39 crc kubenswrapper[4805]: I1203 14:31:39.688064 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/756ba517-3759-4ee7-8718-a5d8c6f1be72-config-data\") pod \"heat-api-69cb5bcd7b-bwpl9\" (UID: \"756ba517-3759-4ee7-8718-a5d8c6f1be72\") " pod="openstack/heat-api-69cb5bcd7b-bwpl9" Dec 03 14:31:39 crc kubenswrapper[4805]: I1203 14:31:39.688604 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/854cde2d-c692-4ab4-9172-02ad6a2b2f6a-combined-ca-bundle\") pod \"heat-engine-76dcfc959c-swfs4\" (UID: \"854cde2d-c692-4ab4-9172-02ad6a2b2f6a\") " pod="openstack/heat-engine-76dcfc959c-swfs4" Dec 03 14:31:39 crc kubenswrapper[4805]: I1203 14:31:39.688820 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-z7p7f\" (UniqueName: \"kubernetes.io/projected/d1d26c95-d05a-4551-8d00-655941ce122b-kube-api-access-z7p7f\") pod \"heat-cfnapi-6cd94c587b-nshpc\" (UID: \"d1d26c95-d05a-4551-8d00-655941ce122b\") " pod="openstack/heat-cfnapi-6cd94c587b-nshpc" Dec 03 14:31:39 crc kubenswrapper[4805]: I1203 14:31:39.691639 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/854cde2d-c692-4ab4-9172-02ad6a2b2f6a-config-data-custom\") pod \"heat-engine-76dcfc959c-swfs4\" (UID: \"854cde2d-c692-4ab4-9172-02ad6a2b2f6a\") " pod="openstack/heat-engine-76dcfc959c-swfs4" Dec 03 14:31:39 crc kubenswrapper[4805]: I1203 14:31:39.696524 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d1d26c95-d05a-4551-8d00-655941ce122b-config-data\") pod \"heat-cfnapi-6cd94c587b-nshpc\" (UID: \"d1d26c95-d05a-4551-8d00-655941ce122b\") " pod="openstack/heat-cfnapi-6cd94c587b-nshpc" Dec 03 14:31:39 crc kubenswrapper[4805]: I1203 14:31:39.703004 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/d1d26c95-d05a-4551-8d00-655941ce122b-config-data-custom\") pod \"heat-cfnapi-6cd94c587b-nshpc\" (UID: \"d1d26c95-d05a-4551-8d00-655941ce122b\") " pod="openstack/heat-cfnapi-6cd94c587b-nshpc" Dec 03 14:31:39 crc kubenswrapper[4805]: I1203 14:31:39.706090 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-p2m4h\" (UniqueName: \"kubernetes.io/projected/854cde2d-c692-4ab4-9172-02ad6a2b2f6a-kube-api-access-p2m4h\") pod \"heat-engine-76dcfc959c-swfs4\" (UID: \"854cde2d-c692-4ab4-9172-02ad6a2b2f6a\") " pod="openstack/heat-engine-76dcfc959c-swfs4" Dec 03 14:31:39 crc kubenswrapper[4805]: I1203 14:31:39.707554 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ckx8m\" (UniqueName: \"kubernetes.io/projected/756ba517-3759-4ee7-8718-a5d8c6f1be72-kube-api-access-ckx8m\") pod \"heat-api-69cb5bcd7b-bwpl9\" (UID: \"756ba517-3759-4ee7-8718-a5d8c6f1be72\") " pod="openstack/heat-api-69cb5bcd7b-bwpl9" Dec 03 14:31:39 crc kubenswrapper[4805]: I1203 14:31:39.838066 4805 scope.go:117] "RemoveContainer" containerID="48672ddefb5f29199ffd9090e43f5c584cf661e8cf1e4313d963fb96b28729b6" Dec 03 14:31:39 crc kubenswrapper[4805]: I1203 14:31:39.862986 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-engine-76dcfc959c-swfs4" Dec 03 14:31:39 crc kubenswrapper[4805]: I1203 14:31:39.882178 4805 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-845f47d448-thqcr" Dec 03 14:31:39 crc kubenswrapper[4805]: I1203 14:31:39.885225 4805 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5784cf869f-l4bt9" Dec 03 14:31:39 crc kubenswrapper[4805]: I1203 14:31:39.920768 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-cfnapi-6cd94c587b-nshpc" Dec 03 14:31:39 crc kubenswrapper[4805]: I1203 14:31:39.963430 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-api-69cb5bcd7b-bwpl9" Dec 03 14:31:40 crc kubenswrapper[4805]: I1203 14:31:40.066208 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/bc7fd7a7-02bc-48e1-9bde-d1680876da37-ovsdbserver-sb\") pod \"bc7fd7a7-02bc-48e1-9bde-d1680876da37\" (UID: \"bc7fd7a7-02bc-48e1-9bde-d1680876da37\") " Dec 03 14:31:40 crc kubenswrapper[4805]: I1203 14:31:40.066324 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/37658956-d75a-43ce-81d8-38f9bca42957-logs\") pod \"37658956-d75a-43ce-81d8-38f9bca42957\" (UID: \"37658956-d75a-43ce-81d8-38f9bca42957\") " Dec 03 14:31:40 crc kubenswrapper[4805]: I1203 14:31:40.066354 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/bc7fd7a7-02bc-48e1-9bde-d1680876da37-config\") pod \"bc7fd7a7-02bc-48e1-9bde-d1680876da37\" (UID: \"bc7fd7a7-02bc-48e1-9bde-d1680876da37\") " Dec 03 14:31:40 crc kubenswrapper[4805]: I1203 14:31:40.066402 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/bc7fd7a7-02bc-48e1-9bde-d1680876da37-ovsdbserver-nb\") pod \"bc7fd7a7-02bc-48e1-9bde-d1680876da37\" (UID: \"bc7fd7a7-02bc-48e1-9bde-d1680876da37\") " Dec 03 14:31:40 crc kubenswrapper[4805]: I1203 14:31:40.066483 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/bc7fd7a7-02bc-48e1-9bde-d1680876da37-dns-svc\") pod \"bc7fd7a7-02bc-48e1-9bde-d1680876da37\" (UID: \"bc7fd7a7-02bc-48e1-9bde-d1680876da37\") " Dec 03 14:31:40 crc kubenswrapper[4805]: I1203 14:31:40.066516 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/bc7fd7a7-02bc-48e1-9bde-d1680876da37-dns-swift-storage-0\") pod \"bc7fd7a7-02bc-48e1-9bde-d1680876da37\" (UID: \"bc7fd7a7-02bc-48e1-9bde-d1680876da37\") " Dec 03 14:31:40 crc kubenswrapper[4805]: I1203 14:31:40.066555 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dz2wl\" (UniqueName: \"kubernetes.io/projected/bc7fd7a7-02bc-48e1-9bde-d1680876da37-kube-api-access-dz2wl\") pod \"bc7fd7a7-02bc-48e1-9bde-d1680876da37\" (UID: \"bc7fd7a7-02bc-48e1-9bde-d1680876da37\") " Dec 03 14:31:40 crc kubenswrapper[4805]: I1203 14:31:40.066591 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wdmrf\" (UniqueName: \"kubernetes.io/projected/37658956-d75a-43ce-81d8-38f9bca42957-kube-api-access-wdmrf\") pod \"37658956-d75a-43ce-81d8-38f9bca42957\" (UID: \"37658956-d75a-43ce-81d8-38f9bca42957\") " Dec 03 14:31:40 crc kubenswrapper[4805]: I1203 14:31:40.066626 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/37658956-d75a-43ce-81d8-38f9bca42957-config-data\") pod \"37658956-d75a-43ce-81d8-38f9bca42957\" (UID: \"37658956-d75a-43ce-81d8-38f9bca42957\") " Dec 03 14:31:40 crc kubenswrapper[4805]: I1203 14:31:40.066762 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/37658956-d75a-43ce-81d8-38f9bca42957-config-data-custom\") pod \"37658956-d75a-43ce-81d8-38f9bca42957\" (UID: \"37658956-d75a-43ce-81d8-38f9bca42957\") " Dec 03 14:31:40 crc kubenswrapper[4805]: I1203 14:31:40.066855 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/37658956-d75a-43ce-81d8-38f9bca42957-combined-ca-bundle\") pod \"37658956-d75a-43ce-81d8-38f9bca42957\" (UID: \"37658956-d75a-43ce-81d8-38f9bca42957\") " Dec 03 14:31:40 crc kubenswrapper[4805]: I1203 14:31:40.067649 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/37658956-d75a-43ce-81d8-38f9bca42957-logs" (OuterVolumeSpecName: "logs") pod "37658956-d75a-43ce-81d8-38f9bca42957" (UID: "37658956-d75a-43ce-81d8-38f9bca42957"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 14:31:40 crc kubenswrapper[4805]: I1203 14:31:40.082069 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/37658956-d75a-43ce-81d8-38f9bca42957-kube-api-access-wdmrf" (OuterVolumeSpecName: "kube-api-access-wdmrf") pod "37658956-d75a-43ce-81d8-38f9bca42957" (UID: "37658956-d75a-43ce-81d8-38f9bca42957"). InnerVolumeSpecName "kube-api-access-wdmrf". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 14:31:40 crc kubenswrapper[4805]: I1203 14:31:40.097701 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/37658956-d75a-43ce-81d8-38f9bca42957-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "37658956-d75a-43ce-81d8-38f9bca42957" (UID: "37658956-d75a-43ce-81d8-38f9bca42957"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 14:31:40 crc kubenswrapper[4805]: I1203 14:31:40.144805 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"7b5b9f5a-5af6-41d7-be26-3f4c091098b7","Type":"ContainerStarted","Data":"f6cc1a7c62128f0e7fb8cab5df7f53433c5a44d654fa753a7e9430e505bb0561"} Dec 03 14:31:40 crc kubenswrapper[4805]: I1203 14:31:40.172707 4805 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/37658956-d75a-43ce-81d8-38f9bca42957-logs\") on node \"crc\" DevicePath \"\"" Dec 03 14:31:40 crc kubenswrapper[4805]: I1203 14:31:40.172751 4805 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wdmrf\" (UniqueName: \"kubernetes.io/projected/37658956-d75a-43ce-81d8-38f9bca42957-kube-api-access-wdmrf\") on node \"crc\" DevicePath \"\"" Dec 03 14:31:40 crc kubenswrapper[4805]: I1203 14:31:40.172768 4805 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/37658956-d75a-43ce-81d8-38f9bca42957-config-data-custom\") on node \"crc\" DevicePath \"\"" Dec 03 14:31:40 crc kubenswrapper[4805]: I1203 14:31:40.176463 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-845f47d448-thqcr" event={"ID":"37658956-d75a-43ce-81d8-38f9bca42957","Type":"ContainerDied","Data":"49d130dec5d88f2be0467ebb81763437af671eee08cac357241332cdcd156538"} Dec 03 14:31:40 crc kubenswrapper[4805]: I1203 14:31:40.176645 4805 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-845f47d448-thqcr" Dec 03 14:31:40 crc kubenswrapper[4805]: I1203 14:31:40.187025 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bc7fd7a7-02bc-48e1-9bde-d1680876da37-kube-api-access-dz2wl" (OuterVolumeSpecName: "kube-api-access-dz2wl") pod "bc7fd7a7-02bc-48e1-9bde-d1680876da37" (UID: "bc7fd7a7-02bc-48e1-9bde-d1680876da37"). InnerVolumeSpecName "kube-api-access-dz2wl". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 14:31:40 crc kubenswrapper[4805]: I1203 14:31:40.196266 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/37658956-d75a-43ce-81d8-38f9bca42957-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "37658956-d75a-43ce-81d8-38f9bca42957" (UID: "37658956-d75a-43ce-81d8-38f9bca42957"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 14:31:40 crc kubenswrapper[4805]: I1203 14:31:40.200097 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5784cf869f-l4bt9" event={"ID":"bc7fd7a7-02bc-48e1-9bde-d1680876da37","Type":"ContainerDied","Data":"09a0e37d64b74a252a57f21a3a4a49eec5dc40794afbee941035747c1661f72e"} Dec 03 14:31:40 crc kubenswrapper[4805]: I1203 14:31:40.200116 4805 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5784cf869f-l4bt9" Dec 03 14:31:40 crc kubenswrapper[4805]: I1203 14:31:40.232603 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/bc7fd7a7-02bc-48e1-9bde-d1680876da37-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "bc7fd7a7-02bc-48e1-9bde-d1680876da37" (UID: "bc7fd7a7-02bc-48e1-9bde-d1680876da37"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 14:31:40 crc kubenswrapper[4805]: I1203 14:31:40.261144 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/bc7fd7a7-02bc-48e1-9bde-d1680876da37-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "bc7fd7a7-02bc-48e1-9bde-d1680876da37" (UID: "bc7fd7a7-02bc-48e1-9bde-d1680876da37"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 14:31:40 crc kubenswrapper[4805]: I1203 14:31:40.270650 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/37658956-d75a-43ce-81d8-38f9bca42957-config-data" (OuterVolumeSpecName: "config-data") pod "37658956-d75a-43ce-81d8-38f9bca42957" (UID: "37658956-d75a-43ce-81d8-38f9bca42957"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 14:31:40 crc kubenswrapper[4805]: I1203 14:31:40.274706 4805 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dz2wl\" (UniqueName: \"kubernetes.io/projected/bc7fd7a7-02bc-48e1-9bde-d1680876da37-kube-api-access-dz2wl\") on node \"crc\" DevicePath \"\"" Dec 03 14:31:40 crc kubenswrapper[4805]: I1203 14:31:40.274735 4805 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/37658956-d75a-43ce-81d8-38f9bca42957-config-data\") on node \"crc\" DevicePath \"\"" Dec 03 14:31:40 crc kubenswrapper[4805]: I1203 14:31:40.274744 4805 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/37658956-d75a-43ce-81d8-38f9bca42957-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 14:31:40 crc kubenswrapper[4805]: I1203 14:31:40.274754 4805 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/bc7fd7a7-02bc-48e1-9bde-d1680876da37-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Dec 03 14:31:40 crc kubenswrapper[4805]: I1203 14:31:40.274762 4805 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/bc7fd7a7-02bc-48e1-9bde-d1680876da37-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Dec 03 14:31:40 crc kubenswrapper[4805]: I1203 14:31:40.279102 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/bc7fd7a7-02bc-48e1-9bde-d1680876da37-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "bc7fd7a7-02bc-48e1-9bde-d1680876da37" (UID: "bc7fd7a7-02bc-48e1-9bde-d1680876da37"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 14:31:40 crc kubenswrapper[4805]: I1203 14:31:40.288389 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/bc7fd7a7-02bc-48e1-9bde-d1680876da37-config" (OuterVolumeSpecName: "config") pod "bc7fd7a7-02bc-48e1-9bde-d1680876da37" (UID: "bc7fd7a7-02bc-48e1-9bde-d1680876da37"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 14:31:40 crc kubenswrapper[4805]: I1203 14:31:40.293160 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/bc7fd7a7-02bc-48e1-9bde-d1680876da37-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "bc7fd7a7-02bc-48e1-9bde-d1680876da37" (UID: "bc7fd7a7-02bc-48e1-9bde-d1680876da37"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 14:31:40 crc kubenswrapper[4805]: I1203 14:31:40.316641 4805 scope.go:117] "RemoveContainer" containerID="7adb8b3c5e068f3ee12e9581aecb5bdfa43066e3e1f0991bc28977ea2af801da" Dec 03 14:31:40 crc kubenswrapper[4805]: I1203 14:31:40.377291 4805 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/bc7fd7a7-02bc-48e1-9bde-d1680876da37-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Dec 03 14:31:40 crc kubenswrapper[4805]: I1203 14:31:40.379873 4805 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/bc7fd7a7-02bc-48e1-9bde-d1680876da37-config\") on node \"crc\" DevicePath \"\"" Dec 03 14:31:40 crc kubenswrapper[4805]: I1203 14:31:40.379901 4805 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/bc7fd7a7-02bc-48e1-9bde-d1680876da37-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 03 14:31:40 crc kubenswrapper[4805]: I1203 14:31:40.395296 4805 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 03 14:31:40 crc kubenswrapper[4805]: I1203 14:31:40.562049 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-db-create-b8n7x"] Dec 03 14:31:40 crc kubenswrapper[4805]: I1203 14:31:40.596382 4805 scope.go:117] "RemoveContainer" containerID="a54729ca2abce4df438fff733d8adb4838b1bcfb1e74dd34f90719edf58bd259" Dec 03 14:31:40 crc kubenswrapper[4805]: I1203 14:31:40.605907 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/58c98434-0b4d-4c2d-ac35-109e38f96803-scripts\") pod \"58c98434-0b4d-4c2d-ac35-109e38f96803\" (UID: \"58c98434-0b4d-4c2d-ac35-109e38f96803\") " Dec 03 14:31:40 crc kubenswrapper[4805]: I1203 14:31:40.605988 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8d5p7\" (UniqueName: \"kubernetes.io/projected/58c98434-0b4d-4c2d-ac35-109e38f96803-kube-api-access-8d5p7\") pod \"58c98434-0b4d-4c2d-ac35-109e38f96803\" (UID: \"58c98434-0b4d-4c2d-ac35-109e38f96803\") " Dec 03 14:31:40 crc kubenswrapper[4805]: I1203 14:31:40.606052 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/58c98434-0b4d-4c2d-ac35-109e38f96803-sg-core-conf-yaml\") pod \"58c98434-0b4d-4c2d-ac35-109e38f96803\" (UID: \"58c98434-0b4d-4c2d-ac35-109e38f96803\") " Dec 03 14:31:40 crc kubenswrapper[4805]: I1203 14:31:40.606074 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/58c98434-0b4d-4c2d-ac35-109e38f96803-combined-ca-bundle\") pod \"58c98434-0b4d-4c2d-ac35-109e38f96803\" (UID: \"58c98434-0b4d-4c2d-ac35-109e38f96803\") " Dec 03 14:31:40 crc kubenswrapper[4805]: I1203 14:31:40.606097 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/58c98434-0b4d-4c2d-ac35-109e38f96803-log-httpd\") pod \"58c98434-0b4d-4c2d-ac35-109e38f96803\" (UID: \"58c98434-0b4d-4c2d-ac35-109e38f96803\") " Dec 03 14:31:40 crc kubenswrapper[4805]: I1203 14:31:40.606192 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/58c98434-0b4d-4c2d-ac35-109e38f96803-run-httpd\") pod \"58c98434-0b4d-4c2d-ac35-109e38f96803\" (UID: \"58c98434-0b4d-4c2d-ac35-109e38f96803\") " Dec 03 14:31:40 crc kubenswrapper[4805]: I1203 14:31:40.606271 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/58c98434-0b4d-4c2d-ac35-109e38f96803-config-data\") pod \"58c98434-0b4d-4c2d-ac35-109e38f96803\" (UID: \"58c98434-0b4d-4c2d-ac35-109e38f96803\") " Dec 03 14:31:40 crc kubenswrapper[4805]: I1203 14:31:40.613493 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/58c98434-0b4d-4c2d-ac35-109e38f96803-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "58c98434-0b4d-4c2d-ac35-109e38f96803" (UID: "58c98434-0b4d-4c2d-ac35-109e38f96803"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 14:31:40 crc kubenswrapper[4805]: I1203 14:31:40.614035 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/58c98434-0b4d-4c2d-ac35-109e38f96803-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "58c98434-0b4d-4c2d-ac35-109e38f96803" (UID: "58c98434-0b4d-4c2d-ac35-109e38f96803"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 14:31:40 crc kubenswrapper[4805]: I1203 14:31:40.646695 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/58c98434-0b4d-4c2d-ac35-109e38f96803-scripts" (OuterVolumeSpecName: "scripts") pod "58c98434-0b4d-4c2d-ac35-109e38f96803" (UID: "58c98434-0b4d-4c2d-ac35-109e38f96803"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 14:31:40 crc kubenswrapper[4805]: I1203 14:31:40.651672 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/58c98434-0b4d-4c2d-ac35-109e38f96803-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "58c98434-0b4d-4c2d-ac35-109e38f96803" (UID: "58c98434-0b4d-4c2d-ac35-109e38f96803"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 14:31:40 crc kubenswrapper[4805]: I1203 14:31:40.657526 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/58c98434-0b4d-4c2d-ac35-109e38f96803-kube-api-access-8d5p7" (OuterVolumeSpecName: "kube-api-access-8d5p7") pod "58c98434-0b4d-4c2d-ac35-109e38f96803" (UID: "58c98434-0b4d-4c2d-ac35-109e38f96803"). InnerVolumeSpecName "kube-api-access-8d5p7". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 14:31:40 crc kubenswrapper[4805]: I1203 14:31:40.679203 4805 scope.go:117] "RemoveContainer" containerID="d957799f184adf285454f33a4f7aacfb2a9823b898bb6d66acdb0d27983681d6" Dec 03 14:31:40 crc kubenswrapper[4805]: I1203 14:31:40.707622 4805 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/58c98434-0b4d-4c2d-ac35-109e38f96803-scripts\") on node \"crc\" DevicePath \"\"" Dec 03 14:31:40 crc kubenswrapper[4805]: I1203 14:31:40.707649 4805 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8d5p7\" (UniqueName: \"kubernetes.io/projected/58c98434-0b4d-4c2d-ac35-109e38f96803-kube-api-access-8d5p7\") on node \"crc\" DevicePath \"\"" Dec 03 14:31:40 crc kubenswrapper[4805]: I1203 14:31:40.707659 4805 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/58c98434-0b4d-4c2d-ac35-109e38f96803-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Dec 03 14:31:40 crc kubenswrapper[4805]: I1203 14:31:40.707668 4805 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/58c98434-0b4d-4c2d-ac35-109e38f96803-log-httpd\") on node \"crc\" DevicePath \"\"" Dec 03 14:31:40 crc kubenswrapper[4805]: I1203 14:31:40.707676 4805 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/58c98434-0b4d-4c2d-ac35-109e38f96803-run-httpd\") on node \"crc\" DevicePath \"\"" Dec 03 14:31:40 crc kubenswrapper[4805]: I1203 14:31:40.751553 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/58c98434-0b4d-4c2d-ac35-109e38f96803-config-data" (OuterVolumeSpecName: "config-data") pod "58c98434-0b4d-4c2d-ac35-109e38f96803" (UID: "58c98434-0b4d-4c2d-ac35-109e38f96803"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 14:31:40 crc kubenswrapper[4805]: I1203 14:31:40.760505 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/58c98434-0b4d-4c2d-ac35-109e38f96803-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "58c98434-0b4d-4c2d-ac35-109e38f96803" (UID: "58c98434-0b4d-4c2d-ac35-109e38f96803"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 14:31:40 crc kubenswrapper[4805]: I1203 14:31:40.800851 4805 scope.go:117] "RemoveContainer" containerID="f31912c14d4c8f23d67c323b6a6401e02aee26fb20bd559099f3c3aada4dffa3" Dec 03 14:31:40 crc kubenswrapper[4805]: I1203 14:31:40.809086 4805 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/58c98434-0b4d-4c2d-ac35-109e38f96803-config-data\") on node \"crc\" DevicePath \"\"" Dec 03 14:31:40 crc kubenswrapper[4805]: I1203 14:31:40.809112 4805 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/58c98434-0b4d-4c2d-ac35-109e38f96803-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 14:31:40 crc kubenswrapper[4805]: I1203 14:31:40.848280 4805 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-api-845f47d448-thqcr"] Dec 03 14:31:40 crc kubenswrapper[4805]: I1203 14:31:40.860671 4805 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-api-845f47d448-thqcr"] Dec 03 14:31:40 crc kubenswrapper[4805]: I1203 14:31:40.929909 4805 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5784cf869f-l4bt9"] Dec 03 14:31:40 crc kubenswrapper[4805]: I1203 14:31:40.971822 4805 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-5784cf869f-l4bt9"] Dec 03 14:31:41 crc kubenswrapper[4805]: I1203 14:31:41.223550 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-db-create-b8n7x" event={"ID":"596937d0-4958-4d99-a677-698ca4b7d4eb","Type":"ContainerStarted","Data":"24391668ed2d2228fb3f9b78b7218ec1e72bcd0841b46649a97e17468974fb65"} Dec 03 14:31:41 crc kubenswrapper[4805]: I1203 14:31:41.229454 4805 generic.go:334] "Generic (PLEG): container finished" podID="29710b09-a1ad-468f-af26-b5d5eb516898" containerID="cb75dea47e869ccbdcc4f2f309b4ef5bddded1bea1f31601b29cb63b5646b269" exitCode=0 Dec 03 14:31:41 crc kubenswrapper[4805]: I1203 14:31:41.229824 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"29710b09-a1ad-468f-af26-b5d5eb516898","Type":"ContainerDied","Data":"cb75dea47e869ccbdcc4f2f309b4ef5bddded1bea1f31601b29cb63b5646b269"} Dec 03 14:31:41 crc kubenswrapper[4805]: I1203 14:31:41.236196 4805 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 03 14:31:41 crc kubenswrapper[4805]: I1203 14:31:41.236967 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"58c98434-0b4d-4c2d-ac35-109e38f96803","Type":"ContainerDied","Data":"b98faa6bdfacd7a3dbe424082fb5dc2be3f9f63287fa6fd24e34505f98826444"} Dec 03 14:31:41 crc kubenswrapper[4805]: I1203 14:31:41.237003 4805 scope.go:117] "RemoveContainer" containerID="126ffcd2cdaa24e89e915efb3598e7f49248d24b6884776c5b6848a1429ab5c0" Dec 03 14:31:41 crc kubenswrapper[4805]: I1203 14:31:41.300653 4805 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 03 14:31:41 crc kubenswrapper[4805]: I1203 14:31:41.301376 4805 scope.go:117] "RemoveContainer" containerID="20bba63c0e20fac3641cd24a4df07f94b5cb20bb86b1e7fd6adcb5e5071e51b0" Dec 03 14:31:41 crc kubenswrapper[4805]: I1203 14:31:41.332896 4805 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Dec 03 14:31:41 crc kubenswrapper[4805]: I1203 14:31:41.355665 4805 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Dec 03 14:31:41 crc kubenswrapper[4805]: E1203 14:31:41.356105 4805 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="58c98434-0b4d-4c2d-ac35-109e38f96803" containerName="ceilometer-notification-agent" Dec 03 14:31:41 crc kubenswrapper[4805]: I1203 14:31:41.356122 4805 state_mem.go:107] "Deleted CPUSet assignment" podUID="58c98434-0b4d-4c2d-ac35-109e38f96803" containerName="ceilometer-notification-agent" Dec 03 14:31:41 crc kubenswrapper[4805]: E1203 14:31:41.356136 4805 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="37658956-d75a-43ce-81d8-38f9bca42957" containerName="barbican-api-log" Dec 03 14:31:41 crc kubenswrapper[4805]: I1203 14:31:41.356143 4805 state_mem.go:107] "Deleted CPUSet assignment" podUID="37658956-d75a-43ce-81d8-38f9bca42957" containerName="barbican-api-log" Dec 03 14:31:41 crc kubenswrapper[4805]: E1203 14:31:41.356151 4805 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="37658956-d75a-43ce-81d8-38f9bca42957" containerName="barbican-api" Dec 03 14:31:41 crc kubenswrapper[4805]: I1203 14:31:41.356157 4805 state_mem.go:107] "Deleted CPUSet assignment" podUID="37658956-d75a-43ce-81d8-38f9bca42957" containerName="barbican-api" Dec 03 14:31:41 crc kubenswrapper[4805]: E1203 14:31:41.356167 4805 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bc7fd7a7-02bc-48e1-9bde-d1680876da37" containerName="init" Dec 03 14:31:41 crc kubenswrapper[4805]: I1203 14:31:41.356173 4805 state_mem.go:107] "Deleted CPUSet assignment" podUID="bc7fd7a7-02bc-48e1-9bde-d1680876da37" containerName="init" Dec 03 14:31:41 crc kubenswrapper[4805]: E1203 14:31:41.356180 4805 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="58c98434-0b4d-4c2d-ac35-109e38f96803" containerName="proxy-httpd" Dec 03 14:31:41 crc kubenswrapper[4805]: I1203 14:31:41.356185 4805 state_mem.go:107] "Deleted CPUSet assignment" podUID="58c98434-0b4d-4c2d-ac35-109e38f96803" containerName="proxy-httpd" Dec 03 14:31:41 crc kubenswrapper[4805]: E1203 14:31:41.356197 4805 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bc7fd7a7-02bc-48e1-9bde-d1680876da37" containerName="dnsmasq-dns" Dec 03 14:31:41 crc kubenswrapper[4805]: I1203 14:31:41.356202 4805 state_mem.go:107] "Deleted CPUSet assignment" podUID="bc7fd7a7-02bc-48e1-9bde-d1680876da37" containerName="dnsmasq-dns" Dec 03 14:31:41 crc kubenswrapper[4805]: E1203 14:31:41.356220 4805 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="58c98434-0b4d-4c2d-ac35-109e38f96803" containerName="ceilometer-central-agent" Dec 03 14:31:41 crc kubenswrapper[4805]: I1203 14:31:41.356228 4805 state_mem.go:107] "Deleted CPUSet assignment" podUID="58c98434-0b4d-4c2d-ac35-109e38f96803" containerName="ceilometer-central-agent" Dec 03 14:31:41 crc kubenswrapper[4805]: E1203 14:31:41.356266 4805 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="58c98434-0b4d-4c2d-ac35-109e38f96803" containerName="sg-core" Dec 03 14:31:41 crc kubenswrapper[4805]: I1203 14:31:41.356282 4805 state_mem.go:107] "Deleted CPUSet assignment" podUID="58c98434-0b4d-4c2d-ac35-109e38f96803" containerName="sg-core" Dec 03 14:31:41 crc kubenswrapper[4805]: I1203 14:31:41.356463 4805 memory_manager.go:354] "RemoveStaleState removing state" podUID="37658956-d75a-43ce-81d8-38f9bca42957" containerName="barbican-api-log" Dec 03 14:31:41 crc kubenswrapper[4805]: I1203 14:31:41.356477 4805 memory_manager.go:354] "RemoveStaleState removing state" podUID="bc7fd7a7-02bc-48e1-9bde-d1680876da37" containerName="dnsmasq-dns" Dec 03 14:31:41 crc kubenswrapper[4805]: I1203 14:31:41.356488 4805 memory_manager.go:354] "RemoveStaleState removing state" podUID="58c98434-0b4d-4c2d-ac35-109e38f96803" containerName="proxy-httpd" Dec 03 14:31:41 crc kubenswrapper[4805]: I1203 14:31:41.356509 4805 memory_manager.go:354] "RemoveStaleState removing state" podUID="58c98434-0b4d-4c2d-ac35-109e38f96803" containerName="sg-core" Dec 03 14:31:41 crc kubenswrapper[4805]: I1203 14:31:41.356519 4805 memory_manager.go:354] "RemoveStaleState removing state" podUID="37658956-d75a-43ce-81d8-38f9bca42957" containerName="barbican-api" Dec 03 14:31:41 crc kubenswrapper[4805]: I1203 14:31:41.356527 4805 memory_manager.go:354] "RemoveStaleState removing state" podUID="58c98434-0b4d-4c2d-ac35-109e38f96803" containerName="ceilometer-notification-agent" Dec 03 14:31:41 crc kubenswrapper[4805]: I1203 14:31:41.356536 4805 memory_manager.go:354] "RemoveStaleState removing state" podUID="58c98434-0b4d-4c2d-ac35-109e38f96803" containerName="ceilometer-central-agent" Dec 03 14:31:41 crc kubenswrapper[4805]: I1203 14:31:41.358244 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 03 14:31:41 crc kubenswrapper[4805]: I1203 14:31:41.360440 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Dec 03 14:31:41 crc kubenswrapper[4805]: I1203 14:31:41.366563 4805 scope.go:117] "RemoveContainer" containerID="9cf635a2329565cc994a3a2337944c2eca2f7047534f27d7eeb5f3ceaa96458b" Dec 03 14:31:41 crc kubenswrapper[4805]: I1203 14:31:41.366616 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Dec 03 14:31:41 crc kubenswrapper[4805]: I1203 14:31:41.380402 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 03 14:31:41 crc kubenswrapper[4805]: I1203 14:31:41.425392 4805 scope.go:117] "RemoveContainer" containerID="f98e610e128d27a8cf1f7ca39fe6a3f759d8784ae234f8e1a4c4d5e228950639" Dec 03 14:31:41 crc kubenswrapper[4805]: I1203 14:31:41.558160 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xcs9q\" (UniqueName: \"kubernetes.io/projected/5979c218-e720-4ce6-aebc-993a3676abb5-kube-api-access-xcs9q\") pod \"ceilometer-0\" (UID: \"5979c218-e720-4ce6-aebc-993a3676abb5\") " pod="openstack/ceilometer-0" Dec 03 14:31:41 crc kubenswrapper[4805]: I1203 14:31:41.558607 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5979c218-e720-4ce6-aebc-993a3676abb5-config-data\") pod \"ceilometer-0\" (UID: \"5979c218-e720-4ce6-aebc-993a3676abb5\") " pod="openstack/ceilometer-0" Dec 03 14:31:41 crc kubenswrapper[4805]: I1203 14:31:41.558753 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/5979c218-e720-4ce6-aebc-993a3676abb5-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"5979c218-e720-4ce6-aebc-993a3676abb5\") " pod="openstack/ceilometer-0" Dec 03 14:31:41 crc kubenswrapper[4805]: I1203 14:31:41.558890 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/5979c218-e720-4ce6-aebc-993a3676abb5-run-httpd\") pod \"ceilometer-0\" (UID: \"5979c218-e720-4ce6-aebc-993a3676abb5\") " pod="openstack/ceilometer-0" Dec 03 14:31:41 crc kubenswrapper[4805]: I1203 14:31:41.559051 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5979c218-e720-4ce6-aebc-993a3676abb5-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"5979c218-e720-4ce6-aebc-993a3676abb5\") " pod="openstack/ceilometer-0" Dec 03 14:31:41 crc kubenswrapper[4805]: I1203 14:31:41.559306 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5979c218-e720-4ce6-aebc-993a3676abb5-scripts\") pod \"ceilometer-0\" (UID: \"5979c218-e720-4ce6-aebc-993a3676abb5\") " pod="openstack/ceilometer-0" Dec 03 14:31:41 crc kubenswrapper[4805]: I1203 14:31:41.559447 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/5979c218-e720-4ce6-aebc-993a3676abb5-log-httpd\") pod \"ceilometer-0\" (UID: \"5979c218-e720-4ce6-aebc-993a3676abb5\") " pod="openstack/ceilometer-0" Dec 03 14:31:41 crc kubenswrapper[4805]: I1203 14:31:41.660954 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xcs9q\" (UniqueName: \"kubernetes.io/projected/5979c218-e720-4ce6-aebc-993a3676abb5-kube-api-access-xcs9q\") pod \"ceilometer-0\" (UID: \"5979c218-e720-4ce6-aebc-993a3676abb5\") " pod="openstack/ceilometer-0" Dec 03 14:31:41 crc kubenswrapper[4805]: I1203 14:31:41.662259 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5979c218-e720-4ce6-aebc-993a3676abb5-config-data\") pod \"ceilometer-0\" (UID: \"5979c218-e720-4ce6-aebc-993a3676abb5\") " pod="openstack/ceilometer-0" Dec 03 14:31:41 crc kubenswrapper[4805]: I1203 14:31:41.663112 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/5979c218-e720-4ce6-aebc-993a3676abb5-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"5979c218-e720-4ce6-aebc-993a3676abb5\") " pod="openstack/ceilometer-0" Dec 03 14:31:41 crc kubenswrapper[4805]: I1203 14:31:41.663266 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/5979c218-e720-4ce6-aebc-993a3676abb5-run-httpd\") pod \"ceilometer-0\" (UID: \"5979c218-e720-4ce6-aebc-993a3676abb5\") " pod="openstack/ceilometer-0" Dec 03 14:31:41 crc kubenswrapper[4805]: I1203 14:31:41.663438 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5979c218-e720-4ce6-aebc-993a3676abb5-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"5979c218-e720-4ce6-aebc-993a3676abb5\") " pod="openstack/ceilometer-0" Dec 03 14:31:41 crc kubenswrapper[4805]: I1203 14:31:41.663732 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5979c218-e720-4ce6-aebc-993a3676abb5-scripts\") pod \"ceilometer-0\" (UID: \"5979c218-e720-4ce6-aebc-993a3676abb5\") " pod="openstack/ceilometer-0" Dec 03 14:31:41 crc kubenswrapper[4805]: I1203 14:31:41.663905 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/5979c218-e720-4ce6-aebc-993a3676abb5-run-httpd\") pod \"ceilometer-0\" (UID: \"5979c218-e720-4ce6-aebc-993a3676abb5\") " pod="openstack/ceilometer-0" Dec 03 14:31:41 crc kubenswrapper[4805]: I1203 14:31:41.664026 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/5979c218-e720-4ce6-aebc-993a3676abb5-log-httpd\") pod \"ceilometer-0\" (UID: \"5979c218-e720-4ce6-aebc-993a3676abb5\") " pod="openstack/ceilometer-0" Dec 03 14:31:41 crc kubenswrapper[4805]: I1203 14:31:41.664760 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/5979c218-e720-4ce6-aebc-993a3676abb5-log-httpd\") pod \"ceilometer-0\" (UID: \"5979c218-e720-4ce6-aebc-993a3676abb5\") " pod="openstack/ceilometer-0" Dec 03 14:31:41 crc kubenswrapper[4805]: I1203 14:31:41.670279 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5979c218-e720-4ce6-aebc-993a3676abb5-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"5979c218-e720-4ce6-aebc-993a3676abb5\") " pod="openstack/ceilometer-0" Dec 03 14:31:41 crc kubenswrapper[4805]: I1203 14:31:41.689157 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5979c218-e720-4ce6-aebc-993a3676abb5-scripts\") pod \"ceilometer-0\" (UID: \"5979c218-e720-4ce6-aebc-993a3676abb5\") " pod="openstack/ceilometer-0" Dec 03 14:31:41 crc kubenswrapper[4805]: I1203 14:31:41.689750 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5979c218-e720-4ce6-aebc-993a3676abb5-config-data\") pod \"ceilometer-0\" (UID: \"5979c218-e720-4ce6-aebc-993a3676abb5\") " pod="openstack/ceilometer-0" Dec 03 14:31:41 crc kubenswrapper[4805]: I1203 14:31:41.690491 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/5979c218-e720-4ce6-aebc-993a3676abb5-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"5979c218-e720-4ce6-aebc-993a3676abb5\") " pod="openstack/ceilometer-0" Dec 03 14:31:41 crc kubenswrapper[4805]: I1203 14:31:41.698723 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xcs9q\" (UniqueName: \"kubernetes.io/projected/5979c218-e720-4ce6-aebc-993a3676abb5-kube-api-access-xcs9q\") pod \"ceilometer-0\" (UID: \"5979c218-e720-4ce6-aebc-993a3676abb5\") " pod="openstack/ceilometer-0" Dec 03 14:31:41 crc kubenswrapper[4805]: I1203 14:31:41.766879 4805 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/heat-cfnapi-795fdbdfbf-8zpm7"] Dec 03 14:31:41 crc kubenswrapper[4805]: I1203 14:31:41.798523 4805 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/heat-api-6bd4dc884b-6qst5"] Dec 03 14:31:41 crc kubenswrapper[4805]: I1203 14:31:41.818419 4805 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/heat-api-84746cbddc-kc7hm"] Dec 03 14:31:41 crc kubenswrapper[4805]: I1203 14:31:41.820247 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-api-84746cbddc-kc7hm" Dec 03 14:31:41 crc kubenswrapper[4805]: I1203 14:31:41.823675 4805 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/heat-cfnapi-68589fbcdd-snszz"] Dec 03 14:31:41 crc kubenswrapper[4805]: I1203 14:31:41.824831 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-cfnapi-68589fbcdd-snszz" Dec 03 14:31:41 crc kubenswrapper[4805]: I1203 14:31:41.829437 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-heat-cfnapi-public-svc" Dec 03 14:31:41 crc kubenswrapper[4805]: I1203 14:31:41.829580 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-heat-api-public-svc" Dec 03 14:31:41 crc kubenswrapper[4805]: I1203 14:31:41.829798 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-heat-cfnapi-internal-svc" Dec 03 14:31:41 crc kubenswrapper[4805]: I1203 14:31:41.829707 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-heat-api-internal-svc" Dec 03 14:31:41 crc kubenswrapper[4805]: I1203 14:31:41.833883 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-cfnapi-68589fbcdd-snszz"] Dec 03 14:31:41 crc kubenswrapper[4805]: I1203 14:31:41.850699 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-api-84746cbddc-kc7hm"] Dec 03 14:31:41 crc kubenswrapper[4805]: I1203 14:31:41.879870 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/cde129c0-8757-4996-8974-c553a5a48f17-config-data-custom\") pod \"heat-cfnapi-68589fbcdd-snszz\" (UID: \"cde129c0-8757-4996-8974-c553a5a48f17\") " pod="openstack/heat-cfnapi-68589fbcdd-snszz" Dec 03 14:31:41 crc kubenswrapper[4805]: I1203 14:31:41.879906 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/078bbd61-70d1-4e2c-ab8c-eeba4127f662-public-tls-certs\") pod \"heat-api-84746cbddc-kc7hm\" (UID: \"078bbd61-70d1-4e2c-ab8c-eeba4127f662\") " pod="openstack/heat-api-84746cbddc-kc7hm" Dec 03 14:31:41 crc kubenswrapper[4805]: I1203 14:31:41.879954 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/078bbd61-70d1-4e2c-ab8c-eeba4127f662-config-data\") pod \"heat-api-84746cbddc-kc7hm\" (UID: \"078bbd61-70d1-4e2c-ab8c-eeba4127f662\") " pod="openstack/heat-api-84746cbddc-kc7hm" Dec 03 14:31:41 crc kubenswrapper[4805]: I1203 14:31:41.879980 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cde129c0-8757-4996-8974-c553a5a48f17-combined-ca-bundle\") pod \"heat-cfnapi-68589fbcdd-snszz\" (UID: \"cde129c0-8757-4996-8974-c553a5a48f17\") " pod="openstack/heat-cfnapi-68589fbcdd-snszz" Dec 03 14:31:41 crc kubenswrapper[4805]: I1203 14:31:41.880002 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/078bbd61-70d1-4e2c-ab8c-eeba4127f662-config-data-custom\") pod \"heat-api-84746cbddc-kc7hm\" (UID: \"078bbd61-70d1-4e2c-ab8c-eeba4127f662\") " pod="openstack/heat-api-84746cbddc-kc7hm" Dec 03 14:31:41 crc kubenswrapper[4805]: I1203 14:31:41.880066 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-l6gbf\" (UniqueName: \"kubernetes.io/projected/cde129c0-8757-4996-8974-c553a5a48f17-kube-api-access-l6gbf\") pod \"heat-cfnapi-68589fbcdd-snszz\" (UID: \"cde129c0-8757-4996-8974-c553a5a48f17\") " pod="openstack/heat-cfnapi-68589fbcdd-snszz" Dec 03 14:31:41 crc kubenswrapper[4805]: I1203 14:31:41.880098 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/cde129c0-8757-4996-8974-c553a5a48f17-public-tls-certs\") pod \"heat-cfnapi-68589fbcdd-snszz\" (UID: \"cde129c0-8757-4996-8974-c553a5a48f17\") " pod="openstack/heat-cfnapi-68589fbcdd-snszz" Dec 03 14:31:41 crc kubenswrapper[4805]: I1203 14:31:41.880117 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cde129c0-8757-4996-8974-c553a5a48f17-config-data\") pod \"heat-cfnapi-68589fbcdd-snszz\" (UID: \"cde129c0-8757-4996-8974-c553a5a48f17\") " pod="openstack/heat-cfnapi-68589fbcdd-snszz" Dec 03 14:31:41 crc kubenswrapper[4805]: I1203 14:31:41.880143 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/cde129c0-8757-4996-8974-c553a5a48f17-internal-tls-certs\") pod \"heat-cfnapi-68589fbcdd-snszz\" (UID: \"cde129c0-8757-4996-8974-c553a5a48f17\") " pod="openstack/heat-cfnapi-68589fbcdd-snszz" Dec 03 14:31:41 crc kubenswrapper[4805]: I1203 14:31:41.880167 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/078bbd61-70d1-4e2c-ab8c-eeba4127f662-combined-ca-bundle\") pod \"heat-api-84746cbddc-kc7hm\" (UID: \"078bbd61-70d1-4e2c-ab8c-eeba4127f662\") " pod="openstack/heat-api-84746cbddc-kc7hm" Dec 03 14:31:41 crc kubenswrapper[4805]: I1203 14:31:41.880184 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4m849\" (UniqueName: \"kubernetes.io/projected/078bbd61-70d1-4e2c-ab8c-eeba4127f662-kube-api-access-4m849\") pod \"heat-api-84746cbddc-kc7hm\" (UID: \"078bbd61-70d1-4e2c-ab8c-eeba4127f662\") " pod="openstack/heat-api-84746cbddc-kc7hm" Dec 03 14:31:41 crc kubenswrapper[4805]: I1203 14:31:41.880219 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/078bbd61-70d1-4e2c-ab8c-eeba4127f662-internal-tls-certs\") pod \"heat-api-84746cbddc-kc7hm\" (UID: \"078bbd61-70d1-4e2c-ab8c-eeba4127f662\") " pod="openstack/heat-api-84746cbddc-kc7hm" Dec 03 14:31:41 crc kubenswrapper[4805]: I1203 14:31:41.983981 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/078bbd61-70d1-4e2c-ab8c-eeba4127f662-internal-tls-certs\") pod \"heat-api-84746cbddc-kc7hm\" (UID: \"078bbd61-70d1-4e2c-ab8c-eeba4127f662\") " pod="openstack/heat-api-84746cbddc-kc7hm" Dec 03 14:31:41 crc kubenswrapper[4805]: I1203 14:31:41.984052 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/cde129c0-8757-4996-8974-c553a5a48f17-config-data-custom\") pod \"heat-cfnapi-68589fbcdd-snszz\" (UID: \"cde129c0-8757-4996-8974-c553a5a48f17\") " pod="openstack/heat-cfnapi-68589fbcdd-snszz" Dec 03 14:31:41 crc kubenswrapper[4805]: I1203 14:31:41.984072 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/078bbd61-70d1-4e2c-ab8c-eeba4127f662-public-tls-certs\") pod \"heat-api-84746cbddc-kc7hm\" (UID: \"078bbd61-70d1-4e2c-ab8c-eeba4127f662\") " pod="openstack/heat-api-84746cbddc-kc7hm" Dec 03 14:31:41 crc kubenswrapper[4805]: I1203 14:31:41.984113 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/078bbd61-70d1-4e2c-ab8c-eeba4127f662-config-data\") pod \"heat-api-84746cbddc-kc7hm\" (UID: \"078bbd61-70d1-4e2c-ab8c-eeba4127f662\") " pod="openstack/heat-api-84746cbddc-kc7hm" Dec 03 14:31:41 crc kubenswrapper[4805]: I1203 14:31:41.984135 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cde129c0-8757-4996-8974-c553a5a48f17-combined-ca-bundle\") pod \"heat-cfnapi-68589fbcdd-snszz\" (UID: \"cde129c0-8757-4996-8974-c553a5a48f17\") " pod="openstack/heat-cfnapi-68589fbcdd-snszz" Dec 03 14:31:41 crc kubenswrapper[4805]: I1203 14:31:41.984156 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/078bbd61-70d1-4e2c-ab8c-eeba4127f662-config-data-custom\") pod \"heat-api-84746cbddc-kc7hm\" (UID: \"078bbd61-70d1-4e2c-ab8c-eeba4127f662\") " pod="openstack/heat-api-84746cbddc-kc7hm" Dec 03 14:31:41 crc kubenswrapper[4805]: I1203 14:31:41.984221 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-l6gbf\" (UniqueName: \"kubernetes.io/projected/cde129c0-8757-4996-8974-c553a5a48f17-kube-api-access-l6gbf\") pod \"heat-cfnapi-68589fbcdd-snszz\" (UID: \"cde129c0-8757-4996-8974-c553a5a48f17\") " pod="openstack/heat-cfnapi-68589fbcdd-snszz" Dec 03 14:31:41 crc kubenswrapper[4805]: I1203 14:31:41.984238 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/cde129c0-8757-4996-8974-c553a5a48f17-public-tls-certs\") pod \"heat-cfnapi-68589fbcdd-snszz\" (UID: \"cde129c0-8757-4996-8974-c553a5a48f17\") " pod="openstack/heat-cfnapi-68589fbcdd-snszz" Dec 03 14:31:41 crc kubenswrapper[4805]: I1203 14:31:41.984259 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cde129c0-8757-4996-8974-c553a5a48f17-config-data\") pod \"heat-cfnapi-68589fbcdd-snszz\" (UID: \"cde129c0-8757-4996-8974-c553a5a48f17\") " pod="openstack/heat-cfnapi-68589fbcdd-snszz" Dec 03 14:31:41 crc kubenswrapper[4805]: I1203 14:31:41.984285 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/cde129c0-8757-4996-8974-c553a5a48f17-internal-tls-certs\") pod \"heat-cfnapi-68589fbcdd-snszz\" (UID: \"cde129c0-8757-4996-8974-c553a5a48f17\") " pod="openstack/heat-cfnapi-68589fbcdd-snszz" Dec 03 14:31:41 crc kubenswrapper[4805]: I1203 14:31:41.984308 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/078bbd61-70d1-4e2c-ab8c-eeba4127f662-combined-ca-bundle\") pod \"heat-api-84746cbddc-kc7hm\" (UID: \"078bbd61-70d1-4e2c-ab8c-eeba4127f662\") " pod="openstack/heat-api-84746cbddc-kc7hm" Dec 03 14:31:41 crc kubenswrapper[4805]: I1203 14:31:41.984325 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4m849\" (UniqueName: \"kubernetes.io/projected/078bbd61-70d1-4e2c-ab8c-eeba4127f662-kube-api-access-4m849\") pod \"heat-api-84746cbddc-kc7hm\" (UID: \"078bbd61-70d1-4e2c-ab8c-eeba4127f662\") " pod="openstack/heat-api-84746cbddc-kc7hm" Dec 03 14:31:41 crc kubenswrapper[4805]: I1203 14:31:41.989019 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/cde129c0-8757-4996-8974-c553a5a48f17-internal-tls-certs\") pod \"heat-cfnapi-68589fbcdd-snszz\" (UID: \"cde129c0-8757-4996-8974-c553a5a48f17\") " pod="openstack/heat-cfnapi-68589fbcdd-snszz" Dec 03 14:31:41 crc kubenswrapper[4805]: I1203 14:31:41.991827 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/078bbd61-70d1-4e2c-ab8c-eeba4127f662-config-data-custom\") pod \"heat-api-84746cbddc-kc7hm\" (UID: \"078bbd61-70d1-4e2c-ab8c-eeba4127f662\") " pod="openstack/heat-api-84746cbddc-kc7hm" Dec 03 14:31:41 crc kubenswrapper[4805]: I1203 14:31:41.992565 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 03 14:31:41 crc kubenswrapper[4805]: I1203 14:31:41.998306 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/cde129c0-8757-4996-8974-c553a5a48f17-public-tls-certs\") pod \"heat-cfnapi-68589fbcdd-snszz\" (UID: \"cde129c0-8757-4996-8974-c553a5a48f17\") " pod="openstack/heat-cfnapi-68589fbcdd-snszz" Dec 03 14:31:41 crc kubenswrapper[4805]: I1203 14:31:41.998424 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/078bbd61-70d1-4e2c-ab8c-eeba4127f662-combined-ca-bundle\") pod \"heat-api-84746cbddc-kc7hm\" (UID: \"078bbd61-70d1-4e2c-ab8c-eeba4127f662\") " pod="openstack/heat-api-84746cbddc-kc7hm" Dec 03 14:31:41 crc kubenswrapper[4805]: I1203 14:31:41.998470 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cde129c0-8757-4996-8974-c553a5a48f17-config-data\") pod \"heat-cfnapi-68589fbcdd-snszz\" (UID: \"cde129c0-8757-4996-8974-c553a5a48f17\") " pod="openstack/heat-cfnapi-68589fbcdd-snszz" Dec 03 14:31:41 crc kubenswrapper[4805]: I1203 14:31:41.998589 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/078bbd61-70d1-4e2c-ab8c-eeba4127f662-internal-tls-certs\") pod \"heat-api-84746cbddc-kc7hm\" (UID: \"078bbd61-70d1-4e2c-ab8c-eeba4127f662\") " pod="openstack/heat-api-84746cbddc-kc7hm" Dec 03 14:31:42 crc kubenswrapper[4805]: I1203 14:31:42.001432 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/078bbd61-70d1-4e2c-ab8c-eeba4127f662-config-data\") pod \"heat-api-84746cbddc-kc7hm\" (UID: \"078bbd61-70d1-4e2c-ab8c-eeba4127f662\") " pod="openstack/heat-api-84746cbddc-kc7hm" Dec 03 14:31:42 crc kubenswrapper[4805]: I1203 14:31:42.001571 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/078bbd61-70d1-4e2c-ab8c-eeba4127f662-public-tls-certs\") pod \"heat-api-84746cbddc-kc7hm\" (UID: \"078bbd61-70d1-4e2c-ab8c-eeba4127f662\") " pod="openstack/heat-api-84746cbddc-kc7hm" Dec 03 14:31:42 crc kubenswrapper[4805]: I1203 14:31:42.002249 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/cde129c0-8757-4996-8974-c553a5a48f17-config-data-custom\") pod \"heat-cfnapi-68589fbcdd-snszz\" (UID: \"cde129c0-8757-4996-8974-c553a5a48f17\") " pod="openstack/heat-cfnapi-68589fbcdd-snszz" Dec 03 14:31:42 crc kubenswrapper[4805]: I1203 14:31:42.004895 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4m849\" (UniqueName: \"kubernetes.io/projected/078bbd61-70d1-4e2c-ab8c-eeba4127f662-kube-api-access-4m849\") pod \"heat-api-84746cbddc-kc7hm\" (UID: \"078bbd61-70d1-4e2c-ab8c-eeba4127f662\") " pod="openstack/heat-api-84746cbddc-kc7hm" Dec 03 14:31:42 crc kubenswrapper[4805]: I1203 14:31:42.005109 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cde129c0-8757-4996-8974-c553a5a48f17-combined-ca-bundle\") pod \"heat-cfnapi-68589fbcdd-snszz\" (UID: \"cde129c0-8757-4996-8974-c553a5a48f17\") " pod="openstack/heat-cfnapi-68589fbcdd-snszz" Dec 03 14:31:42 crc kubenswrapper[4805]: I1203 14:31:42.013597 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-l6gbf\" (UniqueName: \"kubernetes.io/projected/cde129c0-8757-4996-8974-c553a5a48f17-kube-api-access-l6gbf\") pod \"heat-cfnapi-68589fbcdd-snszz\" (UID: \"cde129c0-8757-4996-8974-c553a5a48f17\") " pod="openstack/heat-cfnapi-68589fbcdd-snszz" Dec 03 14:31:42 crc kubenswrapper[4805]: I1203 14:31:42.150153 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-api-84746cbddc-kc7hm" Dec 03 14:31:42 crc kubenswrapper[4805]: I1203 14:31:42.171211 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-cfnapi-68589fbcdd-snszz" Dec 03 14:31:42 crc kubenswrapper[4805]: I1203 14:31:42.358069 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstackclient" event={"ID":"29c53624-ef4b-43ab-8922-181a78908d24","Type":"ContainerStarted","Data":"26262b91874fa88789d54d9ed21dcbd2e4122b6fb279633aaa0a8892e3a82995"} Dec 03 14:31:42 crc kubenswrapper[4805]: W1203 14:31:42.375514 4805 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod1082c862_47c9_4382_b3e8_09e58fb5d85d.slice/crio-6e03529f359e02161466a41b1b8a6d032878bfc2dedca9892e2cfd3e88d733c0 WatchSource:0}: Error finding container 6e03529f359e02161466a41b1b8a6d032878bfc2dedca9892e2cfd3e88d733c0: Status 404 returned error can't find the container with id 6e03529f359e02161466a41b1b8a6d032878bfc2dedca9892e2cfd3e88d733c0 Dec 03 14:31:42 crc kubenswrapper[4805]: I1203 14:31:42.382278 4805 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/heat-cfnapi-795fdbdfbf-8zpm7"] Dec 03 14:31:42 crc kubenswrapper[4805]: I1203 14:31:42.383140 4805 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Dec 03 14:31:42 crc kubenswrapper[4805]: I1203 14:31:42.383377 4805 generic.go:334] "Generic (PLEG): container finished" podID="596937d0-4958-4d99-a677-698ca4b7d4eb" containerID="d095c57ea2be9ec1e49b311636249f5bb64be63c03bb0fa7f97830a9b479f1ab" exitCode=0 Dec 03 14:31:42 crc kubenswrapper[4805]: I1203 14:31:42.383421 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-db-create-b8n7x" event={"ID":"596937d0-4958-4d99-a677-698ca4b7d4eb","Type":"ContainerDied","Data":"d095c57ea2be9ec1e49b311636249f5bb64be63c03bb0fa7f97830a9b479f1ab"} Dec 03 14:31:42 crc kubenswrapper[4805]: I1203 14:31:42.406502 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/29710b09-a1ad-468f-af26-b5d5eb516898-config-data-custom\") pod \"29710b09-a1ad-468f-af26-b5d5eb516898\" (UID: \"29710b09-a1ad-468f-af26-b5d5eb516898\") " Dec 03 14:31:42 crc kubenswrapper[4805]: I1203 14:31:42.406607 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/29710b09-a1ad-468f-af26-b5d5eb516898-config-data\") pod \"29710b09-a1ad-468f-af26-b5d5eb516898\" (UID: \"29710b09-a1ad-468f-af26-b5d5eb516898\") " Dec 03 14:31:42 crc kubenswrapper[4805]: I1203 14:31:42.406628 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/29710b09-a1ad-468f-af26-b5d5eb516898-etc-machine-id\") pod \"29710b09-a1ad-468f-af26-b5d5eb516898\" (UID: \"29710b09-a1ad-468f-af26-b5d5eb516898\") " Dec 03 14:31:42 crc kubenswrapper[4805]: I1203 14:31:42.406774 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qfmb5\" (UniqueName: \"kubernetes.io/projected/29710b09-a1ad-468f-af26-b5d5eb516898-kube-api-access-qfmb5\") pod \"29710b09-a1ad-468f-af26-b5d5eb516898\" (UID: \"29710b09-a1ad-468f-af26-b5d5eb516898\") " Dec 03 14:31:42 crc kubenswrapper[4805]: I1203 14:31:42.406845 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/29710b09-a1ad-468f-af26-b5d5eb516898-scripts\") pod \"29710b09-a1ad-468f-af26-b5d5eb516898\" (UID: \"29710b09-a1ad-468f-af26-b5d5eb516898\") " Dec 03 14:31:42 crc kubenswrapper[4805]: I1203 14:31:42.407202 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/29710b09-a1ad-468f-af26-b5d5eb516898-combined-ca-bundle\") pod \"29710b09-a1ad-468f-af26-b5d5eb516898\" (UID: \"29710b09-a1ad-468f-af26-b5d5eb516898\") " Dec 03 14:31:42 crc kubenswrapper[4805]: I1203 14:31:42.411764 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/29710b09-a1ad-468f-af26-b5d5eb516898-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "29710b09-a1ad-468f-af26-b5d5eb516898" (UID: "29710b09-a1ad-468f-af26-b5d5eb516898"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 03 14:31:42 crc kubenswrapper[4805]: I1203 14:31:42.438219 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-1be1-account-create-update-58nc2"] Dec 03 14:31:42 crc kubenswrapper[4805]: I1203 14:31:42.485625 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/29710b09-a1ad-468f-af26-b5d5eb516898-scripts" (OuterVolumeSpecName: "scripts") pod "29710b09-a1ad-468f-af26-b5d5eb516898" (UID: "29710b09-a1ad-468f-af26-b5d5eb516898"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 14:31:42 crc kubenswrapper[4805]: I1203 14:31:42.485738 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/29710b09-a1ad-468f-af26-b5d5eb516898-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "29710b09-a1ad-468f-af26-b5d5eb516898" (UID: "29710b09-a1ad-468f-af26-b5d5eb516898"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 14:31:42 crc kubenswrapper[4805]: I1203 14:31:42.485820 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/29710b09-a1ad-468f-af26-b5d5eb516898-kube-api-access-qfmb5" (OuterVolumeSpecName: "kube-api-access-qfmb5") pod "29710b09-a1ad-468f-af26-b5d5eb516898" (UID: "29710b09-a1ad-468f-af26-b5d5eb516898"). InnerVolumeSpecName "kube-api-access-qfmb5". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 14:31:42 crc kubenswrapper[4805]: I1203 14:31:42.488342 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"7b5b9f5a-5af6-41d7-be26-3f4c091098b7","Type":"ContainerStarted","Data":"d06e2eaab85f7eab8fb6d5c2d4d78f2ae752881e6f8968e2d1a24ecd1db7fd33"} Dec 03 14:31:42 crc kubenswrapper[4805]: I1203 14:31:42.498372 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-7801-account-create-update-rjbpl"] Dec 03 14:31:42 crc kubenswrapper[4805]: I1203 14:31:42.512567 4805 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/29710b09-a1ad-468f-af26-b5d5eb516898-config-data-custom\") on node \"crc\" DevicePath \"\"" Dec 03 14:31:42 crc kubenswrapper[4805]: I1203 14:31:42.512593 4805 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/29710b09-a1ad-468f-af26-b5d5eb516898-etc-machine-id\") on node \"crc\" DevicePath \"\"" Dec 03 14:31:42 crc kubenswrapper[4805]: I1203 14:31:42.512603 4805 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qfmb5\" (UniqueName: \"kubernetes.io/projected/29710b09-a1ad-468f-af26-b5d5eb516898-kube-api-access-qfmb5\") on node \"crc\" DevicePath \"\"" Dec 03 14:31:42 crc kubenswrapper[4805]: I1203 14:31:42.512612 4805 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/29710b09-a1ad-468f-af26-b5d5eb516898-scripts\") on node \"crc\" DevicePath \"\"" Dec 03 14:31:42 crc kubenswrapper[4805]: I1203 14:31:42.543793 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"29710b09-a1ad-468f-af26-b5d5eb516898","Type":"ContainerDied","Data":"c6b5100cdc92b5421f6d6c7f7bc48f88da1ec74c305b02d496f9ff09c6110c39"} Dec 03 14:31:42 crc kubenswrapper[4805]: I1203 14:31:42.544153 4805 scope.go:117] "RemoveContainer" containerID="95b1ef4794618699f48a7350bb18cecf536b9ae5725dda4c6c96e5147bb28baf" Dec 03 14:31:42 crc kubenswrapper[4805]: I1203 14:31:42.544258 4805 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Dec 03 14:31:42 crc kubenswrapper[4805]: I1203 14:31:42.569260 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-f6bc4c6c9-x5tdg"] Dec 03 14:31:42 crc kubenswrapper[4805]: I1203 14:31:42.583942 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-engine-5fdfc8886c-v7jx4"] Dec 03 14:31:42 crc kubenswrapper[4805]: I1203 14:31:42.586687 4805 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/openstackclient" podStartSLOduration=4.18673193 podStartE2EDuration="21.586674395s" podCreationTimestamp="2025-12-03 14:31:21 +0000 UTC" firstStartedPulling="2025-12-03 14:31:22.684338745 +0000 UTC m=+1312.347255678" lastFinishedPulling="2025-12-03 14:31:40.08428122 +0000 UTC m=+1329.747198143" observedRunningTime="2025-12-03 14:31:42.406448911 +0000 UTC m=+1332.069365844" watchObservedRunningTime="2025-12-03 14:31:42.586674395 +0000 UTC m=+1332.249591318" Dec 03 14:31:42 crc kubenswrapper[4805]: I1203 14:31:42.605099 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/29710b09-a1ad-468f-af26-b5d5eb516898-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "29710b09-a1ad-468f-af26-b5d5eb516898" (UID: "29710b09-a1ad-468f-af26-b5d5eb516898"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 14:31:42 crc kubenswrapper[4805]: I1203 14:31:42.607146 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-db-create-z7rqx"] Dec 03 14:31:42 crc kubenswrapper[4805]: I1203 14:31:42.616156 4805 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/29710b09-a1ad-468f-af26-b5d5eb516898-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 14:31:42 crc kubenswrapper[4805]: I1203 14:31:42.642991 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-f183-account-create-update-p7mnx"] Dec 03 14:31:42 crc kubenswrapper[4805]: I1203 14:31:42.803410 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/29710b09-a1ad-468f-af26-b5d5eb516898-config-data" (OuterVolumeSpecName: "config-data") pod "29710b09-a1ad-468f-af26-b5d5eb516898" (UID: "29710b09-a1ad-468f-af26-b5d5eb516898"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 14:31:43 crc kubenswrapper[4805]: I1203 14:31:42.820316 4805 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/29710b09-a1ad-468f-af26-b5d5eb516898-config-data\") on node \"crc\" DevicePath \"\"" Dec 03 14:31:43 crc kubenswrapper[4805]: I1203 14:31:42.995979 4805 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="37658956-d75a-43ce-81d8-38f9bca42957" path="/var/lib/kubelet/pods/37658956-d75a-43ce-81d8-38f9bca42957/volumes" Dec 03 14:31:43 crc kubenswrapper[4805]: I1203 14:31:42.997726 4805 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="58c98434-0b4d-4c2d-ac35-109e38f96803" path="/var/lib/kubelet/pods/58c98434-0b4d-4c2d-ac35-109e38f96803/volumes" Dec 03 14:31:43 crc kubenswrapper[4805]: I1203 14:31:42.998722 4805 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bc7fd7a7-02bc-48e1-9bde-d1680876da37" path="/var/lib/kubelet/pods/bc7fd7a7-02bc-48e1-9bde-d1680876da37/volumes" Dec 03 14:31:43 crc kubenswrapper[4805]: I1203 14:31:43.004211 4805 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/heat-api-6bd4dc884b-6qst5"] Dec 03 14:31:43 crc kubenswrapper[4805]: I1203 14:31:43.004245 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-db-create-bc8gs"] Dec 03 14:31:43 crc kubenswrapper[4805]: I1203 14:31:43.004266 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-engine-76dcfc959c-swfs4"] Dec 03 14:31:43 crc kubenswrapper[4805]: I1203 14:31:43.004278 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-cfnapi-6cd94c587b-nshpc"] Dec 03 14:31:43 crc kubenswrapper[4805]: I1203 14:31:43.004295 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-api-69cb5bcd7b-bwpl9"] Dec 03 14:31:43 crc kubenswrapper[4805]: I1203 14:31:43.004306 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-proxy-78bdd85b87-5tf8n"] Dec 03 14:31:43 crc kubenswrapper[4805]: I1203 14:31:43.004321 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 03 14:31:43 crc kubenswrapper[4805]: I1203 14:31:43.015057 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-api-84746cbddc-kc7hm"] Dec 03 14:31:43 crc kubenswrapper[4805]: W1203 14:31:43.075375 4805 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod5979c218_e720_4ce6_aebc_993a3676abb5.slice/crio-6f9ec724498267419adc91978a0bbeb50431ae3447528e3c8b439db73170246c WatchSource:0}: Error finding container 6f9ec724498267419adc91978a0bbeb50431ae3447528e3c8b439db73170246c: Status 404 returned error can't find the container with id 6f9ec724498267419adc91978a0bbeb50431ae3447528e3c8b439db73170246c Dec 03 14:31:43 crc kubenswrapper[4805]: W1203 14:31:43.092367 4805 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod078bbd61_70d1_4e2c_ab8c_eeba4127f662.slice/crio-313a1792e0aea0e6d75caea1996a240ae203a037fd15b2bc228fa04de0dcaee0 WatchSource:0}: Error finding container 313a1792e0aea0e6d75caea1996a240ae203a037fd15b2bc228fa04de0dcaee0: Status 404 returned error can't find the container with id 313a1792e0aea0e6d75caea1996a240ae203a037fd15b2bc228fa04de0dcaee0 Dec 03 14:31:43 crc kubenswrapper[4805]: W1203 14:31:43.198546 4805 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podcde129c0_8757_4996_8974_c553a5a48f17.slice/crio-5db567d5b1eaed9cca533cffe2466ec896e438f73c2823689fc3faa7c1532c94 WatchSource:0}: Error finding container 5db567d5b1eaed9cca533cffe2466ec896e438f73c2823689fc3faa7c1532c94: Status 404 returned error can't find the container with id 5db567d5b1eaed9cca533cffe2466ec896e438f73c2823689fc3faa7c1532c94 Dec 03 14:31:43 crc kubenswrapper[4805]: I1203 14:31:43.214222 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-cfnapi-68589fbcdd-snszz"] Dec 03 14:31:43 crc kubenswrapper[4805]: I1203 14:31:43.394128 4805 scope.go:117] "RemoveContainer" containerID="cb75dea47e869ccbdcc4f2f309b4ef5bddded1bea1f31601b29cb63b5646b269" Dec 03 14:31:43 crc kubenswrapper[4805]: I1203 14:31:43.673527 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-7801-account-create-update-rjbpl" event={"ID":"5c694d3e-75d4-40d4-924d-cfe9e9ef5ef7","Type":"ContainerStarted","Data":"a8c3fe16a920663ad4e9720c693e14d5b0da7bfafc20f36f482f2d5220183b1d"} Dec 03 14:31:43 crc kubenswrapper[4805]: I1203 14:31:43.682456 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-1be1-account-create-update-58nc2" event={"ID":"1082c862-47c9-4382-b3e8-09e58fb5d85d","Type":"ContainerStarted","Data":"832cf42c86bbb5add1b9f92b1884cf830d1330476cffa86cd81f8731103c155a"} Dec 03 14:31:43 crc kubenswrapper[4805]: I1203 14:31:43.682495 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-1be1-account-create-update-58nc2" event={"ID":"1082c862-47c9-4382-b3e8-09e58fb5d85d","Type":"ContainerStarted","Data":"6e03529f359e02161466a41b1b8a6d032878bfc2dedca9892e2cfd3e88d733c0"} Dec 03 14:31:43 crc kubenswrapper[4805]: I1203 14:31:43.685739 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-f183-account-create-update-p7mnx" event={"ID":"0c5c12d6-db52-4367-8898-e88fc8cdde41","Type":"ContainerStarted","Data":"40bf98d3f171d99179acd99a215011038ee2992502d009f80c701bf38be239b7"} Dec 03 14:31:43 crc kubenswrapper[4805]: I1203 14:31:43.685767 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-f183-account-create-update-p7mnx" event={"ID":"0c5c12d6-db52-4367-8898-e88fc8cdde41","Type":"ContainerStarted","Data":"d3d208eafa428495a749150830f2e68b980b2dcfc0dfdf3a6cf80529fea10dc5"} Dec 03 14:31:43 crc kubenswrapper[4805]: I1203 14:31:43.696422 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-api-84746cbddc-kc7hm" event={"ID":"078bbd61-70d1-4e2c-ab8c-eeba4127f662","Type":"ContainerStarted","Data":"313a1792e0aea0e6d75caea1996a240ae203a037fd15b2bc228fa04de0dcaee0"} Dec 03 14:31:43 crc kubenswrapper[4805]: I1203 14:31:43.700241 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-engine-5fdfc8886c-v7jx4" event={"ID":"36af56b0-4335-48cf-82ba-bd95a50b14f3","Type":"ContainerStarted","Data":"3024bb1954bfb673162d931d82304a21f3d101c3872b3723478056576de1e7b1"} Dec 03 14:31:43 crc kubenswrapper[4805]: I1203 14:31:43.709877 4805 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-1be1-account-create-update-58nc2" podStartSLOduration=9.709857482 podStartE2EDuration="9.709857482s" podCreationTimestamp="2025-12-03 14:31:34 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 14:31:43.701389 +0000 UTC m=+1333.364305923" watchObservedRunningTime="2025-12-03 14:31:43.709857482 +0000 UTC m=+1333.372774405" Dec 03 14:31:43 crc kubenswrapper[4805]: I1203 14:31:43.721297 4805 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-f183-account-create-update-p7mnx" podStartSLOduration=8.721275225 podStartE2EDuration="8.721275225s" podCreationTimestamp="2025-12-03 14:31:35 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 14:31:43.720464573 +0000 UTC m=+1333.383381496" watchObservedRunningTime="2025-12-03 14:31:43.721275225 +0000 UTC m=+1333.384192148" Dec 03 14:31:43 crc kubenswrapper[4805]: I1203 14:31:43.727243 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-cfnapi-68589fbcdd-snszz" event={"ID":"cde129c0-8757-4996-8974-c553a5a48f17","Type":"ContainerStarted","Data":"5db567d5b1eaed9cca533cffe2466ec896e438f73c2823689fc3faa7c1532c94"} Dec 03 14:31:43 crc kubenswrapper[4805]: I1203 14:31:43.738006 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-api-69cb5bcd7b-bwpl9" event={"ID":"756ba517-3759-4ee7-8718-a5d8c6f1be72","Type":"ContainerStarted","Data":"3847c2681419a98a4e00e1e300c9371a8336a1eec4300e25686a7fd2393ebbdb"} Dec 03 14:31:43 crc kubenswrapper[4805]: I1203 14:31:43.743497 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-f6bc4c6c9-x5tdg" event={"ID":"4994372a-ac42-4661-aed6-734070bee008","Type":"ContainerStarted","Data":"32cefd53191c106c90ee4ec6f5b5281b248b8f25bcf06f4ae10cb2966c32886a"} Dec 03 14:31:43 crc kubenswrapper[4805]: I1203 14:31:43.743532 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-f6bc4c6c9-x5tdg" event={"ID":"4994372a-ac42-4661-aed6-734070bee008","Type":"ContainerStarted","Data":"57e94c3359e37674bb98418990700796598ac64a12d8b76df0724f15c0471c86"} Dec 03 14:31:43 crc kubenswrapper[4805]: I1203 14:31:43.765096 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-db-create-bc8gs" event={"ID":"1245f02d-543a-40d6-bc19-745a48249604","Type":"ContainerStarted","Data":"f9d62dfb3ecb9d58cef08a2970ae697f48a75d92ad55d701ecdf1e75087b7aa7"} Dec 03 14:31:43 crc kubenswrapper[4805]: I1203 14:31:43.791745 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-api-6bd4dc884b-6qst5" event={"ID":"c4431968-a92e-4fc0-951f-f42a15942f33","Type":"ContainerStarted","Data":"bf0e3e07687b91d9e6e73348065d3d6132523b5624cd87d8e1623d8a16fb1bfd"} Dec 03 14:31:43 crc kubenswrapper[4805]: I1203 14:31:43.795080 4805 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell0-db-create-bc8gs" podStartSLOduration=9.795059185 podStartE2EDuration="9.795059185s" podCreationTimestamp="2025-12-03 14:31:34 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 14:31:43.78792975 +0000 UTC m=+1333.450846683" watchObservedRunningTime="2025-12-03 14:31:43.795059185 +0000 UTC m=+1333.457976108" Dec 03 14:31:43 crc kubenswrapper[4805]: I1203 14:31:43.803816 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-cfnapi-6cd94c587b-nshpc" event={"ID":"d1d26c95-d05a-4551-8d00-655941ce122b","Type":"ContainerStarted","Data":"df8e2f594cef67002f327c841b44bde459211f47621558cf1d44e38124a4c18a"} Dec 03 14:31:43 crc kubenswrapper[4805]: I1203 14:31:43.816575 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"7b5b9f5a-5af6-41d7-be26-3f4c091098b7","Type":"ContainerStarted","Data":"87dc7be41fa7f3166e62e4b517ff61a5c62f2d797f9004dacd694688dc118eb7"} Dec 03 14:31:43 crc kubenswrapper[4805]: I1203 14:31:43.816979 4805 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/cinder-api-0" Dec 03 14:31:43 crc kubenswrapper[4805]: I1203 14:31:43.841057 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-cfnapi-795fdbdfbf-8zpm7" event={"ID":"33d42113-07b3-4db3-95b8-f43810991ce4","Type":"ContainerStarted","Data":"782aa0b2673f8a838bacd094fec9c0de3f9f91e5da1d5fe1ec5913918509d105"} Dec 03 14:31:43 crc kubenswrapper[4805]: I1203 14:31:43.844552 4805 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-api-0" podStartSLOduration=12.844525849 podStartE2EDuration="12.844525849s" podCreationTimestamp="2025-12-03 14:31:31 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 14:31:43.838497274 +0000 UTC m=+1333.501414207" watchObservedRunningTime="2025-12-03 14:31:43.844525849 +0000 UTC m=+1333.507442792" Dec 03 14:31:43 crc kubenswrapper[4805]: I1203 14:31:43.857624 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"5979c218-e720-4ce6-aebc-993a3676abb5","Type":"ContainerStarted","Data":"6f9ec724498267419adc91978a0bbeb50431ae3447528e3c8b439db73170246c"} Dec 03 14:31:43 crc kubenswrapper[4805]: I1203 14:31:43.878093 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-db-create-z7rqx" event={"ID":"649e0d3a-332c-43d7-90bd-8ad10015268f","Type":"ContainerStarted","Data":"2d5e284cc6cd04fe014f3491c8154c9cb2d92ee7ff0d3f65dde2650adaf2e294"} Dec 03 14:31:43 crc kubenswrapper[4805]: I1203 14:31:43.919276 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-engine-76dcfc959c-swfs4" event={"ID":"854cde2d-c692-4ab4-9172-02ad6a2b2f6a","Type":"ContainerStarted","Data":"1fc6b296d0d28edef1a51855de9dd504816edaee6f87587095146c02053d6e75"} Dec 03 14:31:43 crc kubenswrapper[4805]: I1203 14:31:43.919325 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-engine-76dcfc959c-swfs4" event={"ID":"854cde2d-c692-4ab4-9172-02ad6a2b2f6a","Type":"ContainerStarted","Data":"9a7095781ac670f2203edbec8bae26d2a2f6cf159067bf7e02f94bfddb808d3b"} Dec 03 14:31:43 crc kubenswrapper[4805]: I1203 14:31:43.919370 4805 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/heat-engine-76dcfc959c-swfs4" Dec 03 14:31:43 crc kubenswrapper[4805]: I1203 14:31:43.937034 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-proxy-78bdd85b87-5tf8n" event={"ID":"0fdb9aee-e509-4e04-88b7-61fc216cfc34","Type":"ContainerStarted","Data":"eee22f1c0079a0057bb758ba1f0adb950a211ede3c9e145fa7adf9e4bc6e1bd6"} Dec 03 14:31:44 crc kubenswrapper[4805]: I1203 14:31:44.026390 4805 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-db-create-z7rqx" podStartSLOduration=10.026365837 podStartE2EDuration="10.026365837s" podCreationTimestamp="2025-12-03 14:31:34 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 14:31:43.900783309 +0000 UTC m=+1333.563700242" watchObservedRunningTime="2025-12-03 14:31:44.026365837 +0000 UTC m=+1333.689282760" Dec 03 14:31:44 crc kubenswrapper[4805]: I1203 14:31:44.036175 4805 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/heat-engine-76dcfc959c-swfs4" podStartSLOduration=5.036157525 podStartE2EDuration="5.036157525s" podCreationTimestamp="2025-12-03 14:31:39 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 14:31:43.951478517 +0000 UTC m=+1333.614395440" watchObservedRunningTime="2025-12-03 14:31:44.036157525 +0000 UTC m=+1333.699074448" Dec 03 14:31:44 crc kubenswrapper[4805]: I1203 14:31:44.346497 4805 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-5784cf869f-l4bt9" podUID="bc7fd7a7-02bc-48e1-9bde-d1680876da37" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.163:5353: i/o timeout" Dec 03 14:31:44 crc kubenswrapper[4805]: I1203 14:31:44.504914 4805 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-db-create-b8n7x" Dec 03 14:31:44 crc kubenswrapper[4805]: I1203 14:31:44.565451 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2kg82\" (UniqueName: \"kubernetes.io/projected/596937d0-4958-4d99-a677-698ca4b7d4eb-kube-api-access-2kg82\") pod \"596937d0-4958-4d99-a677-698ca4b7d4eb\" (UID: \"596937d0-4958-4d99-a677-698ca4b7d4eb\") " Dec 03 14:31:44 crc kubenswrapper[4805]: I1203 14:31:44.565544 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/596937d0-4958-4d99-a677-698ca4b7d4eb-operator-scripts\") pod \"596937d0-4958-4d99-a677-698ca4b7d4eb\" (UID: \"596937d0-4958-4d99-a677-698ca4b7d4eb\") " Dec 03 14:31:44 crc kubenswrapper[4805]: I1203 14:31:44.567233 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/596937d0-4958-4d99-a677-698ca4b7d4eb-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "596937d0-4958-4d99-a677-698ca4b7d4eb" (UID: "596937d0-4958-4d99-a677-698ca4b7d4eb"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 14:31:44 crc kubenswrapper[4805]: I1203 14:31:44.579610 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/596937d0-4958-4d99-a677-698ca4b7d4eb-kube-api-access-2kg82" (OuterVolumeSpecName: "kube-api-access-2kg82") pod "596937d0-4958-4d99-a677-698ca4b7d4eb" (UID: "596937d0-4958-4d99-a677-698ca4b7d4eb"). InnerVolumeSpecName "kube-api-access-2kg82". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 14:31:44 crc kubenswrapper[4805]: I1203 14:31:44.667173 4805 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2kg82\" (UniqueName: \"kubernetes.io/projected/596937d0-4958-4d99-a677-698ca4b7d4eb-kube-api-access-2kg82\") on node \"crc\" DevicePath \"\"" Dec 03 14:31:44 crc kubenswrapper[4805]: I1203 14:31:44.667217 4805 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/596937d0-4958-4d99-a677-698ca4b7d4eb-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 03 14:31:44 crc kubenswrapper[4805]: I1203 14:31:44.949448 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-db-create-b8n7x" event={"ID":"596937d0-4958-4d99-a677-698ca4b7d4eb","Type":"ContainerDied","Data":"24391668ed2d2228fb3f9b78b7218ec1e72bcd0841b46649a97e17468974fb65"} Dec 03 14:31:44 crc kubenswrapper[4805]: I1203 14:31:44.949747 4805 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="24391668ed2d2228fb3f9b78b7218ec1e72bcd0841b46649a97e17468974fb65" Dec 03 14:31:44 crc kubenswrapper[4805]: I1203 14:31:44.949465 4805 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-db-create-b8n7x" Dec 03 14:31:44 crc kubenswrapper[4805]: I1203 14:31:44.951813 4805 generic.go:334] "Generic (PLEG): container finished" podID="649e0d3a-332c-43d7-90bd-8ad10015268f" containerID="0ea9106d14a822eebf60173804f34994b649001c2b6d78c9d19a9f0dc68877ff" exitCode=0 Dec 03 14:31:44 crc kubenswrapper[4805]: I1203 14:31:44.951915 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-db-create-z7rqx" event={"ID":"649e0d3a-332c-43d7-90bd-8ad10015268f","Type":"ContainerDied","Data":"0ea9106d14a822eebf60173804f34994b649001c2b6d78c9d19a9f0dc68877ff"} Dec 03 14:31:44 crc kubenswrapper[4805]: I1203 14:31:44.960172 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-proxy-78bdd85b87-5tf8n" event={"ID":"0fdb9aee-e509-4e04-88b7-61fc216cfc34","Type":"ContainerStarted","Data":"b4513578b27d39e5d1b93811d393cca3a43d4d6af773297ccd97bcadecf13295"} Dec 03 14:31:44 crc kubenswrapper[4805]: I1203 14:31:44.960217 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-proxy-78bdd85b87-5tf8n" event={"ID":"0fdb9aee-e509-4e04-88b7-61fc216cfc34","Type":"ContainerStarted","Data":"e764133dd2f70c2989069432891b751c61b34ffe0f11084aab2d13f5feb5b1d5"} Dec 03 14:31:44 crc kubenswrapper[4805]: I1203 14:31:44.960260 4805 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/swift-proxy-78bdd85b87-5tf8n" Dec 03 14:31:44 crc kubenswrapper[4805]: I1203 14:31:44.960301 4805 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/swift-proxy-78bdd85b87-5tf8n" Dec 03 14:31:44 crc kubenswrapper[4805]: I1203 14:31:44.987132 4805 generic.go:334] "Generic (PLEG): container finished" podID="4994372a-ac42-4661-aed6-734070bee008" containerID="32cefd53191c106c90ee4ec6f5b5281b248b8f25bcf06f4ae10cb2966c32886a" exitCode=0 Dec 03 14:31:44 crc kubenswrapper[4805]: I1203 14:31:44.988113 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-f6bc4c6c9-x5tdg" event={"ID":"4994372a-ac42-4661-aed6-734070bee008","Type":"ContainerDied","Data":"32cefd53191c106c90ee4ec6f5b5281b248b8f25bcf06f4ae10cb2966c32886a"} Dec 03 14:31:44 crc kubenswrapper[4805]: I1203 14:31:44.988172 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-f6bc4c6c9-x5tdg" event={"ID":"4994372a-ac42-4661-aed6-734070bee008","Type":"ContainerStarted","Data":"fda7f3c053dd3dd390f549cc83ac2bcd36e8561e014152b04f3277f127e4ceef"} Dec 03 14:31:44 crc kubenswrapper[4805]: I1203 14:31:44.988193 4805 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-f6bc4c6c9-x5tdg" Dec 03 14:31:44 crc kubenswrapper[4805]: I1203 14:31:44.995959 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"5979c218-e720-4ce6-aebc-993a3676abb5","Type":"ContainerStarted","Data":"0766a2a356b4c585109ee5f02dd5def6ec3f30d6e59c15941ad33bbd31cf88ba"} Dec 03 14:31:44 crc kubenswrapper[4805]: I1203 14:31:44.996003 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"5979c218-e720-4ce6-aebc-993a3676abb5","Type":"ContainerStarted","Data":"2bb62034f23ff306eba03cc77fbd7f0b60e195607023bac74cfc090ef18a15c9"} Dec 03 14:31:44 crc kubenswrapper[4805]: I1203 14:31:44.998289 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-engine-5fdfc8886c-v7jx4" event={"ID":"36af56b0-4335-48cf-82ba-bd95a50b14f3","Type":"ContainerStarted","Data":"3d42d71bed5afbdee9635a53cd2836c52522e7333a4cbda0a4d910e814f3ae57"} Dec 03 14:31:44 crc kubenswrapper[4805]: I1203 14:31:44.999080 4805 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/heat-engine-5fdfc8886c-v7jx4" Dec 03 14:31:45 crc kubenswrapper[4805]: I1203 14:31:45.002115 4805 generic.go:334] "Generic (PLEG): container finished" podID="5c694d3e-75d4-40d4-924d-cfe9e9ef5ef7" containerID="eda2ac15752bbe039e0576b6fba2c9ca5e27ae5381c8e8d3815692ad5a460f9b" exitCode=0 Dec 03 14:31:45 crc kubenswrapper[4805]: I1203 14:31:45.002195 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-7801-account-create-update-rjbpl" event={"ID":"5c694d3e-75d4-40d4-924d-cfe9e9ef5ef7","Type":"ContainerDied","Data":"eda2ac15752bbe039e0576b6fba2c9ca5e27ae5381c8e8d3815692ad5a460f9b"} Dec 03 14:31:45 crc kubenswrapper[4805]: I1203 14:31:45.008703 4805 generic.go:334] "Generic (PLEG): container finished" podID="1082c862-47c9-4382-b3e8-09e58fb5d85d" containerID="832cf42c86bbb5add1b9f92b1884cf830d1330476cffa86cd81f8731103c155a" exitCode=0 Dec 03 14:31:45 crc kubenswrapper[4805]: I1203 14:31:45.008775 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-1be1-account-create-update-58nc2" event={"ID":"1082c862-47c9-4382-b3e8-09e58fb5d85d","Type":"ContainerDied","Data":"832cf42c86bbb5add1b9f92b1884cf830d1330476cffa86cd81f8731103c155a"} Dec 03 14:31:45 crc kubenswrapper[4805]: I1203 14:31:45.009604 4805 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/swift-proxy-78bdd85b87-5tf8n" podStartSLOduration=14.009583824 podStartE2EDuration="14.009583824s" podCreationTimestamp="2025-12-03 14:31:31 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 14:31:44.993589256 +0000 UTC m=+1334.656506179" watchObservedRunningTime="2025-12-03 14:31:45.009583824 +0000 UTC m=+1334.672500737" Dec 03 14:31:45 crc kubenswrapper[4805]: I1203 14:31:45.017236 4805 generic.go:334] "Generic (PLEG): container finished" podID="0c5c12d6-db52-4367-8898-e88fc8cdde41" containerID="40bf98d3f171d99179acd99a215011038ee2992502d009f80c701bf38be239b7" exitCode=0 Dec 03 14:31:45 crc kubenswrapper[4805]: I1203 14:31:45.017342 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-f183-account-create-update-p7mnx" event={"ID":"0c5c12d6-db52-4367-8898-e88fc8cdde41","Type":"ContainerDied","Data":"40bf98d3f171d99179acd99a215011038ee2992502d009f80c701bf38be239b7"} Dec 03 14:31:45 crc kubenswrapper[4805]: I1203 14:31:45.020265 4805 generic.go:334] "Generic (PLEG): container finished" podID="1245f02d-543a-40d6-bc19-745a48249604" containerID="1ec90286975a522ff11e0855f75e40651d133383eee6046c19747bea819bb89d" exitCode=0 Dec 03 14:31:45 crc kubenswrapper[4805]: I1203 14:31:45.020389 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-db-create-bc8gs" event={"ID":"1245f02d-543a-40d6-bc19-745a48249604","Type":"ContainerDied","Data":"1ec90286975a522ff11e0855f75e40651d133383eee6046c19747bea819bb89d"} Dec 03 14:31:45 crc kubenswrapper[4805]: I1203 14:31:45.035599 4805 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/heat-engine-5fdfc8886c-v7jx4" podStartSLOduration=12.035574685 podStartE2EDuration="12.035574685s" podCreationTimestamp="2025-12-03 14:31:33 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 14:31:45.017437509 +0000 UTC m=+1334.680354422" watchObservedRunningTime="2025-12-03 14:31:45.035574685 +0000 UTC m=+1334.698491618" Dec 03 14:31:45 crc kubenswrapper[4805]: I1203 14:31:45.052454 4805 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-f6bc4c6c9-x5tdg" podStartSLOduration=12.052433277 podStartE2EDuration="12.052433277s" podCreationTimestamp="2025-12-03 14:31:33 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 14:31:45.034317561 +0000 UTC m=+1334.697234484" watchObservedRunningTime="2025-12-03 14:31:45.052433277 +0000 UTC m=+1334.715350200" Dec 03 14:31:47 crc kubenswrapper[4805]: I1203 14:31:47.706934 4805 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-db-create-bc8gs" Dec 03 14:31:47 crc kubenswrapper[4805]: I1203 14:31:47.767731 4805 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-db-create-z7rqx" Dec 03 14:31:47 crc kubenswrapper[4805]: I1203 14:31:47.829959 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/1245f02d-543a-40d6-bc19-745a48249604-operator-scripts\") pod \"1245f02d-543a-40d6-bc19-745a48249604\" (UID: \"1245f02d-543a-40d6-bc19-745a48249604\") " Dec 03 14:31:47 crc kubenswrapper[4805]: I1203 14:31:47.830094 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nmhpq\" (UniqueName: \"kubernetes.io/projected/1245f02d-543a-40d6-bc19-745a48249604-kube-api-access-nmhpq\") pod \"1245f02d-543a-40d6-bc19-745a48249604\" (UID: \"1245f02d-543a-40d6-bc19-745a48249604\") " Dec 03 14:31:47 crc kubenswrapper[4805]: I1203 14:31:47.830396 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1245f02d-543a-40d6-bc19-745a48249604-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "1245f02d-543a-40d6-bc19-745a48249604" (UID: "1245f02d-543a-40d6-bc19-745a48249604"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 14:31:47 crc kubenswrapper[4805]: I1203 14:31:47.830660 4805 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/1245f02d-543a-40d6-bc19-745a48249604-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 03 14:31:47 crc kubenswrapper[4805]: I1203 14:31:47.837099 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1245f02d-543a-40d6-bc19-745a48249604-kube-api-access-nmhpq" (OuterVolumeSpecName: "kube-api-access-nmhpq") pod "1245f02d-543a-40d6-bc19-745a48249604" (UID: "1245f02d-543a-40d6-bc19-745a48249604"). InnerVolumeSpecName "kube-api-access-nmhpq". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 14:31:47 crc kubenswrapper[4805]: I1203 14:31:47.931818 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fqfx6\" (UniqueName: \"kubernetes.io/projected/649e0d3a-332c-43d7-90bd-8ad10015268f-kube-api-access-fqfx6\") pod \"649e0d3a-332c-43d7-90bd-8ad10015268f\" (UID: \"649e0d3a-332c-43d7-90bd-8ad10015268f\") " Dec 03 14:31:47 crc kubenswrapper[4805]: I1203 14:31:47.931961 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/649e0d3a-332c-43d7-90bd-8ad10015268f-operator-scripts\") pod \"649e0d3a-332c-43d7-90bd-8ad10015268f\" (UID: \"649e0d3a-332c-43d7-90bd-8ad10015268f\") " Dec 03 14:31:47 crc kubenswrapper[4805]: I1203 14:31:47.932409 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/649e0d3a-332c-43d7-90bd-8ad10015268f-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "649e0d3a-332c-43d7-90bd-8ad10015268f" (UID: "649e0d3a-332c-43d7-90bd-8ad10015268f"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 14:31:47 crc kubenswrapper[4805]: I1203 14:31:47.932826 4805 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/649e0d3a-332c-43d7-90bd-8ad10015268f-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 03 14:31:47 crc kubenswrapper[4805]: I1203 14:31:47.932867 4805 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nmhpq\" (UniqueName: \"kubernetes.io/projected/1245f02d-543a-40d6-bc19-745a48249604-kube-api-access-nmhpq\") on node \"crc\" DevicePath \"\"" Dec 03 14:31:47 crc kubenswrapper[4805]: I1203 14:31:47.934950 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/649e0d3a-332c-43d7-90bd-8ad10015268f-kube-api-access-fqfx6" (OuterVolumeSpecName: "kube-api-access-fqfx6") pod "649e0d3a-332c-43d7-90bd-8ad10015268f" (UID: "649e0d3a-332c-43d7-90bd-8ad10015268f"). InnerVolumeSpecName "kube-api-access-fqfx6". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 14:31:48 crc kubenswrapper[4805]: I1203 14:31:48.037113 4805 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fqfx6\" (UniqueName: \"kubernetes.io/projected/649e0d3a-332c-43d7-90bd-8ad10015268f-kube-api-access-fqfx6\") on node \"crc\" DevicePath \"\"" Dec 03 14:31:48 crc kubenswrapper[4805]: I1203 14:31:48.056141 4805 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-7801-account-create-update-rjbpl" Dec 03 14:31:48 crc kubenswrapper[4805]: I1203 14:31:48.063437 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-7801-account-create-update-rjbpl" event={"ID":"5c694d3e-75d4-40d4-924d-cfe9e9ef5ef7","Type":"ContainerDied","Data":"a8c3fe16a920663ad4e9720c693e14d5b0da7bfafc20f36f482f2d5220183b1d"} Dec 03 14:31:48 crc kubenswrapper[4805]: I1203 14:31:48.063472 4805 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="a8c3fe16a920663ad4e9720c693e14d5b0da7bfafc20f36f482f2d5220183b1d" Dec 03 14:31:48 crc kubenswrapper[4805]: I1203 14:31:48.064257 4805 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-f183-account-create-update-p7mnx" Dec 03 14:31:48 crc kubenswrapper[4805]: I1203 14:31:48.068821 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-1be1-account-create-update-58nc2" event={"ID":"1082c862-47c9-4382-b3e8-09e58fb5d85d","Type":"ContainerDied","Data":"6e03529f359e02161466a41b1b8a6d032878bfc2dedca9892e2cfd3e88d733c0"} Dec 03 14:31:48 crc kubenswrapper[4805]: I1203 14:31:48.068878 4805 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="6e03529f359e02161466a41b1b8a6d032878bfc2dedca9892e2cfd3e88d733c0" Dec 03 14:31:48 crc kubenswrapper[4805]: I1203 14:31:48.075343 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-db-create-z7rqx" event={"ID":"649e0d3a-332c-43d7-90bd-8ad10015268f","Type":"ContainerDied","Data":"2d5e284cc6cd04fe014f3491c8154c9cb2d92ee7ff0d3f65dde2650adaf2e294"} Dec 03 14:31:48 crc kubenswrapper[4805]: I1203 14:31:48.075378 4805 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="2d5e284cc6cd04fe014f3491c8154c9cb2d92ee7ff0d3f65dde2650adaf2e294" Dec 03 14:31:48 crc kubenswrapper[4805]: I1203 14:31:48.075435 4805 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-db-create-z7rqx" Dec 03 14:31:48 crc kubenswrapper[4805]: I1203 14:31:48.081545 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-f183-account-create-update-p7mnx" event={"ID":"0c5c12d6-db52-4367-8898-e88fc8cdde41","Type":"ContainerDied","Data":"d3d208eafa428495a749150830f2e68b980b2dcfc0dfdf3a6cf80529fea10dc5"} Dec 03 14:31:48 crc kubenswrapper[4805]: I1203 14:31:48.081674 4805 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="d3d208eafa428495a749150830f2e68b980b2dcfc0dfdf3a6cf80529fea10dc5" Dec 03 14:31:48 crc kubenswrapper[4805]: I1203 14:31:48.081816 4805 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-f183-account-create-update-p7mnx" Dec 03 14:31:48 crc kubenswrapper[4805]: I1203 14:31:48.085311 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-db-create-bc8gs" event={"ID":"1245f02d-543a-40d6-bc19-745a48249604","Type":"ContainerDied","Data":"f9d62dfb3ecb9d58cef08a2970ae697f48a75d92ad55d701ecdf1e75087b7aa7"} Dec 03 14:31:48 crc kubenswrapper[4805]: I1203 14:31:48.085349 4805 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="f9d62dfb3ecb9d58cef08a2970ae697f48a75d92ad55d701ecdf1e75087b7aa7" Dec 03 14:31:48 crc kubenswrapper[4805]: I1203 14:31:48.085404 4805 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-db-create-bc8gs" Dec 03 14:31:48 crc kubenswrapper[4805]: I1203 14:31:48.084997 4805 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-1be1-account-create-update-58nc2" Dec 03 14:31:48 crc kubenswrapper[4805]: I1203 14:31:48.240973 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-k5wks\" (UniqueName: \"kubernetes.io/projected/5c694d3e-75d4-40d4-924d-cfe9e9ef5ef7-kube-api-access-k5wks\") pod \"5c694d3e-75d4-40d4-924d-cfe9e9ef5ef7\" (UID: \"5c694d3e-75d4-40d4-924d-cfe9e9ef5ef7\") " Dec 03 14:31:48 crc kubenswrapper[4805]: I1203 14:31:48.241064 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/0c5c12d6-db52-4367-8898-e88fc8cdde41-operator-scripts\") pod \"0c5c12d6-db52-4367-8898-e88fc8cdde41\" (UID: \"0c5c12d6-db52-4367-8898-e88fc8cdde41\") " Dec 03 14:31:48 crc kubenswrapper[4805]: I1203 14:31:48.241151 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/1082c862-47c9-4382-b3e8-09e58fb5d85d-operator-scripts\") pod \"1082c862-47c9-4382-b3e8-09e58fb5d85d\" (UID: \"1082c862-47c9-4382-b3e8-09e58fb5d85d\") " Dec 03 14:31:48 crc kubenswrapper[4805]: I1203 14:31:48.241290 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-z82gm\" (UniqueName: \"kubernetes.io/projected/0c5c12d6-db52-4367-8898-e88fc8cdde41-kube-api-access-z82gm\") pod \"0c5c12d6-db52-4367-8898-e88fc8cdde41\" (UID: \"0c5c12d6-db52-4367-8898-e88fc8cdde41\") " Dec 03 14:31:48 crc kubenswrapper[4805]: I1203 14:31:48.241360 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/5c694d3e-75d4-40d4-924d-cfe9e9ef5ef7-operator-scripts\") pod \"5c694d3e-75d4-40d4-924d-cfe9e9ef5ef7\" (UID: \"5c694d3e-75d4-40d4-924d-cfe9e9ef5ef7\") " Dec 03 14:31:48 crc kubenswrapper[4805]: I1203 14:31:48.241390 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-n6mc6\" (UniqueName: \"kubernetes.io/projected/1082c862-47c9-4382-b3e8-09e58fb5d85d-kube-api-access-n6mc6\") pod \"1082c862-47c9-4382-b3e8-09e58fb5d85d\" (UID: \"1082c862-47c9-4382-b3e8-09e58fb5d85d\") " Dec 03 14:31:48 crc kubenswrapper[4805]: I1203 14:31:48.243703 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1082c862-47c9-4382-b3e8-09e58fb5d85d-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "1082c862-47c9-4382-b3e8-09e58fb5d85d" (UID: "1082c862-47c9-4382-b3e8-09e58fb5d85d"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 14:31:48 crc kubenswrapper[4805]: I1203 14:31:48.244537 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5c694d3e-75d4-40d4-924d-cfe9e9ef5ef7-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "5c694d3e-75d4-40d4-924d-cfe9e9ef5ef7" (UID: "5c694d3e-75d4-40d4-924d-cfe9e9ef5ef7"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 14:31:48 crc kubenswrapper[4805]: I1203 14:31:48.245302 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0c5c12d6-db52-4367-8898-e88fc8cdde41-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "0c5c12d6-db52-4367-8898-e88fc8cdde41" (UID: "0c5c12d6-db52-4367-8898-e88fc8cdde41"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 14:31:48 crc kubenswrapper[4805]: I1203 14:31:48.246078 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1082c862-47c9-4382-b3e8-09e58fb5d85d-kube-api-access-n6mc6" (OuterVolumeSpecName: "kube-api-access-n6mc6") pod "1082c862-47c9-4382-b3e8-09e58fb5d85d" (UID: "1082c862-47c9-4382-b3e8-09e58fb5d85d"). InnerVolumeSpecName "kube-api-access-n6mc6". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 14:31:48 crc kubenswrapper[4805]: I1203 14:31:48.258598 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5c694d3e-75d4-40d4-924d-cfe9e9ef5ef7-kube-api-access-k5wks" (OuterVolumeSpecName: "kube-api-access-k5wks") pod "5c694d3e-75d4-40d4-924d-cfe9e9ef5ef7" (UID: "5c694d3e-75d4-40d4-924d-cfe9e9ef5ef7"). InnerVolumeSpecName "kube-api-access-k5wks". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 14:31:48 crc kubenswrapper[4805]: I1203 14:31:48.261100 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0c5c12d6-db52-4367-8898-e88fc8cdde41-kube-api-access-z82gm" (OuterVolumeSpecName: "kube-api-access-z82gm") pod "0c5c12d6-db52-4367-8898-e88fc8cdde41" (UID: "0c5c12d6-db52-4367-8898-e88fc8cdde41"). InnerVolumeSpecName "kube-api-access-z82gm". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 14:31:48 crc kubenswrapper[4805]: I1203 14:31:48.344020 4805 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-z82gm\" (UniqueName: \"kubernetes.io/projected/0c5c12d6-db52-4367-8898-e88fc8cdde41-kube-api-access-z82gm\") on node \"crc\" DevicePath \"\"" Dec 03 14:31:48 crc kubenswrapper[4805]: I1203 14:31:48.344059 4805 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/5c694d3e-75d4-40d4-924d-cfe9e9ef5ef7-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 03 14:31:48 crc kubenswrapper[4805]: I1203 14:31:48.344072 4805 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-n6mc6\" (UniqueName: \"kubernetes.io/projected/1082c862-47c9-4382-b3e8-09e58fb5d85d-kube-api-access-n6mc6\") on node \"crc\" DevicePath \"\"" Dec 03 14:31:48 crc kubenswrapper[4805]: I1203 14:31:48.344086 4805 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-k5wks\" (UniqueName: \"kubernetes.io/projected/5c694d3e-75d4-40d4-924d-cfe9e9ef5ef7-kube-api-access-k5wks\") on node \"crc\" DevicePath \"\"" Dec 03 14:31:48 crc kubenswrapper[4805]: I1203 14:31:48.344098 4805 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/0c5c12d6-db52-4367-8898-e88fc8cdde41-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 03 14:31:48 crc kubenswrapper[4805]: I1203 14:31:48.344108 4805 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/1082c862-47c9-4382-b3e8-09e58fb5d85d-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 03 14:31:49 crc kubenswrapper[4805]: I1203 14:31:49.093910 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-api-69cb5bcd7b-bwpl9" event={"ID":"756ba517-3759-4ee7-8718-a5d8c6f1be72","Type":"ContainerStarted","Data":"8eb73b26a6d1c034ac11da46581575c2940d67f72c07a7b5867123e276715fa8"} Dec 03 14:31:49 crc kubenswrapper[4805]: I1203 14:31:49.093979 4805 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-7801-account-create-update-rjbpl" Dec 03 14:31:49 crc kubenswrapper[4805]: I1203 14:31:49.093965 4805 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-1be1-account-create-update-58nc2" Dec 03 14:31:49 crc kubenswrapper[4805]: I1203 14:31:49.094436 4805 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/heat-api-69cb5bcd7b-bwpl9" Dec 03 14:31:49 crc kubenswrapper[4805]: I1203 14:31:49.123589 4805 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/heat-api-69cb5bcd7b-bwpl9" podStartSLOduration=5.224103182 podStartE2EDuration="10.123564077s" podCreationTimestamp="2025-12-03 14:31:39 +0000 UTC" firstStartedPulling="2025-12-03 14:31:43.094518348 +0000 UTC m=+1332.757435271" lastFinishedPulling="2025-12-03 14:31:47.993979243 +0000 UTC m=+1337.656896166" observedRunningTime="2025-12-03 14:31:49.112029341 +0000 UTC m=+1338.774946274" watchObservedRunningTime="2025-12-03 14:31:49.123564077 +0000 UTC m=+1338.786481000" Dec 03 14:31:49 crc kubenswrapper[4805]: I1203 14:31:49.812955 4805 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 03 14:31:51 crc kubenswrapper[4805]: I1203 14:31:51.847275 4805 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-conductor-db-sync-dxrx8"] Dec 03 14:31:51 crc kubenswrapper[4805]: E1203 14:31:51.847963 4805 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="649e0d3a-332c-43d7-90bd-8ad10015268f" containerName="mariadb-database-create" Dec 03 14:31:51 crc kubenswrapper[4805]: I1203 14:31:51.847977 4805 state_mem.go:107] "Deleted CPUSet assignment" podUID="649e0d3a-332c-43d7-90bd-8ad10015268f" containerName="mariadb-database-create" Dec 03 14:31:51 crc kubenswrapper[4805]: E1203 14:31:51.847990 4805 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0c5c12d6-db52-4367-8898-e88fc8cdde41" containerName="mariadb-account-create-update" Dec 03 14:31:51 crc kubenswrapper[4805]: I1203 14:31:51.847997 4805 state_mem.go:107] "Deleted CPUSet assignment" podUID="0c5c12d6-db52-4367-8898-e88fc8cdde41" containerName="mariadb-account-create-update" Dec 03 14:31:51 crc kubenswrapper[4805]: E1203 14:31:51.848019 4805 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1082c862-47c9-4382-b3e8-09e58fb5d85d" containerName="mariadb-account-create-update" Dec 03 14:31:51 crc kubenswrapper[4805]: I1203 14:31:51.848027 4805 state_mem.go:107] "Deleted CPUSet assignment" podUID="1082c862-47c9-4382-b3e8-09e58fb5d85d" containerName="mariadb-account-create-update" Dec 03 14:31:51 crc kubenswrapper[4805]: E1203 14:31:51.848039 4805 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5c694d3e-75d4-40d4-924d-cfe9e9ef5ef7" containerName="mariadb-account-create-update" Dec 03 14:31:51 crc kubenswrapper[4805]: I1203 14:31:51.848046 4805 state_mem.go:107] "Deleted CPUSet assignment" podUID="5c694d3e-75d4-40d4-924d-cfe9e9ef5ef7" containerName="mariadb-account-create-update" Dec 03 14:31:51 crc kubenswrapper[4805]: E1203 14:31:51.848059 4805 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1245f02d-543a-40d6-bc19-745a48249604" containerName="mariadb-database-create" Dec 03 14:31:51 crc kubenswrapper[4805]: I1203 14:31:51.848068 4805 state_mem.go:107] "Deleted CPUSet assignment" podUID="1245f02d-543a-40d6-bc19-745a48249604" containerName="mariadb-database-create" Dec 03 14:31:51 crc kubenswrapper[4805]: E1203 14:31:51.848079 4805 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="29710b09-a1ad-468f-af26-b5d5eb516898" containerName="probe" Dec 03 14:31:51 crc kubenswrapper[4805]: I1203 14:31:51.848086 4805 state_mem.go:107] "Deleted CPUSet assignment" podUID="29710b09-a1ad-468f-af26-b5d5eb516898" containerName="probe" Dec 03 14:31:51 crc kubenswrapper[4805]: E1203 14:31:51.848101 4805 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="596937d0-4958-4d99-a677-698ca4b7d4eb" containerName="mariadb-database-create" Dec 03 14:31:51 crc kubenswrapper[4805]: I1203 14:31:51.848109 4805 state_mem.go:107] "Deleted CPUSet assignment" podUID="596937d0-4958-4d99-a677-698ca4b7d4eb" containerName="mariadb-database-create" Dec 03 14:31:51 crc kubenswrapper[4805]: E1203 14:31:51.848125 4805 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="29710b09-a1ad-468f-af26-b5d5eb516898" containerName="cinder-scheduler" Dec 03 14:31:51 crc kubenswrapper[4805]: I1203 14:31:51.848132 4805 state_mem.go:107] "Deleted CPUSet assignment" podUID="29710b09-a1ad-468f-af26-b5d5eb516898" containerName="cinder-scheduler" Dec 03 14:31:51 crc kubenswrapper[4805]: I1203 14:31:51.848382 4805 memory_manager.go:354] "RemoveStaleState removing state" podUID="29710b09-a1ad-468f-af26-b5d5eb516898" containerName="probe" Dec 03 14:31:51 crc kubenswrapper[4805]: I1203 14:31:51.848396 4805 memory_manager.go:354] "RemoveStaleState removing state" podUID="596937d0-4958-4d99-a677-698ca4b7d4eb" containerName="mariadb-database-create" Dec 03 14:31:51 crc kubenswrapper[4805]: I1203 14:31:51.848409 4805 memory_manager.go:354] "RemoveStaleState removing state" podUID="1245f02d-543a-40d6-bc19-745a48249604" containerName="mariadb-database-create" Dec 03 14:31:51 crc kubenswrapper[4805]: I1203 14:31:51.848423 4805 memory_manager.go:354] "RemoveStaleState removing state" podUID="1082c862-47c9-4382-b3e8-09e58fb5d85d" containerName="mariadb-account-create-update" Dec 03 14:31:51 crc kubenswrapper[4805]: I1203 14:31:51.848435 4805 memory_manager.go:354] "RemoveStaleState removing state" podUID="5c694d3e-75d4-40d4-924d-cfe9e9ef5ef7" containerName="mariadb-account-create-update" Dec 03 14:31:51 crc kubenswrapper[4805]: I1203 14:31:51.848448 4805 memory_manager.go:354] "RemoveStaleState removing state" podUID="649e0d3a-332c-43d7-90bd-8ad10015268f" containerName="mariadb-database-create" Dec 03 14:31:51 crc kubenswrapper[4805]: I1203 14:31:51.848463 4805 memory_manager.go:354] "RemoveStaleState removing state" podUID="0c5c12d6-db52-4367-8898-e88fc8cdde41" containerName="mariadb-account-create-update" Dec 03 14:31:51 crc kubenswrapper[4805]: I1203 14:31:51.848481 4805 memory_manager.go:354] "RemoveStaleState removing state" podUID="29710b09-a1ad-468f-af26-b5d5eb516898" containerName="cinder-scheduler" Dec 03 14:31:51 crc kubenswrapper[4805]: I1203 14:31:51.849247 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-db-sync-dxrx8" Dec 03 14:31:51 crc kubenswrapper[4805]: I1203 14:31:51.851270 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-conductor-scripts" Dec 03 14:31:51 crc kubenswrapper[4805]: I1203 14:31:51.851985 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-conductor-config-data" Dec 03 14:31:51 crc kubenswrapper[4805]: I1203 14:31:51.855982 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-nova-dockercfg-qktk6" Dec 03 14:31:51 crc kubenswrapper[4805]: I1203 14:31:51.869936 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-db-sync-dxrx8"] Dec 03 14:31:52 crc kubenswrapper[4805]: I1203 14:31:52.011869 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e56c5750-e0c6-4427-ae8c-ab4af0c9b7d0-config-data\") pod \"nova-cell0-conductor-db-sync-dxrx8\" (UID: \"e56c5750-e0c6-4427-ae8c-ab4af0c9b7d0\") " pod="openstack/nova-cell0-conductor-db-sync-dxrx8" Dec 03 14:31:52 crc kubenswrapper[4805]: I1203 14:31:52.012126 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e56c5750-e0c6-4427-ae8c-ab4af0c9b7d0-scripts\") pod \"nova-cell0-conductor-db-sync-dxrx8\" (UID: \"e56c5750-e0c6-4427-ae8c-ab4af0c9b7d0\") " pod="openstack/nova-cell0-conductor-db-sync-dxrx8" Dec 03 14:31:52 crc kubenswrapper[4805]: I1203 14:31:52.012230 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e56c5750-e0c6-4427-ae8c-ab4af0c9b7d0-combined-ca-bundle\") pod \"nova-cell0-conductor-db-sync-dxrx8\" (UID: \"e56c5750-e0c6-4427-ae8c-ab4af0c9b7d0\") " pod="openstack/nova-cell0-conductor-db-sync-dxrx8" Dec 03 14:31:52 crc kubenswrapper[4805]: I1203 14:31:52.012296 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9dr74\" (UniqueName: \"kubernetes.io/projected/e56c5750-e0c6-4427-ae8c-ab4af0c9b7d0-kube-api-access-9dr74\") pod \"nova-cell0-conductor-db-sync-dxrx8\" (UID: \"e56c5750-e0c6-4427-ae8c-ab4af0c9b7d0\") " pod="openstack/nova-cell0-conductor-db-sync-dxrx8" Dec 03 14:31:52 crc kubenswrapper[4805]: I1203 14:31:52.113702 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e56c5750-e0c6-4427-ae8c-ab4af0c9b7d0-scripts\") pod \"nova-cell0-conductor-db-sync-dxrx8\" (UID: \"e56c5750-e0c6-4427-ae8c-ab4af0c9b7d0\") " pod="openstack/nova-cell0-conductor-db-sync-dxrx8" Dec 03 14:31:52 crc kubenswrapper[4805]: I1203 14:31:52.113755 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e56c5750-e0c6-4427-ae8c-ab4af0c9b7d0-combined-ca-bundle\") pod \"nova-cell0-conductor-db-sync-dxrx8\" (UID: \"e56c5750-e0c6-4427-ae8c-ab4af0c9b7d0\") " pod="openstack/nova-cell0-conductor-db-sync-dxrx8" Dec 03 14:31:52 crc kubenswrapper[4805]: I1203 14:31:52.113779 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9dr74\" (UniqueName: \"kubernetes.io/projected/e56c5750-e0c6-4427-ae8c-ab4af0c9b7d0-kube-api-access-9dr74\") pod \"nova-cell0-conductor-db-sync-dxrx8\" (UID: \"e56c5750-e0c6-4427-ae8c-ab4af0c9b7d0\") " pod="openstack/nova-cell0-conductor-db-sync-dxrx8" Dec 03 14:31:52 crc kubenswrapper[4805]: I1203 14:31:52.113924 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e56c5750-e0c6-4427-ae8c-ab4af0c9b7d0-config-data\") pod \"nova-cell0-conductor-db-sync-dxrx8\" (UID: \"e56c5750-e0c6-4427-ae8c-ab4af0c9b7d0\") " pod="openstack/nova-cell0-conductor-db-sync-dxrx8" Dec 03 14:31:52 crc kubenswrapper[4805]: I1203 14:31:52.122749 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e56c5750-e0c6-4427-ae8c-ab4af0c9b7d0-scripts\") pod \"nova-cell0-conductor-db-sync-dxrx8\" (UID: \"e56c5750-e0c6-4427-ae8c-ab4af0c9b7d0\") " pod="openstack/nova-cell0-conductor-db-sync-dxrx8" Dec 03 14:31:52 crc kubenswrapper[4805]: I1203 14:31:52.123576 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e56c5750-e0c6-4427-ae8c-ab4af0c9b7d0-config-data\") pod \"nova-cell0-conductor-db-sync-dxrx8\" (UID: \"e56c5750-e0c6-4427-ae8c-ab4af0c9b7d0\") " pod="openstack/nova-cell0-conductor-db-sync-dxrx8" Dec 03 14:31:52 crc kubenswrapper[4805]: I1203 14:31:52.124463 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e56c5750-e0c6-4427-ae8c-ab4af0c9b7d0-combined-ca-bundle\") pod \"nova-cell0-conductor-db-sync-dxrx8\" (UID: \"e56c5750-e0c6-4427-ae8c-ab4af0c9b7d0\") " pod="openstack/nova-cell0-conductor-db-sync-dxrx8" Dec 03 14:31:52 crc kubenswrapper[4805]: I1203 14:31:52.141876 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9dr74\" (UniqueName: \"kubernetes.io/projected/e56c5750-e0c6-4427-ae8c-ab4af0c9b7d0-kube-api-access-9dr74\") pod \"nova-cell0-conductor-db-sync-dxrx8\" (UID: \"e56c5750-e0c6-4427-ae8c-ab4af0c9b7d0\") " pod="openstack/nova-cell0-conductor-db-sync-dxrx8" Dec 03 14:31:52 crc kubenswrapper[4805]: I1203 14:31:52.143275 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-cfnapi-6cd94c587b-nshpc" event={"ID":"d1d26c95-d05a-4551-8d00-655941ce122b","Type":"ContainerStarted","Data":"04fafceedea4c7029772a31c84f7c0fbe34ab906c19eadb096709755e15243c0"} Dec 03 14:31:52 crc kubenswrapper[4805]: I1203 14:31:52.143331 4805 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/heat-cfnapi-6cd94c587b-nshpc" Dec 03 14:31:52 crc kubenswrapper[4805]: I1203 14:31:52.145453 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-api-84746cbddc-kc7hm" event={"ID":"078bbd61-70d1-4e2c-ab8c-eeba4127f662","Type":"ContainerStarted","Data":"89c83b2d552ec1be99173b38d728e49b01167ba94f0713f2be128fc11a44c55a"} Dec 03 14:31:52 crc kubenswrapper[4805]: I1203 14:31:52.145503 4805 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/heat-api-84746cbddc-kc7hm" Dec 03 14:31:52 crc kubenswrapper[4805]: I1203 14:31:52.147646 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-cfnapi-795fdbdfbf-8zpm7" event={"ID":"33d42113-07b3-4db3-95b8-f43810991ce4","Type":"ContainerStarted","Data":"2bde7c7fcdbea9aa9f1f6642978c89d9478ebabcfe3db1f02634badb979b5b32"} Dec 03 14:31:52 crc kubenswrapper[4805]: I1203 14:31:52.147730 4805 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/heat-cfnapi-795fdbdfbf-8zpm7" podUID="33d42113-07b3-4db3-95b8-f43810991ce4" containerName="heat-cfnapi" containerID="cri-o://2bde7c7fcdbea9aa9f1f6642978c89d9478ebabcfe3db1f02634badb979b5b32" gracePeriod=60 Dec 03 14:31:52 crc kubenswrapper[4805]: I1203 14:31:52.147760 4805 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/heat-cfnapi-795fdbdfbf-8zpm7" Dec 03 14:31:52 crc kubenswrapper[4805]: I1203 14:31:52.153870 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"5979c218-e720-4ce6-aebc-993a3676abb5","Type":"ContainerStarted","Data":"61ecf9c7ac9b0bc53e3380d461ce4698dd8336362b53387b5832742de7bdc4f8"} Dec 03 14:31:52 crc kubenswrapper[4805]: I1203 14:31:52.161998 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-cfnapi-68589fbcdd-snszz" event={"ID":"cde129c0-8757-4996-8974-c553a5a48f17","Type":"ContainerStarted","Data":"d130ae9c417ed524204f2839040557ba89d89413bfbcaaf9160e5e3ee18e33ab"} Dec 03 14:31:52 crc kubenswrapper[4805]: I1203 14:31:52.162134 4805 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/heat-cfnapi-68589fbcdd-snszz" Dec 03 14:31:52 crc kubenswrapper[4805]: I1203 14:31:52.164175 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-api-6bd4dc884b-6qst5" event={"ID":"c4431968-a92e-4fc0-951f-f42a15942f33","Type":"ContainerStarted","Data":"0b0c768c0ff7e0144c8b1fd624455711a096a5d37b363a0e79eec761e0566c8b"} Dec 03 14:31:52 crc kubenswrapper[4805]: I1203 14:31:52.164363 4805 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/heat-api-6bd4dc884b-6qst5" Dec 03 14:31:52 crc kubenswrapper[4805]: I1203 14:31:52.164418 4805 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/heat-api-6bd4dc884b-6qst5" podUID="c4431968-a92e-4fc0-951f-f42a15942f33" containerName="heat-api" containerID="cri-o://0b0c768c0ff7e0144c8b1fd624455711a096a5d37b363a0e79eec761e0566c8b" gracePeriod=60 Dec 03 14:31:52 crc kubenswrapper[4805]: I1203 14:31:52.165640 4805 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/heat-cfnapi-6cd94c587b-nshpc" podStartSLOduration=8.176631429 podStartE2EDuration="13.165632025s" podCreationTimestamp="2025-12-03 14:31:39 +0000 UTC" firstStartedPulling="2025-12-03 14:31:43.045818265 +0000 UTC m=+1332.708735188" lastFinishedPulling="2025-12-03 14:31:48.034818861 +0000 UTC m=+1337.697735784" observedRunningTime="2025-12-03 14:31:52.162289404 +0000 UTC m=+1341.825206337" watchObservedRunningTime="2025-12-03 14:31:52.165632025 +0000 UTC m=+1341.828548948" Dec 03 14:31:52 crc kubenswrapper[4805]: I1203 14:31:52.169442 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-db-sync-dxrx8" Dec 03 14:31:52 crc kubenswrapper[4805]: I1203 14:31:52.173701 4805 generic.go:334] "Generic (PLEG): container finished" podID="756ba517-3759-4ee7-8718-a5d8c6f1be72" containerID="8eb73b26a6d1c034ac11da46581575c2940d67f72c07a7b5867123e276715fa8" exitCode=1 Dec 03 14:31:52 crc kubenswrapper[4805]: I1203 14:31:52.173744 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-api-69cb5bcd7b-bwpl9" event={"ID":"756ba517-3759-4ee7-8718-a5d8c6f1be72","Type":"ContainerDied","Data":"8eb73b26a6d1c034ac11da46581575c2940d67f72c07a7b5867123e276715fa8"} Dec 03 14:31:52 crc kubenswrapper[4805]: I1203 14:31:52.174401 4805 scope.go:117] "RemoveContainer" containerID="8eb73b26a6d1c034ac11da46581575c2940d67f72c07a7b5867123e276715fa8" Dec 03 14:31:52 crc kubenswrapper[4805]: I1203 14:31:52.199412 4805 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/heat-cfnapi-68589fbcdd-snszz" podStartSLOduration=6.416811293 podStartE2EDuration="11.199389239s" podCreationTimestamp="2025-12-03 14:31:41 +0000 UTC" firstStartedPulling="2025-12-03 14:31:43.211383057 +0000 UTC m=+1332.874299970" lastFinishedPulling="2025-12-03 14:31:47.993960993 +0000 UTC m=+1337.656877916" observedRunningTime="2025-12-03 14:31:52.196095089 +0000 UTC m=+1341.859012012" watchObservedRunningTime="2025-12-03 14:31:52.199389239 +0000 UTC m=+1341.862306162" Dec 03 14:31:52 crc kubenswrapper[4805]: I1203 14:31:52.256983 4805 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/heat-api-84746cbddc-kc7hm" podStartSLOduration=6.350145039 podStartE2EDuration="11.256958135s" podCreationTimestamp="2025-12-03 14:31:41 +0000 UTC" firstStartedPulling="2025-12-03 14:31:43.111535154 +0000 UTC m=+1332.774452077" lastFinishedPulling="2025-12-03 14:31:48.01834825 +0000 UTC m=+1337.681265173" observedRunningTime="2025-12-03 14:31:52.24032165 +0000 UTC m=+1341.903238573" watchObservedRunningTime="2025-12-03 14:31:52.256958135 +0000 UTC m=+1341.919875058" Dec 03 14:31:52 crc kubenswrapper[4805]: I1203 14:31:52.288805 4805 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/heat-cfnapi-795fdbdfbf-8zpm7" podStartSLOduration=14.337145983 podStartE2EDuration="19.288782887s" podCreationTimestamp="2025-12-03 14:31:33 +0000 UTC" firstStartedPulling="2025-12-03 14:31:43.094931419 +0000 UTC m=+1332.757848342" lastFinishedPulling="2025-12-03 14:31:48.046568323 +0000 UTC m=+1337.709485246" observedRunningTime="2025-12-03 14:31:52.274218138 +0000 UTC m=+1341.937135081" watchObservedRunningTime="2025-12-03 14:31:52.288782887 +0000 UTC m=+1341.951699810" Dec 03 14:31:52 crc kubenswrapper[4805]: I1203 14:31:52.311342 4805 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/swift-proxy-78bdd85b87-5tf8n" Dec 03 14:31:52 crc kubenswrapper[4805]: I1203 14:31:52.312645 4805 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/swift-proxy-78bdd85b87-5tf8n" Dec 03 14:31:52 crc kubenswrapper[4805]: I1203 14:31:52.357009 4805 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/heat-api-6bd4dc884b-6qst5" podStartSLOduration=14.35494376 podStartE2EDuration="19.356987384s" podCreationTimestamp="2025-12-03 14:31:33 +0000 UTC" firstStartedPulling="2025-12-03 14:31:43.042632807 +0000 UTC m=+1332.705549730" lastFinishedPulling="2025-12-03 14:31:48.044676441 +0000 UTC m=+1337.707593354" observedRunningTime="2025-12-03 14:31:52.339415033 +0000 UTC m=+1342.002331956" watchObservedRunningTime="2025-12-03 14:31:52.356987384 +0000 UTC m=+1342.019904307" Dec 03 14:31:52 crc kubenswrapper[4805]: I1203 14:31:52.882134 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-db-sync-dxrx8"] Dec 03 14:31:52 crc kubenswrapper[4805]: W1203 14:31:52.891664 4805 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pode56c5750_e0c6_4427_ae8c_ab4af0c9b7d0.slice/crio-e7b3280c5b3ca26f01d3143c66f2f958b2bef41c241de233af9997ce89459f0a WatchSource:0}: Error finding container e7b3280c5b3ca26f01d3143c66f2f958b2bef41c241de233af9997ce89459f0a: Status 404 returned error can't find the container with id e7b3280c5b3ca26f01d3143c66f2f958b2bef41c241de233af9997ce89459f0a Dec 03 14:31:53 crc kubenswrapper[4805]: I1203 14:31:53.184455 4805 generic.go:334] "Generic (PLEG): container finished" podID="d1d26c95-d05a-4551-8d00-655941ce122b" containerID="04fafceedea4c7029772a31c84f7c0fbe34ab906c19eadb096709755e15243c0" exitCode=1 Dec 03 14:31:53 crc kubenswrapper[4805]: I1203 14:31:53.184525 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-cfnapi-6cd94c587b-nshpc" event={"ID":"d1d26c95-d05a-4551-8d00-655941ce122b","Type":"ContainerDied","Data":"04fafceedea4c7029772a31c84f7c0fbe34ab906c19eadb096709755e15243c0"} Dec 03 14:31:53 crc kubenswrapper[4805]: I1203 14:31:53.185917 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-db-sync-dxrx8" event={"ID":"e56c5750-e0c6-4427-ae8c-ab4af0c9b7d0","Type":"ContainerStarted","Data":"e7b3280c5b3ca26f01d3143c66f2f958b2bef41c241de233af9997ce89459f0a"} Dec 03 14:31:53 crc kubenswrapper[4805]: I1203 14:31:53.186513 4805 scope.go:117] "RemoveContainer" containerID="04fafceedea4c7029772a31c84f7c0fbe34ab906c19eadb096709755e15243c0" Dec 03 14:31:53 crc kubenswrapper[4805]: I1203 14:31:53.570426 4805 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/heat-engine-5fdfc8886c-v7jx4" Dec 03 14:31:53 crc kubenswrapper[4805]: I1203 14:31:53.663114 4805 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-f6bc4c6c9-x5tdg" Dec 03 14:31:53 crc kubenswrapper[4805]: I1203 14:31:53.756579 4805 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-75c8ddd69c-s5r44"] Dec 03 14:31:53 crc kubenswrapper[4805]: I1203 14:31:53.756869 4805 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-75c8ddd69c-s5r44" podUID="9f381959-280e-4eb1-9fb2-2365db2f9c43" containerName="dnsmasq-dns" containerID="cri-o://0f465542ed3b0873cf31f730f8258e4061dbb10f0ebce0003303d33cd5b14544" gracePeriod=10 Dec 03 14:31:54 crc kubenswrapper[4805]: I1203 14:31:54.213872 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-cfnapi-6cd94c587b-nshpc" event={"ID":"d1d26c95-d05a-4551-8d00-655941ce122b","Type":"ContainerStarted","Data":"8cb15466711e9cb141150e29e558fefcad0141bf93d2e6383f7505500faacd21"} Dec 03 14:31:54 crc kubenswrapper[4805]: I1203 14:31:54.214535 4805 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/heat-cfnapi-6cd94c587b-nshpc" Dec 03 14:31:54 crc kubenswrapper[4805]: I1203 14:31:54.223645 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"5979c218-e720-4ce6-aebc-993a3676abb5","Type":"ContainerStarted","Data":"c5b0c5fc24db36ce396804d2d37d6255b90861754bbc2f992c426edf5f800276"} Dec 03 14:31:54 crc kubenswrapper[4805]: I1203 14:31:54.223831 4805 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="5979c218-e720-4ce6-aebc-993a3676abb5" containerName="ceilometer-central-agent" containerID="cri-o://2bb62034f23ff306eba03cc77fbd7f0b60e195607023bac74cfc090ef18a15c9" gracePeriod=30 Dec 03 14:31:54 crc kubenswrapper[4805]: I1203 14:31:54.224183 4805 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Dec 03 14:31:54 crc kubenswrapper[4805]: I1203 14:31:54.224228 4805 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="5979c218-e720-4ce6-aebc-993a3676abb5" containerName="proxy-httpd" containerID="cri-o://c5b0c5fc24db36ce396804d2d37d6255b90861754bbc2f992c426edf5f800276" gracePeriod=30 Dec 03 14:31:54 crc kubenswrapper[4805]: I1203 14:31:54.224277 4805 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="5979c218-e720-4ce6-aebc-993a3676abb5" containerName="sg-core" containerID="cri-o://61ecf9c7ac9b0bc53e3380d461ce4698dd8336362b53387b5832742de7bdc4f8" gracePeriod=30 Dec 03 14:31:54 crc kubenswrapper[4805]: I1203 14:31:54.224309 4805 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="5979c218-e720-4ce6-aebc-993a3676abb5" containerName="ceilometer-notification-agent" containerID="cri-o://0766a2a356b4c585109ee5f02dd5def6ec3f30d6e59c15941ad33bbd31cf88ba" gracePeriod=30 Dec 03 14:31:54 crc kubenswrapper[4805]: I1203 14:31:54.235640 4805 generic.go:334] "Generic (PLEG): container finished" podID="9f381959-280e-4eb1-9fb2-2365db2f9c43" containerID="0f465542ed3b0873cf31f730f8258e4061dbb10f0ebce0003303d33cd5b14544" exitCode=0 Dec 03 14:31:54 crc kubenswrapper[4805]: I1203 14:31:54.235713 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-75c8ddd69c-s5r44" event={"ID":"9f381959-280e-4eb1-9fb2-2365db2f9c43","Type":"ContainerDied","Data":"0f465542ed3b0873cf31f730f8258e4061dbb10f0ebce0003303d33cd5b14544"} Dec 03 14:31:54 crc kubenswrapper[4805]: I1203 14:31:54.252879 4805 generic.go:334] "Generic (PLEG): container finished" podID="756ba517-3759-4ee7-8718-a5d8c6f1be72" containerID="b4f259365efcb4f9557e4b79ee46dfcda1adab8f5bbd6b504a744431ebc41731" exitCode=1 Dec 03 14:31:54 crc kubenswrapper[4805]: I1203 14:31:54.252928 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-api-69cb5bcd7b-bwpl9" event={"ID":"756ba517-3759-4ee7-8718-a5d8c6f1be72","Type":"ContainerDied","Data":"b4f259365efcb4f9557e4b79ee46dfcda1adab8f5bbd6b504a744431ebc41731"} Dec 03 14:31:54 crc kubenswrapper[4805]: I1203 14:31:54.252965 4805 scope.go:117] "RemoveContainer" containerID="8eb73b26a6d1c034ac11da46581575c2940d67f72c07a7b5867123e276715fa8" Dec 03 14:31:54 crc kubenswrapper[4805]: I1203 14:31:54.253738 4805 scope.go:117] "RemoveContainer" containerID="b4f259365efcb4f9557e4b79ee46dfcda1adab8f5bbd6b504a744431ebc41731" Dec 03 14:31:54 crc kubenswrapper[4805]: E1203 14:31:54.253980 4805 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"heat-api\" with CrashLoopBackOff: \"back-off 10s restarting failed container=heat-api pod=heat-api-69cb5bcd7b-bwpl9_openstack(756ba517-3759-4ee7-8718-a5d8c6f1be72)\"" pod="openstack/heat-api-69cb5bcd7b-bwpl9" podUID="756ba517-3759-4ee7-8718-a5d8c6f1be72" Dec 03 14:31:54 crc kubenswrapper[4805]: I1203 14:31:54.293349 4805 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=2.808800441 podStartE2EDuration="13.293310991s" podCreationTimestamp="2025-12-03 14:31:41 +0000 UTC" firstStartedPulling="2025-12-03 14:31:43.128998202 +0000 UTC m=+1332.791915125" lastFinishedPulling="2025-12-03 14:31:53.613508752 +0000 UTC m=+1343.276425675" observedRunningTime="2025-12-03 14:31:54.289726873 +0000 UTC m=+1343.952643796" watchObservedRunningTime="2025-12-03 14:31:54.293310991 +0000 UTC m=+1343.956227914" Dec 03 14:31:54 crc kubenswrapper[4805]: I1203 14:31:54.446196 4805 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-75c8ddd69c-s5r44" Dec 03 14:31:54 crc kubenswrapper[4805]: I1203 14:31:54.608100 4805 prober.go:107] "Probe failed" probeType="Liveness" pod="openstack/cinder-api-0" podUID="7b5b9f5a-5af6-41d7-be26-3f4c091098b7" containerName="cinder-api" probeResult="failure" output="Get \"https://10.217.0.165:8776/healthcheck\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Dec 03 14:31:54 crc kubenswrapper[4805]: I1203 14:31:54.616423 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/9f381959-280e-4eb1-9fb2-2365db2f9c43-dns-swift-storage-0\") pod \"9f381959-280e-4eb1-9fb2-2365db2f9c43\" (UID: \"9f381959-280e-4eb1-9fb2-2365db2f9c43\") " Dec 03 14:31:54 crc kubenswrapper[4805]: I1203 14:31:54.616553 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/9f381959-280e-4eb1-9fb2-2365db2f9c43-ovsdbserver-sb\") pod \"9f381959-280e-4eb1-9fb2-2365db2f9c43\" (UID: \"9f381959-280e-4eb1-9fb2-2365db2f9c43\") " Dec 03 14:31:54 crc kubenswrapper[4805]: I1203 14:31:54.616600 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wjs7b\" (UniqueName: \"kubernetes.io/projected/9f381959-280e-4eb1-9fb2-2365db2f9c43-kube-api-access-wjs7b\") pod \"9f381959-280e-4eb1-9fb2-2365db2f9c43\" (UID: \"9f381959-280e-4eb1-9fb2-2365db2f9c43\") " Dec 03 14:31:54 crc kubenswrapper[4805]: I1203 14:31:54.616667 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9f381959-280e-4eb1-9fb2-2365db2f9c43-config\") pod \"9f381959-280e-4eb1-9fb2-2365db2f9c43\" (UID: \"9f381959-280e-4eb1-9fb2-2365db2f9c43\") " Dec 03 14:31:54 crc kubenswrapper[4805]: I1203 14:31:54.616787 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/9f381959-280e-4eb1-9fb2-2365db2f9c43-ovsdbserver-nb\") pod \"9f381959-280e-4eb1-9fb2-2365db2f9c43\" (UID: \"9f381959-280e-4eb1-9fb2-2365db2f9c43\") " Dec 03 14:31:54 crc kubenswrapper[4805]: I1203 14:31:54.616875 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/9f381959-280e-4eb1-9fb2-2365db2f9c43-dns-svc\") pod \"9f381959-280e-4eb1-9fb2-2365db2f9c43\" (UID: \"9f381959-280e-4eb1-9fb2-2365db2f9c43\") " Dec 03 14:31:54 crc kubenswrapper[4805]: I1203 14:31:54.642438 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9f381959-280e-4eb1-9fb2-2365db2f9c43-kube-api-access-wjs7b" (OuterVolumeSpecName: "kube-api-access-wjs7b") pod "9f381959-280e-4eb1-9fb2-2365db2f9c43" (UID: "9f381959-280e-4eb1-9fb2-2365db2f9c43"). InnerVolumeSpecName "kube-api-access-wjs7b". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 14:31:54 crc kubenswrapper[4805]: I1203 14:31:54.680629 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9f381959-280e-4eb1-9fb2-2365db2f9c43-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "9f381959-280e-4eb1-9fb2-2365db2f9c43" (UID: "9f381959-280e-4eb1-9fb2-2365db2f9c43"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 14:31:54 crc kubenswrapper[4805]: I1203 14:31:54.694105 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9f381959-280e-4eb1-9fb2-2365db2f9c43-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "9f381959-280e-4eb1-9fb2-2365db2f9c43" (UID: "9f381959-280e-4eb1-9fb2-2365db2f9c43"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 14:31:54 crc kubenswrapper[4805]: I1203 14:31:54.724785 4805 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/9f381959-280e-4eb1-9fb2-2365db2f9c43-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Dec 03 14:31:54 crc kubenswrapper[4805]: I1203 14:31:54.724830 4805 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/9f381959-280e-4eb1-9fb2-2365db2f9c43-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Dec 03 14:31:54 crc kubenswrapper[4805]: I1203 14:31:54.726637 4805 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wjs7b\" (UniqueName: \"kubernetes.io/projected/9f381959-280e-4eb1-9fb2-2365db2f9c43-kube-api-access-wjs7b\") on node \"crc\" DevicePath \"\"" Dec 03 14:31:54 crc kubenswrapper[4805]: I1203 14:31:54.728884 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9f381959-280e-4eb1-9fb2-2365db2f9c43-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "9f381959-280e-4eb1-9fb2-2365db2f9c43" (UID: "9f381959-280e-4eb1-9fb2-2365db2f9c43"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 14:31:54 crc kubenswrapper[4805]: I1203 14:31:54.757364 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9f381959-280e-4eb1-9fb2-2365db2f9c43-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "9f381959-280e-4eb1-9fb2-2365db2f9c43" (UID: "9f381959-280e-4eb1-9fb2-2365db2f9c43"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 14:31:54 crc kubenswrapper[4805]: I1203 14:31:54.771038 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9f381959-280e-4eb1-9fb2-2365db2f9c43-config" (OuterVolumeSpecName: "config") pod "9f381959-280e-4eb1-9fb2-2365db2f9c43" (UID: "9f381959-280e-4eb1-9fb2-2365db2f9c43"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 14:31:54 crc kubenswrapper[4805]: I1203 14:31:54.828959 4805 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/9f381959-280e-4eb1-9fb2-2365db2f9c43-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Dec 03 14:31:54 crc kubenswrapper[4805]: I1203 14:31:54.829015 4805 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/9f381959-280e-4eb1-9fb2-2365db2f9c43-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 03 14:31:54 crc kubenswrapper[4805]: I1203 14:31:54.829030 4805 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9f381959-280e-4eb1-9fb2-2365db2f9c43-config\") on node \"crc\" DevicePath \"\"" Dec 03 14:31:54 crc kubenswrapper[4805]: I1203 14:31:54.965580 4805 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openstack/heat-api-69cb5bcd7b-bwpl9" Dec 03 14:31:54 crc kubenswrapper[4805]: I1203 14:31:54.965630 4805 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/heat-api-69cb5bcd7b-bwpl9" Dec 03 14:31:55 crc kubenswrapper[4805]: I1203 14:31:55.267166 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-75c8ddd69c-s5r44" event={"ID":"9f381959-280e-4eb1-9fb2-2365db2f9c43","Type":"ContainerDied","Data":"792a5a7dd26d0a43f6c48455682e32c7fb8958acbc12cea35384caede78c4f97"} Dec 03 14:31:55 crc kubenswrapper[4805]: I1203 14:31:55.267228 4805 scope.go:117] "RemoveContainer" containerID="0f465542ed3b0873cf31f730f8258e4061dbb10f0ebce0003303d33cd5b14544" Dec 03 14:31:55 crc kubenswrapper[4805]: I1203 14:31:55.267426 4805 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-75c8ddd69c-s5r44" Dec 03 14:31:55 crc kubenswrapper[4805]: I1203 14:31:55.284127 4805 scope.go:117] "RemoveContainer" containerID="b4f259365efcb4f9557e4b79ee46dfcda1adab8f5bbd6b504a744431ebc41731" Dec 03 14:31:55 crc kubenswrapper[4805]: E1203 14:31:55.291054 4805 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"heat-api\" with CrashLoopBackOff: \"back-off 10s restarting failed container=heat-api pod=heat-api-69cb5bcd7b-bwpl9_openstack(756ba517-3759-4ee7-8718-a5d8c6f1be72)\"" pod="openstack/heat-api-69cb5bcd7b-bwpl9" podUID="756ba517-3759-4ee7-8718-a5d8c6f1be72" Dec 03 14:31:55 crc kubenswrapper[4805]: I1203 14:31:55.297698 4805 generic.go:334] "Generic (PLEG): container finished" podID="d1d26c95-d05a-4551-8d00-655941ce122b" containerID="8cb15466711e9cb141150e29e558fefcad0141bf93d2e6383f7505500faacd21" exitCode=1 Dec 03 14:31:55 crc kubenswrapper[4805]: I1203 14:31:55.297799 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-cfnapi-6cd94c587b-nshpc" event={"ID":"d1d26c95-d05a-4551-8d00-655941ce122b","Type":"ContainerDied","Data":"8cb15466711e9cb141150e29e558fefcad0141bf93d2e6383f7505500faacd21"} Dec 03 14:31:55 crc kubenswrapper[4805]: I1203 14:31:55.298517 4805 scope.go:117] "RemoveContainer" containerID="8cb15466711e9cb141150e29e558fefcad0141bf93d2e6383f7505500faacd21" Dec 03 14:31:55 crc kubenswrapper[4805]: E1203 14:31:55.298776 4805 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"heat-cfnapi\" with CrashLoopBackOff: \"back-off 10s restarting failed container=heat-cfnapi pod=heat-cfnapi-6cd94c587b-nshpc_openstack(d1d26c95-d05a-4551-8d00-655941ce122b)\"" pod="openstack/heat-cfnapi-6cd94c587b-nshpc" podUID="d1d26c95-d05a-4551-8d00-655941ce122b" Dec 03 14:31:55 crc kubenswrapper[4805]: I1203 14:31:55.311595 4805 generic.go:334] "Generic (PLEG): container finished" podID="5979c218-e720-4ce6-aebc-993a3676abb5" containerID="61ecf9c7ac9b0bc53e3380d461ce4698dd8336362b53387b5832742de7bdc4f8" exitCode=2 Dec 03 14:31:55 crc kubenswrapper[4805]: I1203 14:31:55.311635 4805 generic.go:334] "Generic (PLEG): container finished" podID="5979c218-e720-4ce6-aebc-993a3676abb5" containerID="0766a2a356b4c585109ee5f02dd5def6ec3f30d6e59c15941ad33bbd31cf88ba" exitCode=0 Dec 03 14:31:55 crc kubenswrapper[4805]: I1203 14:31:55.311647 4805 generic.go:334] "Generic (PLEG): container finished" podID="5979c218-e720-4ce6-aebc-993a3676abb5" containerID="2bb62034f23ff306eba03cc77fbd7f0b60e195607023bac74cfc090ef18a15c9" exitCode=0 Dec 03 14:31:55 crc kubenswrapper[4805]: I1203 14:31:55.311663 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"5979c218-e720-4ce6-aebc-993a3676abb5","Type":"ContainerDied","Data":"61ecf9c7ac9b0bc53e3380d461ce4698dd8336362b53387b5832742de7bdc4f8"} Dec 03 14:31:55 crc kubenswrapper[4805]: I1203 14:31:55.311702 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"5979c218-e720-4ce6-aebc-993a3676abb5","Type":"ContainerDied","Data":"0766a2a356b4c585109ee5f02dd5def6ec3f30d6e59c15941ad33bbd31cf88ba"} Dec 03 14:31:55 crc kubenswrapper[4805]: I1203 14:31:55.311718 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"5979c218-e720-4ce6-aebc-993a3676abb5","Type":"ContainerDied","Data":"2bb62034f23ff306eba03cc77fbd7f0b60e195607023bac74cfc090ef18a15c9"} Dec 03 14:31:55 crc kubenswrapper[4805]: I1203 14:31:55.312036 4805 scope.go:117] "RemoveContainer" containerID="dfafb0c8fd4a424b2f580f9c4f5ed8901168d0cee724096075c3a6574fb656a5" Dec 03 14:31:55 crc kubenswrapper[4805]: I1203 14:31:55.333387 4805 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-75c8ddd69c-s5r44"] Dec 03 14:31:55 crc kubenswrapper[4805]: I1203 14:31:55.349041 4805 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-75c8ddd69c-s5r44"] Dec 03 14:31:55 crc kubenswrapper[4805]: I1203 14:31:55.351956 4805 scope.go:117] "RemoveContainer" containerID="04fafceedea4c7029772a31c84f7c0fbe34ab906c19eadb096709755e15243c0" Dec 03 14:31:55 crc kubenswrapper[4805]: I1203 14:31:55.988593 4805 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/cinder-api-0" Dec 03 14:31:56 crc kubenswrapper[4805]: I1203 14:31:56.365542 4805 scope.go:117] "RemoveContainer" containerID="8cb15466711e9cb141150e29e558fefcad0141bf93d2e6383f7505500faacd21" Dec 03 14:31:56 crc kubenswrapper[4805]: E1203 14:31:56.365911 4805 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"heat-cfnapi\" with CrashLoopBackOff: \"back-off 10s restarting failed container=heat-cfnapi pod=heat-cfnapi-6cd94c587b-nshpc_openstack(d1d26c95-d05a-4551-8d00-655941ce122b)\"" pod="openstack/heat-cfnapi-6cd94c587b-nshpc" podUID="d1d26c95-d05a-4551-8d00-655941ce122b" Dec 03 14:31:56 crc kubenswrapper[4805]: I1203 14:31:56.715894 4805 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9f381959-280e-4eb1-9fb2-2365db2f9c43" path="/var/lib/kubelet/pods/9f381959-280e-4eb1-9fb2-2365db2f9c43/volumes" Dec 03 14:31:59 crc kubenswrapper[4805]: I1203 14:31:59.580278 4805 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/heat-cfnapi-68589fbcdd-snszz" Dec 03 14:31:59 crc kubenswrapper[4805]: I1203 14:31:59.609638 4805 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/heat-api-84746cbddc-kc7hm" Dec 03 14:31:59 crc kubenswrapper[4805]: I1203 14:31:59.679092 4805 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/heat-cfnapi-6cd94c587b-nshpc"] Dec 03 14:31:59 crc kubenswrapper[4805]: I1203 14:31:59.725194 4805 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/heat-api-69cb5bcd7b-bwpl9"] Dec 03 14:31:59 crc kubenswrapper[4805]: I1203 14:31:59.915692 4805 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/heat-engine-76dcfc959c-swfs4" Dec 03 14:32:00 crc kubenswrapper[4805]: I1203 14:32:00.006938 4805 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/heat-engine-5fdfc8886c-v7jx4"] Dec 03 14:32:00 crc kubenswrapper[4805]: I1203 14:32:00.008967 4805 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/heat-engine-5fdfc8886c-v7jx4" podUID="36af56b0-4335-48cf-82ba-bd95a50b14f3" containerName="heat-engine" containerID="cri-o://3d42d71bed5afbdee9635a53cd2836c52522e7333a4cbda0a4d910e814f3ae57" gracePeriod=60 Dec 03 14:32:00 crc kubenswrapper[4805]: I1203 14:32:00.180273 4805 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/heat-api-6bd4dc884b-6qst5" Dec 03 14:32:00 crc kubenswrapper[4805]: I1203 14:32:00.653747 4805 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/heat-cfnapi-795fdbdfbf-8zpm7" Dec 03 14:32:03 crc kubenswrapper[4805]: E1203 14:32:03.524818 4805 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="3d42d71bed5afbdee9635a53cd2836c52522e7333a4cbda0a4d910e814f3ae57" cmd=["/usr/bin/pgrep","-r","DRST","heat-engine"] Dec 03 14:32:03 crc kubenswrapper[4805]: E1203 14:32:03.529009 4805 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="3d42d71bed5afbdee9635a53cd2836c52522e7333a4cbda0a4d910e814f3ae57" cmd=["/usr/bin/pgrep","-r","DRST","heat-engine"] Dec 03 14:32:03 crc kubenswrapper[4805]: E1203 14:32:03.532969 4805 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="3d42d71bed5afbdee9635a53cd2836c52522e7333a4cbda0a4d910e814f3ae57" cmd=["/usr/bin/pgrep","-r","DRST","heat-engine"] Dec 03 14:32:03 crc kubenswrapper[4805]: E1203 14:32:03.533044 4805 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openstack/heat-engine-5fdfc8886c-v7jx4" podUID="36af56b0-4335-48cf-82ba-bd95a50b14f3" containerName="heat-engine" Dec 03 14:32:08 crc kubenswrapper[4805]: I1203 14:32:08.645613 4805 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/heat-cfnapi-6cd94c587b-nshpc" Dec 03 14:32:08 crc kubenswrapper[4805]: I1203 14:32:08.769475 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/d1d26c95-d05a-4551-8d00-655941ce122b-config-data-custom\") pod \"d1d26c95-d05a-4551-8d00-655941ce122b\" (UID: \"d1d26c95-d05a-4551-8d00-655941ce122b\") " Dec 03 14:32:08 crc kubenswrapper[4805]: I1203 14:32:08.769625 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d1d26c95-d05a-4551-8d00-655941ce122b-combined-ca-bundle\") pod \"d1d26c95-d05a-4551-8d00-655941ce122b\" (UID: \"d1d26c95-d05a-4551-8d00-655941ce122b\") " Dec 03 14:32:08 crc kubenswrapper[4805]: I1203 14:32:08.769817 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-z7p7f\" (UniqueName: \"kubernetes.io/projected/d1d26c95-d05a-4551-8d00-655941ce122b-kube-api-access-z7p7f\") pod \"d1d26c95-d05a-4551-8d00-655941ce122b\" (UID: \"d1d26c95-d05a-4551-8d00-655941ce122b\") " Dec 03 14:32:08 crc kubenswrapper[4805]: I1203 14:32:08.769992 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d1d26c95-d05a-4551-8d00-655941ce122b-config-data\") pod \"d1d26c95-d05a-4551-8d00-655941ce122b\" (UID: \"d1d26c95-d05a-4551-8d00-655941ce122b\") " Dec 03 14:32:08 crc kubenswrapper[4805]: I1203 14:32:08.780803 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d1d26c95-d05a-4551-8d00-655941ce122b-kube-api-access-z7p7f" (OuterVolumeSpecName: "kube-api-access-z7p7f") pod "d1d26c95-d05a-4551-8d00-655941ce122b" (UID: "d1d26c95-d05a-4551-8d00-655941ce122b"). InnerVolumeSpecName "kube-api-access-z7p7f". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 14:32:08 crc kubenswrapper[4805]: I1203 14:32:08.784282 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d1d26c95-d05a-4551-8d00-655941ce122b-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "d1d26c95-d05a-4551-8d00-655941ce122b" (UID: "d1d26c95-d05a-4551-8d00-655941ce122b"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 14:32:08 crc kubenswrapper[4805]: I1203 14:32:08.825720 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d1d26c95-d05a-4551-8d00-655941ce122b-config-data" (OuterVolumeSpecName: "config-data") pod "d1d26c95-d05a-4551-8d00-655941ce122b" (UID: "d1d26c95-d05a-4551-8d00-655941ce122b"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 14:32:08 crc kubenswrapper[4805]: I1203 14:32:08.844239 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d1d26c95-d05a-4551-8d00-655941ce122b-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "d1d26c95-d05a-4551-8d00-655941ce122b" (UID: "d1d26c95-d05a-4551-8d00-655941ce122b"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 14:32:08 crc kubenswrapper[4805]: I1203 14:32:08.872541 4805 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d1d26c95-d05a-4551-8d00-655941ce122b-config-data\") on node \"crc\" DevicePath \"\"" Dec 03 14:32:08 crc kubenswrapper[4805]: I1203 14:32:08.872569 4805 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/d1d26c95-d05a-4551-8d00-655941ce122b-config-data-custom\") on node \"crc\" DevicePath \"\"" Dec 03 14:32:08 crc kubenswrapper[4805]: I1203 14:32:08.872579 4805 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d1d26c95-d05a-4551-8d00-655941ce122b-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 14:32:08 crc kubenswrapper[4805]: I1203 14:32:08.872587 4805 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-z7p7f\" (UniqueName: \"kubernetes.io/projected/d1d26c95-d05a-4551-8d00-655941ce122b-kube-api-access-z7p7f\") on node \"crc\" DevicePath \"\"" Dec 03 14:32:09 crc kubenswrapper[4805]: I1203 14:32:09.524452 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-cfnapi-6cd94c587b-nshpc" event={"ID":"d1d26c95-d05a-4551-8d00-655941ce122b","Type":"ContainerDied","Data":"df8e2f594cef67002f327c841b44bde459211f47621558cf1d44e38124a4c18a"} Dec 03 14:32:09 crc kubenswrapper[4805]: I1203 14:32:09.524786 4805 scope.go:117] "RemoveContainer" containerID="8cb15466711e9cb141150e29e558fefcad0141bf93d2e6383f7505500faacd21" Dec 03 14:32:09 crc kubenswrapper[4805]: I1203 14:32:09.524896 4805 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/heat-cfnapi-6cd94c587b-nshpc" Dec 03 14:32:09 crc kubenswrapper[4805]: I1203 14:32:09.563967 4805 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/heat-cfnapi-6cd94c587b-nshpc"] Dec 03 14:32:09 crc kubenswrapper[4805]: I1203 14:32:09.575163 4805 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/heat-cfnapi-6cd94c587b-nshpc"] Dec 03 14:32:10 crc kubenswrapper[4805]: I1203 14:32:10.325509 4805 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/heat-api-69cb5bcd7b-bwpl9" Dec 03 14:32:10 crc kubenswrapper[4805]: I1203 14:32:10.400957 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/756ba517-3759-4ee7-8718-a5d8c6f1be72-config-data\") pod \"756ba517-3759-4ee7-8718-a5d8c6f1be72\" (UID: \"756ba517-3759-4ee7-8718-a5d8c6f1be72\") " Dec 03 14:32:10 crc kubenswrapper[4805]: I1203 14:32:10.401169 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ckx8m\" (UniqueName: \"kubernetes.io/projected/756ba517-3759-4ee7-8718-a5d8c6f1be72-kube-api-access-ckx8m\") pod \"756ba517-3759-4ee7-8718-a5d8c6f1be72\" (UID: \"756ba517-3759-4ee7-8718-a5d8c6f1be72\") " Dec 03 14:32:10 crc kubenswrapper[4805]: I1203 14:32:10.401236 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/756ba517-3759-4ee7-8718-a5d8c6f1be72-combined-ca-bundle\") pod \"756ba517-3759-4ee7-8718-a5d8c6f1be72\" (UID: \"756ba517-3759-4ee7-8718-a5d8c6f1be72\") " Dec 03 14:32:10 crc kubenswrapper[4805]: I1203 14:32:10.401345 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/756ba517-3759-4ee7-8718-a5d8c6f1be72-config-data-custom\") pod \"756ba517-3759-4ee7-8718-a5d8c6f1be72\" (UID: \"756ba517-3759-4ee7-8718-a5d8c6f1be72\") " Dec 03 14:32:10 crc kubenswrapper[4805]: I1203 14:32:10.405963 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/756ba517-3759-4ee7-8718-a5d8c6f1be72-kube-api-access-ckx8m" (OuterVolumeSpecName: "kube-api-access-ckx8m") pod "756ba517-3759-4ee7-8718-a5d8c6f1be72" (UID: "756ba517-3759-4ee7-8718-a5d8c6f1be72"). InnerVolumeSpecName "kube-api-access-ckx8m". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 14:32:10 crc kubenswrapper[4805]: I1203 14:32:10.406193 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/756ba517-3759-4ee7-8718-a5d8c6f1be72-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "756ba517-3759-4ee7-8718-a5d8c6f1be72" (UID: "756ba517-3759-4ee7-8718-a5d8c6f1be72"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 14:32:10 crc kubenswrapper[4805]: I1203 14:32:10.434108 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/756ba517-3759-4ee7-8718-a5d8c6f1be72-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "756ba517-3759-4ee7-8718-a5d8c6f1be72" (UID: "756ba517-3759-4ee7-8718-a5d8c6f1be72"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 14:32:10 crc kubenswrapper[4805]: I1203 14:32:10.471220 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/756ba517-3759-4ee7-8718-a5d8c6f1be72-config-data" (OuterVolumeSpecName: "config-data") pod "756ba517-3759-4ee7-8718-a5d8c6f1be72" (UID: "756ba517-3759-4ee7-8718-a5d8c6f1be72"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 14:32:10 crc kubenswrapper[4805]: I1203 14:32:10.503945 4805 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ckx8m\" (UniqueName: \"kubernetes.io/projected/756ba517-3759-4ee7-8718-a5d8c6f1be72-kube-api-access-ckx8m\") on node \"crc\" DevicePath \"\"" Dec 03 14:32:10 crc kubenswrapper[4805]: I1203 14:32:10.503977 4805 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/756ba517-3759-4ee7-8718-a5d8c6f1be72-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 14:32:10 crc kubenswrapper[4805]: I1203 14:32:10.503988 4805 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/756ba517-3759-4ee7-8718-a5d8c6f1be72-config-data-custom\") on node \"crc\" DevicePath \"\"" Dec 03 14:32:10 crc kubenswrapper[4805]: I1203 14:32:10.504001 4805 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/756ba517-3759-4ee7-8718-a5d8c6f1be72-config-data\") on node \"crc\" DevicePath \"\"" Dec 03 14:32:10 crc kubenswrapper[4805]: I1203 14:32:10.535280 4805 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/heat-api-69cb5bcd7b-bwpl9" Dec 03 14:32:10 crc kubenswrapper[4805]: I1203 14:32:10.535329 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-api-69cb5bcd7b-bwpl9" event={"ID":"756ba517-3759-4ee7-8718-a5d8c6f1be72","Type":"ContainerDied","Data":"3847c2681419a98a4e00e1e300c9371a8336a1eec4300e25686a7fd2393ebbdb"} Dec 03 14:32:10 crc kubenswrapper[4805]: I1203 14:32:10.535376 4805 scope.go:117] "RemoveContainer" containerID="b4f259365efcb4f9557e4b79ee46dfcda1adab8f5bbd6b504a744431ebc41731" Dec 03 14:32:10 crc kubenswrapper[4805]: I1203 14:32:10.568899 4805 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/heat-api-69cb5bcd7b-bwpl9"] Dec 03 14:32:10 crc kubenswrapper[4805]: I1203 14:32:10.577363 4805 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/heat-api-69cb5bcd7b-bwpl9"] Dec 03 14:32:10 crc kubenswrapper[4805]: I1203 14:32:10.713399 4805 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="756ba517-3759-4ee7-8718-a5d8c6f1be72" path="/var/lib/kubelet/pods/756ba517-3759-4ee7-8718-a5d8c6f1be72/volumes" Dec 03 14:32:10 crc kubenswrapper[4805]: I1203 14:32:10.714552 4805 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d1d26c95-d05a-4551-8d00-655941ce122b" path="/var/lib/kubelet/pods/d1d26c95-d05a-4551-8d00-655941ce122b/volumes" Dec 03 14:32:11 crc kubenswrapper[4805]: I1203 14:32:11.496114 4805 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Dec 03 14:32:11 crc kubenswrapper[4805]: I1203 14:32:11.496364 4805 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-external-api-0" podUID="9ed6dfc0-82db-4d86-ad92-cfe94fa370a2" containerName="glance-log" containerID="cri-o://18b272180201caa258fce70a3153b9e82a5103c2066cac10db4b226952843d2f" gracePeriod=30 Dec 03 14:32:11 crc kubenswrapper[4805]: I1203 14:32:11.496428 4805 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-external-api-0" podUID="9ed6dfc0-82db-4d86-ad92-cfe94fa370a2" containerName="glance-httpd" containerID="cri-o://b40b68276f1683e9c153b03843673b0fb798f28865161b6b4f97250e1dcc86f9" gracePeriod=30 Dec 03 14:32:11 crc kubenswrapper[4805]: E1203 14:32:11.562306 4805 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-nova-conductor:current-podified" Dec 03 14:32:11 crc kubenswrapper[4805]: E1203 14:32:11.562511 4805 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:nova-cell0-conductor-db-sync,Image:quay.io/podified-antelope-centos9/openstack-nova-conductor:current-podified,Command:[/bin/bash],Args:[-c /usr/local/bin/kolla_start],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CELL_NAME,Value:cell0,ValueFrom:nil,},EnvVar{Name:KOLLA_BOOTSTRAP,Value:true,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config-data,ReadOnly:false,MountPath:/var/lib/openstack/config,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:scripts,ReadOnly:false,MountPath:/var/lib/openstack/bin,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:false,MountPath:/var/lib/kolla/config_files/config.json,SubPath:nova-conductor-dbsync-config.json,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:combined-ca-bundle,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-9dr74,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*42436,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod nova-cell0-conductor-db-sync-dxrx8_openstack(e56c5750-e0c6-4427-ae8c-ab4af0c9b7d0): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 03 14:32:11 crc kubenswrapper[4805]: E1203 14:32:11.563892 4805 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"nova-cell0-conductor-db-sync\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/nova-cell0-conductor-db-sync-dxrx8" podUID="e56c5750-e0c6-4427-ae8c-ab4af0c9b7d0" Dec 03 14:32:11 crc kubenswrapper[4805]: I1203 14:32:11.998726 4805 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/ceilometer-0" podUID="5979c218-e720-4ce6-aebc-993a3676abb5" containerName="proxy-httpd" probeResult="failure" output="HTTP probe failed with statuscode: 503" Dec 03 14:32:12 crc kubenswrapper[4805]: I1203 14:32:12.581996 4805 generic.go:334] "Generic (PLEG): container finished" podID="9ed6dfc0-82db-4d86-ad92-cfe94fa370a2" containerID="18b272180201caa258fce70a3153b9e82a5103c2066cac10db4b226952843d2f" exitCode=143 Dec 03 14:32:12 crc kubenswrapper[4805]: I1203 14:32:12.582069 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"9ed6dfc0-82db-4d86-ad92-cfe94fa370a2","Type":"ContainerDied","Data":"18b272180201caa258fce70a3153b9e82a5103c2066cac10db4b226952843d2f"} Dec 03 14:32:12 crc kubenswrapper[4805]: I1203 14:32:12.583698 4805 generic.go:334] "Generic (PLEG): container finished" podID="36af56b0-4335-48cf-82ba-bd95a50b14f3" containerID="3d42d71bed5afbdee9635a53cd2836c52522e7333a4cbda0a4d910e814f3ae57" exitCode=0 Dec 03 14:32:12 crc kubenswrapper[4805]: I1203 14:32:12.584492 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-engine-5fdfc8886c-v7jx4" event={"ID":"36af56b0-4335-48cf-82ba-bd95a50b14f3","Type":"ContainerDied","Data":"3d42d71bed5afbdee9635a53cd2836c52522e7333a4cbda0a4d910e814f3ae57"} Dec 03 14:32:12 crc kubenswrapper[4805]: E1203 14:32:12.585269 4805 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"nova-cell0-conductor-db-sync\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-nova-conductor:current-podified\\\"\"" pod="openstack/nova-cell0-conductor-db-sync-dxrx8" podUID="e56c5750-e0c6-4427-ae8c-ab4af0c9b7d0" Dec 03 14:32:12 crc kubenswrapper[4805]: I1203 14:32:12.852236 4805 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/heat-engine-5fdfc8886c-v7jx4" Dec 03 14:32:12 crc kubenswrapper[4805]: I1203 14:32:12.948781 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/36af56b0-4335-48cf-82ba-bd95a50b14f3-config-data-custom\") pod \"36af56b0-4335-48cf-82ba-bd95a50b14f3\" (UID: \"36af56b0-4335-48cf-82ba-bd95a50b14f3\") " Dec 03 14:32:12 crc kubenswrapper[4805]: I1203 14:32:12.948858 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/36af56b0-4335-48cf-82ba-bd95a50b14f3-config-data\") pod \"36af56b0-4335-48cf-82ba-bd95a50b14f3\" (UID: \"36af56b0-4335-48cf-82ba-bd95a50b14f3\") " Dec 03 14:32:12 crc kubenswrapper[4805]: I1203 14:32:12.948888 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-sb4qs\" (UniqueName: \"kubernetes.io/projected/36af56b0-4335-48cf-82ba-bd95a50b14f3-kube-api-access-sb4qs\") pod \"36af56b0-4335-48cf-82ba-bd95a50b14f3\" (UID: \"36af56b0-4335-48cf-82ba-bd95a50b14f3\") " Dec 03 14:32:12 crc kubenswrapper[4805]: I1203 14:32:12.948946 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/36af56b0-4335-48cf-82ba-bd95a50b14f3-combined-ca-bundle\") pod \"36af56b0-4335-48cf-82ba-bd95a50b14f3\" (UID: \"36af56b0-4335-48cf-82ba-bd95a50b14f3\") " Dec 03 14:32:12 crc kubenswrapper[4805]: I1203 14:32:12.956274 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/36af56b0-4335-48cf-82ba-bd95a50b14f3-kube-api-access-sb4qs" (OuterVolumeSpecName: "kube-api-access-sb4qs") pod "36af56b0-4335-48cf-82ba-bd95a50b14f3" (UID: "36af56b0-4335-48cf-82ba-bd95a50b14f3"). InnerVolumeSpecName "kube-api-access-sb4qs". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 14:32:12 crc kubenswrapper[4805]: I1203 14:32:12.957989 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/36af56b0-4335-48cf-82ba-bd95a50b14f3-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "36af56b0-4335-48cf-82ba-bd95a50b14f3" (UID: "36af56b0-4335-48cf-82ba-bd95a50b14f3"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 14:32:12 crc kubenswrapper[4805]: I1203 14:32:12.981078 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/36af56b0-4335-48cf-82ba-bd95a50b14f3-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "36af56b0-4335-48cf-82ba-bd95a50b14f3" (UID: "36af56b0-4335-48cf-82ba-bd95a50b14f3"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 14:32:12 crc kubenswrapper[4805]: I1203 14:32:12.998264 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/36af56b0-4335-48cf-82ba-bd95a50b14f3-config-data" (OuterVolumeSpecName: "config-data") pod "36af56b0-4335-48cf-82ba-bd95a50b14f3" (UID: "36af56b0-4335-48cf-82ba-bd95a50b14f3"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 14:32:13 crc kubenswrapper[4805]: I1203 14:32:13.054693 4805 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/36af56b0-4335-48cf-82ba-bd95a50b14f3-config-data\") on node \"crc\" DevicePath \"\"" Dec 03 14:32:13 crc kubenswrapper[4805]: I1203 14:32:13.054746 4805 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-sb4qs\" (UniqueName: \"kubernetes.io/projected/36af56b0-4335-48cf-82ba-bd95a50b14f3-kube-api-access-sb4qs\") on node \"crc\" DevicePath \"\"" Dec 03 14:32:13 crc kubenswrapper[4805]: I1203 14:32:13.054762 4805 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/36af56b0-4335-48cf-82ba-bd95a50b14f3-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 14:32:13 crc kubenswrapper[4805]: I1203 14:32:13.054801 4805 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/36af56b0-4335-48cf-82ba-bd95a50b14f3-config-data-custom\") on node \"crc\" DevicePath \"\"" Dec 03 14:32:13 crc kubenswrapper[4805]: I1203 14:32:13.256968 4805 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 03 14:32:13 crc kubenswrapper[4805]: I1203 14:32:13.258556 4805 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-internal-api-0" podUID="10877f19-dd2a-4eb2-99e0-2d0cf82d5ad5" containerName="glance-log" containerID="cri-o://24356197f69f0bfd999f1654211fef34606d3c50328026f4b3c6d2eda195db85" gracePeriod=30 Dec 03 14:32:13 crc kubenswrapper[4805]: I1203 14:32:13.258914 4805 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-internal-api-0" podUID="10877f19-dd2a-4eb2-99e0-2d0cf82d5ad5" containerName="glance-httpd" containerID="cri-o://61aa5396071fa72fd8a539334f4fa1f4982e07e997aee839e400e0b116fff32e" gracePeriod=30 Dec 03 14:32:13 crc kubenswrapper[4805]: I1203 14:32:13.352554 4805 pod_container_manager_linux.go:210] "Failed to delete cgroup paths" cgroupName=["kubepods","besteffort","pod29710b09-a1ad-468f-af26-b5d5eb516898"] err="unable to destroy cgroup paths for cgroup [kubepods besteffort pod29710b09-a1ad-468f-af26-b5d5eb516898] : Timed out while waiting for systemd to remove kubepods-besteffort-pod29710b09_a1ad_468f_af26_b5d5eb516898.slice" Dec 03 14:32:13 crc kubenswrapper[4805]: E1203 14:32:13.352602 4805 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to delete cgroup paths for [kubepods besteffort pod29710b09-a1ad-468f-af26-b5d5eb516898] : unable to destroy cgroup paths for cgroup [kubepods besteffort pod29710b09-a1ad-468f-af26-b5d5eb516898] : Timed out while waiting for systemd to remove kubepods-besteffort-pod29710b09_a1ad_468f_af26_b5d5eb516898.slice" pod="openstack/cinder-scheduler-0" podUID="29710b09-a1ad-468f-af26-b5d5eb516898" Dec 03 14:32:13 crc kubenswrapper[4805]: I1203 14:32:13.593022 4805 generic.go:334] "Generic (PLEG): container finished" podID="10877f19-dd2a-4eb2-99e0-2d0cf82d5ad5" containerID="24356197f69f0bfd999f1654211fef34606d3c50328026f4b3c6d2eda195db85" exitCode=143 Dec 03 14:32:13 crc kubenswrapper[4805]: I1203 14:32:13.593094 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"10877f19-dd2a-4eb2-99e0-2d0cf82d5ad5","Type":"ContainerDied","Data":"24356197f69f0bfd999f1654211fef34606d3c50328026f4b3c6d2eda195db85"} Dec 03 14:32:13 crc kubenswrapper[4805]: I1203 14:32:13.595366 4805 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Dec 03 14:32:13 crc kubenswrapper[4805]: I1203 14:32:13.595373 4805 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/heat-engine-5fdfc8886c-v7jx4" Dec 03 14:32:13 crc kubenswrapper[4805]: I1203 14:32:13.595358 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-engine-5fdfc8886c-v7jx4" event={"ID":"36af56b0-4335-48cf-82ba-bd95a50b14f3","Type":"ContainerDied","Data":"3024bb1954bfb673162d931d82304a21f3d101c3872b3723478056576de1e7b1"} Dec 03 14:32:13 crc kubenswrapper[4805]: I1203 14:32:13.595680 4805 scope.go:117] "RemoveContainer" containerID="3d42d71bed5afbdee9635a53cd2836c52522e7333a4cbda0a4d910e814f3ae57" Dec 03 14:32:13 crc kubenswrapper[4805]: I1203 14:32:13.624595 4805 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-scheduler-0"] Dec 03 14:32:13 crc kubenswrapper[4805]: I1203 14:32:13.632133 4805 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-scheduler-0"] Dec 03 14:32:13 crc kubenswrapper[4805]: I1203 14:32:13.643954 4805 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/heat-engine-5fdfc8886c-v7jx4"] Dec 03 14:32:13 crc kubenswrapper[4805]: I1203 14:32:13.651576 4805 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/heat-engine-5fdfc8886c-v7jx4"] Dec 03 14:32:13 crc kubenswrapper[4805]: I1203 14:32:13.667928 4805 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-scheduler-0"] Dec 03 14:32:13 crc kubenswrapper[4805]: E1203 14:32:13.668349 4805 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="36af56b0-4335-48cf-82ba-bd95a50b14f3" containerName="heat-engine" Dec 03 14:32:13 crc kubenswrapper[4805]: I1203 14:32:13.668373 4805 state_mem.go:107] "Deleted CPUSet assignment" podUID="36af56b0-4335-48cf-82ba-bd95a50b14f3" containerName="heat-engine" Dec 03 14:32:13 crc kubenswrapper[4805]: E1203 14:32:13.668395 4805 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9f381959-280e-4eb1-9fb2-2365db2f9c43" containerName="init" Dec 03 14:32:13 crc kubenswrapper[4805]: I1203 14:32:13.668404 4805 state_mem.go:107] "Deleted CPUSet assignment" podUID="9f381959-280e-4eb1-9fb2-2365db2f9c43" containerName="init" Dec 03 14:32:13 crc kubenswrapper[4805]: E1203 14:32:13.668420 4805 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9f381959-280e-4eb1-9fb2-2365db2f9c43" containerName="dnsmasq-dns" Dec 03 14:32:13 crc kubenswrapper[4805]: I1203 14:32:13.668430 4805 state_mem.go:107] "Deleted CPUSet assignment" podUID="9f381959-280e-4eb1-9fb2-2365db2f9c43" containerName="dnsmasq-dns" Dec 03 14:32:13 crc kubenswrapper[4805]: E1203 14:32:13.668451 4805 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d1d26c95-d05a-4551-8d00-655941ce122b" containerName="heat-cfnapi" Dec 03 14:32:13 crc kubenswrapper[4805]: I1203 14:32:13.668461 4805 state_mem.go:107] "Deleted CPUSet assignment" podUID="d1d26c95-d05a-4551-8d00-655941ce122b" containerName="heat-cfnapi" Dec 03 14:32:13 crc kubenswrapper[4805]: E1203 14:32:13.668470 4805 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="756ba517-3759-4ee7-8718-a5d8c6f1be72" containerName="heat-api" Dec 03 14:32:13 crc kubenswrapper[4805]: I1203 14:32:13.668477 4805 state_mem.go:107] "Deleted CPUSet assignment" podUID="756ba517-3759-4ee7-8718-a5d8c6f1be72" containerName="heat-api" Dec 03 14:32:13 crc kubenswrapper[4805]: E1203 14:32:13.668491 4805 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="756ba517-3759-4ee7-8718-a5d8c6f1be72" containerName="heat-api" Dec 03 14:32:13 crc kubenswrapper[4805]: I1203 14:32:13.668498 4805 state_mem.go:107] "Deleted CPUSet assignment" podUID="756ba517-3759-4ee7-8718-a5d8c6f1be72" containerName="heat-api" Dec 03 14:32:13 crc kubenswrapper[4805]: I1203 14:32:13.668729 4805 memory_manager.go:354] "RemoveStaleState removing state" podUID="756ba517-3759-4ee7-8718-a5d8c6f1be72" containerName="heat-api" Dec 03 14:32:13 crc kubenswrapper[4805]: I1203 14:32:13.668747 4805 memory_manager.go:354] "RemoveStaleState removing state" podUID="756ba517-3759-4ee7-8718-a5d8c6f1be72" containerName="heat-api" Dec 03 14:32:13 crc kubenswrapper[4805]: I1203 14:32:13.668758 4805 memory_manager.go:354] "RemoveStaleState removing state" podUID="d1d26c95-d05a-4551-8d00-655941ce122b" containerName="heat-cfnapi" Dec 03 14:32:13 crc kubenswrapper[4805]: I1203 14:32:13.668772 4805 memory_manager.go:354] "RemoveStaleState removing state" podUID="36af56b0-4335-48cf-82ba-bd95a50b14f3" containerName="heat-engine" Dec 03 14:32:13 crc kubenswrapper[4805]: I1203 14:32:13.668788 4805 memory_manager.go:354] "RemoveStaleState removing state" podUID="9f381959-280e-4eb1-9fb2-2365db2f9c43" containerName="dnsmasq-dns" Dec 03 14:32:13 crc kubenswrapper[4805]: E1203 14:32:13.669026 4805 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d1d26c95-d05a-4551-8d00-655941ce122b" containerName="heat-cfnapi" Dec 03 14:32:13 crc kubenswrapper[4805]: I1203 14:32:13.669037 4805 state_mem.go:107] "Deleted CPUSet assignment" podUID="d1d26c95-d05a-4551-8d00-655941ce122b" containerName="heat-cfnapi" Dec 03 14:32:13 crc kubenswrapper[4805]: I1203 14:32:13.669236 4805 memory_manager.go:354] "RemoveStaleState removing state" podUID="d1d26c95-d05a-4551-8d00-655941ce122b" containerName="heat-cfnapi" Dec 03 14:32:13 crc kubenswrapper[4805]: I1203 14:32:13.669878 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Dec 03 14:32:13 crc kubenswrapper[4805]: I1203 14:32:13.672094 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-scheduler-config-data" Dec 03 14:32:13 crc kubenswrapper[4805]: I1203 14:32:13.675255 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-scheduler-0"] Dec 03 14:32:13 crc kubenswrapper[4805]: I1203 14:32:13.868114 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lnw2d\" (UniqueName: \"kubernetes.io/projected/6d34a555-7003-4796-a15a-85a71c489bc9-kube-api-access-lnw2d\") pod \"cinder-scheduler-0\" (UID: \"6d34a555-7003-4796-a15a-85a71c489bc9\") " pod="openstack/cinder-scheduler-0" Dec 03 14:32:13 crc kubenswrapper[4805]: I1203 14:32:13.868193 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/6d34a555-7003-4796-a15a-85a71c489bc9-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"6d34a555-7003-4796-a15a-85a71c489bc9\") " pod="openstack/cinder-scheduler-0" Dec 03 14:32:13 crc kubenswrapper[4805]: I1203 14:32:13.868217 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/6d34a555-7003-4796-a15a-85a71c489bc9-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"6d34a555-7003-4796-a15a-85a71c489bc9\") " pod="openstack/cinder-scheduler-0" Dec 03 14:32:13 crc kubenswrapper[4805]: I1203 14:32:13.868237 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6d34a555-7003-4796-a15a-85a71c489bc9-config-data\") pod \"cinder-scheduler-0\" (UID: \"6d34a555-7003-4796-a15a-85a71c489bc9\") " pod="openstack/cinder-scheduler-0" Dec 03 14:32:13 crc kubenswrapper[4805]: I1203 14:32:13.868255 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6d34a555-7003-4796-a15a-85a71c489bc9-scripts\") pod \"cinder-scheduler-0\" (UID: \"6d34a555-7003-4796-a15a-85a71c489bc9\") " pod="openstack/cinder-scheduler-0" Dec 03 14:32:13 crc kubenswrapper[4805]: I1203 14:32:13.868284 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6d34a555-7003-4796-a15a-85a71c489bc9-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"6d34a555-7003-4796-a15a-85a71c489bc9\") " pod="openstack/cinder-scheduler-0" Dec 03 14:32:13 crc kubenswrapper[4805]: I1203 14:32:13.969061 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/6d34a555-7003-4796-a15a-85a71c489bc9-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"6d34a555-7003-4796-a15a-85a71c489bc9\") " pod="openstack/cinder-scheduler-0" Dec 03 14:32:13 crc kubenswrapper[4805]: I1203 14:32:13.969113 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/6d34a555-7003-4796-a15a-85a71c489bc9-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"6d34a555-7003-4796-a15a-85a71c489bc9\") " pod="openstack/cinder-scheduler-0" Dec 03 14:32:13 crc kubenswrapper[4805]: I1203 14:32:13.969134 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6d34a555-7003-4796-a15a-85a71c489bc9-config-data\") pod \"cinder-scheduler-0\" (UID: \"6d34a555-7003-4796-a15a-85a71c489bc9\") " pod="openstack/cinder-scheduler-0" Dec 03 14:32:13 crc kubenswrapper[4805]: I1203 14:32:13.969153 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6d34a555-7003-4796-a15a-85a71c489bc9-scripts\") pod \"cinder-scheduler-0\" (UID: \"6d34a555-7003-4796-a15a-85a71c489bc9\") " pod="openstack/cinder-scheduler-0" Dec 03 14:32:13 crc kubenswrapper[4805]: I1203 14:32:13.969180 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6d34a555-7003-4796-a15a-85a71c489bc9-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"6d34a555-7003-4796-a15a-85a71c489bc9\") " pod="openstack/cinder-scheduler-0" Dec 03 14:32:13 crc kubenswrapper[4805]: I1203 14:32:13.969276 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lnw2d\" (UniqueName: \"kubernetes.io/projected/6d34a555-7003-4796-a15a-85a71c489bc9-kube-api-access-lnw2d\") pod \"cinder-scheduler-0\" (UID: \"6d34a555-7003-4796-a15a-85a71c489bc9\") " pod="openstack/cinder-scheduler-0" Dec 03 14:32:13 crc kubenswrapper[4805]: I1203 14:32:13.969631 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/6d34a555-7003-4796-a15a-85a71c489bc9-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"6d34a555-7003-4796-a15a-85a71c489bc9\") " pod="openstack/cinder-scheduler-0" Dec 03 14:32:13 crc kubenswrapper[4805]: I1203 14:32:13.973342 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6d34a555-7003-4796-a15a-85a71c489bc9-scripts\") pod \"cinder-scheduler-0\" (UID: \"6d34a555-7003-4796-a15a-85a71c489bc9\") " pod="openstack/cinder-scheduler-0" Dec 03 14:32:13 crc kubenswrapper[4805]: I1203 14:32:13.973900 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/6d34a555-7003-4796-a15a-85a71c489bc9-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"6d34a555-7003-4796-a15a-85a71c489bc9\") " pod="openstack/cinder-scheduler-0" Dec 03 14:32:13 crc kubenswrapper[4805]: I1203 14:32:13.975828 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6d34a555-7003-4796-a15a-85a71c489bc9-config-data\") pod \"cinder-scheduler-0\" (UID: \"6d34a555-7003-4796-a15a-85a71c489bc9\") " pod="openstack/cinder-scheduler-0" Dec 03 14:32:13 crc kubenswrapper[4805]: I1203 14:32:13.983317 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6d34a555-7003-4796-a15a-85a71c489bc9-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"6d34a555-7003-4796-a15a-85a71c489bc9\") " pod="openstack/cinder-scheduler-0" Dec 03 14:32:13 crc kubenswrapper[4805]: I1203 14:32:13.987016 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lnw2d\" (UniqueName: \"kubernetes.io/projected/6d34a555-7003-4796-a15a-85a71c489bc9-kube-api-access-lnw2d\") pod \"cinder-scheduler-0\" (UID: \"6d34a555-7003-4796-a15a-85a71c489bc9\") " pod="openstack/cinder-scheduler-0" Dec 03 14:32:13 crc kubenswrapper[4805]: I1203 14:32:13.992621 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Dec 03 14:32:14 crc kubenswrapper[4805]: I1203 14:32:14.510130 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-scheduler-0"] Dec 03 14:32:14 crc kubenswrapper[4805]: I1203 14:32:14.605629 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"6d34a555-7003-4796-a15a-85a71c489bc9","Type":"ContainerStarted","Data":"700fd102080259047401e7561304b50ceaf9979b902f15b659015dbbd591a232"} Dec 03 14:32:14 crc kubenswrapper[4805]: I1203 14:32:14.687631 4805 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/glance-default-external-api-0" podUID="9ed6dfc0-82db-4d86-ad92-cfe94fa370a2" containerName="glance-httpd" probeResult="failure" output="Get \"https://10.217.0.147:9292/healthcheck\": read tcp 10.217.0.2:43292->10.217.0.147:9292: read: connection reset by peer" Dec 03 14:32:14 crc kubenswrapper[4805]: I1203 14:32:14.687648 4805 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/glance-default-external-api-0" podUID="9ed6dfc0-82db-4d86-ad92-cfe94fa370a2" containerName="glance-log" probeResult="failure" output="Get \"https://10.217.0.147:9292/healthcheck\": read tcp 10.217.0.2:43294->10.217.0.147:9292: read: connection reset by peer" Dec 03 14:32:14 crc kubenswrapper[4805]: I1203 14:32:14.708667 4805 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="29710b09-a1ad-468f-af26-b5d5eb516898" path="/var/lib/kubelet/pods/29710b09-a1ad-468f-af26-b5d5eb516898/volumes" Dec 03 14:32:14 crc kubenswrapper[4805]: I1203 14:32:14.709659 4805 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="36af56b0-4335-48cf-82ba-bd95a50b14f3" path="/var/lib/kubelet/pods/36af56b0-4335-48cf-82ba-bd95a50b14f3/volumes" Dec 03 14:32:15 crc kubenswrapper[4805]: I1203 14:32:15.304497 4805 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Dec 03 14:32:15 crc kubenswrapper[4805]: I1203 14:32:15.403435 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/9ed6dfc0-82db-4d86-ad92-cfe94fa370a2-httpd-run\") pod \"9ed6dfc0-82db-4d86-ad92-cfe94fa370a2\" (UID: \"9ed6dfc0-82db-4d86-ad92-cfe94fa370a2\") " Dec 03 14:32:15 crc kubenswrapper[4805]: I1203 14:32:15.403501 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9ed6dfc0-82db-4d86-ad92-cfe94fa370a2-combined-ca-bundle\") pod \"9ed6dfc0-82db-4d86-ad92-cfe94fa370a2\" (UID: \"9ed6dfc0-82db-4d86-ad92-cfe94fa370a2\") " Dec 03 14:32:15 crc kubenswrapper[4805]: I1203 14:32:15.403521 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-c5677\" (UniqueName: \"kubernetes.io/projected/9ed6dfc0-82db-4d86-ad92-cfe94fa370a2-kube-api-access-c5677\") pod \"9ed6dfc0-82db-4d86-ad92-cfe94fa370a2\" (UID: \"9ed6dfc0-82db-4d86-ad92-cfe94fa370a2\") " Dec 03 14:32:15 crc kubenswrapper[4805]: I1203 14:32:15.403586 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/9ed6dfc0-82db-4d86-ad92-cfe94fa370a2-public-tls-certs\") pod \"9ed6dfc0-82db-4d86-ad92-cfe94fa370a2\" (UID: \"9ed6dfc0-82db-4d86-ad92-cfe94fa370a2\") " Dec 03 14:32:15 crc kubenswrapper[4805]: I1203 14:32:15.403663 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9ed6dfc0-82db-4d86-ad92-cfe94fa370a2-config-data\") pod \"9ed6dfc0-82db-4d86-ad92-cfe94fa370a2\" (UID: \"9ed6dfc0-82db-4d86-ad92-cfe94fa370a2\") " Dec 03 14:32:15 crc kubenswrapper[4805]: I1203 14:32:15.403712 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/9ed6dfc0-82db-4d86-ad92-cfe94fa370a2-logs\") pod \"9ed6dfc0-82db-4d86-ad92-cfe94fa370a2\" (UID: \"9ed6dfc0-82db-4d86-ad92-cfe94fa370a2\") " Dec 03 14:32:15 crc kubenswrapper[4805]: I1203 14:32:15.403738 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"9ed6dfc0-82db-4d86-ad92-cfe94fa370a2\" (UID: \"9ed6dfc0-82db-4d86-ad92-cfe94fa370a2\") " Dec 03 14:32:15 crc kubenswrapper[4805]: I1203 14:32:15.403760 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9ed6dfc0-82db-4d86-ad92-cfe94fa370a2-scripts\") pod \"9ed6dfc0-82db-4d86-ad92-cfe94fa370a2\" (UID: \"9ed6dfc0-82db-4d86-ad92-cfe94fa370a2\") " Dec 03 14:32:15 crc kubenswrapper[4805]: I1203 14:32:15.405735 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9ed6dfc0-82db-4d86-ad92-cfe94fa370a2-logs" (OuterVolumeSpecName: "logs") pod "9ed6dfc0-82db-4d86-ad92-cfe94fa370a2" (UID: "9ed6dfc0-82db-4d86-ad92-cfe94fa370a2"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 14:32:15 crc kubenswrapper[4805]: I1203 14:32:15.407212 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9ed6dfc0-82db-4d86-ad92-cfe94fa370a2-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "9ed6dfc0-82db-4d86-ad92-cfe94fa370a2" (UID: "9ed6dfc0-82db-4d86-ad92-cfe94fa370a2"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 14:32:15 crc kubenswrapper[4805]: I1203 14:32:15.416155 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9ed6dfc0-82db-4d86-ad92-cfe94fa370a2-scripts" (OuterVolumeSpecName: "scripts") pod "9ed6dfc0-82db-4d86-ad92-cfe94fa370a2" (UID: "9ed6dfc0-82db-4d86-ad92-cfe94fa370a2"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 14:32:15 crc kubenswrapper[4805]: I1203 14:32:15.423201 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9ed6dfc0-82db-4d86-ad92-cfe94fa370a2-kube-api-access-c5677" (OuterVolumeSpecName: "kube-api-access-c5677") pod "9ed6dfc0-82db-4d86-ad92-cfe94fa370a2" (UID: "9ed6dfc0-82db-4d86-ad92-cfe94fa370a2"). InnerVolumeSpecName "kube-api-access-c5677". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 14:32:15 crc kubenswrapper[4805]: I1203 14:32:15.423700 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage07-crc" (OuterVolumeSpecName: "glance") pod "9ed6dfc0-82db-4d86-ad92-cfe94fa370a2" (UID: "9ed6dfc0-82db-4d86-ad92-cfe94fa370a2"). InnerVolumeSpecName "local-storage07-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Dec 03 14:32:15 crc kubenswrapper[4805]: I1203 14:32:15.452402 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9ed6dfc0-82db-4d86-ad92-cfe94fa370a2-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "9ed6dfc0-82db-4d86-ad92-cfe94fa370a2" (UID: "9ed6dfc0-82db-4d86-ad92-cfe94fa370a2"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 14:32:15 crc kubenswrapper[4805]: I1203 14:32:15.504453 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9ed6dfc0-82db-4d86-ad92-cfe94fa370a2-config-data" (OuterVolumeSpecName: "config-data") pod "9ed6dfc0-82db-4d86-ad92-cfe94fa370a2" (UID: "9ed6dfc0-82db-4d86-ad92-cfe94fa370a2"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 14:32:15 crc kubenswrapper[4805]: I1203 14:32:15.505951 4805 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/9ed6dfc0-82db-4d86-ad92-cfe94fa370a2-httpd-run\") on node \"crc\" DevicePath \"\"" Dec 03 14:32:15 crc kubenswrapper[4805]: I1203 14:32:15.506068 4805 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9ed6dfc0-82db-4d86-ad92-cfe94fa370a2-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 14:32:15 crc kubenswrapper[4805]: I1203 14:32:15.506160 4805 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-c5677\" (UniqueName: \"kubernetes.io/projected/9ed6dfc0-82db-4d86-ad92-cfe94fa370a2-kube-api-access-c5677\") on node \"crc\" DevicePath \"\"" Dec 03 14:32:15 crc kubenswrapper[4805]: I1203 14:32:15.506248 4805 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9ed6dfc0-82db-4d86-ad92-cfe94fa370a2-config-data\") on node \"crc\" DevicePath \"\"" Dec 03 14:32:15 crc kubenswrapper[4805]: I1203 14:32:15.506332 4805 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/9ed6dfc0-82db-4d86-ad92-cfe94fa370a2-logs\") on node \"crc\" DevicePath \"\"" Dec 03 14:32:15 crc kubenswrapper[4805]: I1203 14:32:15.506441 4805 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") on node \"crc\" " Dec 03 14:32:15 crc kubenswrapper[4805]: I1203 14:32:15.506544 4805 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9ed6dfc0-82db-4d86-ad92-cfe94fa370a2-scripts\") on node \"crc\" DevicePath \"\"" Dec 03 14:32:15 crc kubenswrapper[4805]: I1203 14:32:15.511116 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9ed6dfc0-82db-4d86-ad92-cfe94fa370a2-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "9ed6dfc0-82db-4d86-ad92-cfe94fa370a2" (UID: "9ed6dfc0-82db-4d86-ad92-cfe94fa370a2"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 14:32:15 crc kubenswrapper[4805]: I1203 14:32:15.563406 4805 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage07-crc" (UniqueName: "kubernetes.io/local-volume/local-storage07-crc") on node "crc" Dec 03 14:32:15 crc kubenswrapper[4805]: I1203 14:32:15.608619 4805 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/9ed6dfc0-82db-4d86-ad92-cfe94fa370a2-public-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 03 14:32:15 crc kubenswrapper[4805]: I1203 14:32:15.608660 4805 reconciler_common.go:293] "Volume detached for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") on node \"crc\" DevicePath \"\"" Dec 03 14:32:15 crc kubenswrapper[4805]: I1203 14:32:15.631311 4805 generic.go:334] "Generic (PLEG): container finished" podID="9ed6dfc0-82db-4d86-ad92-cfe94fa370a2" containerID="b40b68276f1683e9c153b03843673b0fb798f28865161b6b4f97250e1dcc86f9" exitCode=0 Dec 03 14:32:15 crc kubenswrapper[4805]: I1203 14:32:15.631650 4805 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Dec 03 14:32:15 crc kubenswrapper[4805]: I1203 14:32:15.631969 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"9ed6dfc0-82db-4d86-ad92-cfe94fa370a2","Type":"ContainerDied","Data":"b40b68276f1683e9c153b03843673b0fb798f28865161b6b4f97250e1dcc86f9"} Dec 03 14:32:15 crc kubenswrapper[4805]: I1203 14:32:15.632121 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"9ed6dfc0-82db-4d86-ad92-cfe94fa370a2","Type":"ContainerDied","Data":"fb9cf095ff0fd5118500bc41d0e009beeadf259bbfe428a14abc1cdf0703dbfa"} Dec 03 14:32:15 crc kubenswrapper[4805]: I1203 14:32:15.632146 4805 scope.go:117] "RemoveContainer" containerID="b40b68276f1683e9c153b03843673b0fb798f28865161b6b4f97250e1dcc86f9" Dec 03 14:32:15 crc kubenswrapper[4805]: I1203 14:32:15.635322 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"6d34a555-7003-4796-a15a-85a71c489bc9","Type":"ContainerStarted","Data":"530845f60df6d78fe435f6905ab3eb744ea88bc262f9f045084d6a3083f12865"} Dec 03 14:32:15 crc kubenswrapper[4805]: I1203 14:32:15.680929 4805 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Dec 03 14:32:15 crc kubenswrapper[4805]: I1203 14:32:15.706441 4805 scope.go:117] "RemoveContainer" containerID="18b272180201caa258fce70a3153b9e82a5103c2066cac10db4b226952843d2f" Dec 03 14:32:15 crc kubenswrapper[4805]: I1203 14:32:15.706566 4805 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-default-external-api-0"] Dec 03 14:32:15 crc kubenswrapper[4805]: I1203 14:32:15.715485 4805 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-external-api-0"] Dec 03 14:32:15 crc kubenswrapper[4805]: E1203 14:32:15.715999 4805 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9ed6dfc0-82db-4d86-ad92-cfe94fa370a2" containerName="glance-httpd" Dec 03 14:32:15 crc kubenswrapper[4805]: I1203 14:32:15.716021 4805 state_mem.go:107] "Deleted CPUSet assignment" podUID="9ed6dfc0-82db-4d86-ad92-cfe94fa370a2" containerName="glance-httpd" Dec 03 14:32:15 crc kubenswrapper[4805]: E1203 14:32:15.716051 4805 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9ed6dfc0-82db-4d86-ad92-cfe94fa370a2" containerName="glance-log" Dec 03 14:32:15 crc kubenswrapper[4805]: I1203 14:32:15.716059 4805 state_mem.go:107] "Deleted CPUSet assignment" podUID="9ed6dfc0-82db-4d86-ad92-cfe94fa370a2" containerName="glance-log" Dec 03 14:32:15 crc kubenswrapper[4805]: I1203 14:32:15.716301 4805 memory_manager.go:354] "RemoveStaleState removing state" podUID="9ed6dfc0-82db-4d86-ad92-cfe94fa370a2" containerName="glance-log" Dec 03 14:32:15 crc kubenswrapper[4805]: I1203 14:32:15.716317 4805 memory_manager.go:354] "RemoveStaleState removing state" podUID="9ed6dfc0-82db-4d86-ad92-cfe94fa370a2" containerName="glance-httpd" Dec 03 14:32:15 crc kubenswrapper[4805]: I1203 14:32:15.717567 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Dec 03 14:32:15 crc kubenswrapper[4805]: I1203 14:32:15.721890 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-public-svc" Dec 03 14:32:15 crc kubenswrapper[4805]: I1203 14:32:15.723392 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Dec 03 14:32:15 crc kubenswrapper[4805]: I1203 14:32:15.724258 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-external-config-data" Dec 03 14:32:15 crc kubenswrapper[4805]: I1203 14:32:15.779999 4805 scope.go:117] "RemoveContainer" containerID="b40b68276f1683e9c153b03843673b0fb798f28865161b6b4f97250e1dcc86f9" Dec 03 14:32:15 crc kubenswrapper[4805]: E1203 14:32:15.781620 4805 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b40b68276f1683e9c153b03843673b0fb798f28865161b6b4f97250e1dcc86f9\": container with ID starting with b40b68276f1683e9c153b03843673b0fb798f28865161b6b4f97250e1dcc86f9 not found: ID does not exist" containerID="b40b68276f1683e9c153b03843673b0fb798f28865161b6b4f97250e1dcc86f9" Dec 03 14:32:15 crc kubenswrapper[4805]: I1203 14:32:15.781684 4805 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b40b68276f1683e9c153b03843673b0fb798f28865161b6b4f97250e1dcc86f9"} err="failed to get container status \"b40b68276f1683e9c153b03843673b0fb798f28865161b6b4f97250e1dcc86f9\": rpc error: code = NotFound desc = could not find container \"b40b68276f1683e9c153b03843673b0fb798f28865161b6b4f97250e1dcc86f9\": container with ID starting with b40b68276f1683e9c153b03843673b0fb798f28865161b6b4f97250e1dcc86f9 not found: ID does not exist" Dec 03 14:32:15 crc kubenswrapper[4805]: I1203 14:32:15.781712 4805 scope.go:117] "RemoveContainer" containerID="18b272180201caa258fce70a3153b9e82a5103c2066cac10db4b226952843d2f" Dec 03 14:32:15 crc kubenswrapper[4805]: E1203 14:32:15.786253 4805 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"18b272180201caa258fce70a3153b9e82a5103c2066cac10db4b226952843d2f\": container with ID starting with 18b272180201caa258fce70a3153b9e82a5103c2066cac10db4b226952843d2f not found: ID does not exist" containerID="18b272180201caa258fce70a3153b9e82a5103c2066cac10db4b226952843d2f" Dec 03 14:32:15 crc kubenswrapper[4805]: I1203 14:32:15.786308 4805 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"18b272180201caa258fce70a3153b9e82a5103c2066cac10db4b226952843d2f"} err="failed to get container status \"18b272180201caa258fce70a3153b9e82a5103c2066cac10db4b226952843d2f\": rpc error: code = NotFound desc = could not find container \"18b272180201caa258fce70a3153b9e82a5103c2066cac10db4b226952843d2f\": container with ID starting with 18b272180201caa258fce70a3153b9e82a5103c2066cac10db4b226952843d2f not found: ID does not exist" Dec 03 14:32:15 crc kubenswrapper[4805]: I1203 14:32:15.913801 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mfjqc\" (UniqueName: \"kubernetes.io/projected/6a4ca0f4-cff3-4c57-ab1b-3a01d7d6576c-kube-api-access-mfjqc\") pod \"glance-default-external-api-0\" (UID: \"6a4ca0f4-cff3-4c57-ab1b-3a01d7d6576c\") " pod="openstack/glance-default-external-api-0" Dec 03 14:32:15 crc kubenswrapper[4805]: I1203 14:32:15.913920 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6a4ca0f4-cff3-4c57-ab1b-3a01d7d6576c-scripts\") pod \"glance-default-external-api-0\" (UID: \"6a4ca0f4-cff3-4c57-ab1b-3a01d7d6576c\") " pod="openstack/glance-default-external-api-0" Dec 03 14:32:15 crc kubenswrapper[4805]: I1203 14:32:15.913976 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/6a4ca0f4-cff3-4c57-ab1b-3a01d7d6576c-logs\") pod \"glance-default-external-api-0\" (UID: \"6a4ca0f4-cff3-4c57-ab1b-3a01d7d6576c\") " pod="openstack/glance-default-external-api-0" Dec 03 14:32:15 crc kubenswrapper[4805]: I1203 14:32:15.914024 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/6a4ca0f4-cff3-4c57-ab1b-3a01d7d6576c-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"6a4ca0f4-cff3-4c57-ab1b-3a01d7d6576c\") " pod="openstack/glance-default-external-api-0" Dec 03 14:32:15 crc kubenswrapper[4805]: I1203 14:32:15.914175 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/6a4ca0f4-cff3-4c57-ab1b-3a01d7d6576c-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"6a4ca0f4-cff3-4c57-ab1b-3a01d7d6576c\") " pod="openstack/glance-default-external-api-0" Dec 03 14:32:15 crc kubenswrapper[4805]: I1203 14:32:15.914217 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"glance-default-external-api-0\" (UID: \"6a4ca0f4-cff3-4c57-ab1b-3a01d7d6576c\") " pod="openstack/glance-default-external-api-0" Dec 03 14:32:15 crc kubenswrapper[4805]: I1203 14:32:15.914376 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6a4ca0f4-cff3-4c57-ab1b-3a01d7d6576c-config-data\") pod \"glance-default-external-api-0\" (UID: \"6a4ca0f4-cff3-4c57-ab1b-3a01d7d6576c\") " pod="openstack/glance-default-external-api-0" Dec 03 14:32:15 crc kubenswrapper[4805]: I1203 14:32:15.914431 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6a4ca0f4-cff3-4c57-ab1b-3a01d7d6576c-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"6a4ca0f4-cff3-4c57-ab1b-3a01d7d6576c\") " pod="openstack/glance-default-external-api-0" Dec 03 14:32:16 crc kubenswrapper[4805]: I1203 14:32:16.016146 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/6a4ca0f4-cff3-4c57-ab1b-3a01d7d6576c-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"6a4ca0f4-cff3-4c57-ab1b-3a01d7d6576c\") " pod="openstack/glance-default-external-api-0" Dec 03 14:32:16 crc kubenswrapper[4805]: I1203 14:32:16.016189 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"glance-default-external-api-0\" (UID: \"6a4ca0f4-cff3-4c57-ab1b-3a01d7d6576c\") " pod="openstack/glance-default-external-api-0" Dec 03 14:32:16 crc kubenswrapper[4805]: I1203 14:32:16.016221 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6a4ca0f4-cff3-4c57-ab1b-3a01d7d6576c-config-data\") pod \"glance-default-external-api-0\" (UID: \"6a4ca0f4-cff3-4c57-ab1b-3a01d7d6576c\") " pod="openstack/glance-default-external-api-0" Dec 03 14:32:16 crc kubenswrapper[4805]: I1203 14:32:16.016251 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6a4ca0f4-cff3-4c57-ab1b-3a01d7d6576c-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"6a4ca0f4-cff3-4c57-ab1b-3a01d7d6576c\") " pod="openstack/glance-default-external-api-0" Dec 03 14:32:16 crc kubenswrapper[4805]: I1203 14:32:16.016320 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mfjqc\" (UniqueName: \"kubernetes.io/projected/6a4ca0f4-cff3-4c57-ab1b-3a01d7d6576c-kube-api-access-mfjqc\") pod \"glance-default-external-api-0\" (UID: \"6a4ca0f4-cff3-4c57-ab1b-3a01d7d6576c\") " pod="openstack/glance-default-external-api-0" Dec 03 14:32:16 crc kubenswrapper[4805]: I1203 14:32:16.016345 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6a4ca0f4-cff3-4c57-ab1b-3a01d7d6576c-scripts\") pod \"glance-default-external-api-0\" (UID: \"6a4ca0f4-cff3-4c57-ab1b-3a01d7d6576c\") " pod="openstack/glance-default-external-api-0" Dec 03 14:32:16 crc kubenswrapper[4805]: I1203 14:32:16.016381 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/6a4ca0f4-cff3-4c57-ab1b-3a01d7d6576c-logs\") pod \"glance-default-external-api-0\" (UID: \"6a4ca0f4-cff3-4c57-ab1b-3a01d7d6576c\") " pod="openstack/glance-default-external-api-0" Dec 03 14:32:16 crc kubenswrapper[4805]: I1203 14:32:16.016412 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/6a4ca0f4-cff3-4c57-ab1b-3a01d7d6576c-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"6a4ca0f4-cff3-4c57-ab1b-3a01d7d6576c\") " pod="openstack/glance-default-external-api-0" Dec 03 14:32:16 crc kubenswrapper[4805]: I1203 14:32:16.016755 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/6a4ca0f4-cff3-4c57-ab1b-3a01d7d6576c-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"6a4ca0f4-cff3-4c57-ab1b-3a01d7d6576c\") " pod="openstack/glance-default-external-api-0" Dec 03 14:32:16 crc kubenswrapper[4805]: I1203 14:32:16.016797 4805 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"glance-default-external-api-0\" (UID: \"6a4ca0f4-cff3-4c57-ab1b-3a01d7d6576c\") device mount path \"/mnt/openstack/pv07\"" pod="openstack/glance-default-external-api-0" Dec 03 14:32:16 crc kubenswrapper[4805]: I1203 14:32:16.017160 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/6a4ca0f4-cff3-4c57-ab1b-3a01d7d6576c-logs\") pod \"glance-default-external-api-0\" (UID: \"6a4ca0f4-cff3-4c57-ab1b-3a01d7d6576c\") " pod="openstack/glance-default-external-api-0" Dec 03 14:32:16 crc kubenswrapper[4805]: I1203 14:32:16.024139 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6a4ca0f4-cff3-4c57-ab1b-3a01d7d6576c-config-data\") pod \"glance-default-external-api-0\" (UID: \"6a4ca0f4-cff3-4c57-ab1b-3a01d7d6576c\") " pod="openstack/glance-default-external-api-0" Dec 03 14:32:16 crc kubenswrapper[4805]: I1203 14:32:16.024692 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/6a4ca0f4-cff3-4c57-ab1b-3a01d7d6576c-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"6a4ca0f4-cff3-4c57-ab1b-3a01d7d6576c\") " pod="openstack/glance-default-external-api-0" Dec 03 14:32:16 crc kubenswrapper[4805]: I1203 14:32:16.027314 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6a4ca0f4-cff3-4c57-ab1b-3a01d7d6576c-scripts\") pod \"glance-default-external-api-0\" (UID: \"6a4ca0f4-cff3-4c57-ab1b-3a01d7d6576c\") " pod="openstack/glance-default-external-api-0" Dec 03 14:32:16 crc kubenswrapper[4805]: I1203 14:32:16.035947 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6a4ca0f4-cff3-4c57-ab1b-3a01d7d6576c-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"6a4ca0f4-cff3-4c57-ab1b-3a01d7d6576c\") " pod="openstack/glance-default-external-api-0" Dec 03 14:32:16 crc kubenswrapper[4805]: I1203 14:32:16.036474 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mfjqc\" (UniqueName: \"kubernetes.io/projected/6a4ca0f4-cff3-4c57-ab1b-3a01d7d6576c-kube-api-access-mfjqc\") pod \"glance-default-external-api-0\" (UID: \"6a4ca0f4-cff3-4c57-ab1b-3a01d7d6576c\") " pod="openstack/glance-default-external-api-0" Dec 03 14:32:16 crc kubenswrapper[4805]: I1203 14:32:16.057735 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"glance-default-external-api-0\" (UID: \"6a4ca0f4-cff3-4c57-ab1b-3a01d7d6576c\") " pod="openstack/glance-default-external-api-0" Dec 03 14:32:16 crc kubenswrapper[4805]: I1203 14:32:16.340790 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Dec 03 14:32:16 crc kubenswrapper[4805]: I1203 14:32:16.676214 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"6d34a555-7003-4796-a15a-85a71c489bc9","Type":"ContainerStarted","Data":"671ce8c0950cc229937f19ea3fa0fddba97d6701d80adfe4d77e8b1192af083a"} Dec 03 14:32:16 crc kubenswrapper[4805]: I1203 14:32:16.682891 4805 generic.go:334] "Generic (PLEG): container finished" podID="10877f19-dd2a-4eb2-99e0-2d0cf82d5ad5" containerID="61aa5396071fa72fd8a539334f4fa1f4982e07e997aee839e400e0b116fff32e" exitCode=0 Dec 03 14:32:16 crc kubenswrapper[4805]: I1203 14:32:16.683025 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"10877f19-dd2a-4eb2-99e0-2d0cf82d5ad5","Type":"ContainerDied","Data":"61aa5396071fa72fd8a539334f4fa1f4982e07e997aee839e400e0b116fff32e"} Dec 03 14:32:16 crc kubenswrapper[4805]: I1203 14:32:16.694798 4805 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-scheduler-0" podStartSLOduration=3.694763176 podStartE2EDuration="3.694763176s" podCreationTimestamp="2025-12-03 14:32:13 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 14:32:16.691894028 +0000 UTC m=+1366.354810951" watchObservedRunningTime="2025-12-03 14:32:16.694763176 +0000 UTC m=+1366.357680099" Dec 03 14:32:16 crc kubenswrapper[4805]: I1203 14:32:16.714211 4805 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9ed6dfc0-82db-4d86-ad92-cfe94fa370a2" path="/var/lib/kubelet/pods/9ed6dfc0-82db-4d86-ad92-cfe94fa370a2/volumes" Dec 03 14:32:17 crc kubenswrapper[4805]: I1203 14:32:17.012648 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Dec 03 14:32:17 crc kubenswrapper[4805]: W1203 14:32:17.019867 4805 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod6a4ca0f4_cff3_4c57_ab1b_3a01d7d6576c.slice/crio-16a9627c25406295cf5584c868d22ee03909038bb1d39487d1d910d12133c1e3 WatchSource:0}: Error finding container 16a9627c25406295cf5584c868d22ee03909038bb1d39487d1d910d12133c1e3: Status 404 returned error can't find the container with id 16a9627c25406295cf5584c868d22ee03909038bb1d39487d1d910d12133c1e3 Dec 03 14:32:17 crc kubenswrapper[4805]: I1203 14:32:17.142216 4805 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Dec 03 14:32:17 crc kubenswrapper[4805]: I1203 14:32:17.247435 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/10877f19-dd2a-4eb2-99e0-2d0cf82d5ad5-logs\") pod \"10877f19-dd2a-4eb2-99e0-2d0cf82d5ad5\" (UID: \"10877f19-dd2a-4eb2-99e0-2d0cf82d5ad5\") " Dec 03 14:32:17 crc kubenswrapper[4805]: I1203 14:32:17.247487 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/10877f19-dd2a-4eb2-99e0-2d0cf82d5ad5-internal-tls-certs\") pod \"10877f19-dd2a-4eb2-99e0-2d0cf82d5ad5\" (UID: \"10877f19-dd2a-4eb2-99e0-2d0cf82d5ad5\") " Dec 03 14:32:17 crc kubenswrapper[4805]: I1203 14:32:17.247550 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/10877f19-dd2a-4eb2-99e0-2d0cf82d5ad5-config-data\") pod \"10877f19-dd2a-4eb2-99e0-2d0cf82d5ad5\" (UID: \"10877f19-dd2a-4eb2-99e0-2d0cf82d5ad5\") " Dec 03 14:32:17 crc kubenswrapper[4805]: I1203 14:32:17.247616 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mf6bl\" (UniqueName: \"kubernetes.io/projected/10877f19-dd2a-4eb2-99e0-2d0cf82d5ad5-kube-api-access-mf6bl\") pod \"10877f19-dd2a-4eb2-99e0-2d0cf82d5ad5\" (UID: \"10877f19-dd2a-4eb2-99e0-2d0cf82d5ad5\") " Dec 03 14:32:17 crc kubenswrapper[4805]: I1203 14:32:17.247637 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"10877f19-dd2a-4eb2-99e0-2d0cf82d5ad5\" (UID: \"10877f19-dd2a-4eb2-99e0-2d0cf82d5ad5\") " Dec 03 14:32:17 crc kubenswrapper[4805]: I1203 14:32:17.247655 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/10877f19-dd2a-4eb2-99e0-2d0cf82d5ad5-httpd-run\") pod \"10877f19-dd2a-4eb2-99e0-2d0cf82d5ad5\" (UID: \"10877f19-dd2a-4eb2-99e0-2d0cf82d5ad5\") " Dec 03 14:32:17 crc kubenswrapper[4805]: I1203 14:32:17.247685 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/10877f19-dd2a-4eb2-99e0-2d0cf82d5ad5-combined-ca-bundle\") pod \"10877f19-dd2a-4eb2-99e0-2d0cf82d5ad5\" (UID: \"10877f19-dd2a-4eb2-99e0-2d0cf82d5ad5\") " Dec 03 14:32:17 crc kubenswrapper[4805]: I1203 14:32:17.247728 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/10877f19-dd2a-4eb2-99e0-2d0cf82d5ad5-scripts\") pod \"10877f19-dd2a-4eb2-99e0-2d0cf82d5ad5\" (UID: \"10877f19-dd2a-4eb2-99e0-2d0cf82d5ad5\") " Dec 03 14:32:17 crc kubenswrapper[4805]: I1203 14:32:17.248650 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/10877f19-dd2a-4eb2-99e0-2d0cf82d5ad5-logs" (OuterVolumeSpecName: "logs") pod "10877f19-dd2a-4eb2-99e0-2d0cf82d5ad5" (UID: "10877f19-dd2a-4eb2-99e0-2d0cf82d5ad5"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 14:32:17 crc kubenswrapper[4805]: I1203 14:32:17.249470 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/10877f19-dd2a-4eb2-99e0-2d0cf82d5ad5-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "10877f19-dd2a-4eb2-99e0-2d0cf82d5ad5" (UID: "10877f19-dd2a-4eb2-99e0-2d0cf82d5ad5"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 14:32:17 crc kubenswrapper[4805]: I1203 14:32:17.256189 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/10877f19-dd2a-4eb2-99e0-2d0cf82d5ad5-scripts" (OuterVolumeSpecName: "scripts") pod "10877f19-dd2a-4eb2-99e0-2d0cf82d5ad5" (UID: "10877f19-dd2a-4eb2-99e0-2d0cf82d5ad5"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 14:32:17 crc kubenswrapper[4805]: I1203 14:32:17.259987 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage11-crc" (OuterVolumeSpecName: "glance") pod "10877f19-dd2a-4eb2-99e0-2d0cf82d5ad5" (UID: "10877f19-dd2a-4eb2-99e0-2d0cf82d5ad5"). InnerVolumeSpecName "local-storage11-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Dec 03 14:32:17 crc kubenswrapper[4805]: I1203 14:32:17.274113 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/10877f19-dd2a-4eb2-99e0-2d0cf82d5ad5-kube-api-access-mf6bl" (OuterVolumeSpecName: "kube-api-access-mf6bl") pod "10877f19-dd2a-4eb2-99e0-2d0cf82d5ad5" (UID: "10877f19-dd2a-4eb2-99e0-2d0cf82d5ad5"). InnerVolumeSpecName "kube-api-access-mf6bl". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 14:32:17 crc kubenswrapper[4805]: I1203 14:32:17.354452 4805 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mf6bl\" (UniqueName: \"kubernetes.io/projected/10877f19-dd2a-4eb2-99e0-2d0cf82d5ad5-kube-api-access-mf6bl\") on node \"crc\" DevicePath \"\"" Dec 03 14:32:17 crc kubenswrapper[4805]: I1203 14:32:17.354518 4805 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") on node \"crc\" " Dec 03 14:32:17 crc kubenswrapper[4805]: I1203 14:32:17.354532 4805 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/10877f19-dd2a-4eb2-99e0-2d0cf82d5ad5-httpd-run\") on node \"crc\" DevicePath \"\"" Dec 03 14:32:17 crc kubenswrapper[4805]: I1203 14:32:17.354546 4805 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/10877f19-dd2a-4eb2-99e0-2d0cf82d5ad5-scripts\") on node \"crc\" DevicePath \"\"" Dec 03 14:32:17 crc kubenswrapper[4805]: I1203 14:32:17.354556 4805 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/10877f19-dd2a-4eb2-99e0-2d0cf82d5ad5-logs\") on node \"crc\" DevicePath \"\"" Dec 03 14:32:17 crc kubenswrapper[4805]: I1203 14:32:17.357903 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/10877f19-dd2a-4eb2-99e0-2d0cf82d5ad5-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "10877f19-dd2a-4eb2-99e0-2d0cf82d5ad5" (UID: "10877f19-dd2a-4eb2-99e0-2d0cf82d5ad5"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 14:32:17 crc kubenswrapper[4805]: I1203 14:32:17.375005 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/10877f19-dd2a-4eb2-99e0-2d0cf82d5ad5-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "10877f19-dd2a-4eb2-99e0-2d0cf82d5ad5" (UID: "10877f19-dd2a-4eb2-99e0-2d0cf82d5ad5"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 14:32:17 crc kubenswrapper[4805]: I1203 14:32:17.381489 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/10877f19-dd2a-4eb2-99e0-2d0cf82d5ad5-config-data" (OuterVolumeSpecName: "config-data") pod "10877f19-dd2a-4eb2-99e0-2d0cf82d5ad5" (UID: "10877f19-dd2a-4eb2-99e0-2d0cf82d5ad5"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 14:32:17 crc kubenswrapper[4805]: I1203 14:32:17.399461 4805 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage11-crc" (UniqueName: "kubernetes.io/local-volume/local-storage11-crc") on node "crc" Dec 03 14:32:17 crc kubenswrapper[4805]: I1203 14:32:17.455930 4805 reconciler_common.go:293] "Volume detached for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") on node \"crc\" DevicePath \"\"" Dec 03 14:32:17 crc kubenswrapper[4805]: I1203 14:32:17.456323 4805 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/10877f19-dd2a-4eb2-99e0-2d0cf82d5ad5-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 14:32:17 crc kubenswrapper[4805]: I1203 14:32:17.456336 4805 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/10877f19-dd2a-4eb2-99e0-2d0cf82d5ad5-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 03 14:32:17 crc kubenswrapper[4805]: I1203 14:32:17.456348 4805 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/10877f19-dd2a-4eb2-99e0-2d0cf82d5ad5-config-data\") on node \"crc\" DevicePath \"\"" Dec 03 14:32:17 crc kubenswrapper[4805]: I1203 14:32:17.697906 4805 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Dec 03 14:32:17 crc kubenswrapper[4805]: I1203 14:32:17.697905 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"10877f19-dd2a-4eb2-99e0-2d0cf82d5ad5","Type":"ContainerDied","Data":"53cfaf6021f91b400709e52ab56b25b8a1e47a90e0bf4947947db07aec4309ef"} Dec 03 14:32:17 crc kubenswrapper[4805]: I1203 14:32:17.698029 4805 scope.go:117] "RemoveContainer" containerID="61aa5396071fa72fd8a539334f4fa1f4982e07e997aee839e400e0b116fff32e" Dec 03 14:32:17 crc kubenswrapper[4805]: I1203 14:32:17.701125 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"6a4ca0f4-cff3-4c57-ab1b-3a01d7d6576c","Type":"ContainerStarted","Data":"16a9627c25406295cf5584c868d22ee03909038bb1d39487d1d910d12133c1e3"} Dec 03 14:32:17 crc kubenswrapper[4805]: I1203 14:32:17.730059 4805 scope.go:117] "RemoveContainer" containerID="24356197f69f0bfd999f1654211fef34606d3c50328026f4b3c6d2eda195db85" Dec 03 14:32:17 crc kubenswrapper[4805]: I1203 14:32:17.741936 4805 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 03 14:32:17 crc kubenswrapper[4805]: I1203 14:32:17.750017 4805 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 03 14:32:17 crc kubenswrapper[4805]: I1203 14:32:17.784322 4805 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 03 14:32:17 crc kubenswrapper[4805]: E1203 14:32:17.784680 4805 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="10877f19-dd2a-4eb2-99e0-2d0cf82d5ad5" containerName="glance-log" Dec 03 14:32:17 crc kubenswrapper[4805]: I1203 14:32:17.784692 4805 state_mem.go:107] "Deleted CPUSet assignment" podUID="10877f19-dd2a-4eb2-99e0-2d0cf82d5ad5" containerName="glance-log" Dec 03 14:32:17 crc kubenswrapper[4805]: E1203 14:32:17.784705 4805 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="10877f19-dd2a-4eb2-99e0-2d0cf82d5ad5" containerName="glance-httpd" Dec 03 14:32:17 crc kubenswrapper[4805]: I1203 14:32:17.784711 4805 state_mem.go:107] "Deleted CPUSet assignment" podUID="10877f19-dd2a-4eb2-99e0-2d0cf82d5ad5" containerName="glance-httpd" Dec 03 14:32:17 crc kubenswrapper[4805]: I1203 14:32:17.784996 4805 memory_manager.go:354] "RemoveStaleState removing state" podUID="10877f19-dd2a-4eb2-99e0-2d0cf82d5ad5" containerName="glance-httpd" Dec 03 14:32:17 crc kubenswrapper[4805]: I1203 14:32:17.785023 4805 memory_manager.go:354] "RemoveStaleState removing state" podUID="10877f19-dd2a-4eb2-99e0-2d0cf82d5ad5" containerName="glance-log" Dec 03 14:32:17 crc kubenswrapper[4805]: I1203 14:32:17.786132 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Dec 03 14:32:17 crc kubenswrapper[4805]: I1203 14:32:17.808449 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-internal-svc" Dec 03 14:32:17 crc kubenswrapper[4805]: I1203 14:32:17.809113 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-internal-config-data" Dec 03 14:32:17 crc kubenswrapper[4805]: I1203 14:32:17.878457 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"glance-default-internal-api-0\" (UID: \"cbb9e604-033b-4586-b0fd-7300a5e30897\") " pod="openstack/glance-default-internal-api-0" Dec 03 14:32:17 crc kubenswrapper[4805]: I1203 14:32:17.878597 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/cbb9e604-033b-4586-b0fd-7300a5e30897-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"cbb9e604-033b-4586-b0fd-7300a5e30897\") " pod="openstack/glance-default-internal-api-0" Dec 03 14:32:17 crc kubenswrapper[4805]: I1203 14:32:17.878713 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lkg52\" (UniqueName: \"kubernetes.io/projected/cbb9e604-033b-4586-b0fd-7300a5e30897-kube-api-access-lkg52\") pod \"glance-default-internal-api-0\" (UID: \"cbb9e604-033b-4586-b0fd-7300a5e30897\") " pod="openstack/glance-default-internal-api-0" Dec 03 14:32:17 crc kubenswrapper[4805]: I1203 14:32:17.878748 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cbb9e604-033b-4586-b0fd-7300a5e30897-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"cbb9e604-033b-4586-b0fd-7300a5e30897\") " pod="openstack/glance-default-internal-api-0" Dec 03 14:32:17 crc kubenswrapper[4805]: I1203 14:32:17.878793 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/cbb9e604-033b-4586-b0fd-7300a5e30897-scripts\") pod \"glance-default-internal-api-0\" (UID: \"cbb9e604-033b-4586-b0fd-7300a5e30897\") " pod="openstack/glance-default-internal-api-0" Dec 03 14:32:17 crc kubenswrapper[4805]: I1203 14:32:17.878934 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cbb9e604-033b-4586-b0fd-7300a5e30897-config-data\") pod \"glance-default-internal-api-0\" (UID: \"cbb9e604-033b-4586-b0fd-7300a5e30897\") " pod="openstack/glance-default-internal-api-0" Dec 03 14:32:17 crc kubenswrapper[4805]: I1203 14:32:17.879063 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/cbb9e604-033b-4586-b0fd-7300a5e30897-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"cbb9e604-033b-4586-b0fd-7300a5e30897\") " pod="openstack/glance-default-internal-api-0" Dec 03 14:32:17 crc kubenswrapper[4805]: I1203 14:32:17.879133 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/cbb9e604-033b-4586-b0fd-7300a5e30897-logs\") pod \"glance-default-internal-api-0\" (UID: \"cbb9e604-033b-4586-b0fd-7300a5e30897\") " pod="openstack/glance-default-internal-api-0" Dec 03 14:32:17 crc kubenswrapper[4805]: I1203 14:32:17.913433 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 03 14:32:17 crc kubenswrapper[4805]: I1203 14:32:17.981055 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/cbb9e604-033b-4586-b0fd-7300a5e30897-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"cbb9e604-033b-4586-b0fd-7300a5e30897\") " pod="openstack/glance-default-internal-api-0" Dec 03 14:32:17 crc kubenswrapper[4805]: I1203 14:32:17.981130 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/cbb9e604-033b-4586-b0fd-7300a5e30897-logs\") pod \"glance-default-internal-api-0\" (UID: \"cbb9e604-033b-4586-b0fd-7300a5e30897\") " pod="openstack/glance-default-internal-api-0" Dec 03 14:32:17 crc kubenswrapper[4805]: I1203 14:32:17.981176 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"glance-default-internal-api-0\" (UID: \"cbb9e604-033b-4586-b0fd-7300a5e30897\") " pod="openstack/glance-default-internal-api-0" Dec 03 14:32:17 crc kubenswrapper[4805]: I1203 14:32:17.981227 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/cbb9e604-033b-4586-b0fd-7300a5e30897-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"cbb9e604-033b-4586-b0fd-7300a5e30897\") " pod="openstack/glance-default-internal-api-0" Dec 03 14:32:17 crc kubenswrapper[4805]: I1203 14:32:17.981275 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lkg52\" (UniqueName: \"kubernetes.io/projected/cbb9e604-033b-4586-b0fd-7300a5e30897-kube-api-access-lkg52\") pod \"glance-default-internal-api-0\" (UID: \"cbb9e604-033b-4586-b0fd-7300a5e30897\") " pod="openstack/glance-default-internal-api-0" Dec 03 14:32:17 crc kubenswrapper[4805]: I1203 14:32:17.981297 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cbb9e604-033b-4586-b0fd-7300a5e30897-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"cbb9e604-033b-4586-b0fd-7300a5e30897\") " pod="openstack/glance-default-internal-api-0" Dec 03 14:32:17 crc kubenswrapper[4805]: I1203 14:32:17.981320 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/cbb9e604-033b-4586-b0fd-7300a5e30897-scripts\") pod \"glance-default-internal-api-0\" (UID: \"cbb9e604-033b-4586-b0fd-7300a5e30897\") " pod="openstack/glance-default-internal-api-0" Dec 03 14:32:17 crc kubenswrapper[4805]: I1203 14:32:17.981373 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cbb9e604-033b-4586-b0fd-7300a5e30897-config-data\") pod \"glance-default-internal-api-0\" (UID: \"cbb9e604-033b-4586-b0fd-7300a5e30897\") " pod="openstack/glance-default-internal-api-0" Dec 03 14:32:17 crc kubenswrapper[4805]: I1203 14:32:17.982421 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/cbb9e604-033b-4586-b0fd-7300a5e30897-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"cbb9e604-033b-4586-b0fd-7300a5e30897\") " pod="openstack/glance-default-internal-api-0" Dec 03 14:32:17 crc kubenswrapper[4805]: I1203 14:32:17.982431 4805 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"glance-default-internal-api-0\" (UID: \"cbb9e604-033b-4586-b0fd-7300a5e30897\") device mount path \"/mnt/openstack/pv11\"" pod="openstack/glance-default-internal-api-0" Dec 03 14:32:17 crc kubenswrapper[4805]: I1203 14:32:17.983077 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/cbb9e604-033b-4586-b0fd-7300a5e30897-logs\") pod \"glance-default-internal-api-0\" (UID: \"cbb9e604-033b-4586-b0fd-7300a5e30897\") " pod="openstack/glance-default-internal-api-0" Dec 03 14:32:17 crc kubenswrapper[4805]: I1203 14:32:17.987159 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cbb9e604-033b-4586-b0fd-7300a5e30897-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"cbb9e604-033b-4586-b0fd-7300a5e30897\") " pod="openstack/glance-default-internal-api-0" Dec 03 14:32:17 crc kubenswrapper[4805]: I1203 14:32:17.992466 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/cbb9e604-033b-4586-b0fd-7300a5e30897-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"cbb9e604-033b-4586-b0fd-7300a5e30897\") " pod="openstack/glance-default-internal-api-0" Dec 03 14:32:17 crc kubenswrapper[4805]: I1203 14:32:17.992498 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/cbb9e604-033b-4586-b0fd-7300a5e30897-scripts\") pod \"glance-default-internal-api-0\" (UID: \"cbb9e604-033b-4586-b0fd-7300a5e30897\") " pod="openstack/glance-default-internal-api-0" Dec 03 14:32:17 crc kubenswrapper[4805]: I1203 14:32:17.998517 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cbb9e604-033b-4586-b0fd-7300a5e30897-config-data\") pod \"glance-default-internal-api-0\" (UID: \"cbb9e604-033b-4586-b0fd-7300a5e30897\") " pod="openstack/glance-default-internal-api-0" Dec 03 14:32:18 crc kubenswrapper[4805]: I1203 14:32:18.022539 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lkg52\" (UniqueName: \"kubernetes.io/projected/cbb9e604-033b-4586-b0fd-7300a5e30897-kube-api-access-lkg52\") pod \"glance-default-internal-api-0\" (UID: \"cbb9e604-033b-4586-b0fd-7300a5e30897\") " pod="openstack/glance-default-internal-api-0" Dec 03 14:32:18 crc kubenswrapper[4805]: I1203 14:32:18.052607 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"glance-default-internal-api-0\" (UID: \"cbb9e604-033b-4586-b0fd-7300a5e30897\") " pod="openstack/glance-default-internal-api-0" Dec 03 14:32:18 crc kubenswrapper[4805]: I1203 14:32:18.141287 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Dec 03 14:32:18 crc kubenswrapper[4805]: I1203 14:32:18.708448 4805 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="10877f19-dd2a-4eb2-99e0-2d0cf82d5ad5" path="/var/lib/kubelet/pods/10877f19-dd2a-4eb2-99e0-2d0cf82d5ad5/volumes" Dec 03 14:32:18 crc kubenswrapper[4805]: I1203 14:32:18.712122 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"6a4ca0f4-cff3-4c57-ab1b-3a01d7d6576c","Type":"ContainerStarted","Data":"c4900815b9c61f28a4bd39272d912bd983ced8d727c9e1e38544e9d2f43a8da8"} Dec 03 14:32:18 crc kubenswrapper[4805]: I1203 14:32:18.712172 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"6a4ca0f4-cff3-4c57-ab1b-3a01d7d6576c","Type":"ContainerStarted","Data":"2379fb0a74cd3824dd63264a55ceb267423b632a6159181813b3d3c02f9c853f"} Dec 03 14:32:18 crc kubenswrapper[4805]: I1203 14:32:18.742945 4805 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-external-api-0" podStartSLOduration=3.742924587 podStartE2EDuration="3.742924587s" podCreationTimestamp="2025-12-03 14:32:15 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 14:32:18.733531539 +0000 UTC m=+1368.396448462" watchObservedRunningTime="2025-12-03 14:32:18.742924587 +0000 UTC m=+1368.405841520" Dec 03 14:32:18 crc kubenswrapper[4805]: W1203 14:32:18.755052 4805 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podcbb9e604_033b_4586_b0fd_7300a5e30897.slice/crio-1cdb6fa1d5e606b9ce0ea3de26b761dd8bd7f7a6dd17da8b5595e9b6d062e428 WatchSource:0}: Error finding container 1cdb6fa1d5e606b9ce0ea3de26b761dd8bd7f7a6dd17da8b5595e9b6d062e428: Status 404 returned error can't find the container with id 1cdb6fa1d5e606b9ce0ea3de26b761dd8bd7f7a6dd17da8b5595e9b6d062e428 Dec 03 14:32:18 crc kubenswrapper[4805]: I1203 14:32:18.759343 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 03 14:32:18 crc kubenswrapper[4805]: I1203 14:32:18.994134 4805 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/cinder-scheduler-0" Dec 03 14:32:19 crc kubenswrapper[4805]: I1203 14:32:19.734571 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"cbb9e604-033b-4586-b0fd-7300a5e30897","Type":"ContainerStarted","Data":"b0159fa3b276577a076d6e3b47a7fa01e6cc47e4f8325ff10be885215a28902b"} Dec 03 14:32:19 crc kubenswrapper[4805]: I1203 14:32:19.734932 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"cbb9e604-033b-4586-b0fd-7300a5e30897","Type":"ContainerStarted","Data":"1cdb6fa1d5e606b9ce0ea3de26b761dd8bd7f7a6dd17da8b5595e9b6d062e428"} Dec 03 14:32:20 crc kubenswrapper[4805]: I1203 14:32:20.743329 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"cbb9e604-033b-4586-b0fd-7300a5e30897","Type":"ContainerStarted","Data":"9850dfaf084f4a9526a85340ab3c4d56c1bd8cc9dfd908de767c8b20c0e06a5f"} Dec 03 14:32:20 crc kubenswrapper[4805]: I1203 14:32:20.778445 4805 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-internal-api-0" podStartSLOduration=3.778416849 podStartE2EDuration="3.778416849s" podCreationTimestamp="2025-12-03 14:32:17 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 14:32:20.766159484 +0000 UTC m=+1370.429076417" watchObservedRunningTime="2025-12-03 14:32:20.778416849 +0000 UTC m=+1370.441333782" Dec 03 14:32:24 crc kubenswrapper[4805]: I1203 14:32:24.209297 4805 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/cinder-scheduler-0" Dec 03 14:32:24 crc kubenswrapper[4805]: E1203 14:32:24.406593 4805 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod5979c218_e720_4ce6_aebc_993a3676abb5.slice/crio-c5b0c5fc24db36ce396804d2d37d6255b90861754bbc2f992c426edf5f800276.scope\": RecentStats: unable to find data in memory cache]" Dec 03 14:32:24 crc kubenswrapper[4805]: I1203 14:32:24.696246 4805 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 03 14:32:24 crc kubenswrapper[4805]: I1203 14:32:24.783755 4805 generic.go:334] "Generic (PLEG): container finished" podID="5979c218-e720-4ce6-aebc-993a3676abb5" containerID="c5b0c5fc24db36ce396804d2d37d6255b90861754bbc2f992c426edf5f800276" exitCode=137 Dec 03 14:32:24 crc kubenswrapper[4805]: I1203 14:32:24.783872 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"5979c218-e720-4ce6-aebc-993a3676abb5","Type":"ContainerDied","Data":"c5b0c5fc24db36ce396804d2d37d6255b90861754bbc2f992c426edf5f800276"} Dec 03 14:32:24 crc kubenswrapper[4805]: I1203 14:32:24.783950 4805 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 03 14:32:24 crc kubenswrapper[4805]: I1203 14:32:24.784093 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"5979c218-e720-4ce6-aebc-993a3676abb5","Type":"ContainerDied","Data":"6f9ec724498267419adc91978a0bbeb50431ae3447528e3c8b439db73170246c"} Dec 03 14:32:24 crc kubenswrapper[4805]: I1203 14:32:24.784135 4805 scope.go:117] "RemoveContainer" containerID="c5b0c5fc24db36ce396804d2d37d6255b90861754bbc2f992c426edf5f800276" Dec 03 14:32:24 crc kubenswrapper[4805]: I1203 14:32:24.809890 4805 scope.go:117] "RemoveContainer" containerID="61ecf9c7ac9b0bc53e3380d461ce4698dd8336362b53387b5832742de7bdc4f8" Dec 03 14:32:24 crc kubenswrapper[4805]: I1203 14:32:24.819262 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5979c218-e720-4ce6-aebc-993a3676abb5-scripts\") pod \"5979c218-e720-4ce6-aebc-993a3676abb5\" (UID: \"5979c218-e720-4ce6-aebc-993a3676abb5\") " Dec 03 14:32:24 crc kubenswrapper[4805]: I1203 14:32:24.819407 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/5979c218-e720-4ce6-aebc-993a3676abb5-run-httpd\") pod \"5979c218-e720-4ce6-aebc-993a3676abb5\" (UID: \"5979c218-e720-4ce6-aebc-993a3676abb5\") " Dec 03 14:32:24 crc kubenswrapper[4805]: I1203 14:32:24.819449 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/5979c218-e720-4ce6-aebc-993a3676abb5-sg-core-conf-yaml\") pod \"5979c218-e720-4ce6-aebc-993a3676abb5\" (UID: \"5979c218-e720-4ce6-aebc-993a3676abb5\") " Dec 03 14:32:24 crc kubenswrapper[4805]: I1203 14:32:24.819471 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5979c218-e720-4ce6-aebc-993a3676abb5-combined-ca-bundle\") pod \"5979c218-e720-4ce6-aebc-993a3676abb5\" (UID: \"5979c218-e720-4ce6-aebc-993a3676abb5\") " Dec 03 14:32:24 crc kubenswrapper[4805]: I1203 14:32:24.819542 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5979c218-e720-4ce6-aebc-993a3676abb5-config-data\") pod \"5979c218-e720-4ce6-aebc-993a3676abb5\" (UID: \"5979c218-e720-4ce6-aebc-993a3676abb5\") " Dec 03 14:32:24 crc kubenswrapper[4805]: I1203 14:32:24.819631 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/5979c218-e720-4ce6-aebc-993a3676abb5-log-httpd\") pod \"5979c218-e720-4ce6-aebc-993a3676abb5\" (UID: \"5979c218-e720-4ce6-aebc-993a3676abb5\") " Dec 03 14:32:24 crc kubenswrapper[4805]: I1203 14:32:24.819667 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xcs9q\" (UniqueName: \"kubernetes.io/projected/5979c218-e720-4ce6-aebc-993a3676abb5-kube-api-access-xcs9q\") pod \"5979c218-e720-4ce6-aebc-993a3676abb5\" (UID: \"5979c218-e720-4ce6-aebc-993a3676abb5\") " Dec 03 14:32:24 crc kubenswrapper[4805]: I1203 14:32:24.820796 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5979c218-e720-4ce6-aebc-993a3676abb5-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "5979c218-e720-4ce6-aebc-993a3676abb5" (UID: "5979c218-e720-4ce6-aebc-993a3676abb5"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 14:32:24 crc kubenswrapper[4805]: I1203 14:32:24.821327 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5979c218-e720-4ce6-aebc-993a3676abb5-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "5979c218-e720-4ce6-aebc-993a3676abb5" (UID: "5979c218-e720-4ce6-aebc-993a3676abb5"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 14:32:24 crc kubenswrapper[4805]: I1203 14:32:24.826053 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5979c218-e720-4ce6-aebc-993a3676abb5-scripts" (OuterVolumeSpecName: "scripts") pod "5979c218-e720-4ce6-aebc-993a3676abb5" (UID: "5979c218-e720-4ce6-aebc-993a3676abb5"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 14:32:24 crc kubenswrapper[4805]: I1203 14:32:24.833540 4805 scope.go:117] "RemoveContainer" containerID="0766a2a356b4c585109ee5f02dd5def6ec3f30d6e59c15941ad33bbd31cf88ba" Dec 03 14:32:24 crc kubenswrapper[4805]: I1203 14:32:24.845025 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5979c218-e720-4ce6-aebc-993a3676abb5-kube-api-access-xcs9q" (OuterVolumeSpecName: "kube-api-access-xcs9q") pod "5979c218-e720-4ce6-aebc-993a3676abb5" (UID: "5979c218-e720-4ce6-aebc-993a3676abb5"). InnerVolumeSpecName "kube-api-access-xcs9q". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 14:32:24 crc kubenswrapper[4805]: I1203 14:32:24.848985 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5979c218-e720-4ce6-aebc-993a3676abb5-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "5979c218-e720-4ce6-aebc-993a3676abb5" (UID: "5979c218-e720-4ce6-aebc-993a3676abb5"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 14:32:24 crc kubenswrapper[4805]: I1203 14:32:24.913421 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5979c218-e720-4ce6-aebc-993a3676abb5-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "5979c218-e720-4ce6-aebc-993a3676abb5" (UID: "5979c218-e720-4ce6-aebc-993a3676abb5"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 14:32:24 crc kubenswrapper[4805]: I1203 14:32:24.925059 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5979c218-e720-4ce6-aebc-993a3676abb5-config-data" (OuterVolumeSpecName: "config-data") pod "5979c218-e720-4ce6-aebc-993a3676abb5" (UID: "5979c218-e720-4ce6-aebc-993a3676abb5"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 14:32:24 crc kubenswrapper[4805]: I1203 14:32:24.926258 4805 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5979c218-e720-4ce6-aebc-993a3676abb5-config-data\") on node \"crc\" DevicePath \"\"" Dec 03 14:32:24 crc kubenswrapper[4805]: I1203 14:32:24.926300 4805 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/5979c218-e720-4ce6-aebc-993a3676abb5-log-httpd\") on node \"crc\" DevicePath \"\"" Dec 03 14:32:24 crc kubenswrapper[4805]: I1203 14:32:24.926313 4805 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xcs9q\" (UniqueName: \"kubernetes.io/projected/5979c218-e720-4ce6-aebc-993a3676abb5-kube-api-access-xcs9q\") on node \"crc\" DevicePath \"\"" Dec 03 14:32:24 crc kubenswrapper[4805]: I1203 14:32:24.926332 4805 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5979c218-e720-4ce6-aebc-993a3676abb5-scripts\") on node \"crc\" DevicePath \"\"" Dec 03 14:32:24 crc kubenswrapper[4805]: I1203 14:32:24.926344 4805 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/5979c218-e720-4ce6-aebc-993a3676abb5-run-httpd\") on node \"crc\" DevicePath \"\"" Dec 03 14:32:24 crc kubenswrapper[4805]: I1203 14:32:24.926355 4805 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/5979c218-e720-4ce6-aebc-993a3676abb5-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Dec 03 14:32:24 crc kubenswrapper[4805]: I1203 14:32:24.926367 4805 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5979c218-e720-4ce6-aebc-993a3676abb5-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 14:32:24 crc kubenswrapper[4805]: I1203 14:32:24.988731 4805 scope.go:117] "RemoveContainer" containerID="2bb62034f23ff306eba03cc77fbd7f0b60e195607023bac74cfc090ef18a15c9" Dec 03 14:32:25 crc kubenswrapper[4805]: I1203 14:32:25.014436 4805 scope.go:117] "RemoveContainer" containerID="c5b0c5fc24db36ce396804d2d37d6255b90861754bbc2f992c426edf5f800276" Dec 03 14:32:25 crc kubenswrapper[4805]: E1203 14:32:25.017367 4805 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c5b0c5fc24db36ce396804d2d37d6255b90861754bbc2f992c426edf5f800276\": container with ID starting with c5b0c5fc24db36ce396804d2d37d6255b90861754bbc2f992c426edf5f800276 not found: ID does not exist" containerID="c5b0c5fc24db36ce396804d2d37d6255b90861754bbc2f992c426edf5f800276" Dec 03 14:32:25 crc kubenswrapper[4805]: I1203 14:32:25.017413 4805 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c5b0c5fc24db36ce396804d2d37d6255b90861754bbc2f992c426edf5f800276"} err="failed to get container status \"c5b0c5fc24db36ce396804d2d37d6255b90861754bbc2f992c426edf5f800276\": rpc error: code = NotFound desc = could not find container \"c5b0c5fc24db36ce396804d2d37d6255b90861754bbc2f992c426edf5f800276\": container with ID starting with c5b0c5fc24db36ce396804d2d37d6255b90861754bbc2f992c426edf5f800276 not found: ID does not exist" Dec 03 14:32:25 crc kubenswrapper[4805]: I1203 14:32:25.017443 4805 scope.go:117] "RemoveContainer" containerID="61ecf9c7ac9b0bc53e3380d461ce4698dd8336362b53387b5832742de7bdc4f8" Dec 03 14:32:25 crc kubenswrapper[4805]: E1203 14:32:25.018511 4805 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"61ecf9c7ac9b0bc53e3380d461ce4698dd8336362b53387b5832742de7bdc4f8\": container with ID starting with 61ecf9c7ac9b0bc53e3380d461ce4698dd8336362b53387b5832742de7bdc4f8 not found: ID does not exist" containerID="61ecf9c7ac9b0bc53e3380d461ce4698dd8336362b53387b5832742de7bdc4f8" Dec 03 14:32:25 crc kubenswrapper[4805]: I1203 14:32:25.018570 4805 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"61ecf9c7ac9b0bc53e3380d461ce4698dd8336362b53387b5832742de7bdc4f8"} err="failed to get container status \"61ecf9c7ac9b0bc53e3380d461ce4698dd8336362b53387b5832742de7bdc4f8\": rpc error: code = NotFound desc = could not find container \"61ecf9c7ac9b0bc53e3380d461ce4698dd8336362b53387b5832742de7bdc4f8\": container with ID starting with 61ecf9c7ac9b0bc53e3380d461ce4698dd8336362b53387b5832742de7bdc4f8 not found: ID does not exist" Dec 03 14:32:25 crc kubenswrapper[4805]: I1203 14:32:25.018601 4805 scope.go:117] "RemoveContainer" containerID="0766a2a356b4c585109ee5f02dd5def6ec3f30d6e59c15941ad33bbd31cf88ba" Dec 03 14:32:25 crc kubenswrapper[4805]: E1203 14:32:25.019676 4805 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0766a2a356b4c585109ee5f02dd5def6ec3f30d6e59c15941ad33bbd31cf88ba\": container with ID starting with 0766a2a356b4c585109ee5f02dd5def6ec3f30d6e59c15941ad33bbd31cf88ba not found: ID does not exist" containerID="0766a2a356b4c585109ee5f02dd5def6ec3f30d6e59c15941ad33bbd31cf88ba" Dec 03 14:32:25 crc kubenswrapper[4805]: I1203 14:32:25.019727 4805 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0766a2a356b4c585109ee5f02dd5def6ec3f30d6e59c15941ad33bbd31cf88ba"} err="failed to get container status \"0766a2a356b4c585109ee5f02dd5def6ec3f30d6e59c15941ad33bbd31cf88ba\": rpc error: code = NotFound desc = could not find container \"0766a2a356b4c585109ee5f02dd5def6ec3f30d6e59c15941ad33bbd31cf88ba\": container with ID starting with 0766a2a356b4c585109ee5f02dd5def6ec3f30d6e59c15941ad33bbd31cf88ba not found: ID does not exist" Dec 03 14:32:25 crc kubenswrapper[4805]: I1203 14:32:25.019758 4805 scope.go:117] "RemoveContainer" containerID="2bb62034f23ff306eba03cc77fbd7f0b60e195607023bac74cfc090ef18a15c9" Dec 03 14:32:25 crc kubenswrapper[4805]: E1203 14:32:25.020141 4805 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2bb62034f23ff306eba03cc77fbd7f0b60e195607023bac74cfc090ef18a15c9\": container with ID starting with 2bb62034f23ff306eba03cc77fbd7f0b60e195607023bac74cfc090ef18a15c9 not found: ID does not exist" containerID="2bb62034f23ff306eba03cc77fbd7f0b60e195607023bac74cfc090ef18a15c9" Dec 03 14:32:25 crc kubenswrapper[4805]: I1203 14:32:25.020173 4805 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2bb62034f23ff306eba03cc77fbd7f0b60e195607023bac74cfc090ef18a15c9"} err="failed to get container status \"2bb62034f23ff306eba03cc77fbd7f0b60e195607023bac74cfc090ef18a15c9\": rpc error: code = NotFound desc = could not find container \"2bb62034f23ff306eba03cc77fbd7f0b60e195607023bac74cfc090ef18a15c9\": container with ID starting with 2bb62034f23ff306eba03cc77fbd7f0b60e195607023bac74cfc090ef18a15c9 not found: ID does not exist" Dec 03 14:32:25 crc kubenswrapper[4805]: I1203 14:32:25.116824 4805 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 03 14:32:25 crc kubenswrapper[4805]: I1203 14:32:25.134959 4805 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Dec 03 14:32:25 crc kubenswrapper[4805]: I1203 14:32:25.147426 4805 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Dec 03 14:32:25 crc kubenswrapper[4805]: E1203 14:32:25.148028 4805 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5979c218-e720-4ce6-aebc-993a3676abb5" containerName="proxy-httpd" Dec 03 14:32:25 crc kubenswrapper[4805]: I1203 14:32:25.148055 4805 state_mem.go:107] "Deleted CPUSet assignment" podUID="5979c218-e720-4ce6-aebc-993a3676abb5" containerName="proxy-httpd" Dec 03 14:32:25 crc kubenswrapper[4805]: E1203 14:32:25.148073 4805 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5979c218-e720-4ce6-aebc-993a3676abb5" containerName="ceilometer-notification-agent" Dec 03 14:32:25 crc kubenswrapper[4805]: I1203 14:32:25.148081 4805 state_mem.go:107] "Deleted CPUSet assignment" podUID="5979c218-e720-4ce6-aebc-993a3676abb5" containerName="ceilometer-notification-agent" Dec 03 14:32:25 crc kubenswrapper[4805]: E1203 14:32:25.148110 4805 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5979c218-e720-4ce6-aebc-993a3676abb5" containerName="ceilometer-central-agent" Dec 03 14:32:25 crc kubenswrapper[4805]: I1203 14:32:25.148119 4805 state_mem.go:107] "Deleted CPUSet assignment" podUID="5979c218-e720-4ce6-aebc-993a3676abb5" containerName="ceilometer-central-agent" Dec 03 14:32:25 crc kubenswrapper[4805]: E1203 14:32:25.148131 4805 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5979c218-e720-4ce6-aebc-993a3676abb5" containerName="sg-core" Dec 03 14:32:25 crc kubenswrapper[4805]: I1203 14:32:25.148138 4805 state_mem.go:107] "Deleted CPUSet assignment" podUID="5979c218-e720-4ce6-aebc-993a3676abb5" containerName="sg-core" Dec 03 14:32:25 crc kubenswrapper[4805]: I1203 14:32:25.148537 4805 memory_manager.go:354] "RemoveStaleState removing state" podUID="5979c218-e720-4ce6-aebc-993a3676abb5" containerName="sg-core" Dec 03 14:32:25 crc kubenswrapper[4805]: I1203 14:32:25.148561 4805 memory_manager.go:354] "RemoveStaleState removing state" podUID="5979c218-e720-4ce6-aebc-993a3676abb5" containerName="proxy-httpd" Dec 03 14:32:25 crc kubenswrapper[4805]: I1203 14:32:25.148570 4805 memory_manager.go:354] "RemoveStaleState removing state" podUID="5979c218-e720-4ce6-aebc-993a3676abb5" containerName="ceilometer-central-agent" Dec 03 14:32:25 crc kubenswrapper[4805]: I1203 14:32:25.148579 4805 memory_manager.go:354] "RemoveStaleState removing state" podUID="5979c218-e720-4ce6-aebc-993a3676abb5" containerName="ceilometer-notification-agent" Dec 03 14:32:25 crc kubenswrapper[4805]: I1203 14:32:25.150696 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 03 14:32:25 crc kubenswrapper[4805]: I1203 14:32:25.153750 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Dec 03 14:32:25 crc kubenswrapper[4805]: I1203 14:32:25.153945 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Dec 03 14:32:25 crc kubenswrapper[4805]: I1203 14:32:25.169449 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 03 14:32:25 crc kubenswrapper[4805]: I1203 14:32:25.229824 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1ab37f5e-00c9-48b9-bd52-5f28e61ebe4b-config-data\") pod \"ceilometer-0\" (UID: \"1ab37f5e-00c9-48b9-bd52-5f28e61ebe4b\") " pod="openstack/ceilometer-0" Dec 03 14:32:25 crc kubenswrapper[4805]: I1203 14:32:25.229967 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/1ab37f5e-00c9-48b9-bd52-5f28e61ebe4b-log-httpd\") pod \"ceilometer-0\" (UID: \"1ab37f5e-00c9-48b9-bd52-5f28e61ebe4b\") " pod="openstack/ceilometer-0" Dec 03 14:32:25 crc kubenswrapper[4805]: I1203 14:32:25.230009 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/1ab37f5e-00c9-48b9-bd52-5f28e61ebe4b-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"1ab37f5e-00c9-48b9-bd52-5f28e61ebe4b\") " pod="openstack/ceilometer-0" Dec 03 14:32:25 crc kubenswrapper[4805]: I1203 14:32:25.230068 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1ab37f5e-00c9-48b9-bd52-5f28e61ebe4b-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"1ab37f5e-00c9-48b9-bd52-5f28e61ebe4b\") " pod="openstack/ceilometer-0" Dec 03 14:32:25 crc kubenswrapper[4805]: I1203 14:32:25.230106 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1ab37f5e-00c9-48b9-bd52-5f28e61ebe4b-scripts\") pod \"ceilometer-0\" (UID: \"1ab37f5e-00c9-48b9-bd52-5f28e61ebe4b\") " pod="openstack/ceilometer-0" Dec 03 14:32:25 crc kubenswrapper[4805]: I1203 14:32:25.230121 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/1ab37f5e-00c9-48b9-bd52-5f28e61ebe4b-run-httpd\") pod \"ceilometer-0\" (UID: \"1ab37f5e-00c9-48b9-bd52-5f28e61ebe4b\") " pod="openstack/ceilometer-0" Dec 03 14:32:25 crc kubenswrapper[4805]: I1203 14:32:25.230136 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-89mqv\" (UniqueName: \"kubernetes.io/projected/1ab37f5e-00c9-48b9-bd52-5f28e61ebe4b-kube-api-access-89mqv\") pod \"ceilometer-0\" (UID: \"1ab37f5e-00c9-48b9-bd52-5f28e61ebe4b\") " pod="openstack/ceilometer-0" Dec 03 14:32:25 crc kubenswrapper[4805]: I1203 14:32:25.331962 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1ab37f5e-00c9-48b9-bd52-5f28e61ebe4b-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"1ab37f5e-00c9-48b9-bd52-5f28e61ebe4b\") " pod="openstack/ceilometer-0" Dec 03 14:32:25 crc kubenswrapper[4805]: I1203 14:32:25.332040 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1ab37f5e-00c9-48b9-bd52-5f28e61ebe4b-scripts\") pod \"ceilometer-0\" (UID: \"1ab37f5e-00c9-48b9-bd52-5f28e61ebe4b\") " pod="openstack/ceilometer-0" Dec 03 14:32:25 crc kubenswrapper[4805]: I1203 14:32:25.332065 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-89mqv\" (UniqueName: \"kubernetes.io/projected/1ab37f5e-00c9-48b9-bd52-5f28e61ebe4b-kube-api-access-89mqv\") pod \"ceilometer-0\" (UID: \"1ab37f5e-00c9-48b9-bd52-5f28e61ebe4b\") " pod="openstack/ceilometer-0" Dec 03 14:32:25 crc kubenswrapper[4805]: I1203 14:32:25.332088 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/1ab37f5e-00c9-48b9-bd52-5f28e61ebe4b-run-httpd\") pod \"ceilometer-0\" (UID: \"1ab37f5e-00c9-48b9-bd52-5f28e61ebe4b\") " pod="openstack/ceilometer-0" Dec 03 14:32:25 crc kubenswrapper[4805]: I1203 14:32:25.332158 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1ab37f5e-00c9-48b9-bd52-5f28e61ebe4b-config-data\") pod \"ceilometer-0\" (UID: \"1ab37f5e-00c9-48b9-bd52-5f28e61ebe4b\") " pod="openstack/ceilometer-0" Dec 03 14:32:25 crc kubenswrapper[4805]: I1203 14:32:25.332196 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/1ab37f5e-00c9-48b9-bd52-5f28e61ebe4b-log-httpd\") pod \"ceilometer-0\" (UID: \"1ab37f5e-00c9-48b9-bd52-5f28e61ebe4b\") " pod="openstack/ceilometer-0" Dec 03 14:32:25 crc kubenswrapper[4805]: I1203 14:32:25.332243 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/1ab37f5e-00c9-48b9-bd52-5f28e61ebe4b-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"1ab37f5e-00c9-48b9-bd52-5f28e61ebe4b\") " pod="openstack/ceilometer-0" Dec 03 14:32:25 crc kubenswrapper[4805]: I1203 14:32:25.332888 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/1ab37f5e-00c9-48b9-bd52-5f28e61ebe4b-log-httpd\") pod \"ceilometer-0\" (UID: \"1ab37f5e-00c9-48b9-bd52-5f28e61ebe4b\") " pod="openstack/ceilometer-0" Dec 03 14:32:25 crc kubenswrapper[4805]: I1203 14:32:25.332990 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/1ab37f5e-00c9-48b9-bd52-5f28e61ebe4b-run-httpd\") pod \"ceilometer-0\" (UID: \"1ab37f5e-00c9-48b9-bd52-5f28e61ebe4b\") " pod="openstack/ceilometer-0" Dec 03 14:32:25 crc kubenswrapper[4805]: I1203 14:32:25.336579 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1ab37f5e-00c9-48b9-bd52-5f28e61ebe4b-scripts\") pod \"ceilometer-0\" (UID: \"1ab37f5e-00c9-48b9-bd52-5f28e61ebe4b\") " pod="openstack/ceilometer-0" Dec 03 14:32:25 crc kubenswrapper[4805]: I1203 14:32:25.336895 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/1ab37f5e-00c9-48b9-bd52-5f28e61ebe4b-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"1ab37f5e-00c9-48b9-bd52-5f28e61ebe4b\") " pod="openstack/ceilometer-0" Dec 03 14:32:25 crc kubenswrapper[4805]: I1203 14:32:25.338572 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1ab37f5e-00c9-48b9-bd52-5f28e61ebe4b-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"1ab37f5e-00c9-48b9-bd52-5f28e61ebe4b\") " pod="openstack/ceilometer-0" Dec 03 14:32:25 crc kubenswrapper[4805]: I1203 14:32:25.339161 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1ab37f5e-00c9-48b9-bd52-5f28e61ebe4b-config-data\") pod \"ceilometer-0\" (UID: \"1ab37f5e-00c9-48b9-bd52-5f28e61ebe4b\") " pod="openstack/ceilometer-0" Dec 03 14:32:25 crc kubenswrapper[4805]: I1203 14:32:25.352030 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-89mqv\" (UniqueName: \"kubernetes.io/projected/1ab37f5e-00c9-48b9-bd52-5f28e61ebe4b-kube-api-access-89mqv\") pod \"ceilometer-0\" (UID: \"1ab37f5e-00c9-48b9-bd52-5f28e61ebe4b\") " pod="openstack/ceilometer-0" Dec 03 14:32:25 crc kubenswrapper[4805]: I1203 14:32:25.475259 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 03 14:32:25 crc kubenswrapper[4805]: I1203 14:32:25.924671 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 03 14:32:25 crc kubenswrapper[4805]: I1203 14:32:25.930926 4805 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 03 14:32:26 crc kubenswrapper[4805]: I1203 14:32:26.341863 4805 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-external-api-0" Dec 03 14:32:26 crc kubenswrapper[4805]: I1203 14:32:26.342190 4805 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-external-api-0" Dec 03 14:32:26 crc kubenswrapper[4805]: I1203 14:32:26.378133 4805 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-external-api-0" Dec 03 14:32:26 crc kubenswrapper[4805]: I1203 14:32:26.393045 4805 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-external-api-0" Dec 03 14:32:26 crc kubenswrapper[4805]: I1203 14:32:26.707682 4805 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5979c218-e720-4ce6-aebc-993a3676abb5" path="/var/lib/kubelet/pods/5979c218-e720-4ce6-aebc-993a3676abb5/volumes" Dec 03 14:32:26 crc kubenswrapper[4805]: I1203 14:32:26.805026 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"1ab37f5e-00c9-48b9-bd52-5f28e61ebe4b","Type":"ContainerStarted","Data":"8e647e4df9935ecc5cbdc26d98af67739a98402985b6ad653f7551e5b444357e"} Dec 03 14:32:26 crc kubenswrapper[4805]: I1203 14:32:26.805064 4805 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-external-api-0" Dec 03 14:32:26 crc kubenswrapper[4805]: I1203 14:32:26.805077 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"1ab37f5e-00c9-48b9-bd52-5f28e61ebe4b","Type":"ContainerStarted","Data":"a3fadcacc82ebd928d9319d14e87d5f8ac6c152ece5994d53b52f36e62a876c9"} Dec 03 14:32:26 crc kubenswrapper[4805]: I1203 14:32:26.805172 4805 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-external-api-0" Dec 03 14:32:27 crc kubenswrapper[4805]: I1203 14:32:27.055700 4805 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 03 14:32:27 crc kubenswrapper[4805]: I1203 14:32:27.817401 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"1ab37f5e-00c9-48b9-bd52-5f28e61ebe4b","Type":"ContainerStarted","Data":"f7634309afe7db68d9795f782abe922e2f805a99a8bbfbf6d4adac4ff7008304"} Dec 03 14:32:28 crc kubenswrapper[4805]: I1203 14:32:28.143030 4805 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-internal-api-0" Dec 03 14:32:28 crc kubenswrapper[4805]: I1203 14:32:28.143088 4805 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-internal-api-0" Dec 03 14:32:28 crc kubenswrapper[4805]: I1203 14:32:28.187167 4805 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-internal-api-0" Dec 03 14:32:28 crc kubenswrapper[4805]: I1203 14:32:28.234048 4805 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-internal-api-0" Dec 03 14:32:28 crc kubenswrapper[4805]: I1203 14:32:28.827651 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"1ab37f5e-00c9-48b9-bd52-5f28e61ebe4b","Type":"ContainerStarted","Data":"d1ab27011808ebebb712190ef435cc83c4aff57f48b5f18c79d323ed744c8acc"} Dec 03 14:32:28 crc kubenswrapper[4805]: I1203 14:32:28.829209 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-db-sync-dxrx8" event={"ID":"e56c5750-e0c6-4427-ae8c-ab4af0c9b7d0","Type":"ContainerStarted","Data":"8adea027cee74f522316a8abd4ae6407b8f8c77128bc6cb459c352fdf09df968"} Dec 03 14:32:28 crc kubenswrapper[4805]: I1203 14:32:28.829277 4805 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Dec 03 14:32:28 crc kubenswrapper[4805]: I1203 14:32:28.829295 4805 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Dec 03 14:32:28 crc kubenswrapper[4805]: I1203 14:32:28.830275 4805 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-internal-api-0" Dec 03 14:32:28 crc kubenswrapper[4805]: I1203 14:32:28.830310 4805 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-internal-api-0" Dec 03 14:32:28 crc kubenswrapper[4805]: I1203 14:32:28.845519 4805 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell0-conductor-db-sync-dxrx8" podStartSLOduration=2.603628532 podStartE2EDuration="37.845498781s" podCreationTimestamp="2025-12-03 14:31:51 +0000 UTC" firstStartedPulling="2025-12-03 14:31:52.894386436 +0000 UTC m=+1342.557303369" lastFinishedPulling="2025-12-03 14:32:28.136256695 +0000 UTC m=+1377.799173618" observedRunningTime="2025-12-03 14:32:28.844411681 +0000 UTC m=+1378.507328604" watchObservedRunningTime="2025-12-03 14:32:28.845498781 +0000 UTC m=+1378.508415714" Dec 03 14:32:29 crc kubenswrapper[4805]: I1203 14:32:29.189975 4805 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-external-api-0" Dec 03 14:32:29 crc kubenswrapper[4805]: I1203 14:32:29.406870 4805 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-external-api-0" Dec 03 14:32:29 crc kubenswrapper[4805]: I1203 14:32:29.842043 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"1ab37f5e-00c9-48b9-bd52-5f28e61ebe4b","Type":"ContainerStarted","Data":"0c8a405ce922974e8020b42de906448d2a86d9092fd0c2f924f296332c72a3bb"} Dec 03 14:32:29 crc kubenswrapper[4805]: I1203 14:32:29.842585 4805 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="1ab37f5e-00c9-48b9-bd52-5f28e61ebe4b" containerName="ceilometer-central-agent" containerID="cri-o://8e647e4df9935ecc5cbdc26d98af67739a98402985b6ad653f7551e5b444357e" gracePeriod=30 Dec 03 14:32:29 crc kubenswrapper[4805]: I1203 14:32:29.842748 4805 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Dec 03 14:32:29 crc kubenswrapper[4805]: I1203 14:32:29.842998 4805 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="1ab37f5e-00c9-48b9-bd52-5f28e61ebe4b" containerName="proxy-httpd" containerID="cri-o://0c8a405ce922974e8020b42de906448d2a86d9092fd0c2f924f296332c72a3bb" gracePeriod=30 Dec 03 14:32:29 crc kubenswrapper[4805]: I1203 14:32:29.843148 4805 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="1ab37f5e-00c9-48b9-bd52-5f28e61ebe4b" containerName="ceilometer-notification-agent" containerID="cri-o://f7634309afe7db68d9795f782abe922e2f805a99a8bbfbf6d4adac4ff7008304" gracePeriod=30 Dec 03 14:32:29 crc kubenswrapper[4805]: I1203 14:32:29.843215 4805 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="1ab37f5e-00c9-48b9-bd52-5f28e61ebe4b" containerName="sg-core" containerID="cri-o://d1ab27011808ebebb712190ef435cc83c4aff57f48b5f18c79d323ed744c8acc" gracePeriod=30 Dec 03 14:32:29 crc kubenswrapper[4805]: I1203 14:32:29.871852 4805 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=1.513915521 podStartE2EDuration="4.871813706s" podCreationTimestamp="2025-12-03 14:32:25 +0000 UTC" firstStartedPulling="2025-12-03 14:32:25.930674876 +0000 UTC m=+1375.593591799" lastFinishedPulling="2025-12-03 14:32:29.288573071 +0000 UTC m=+1378.951489984" observedRunningTime="2025-12-03 14:32:29.864111486 +0000 UTC m=+1379.527028409" watchObservedRunningTime="2025-12-03 14:32:29.871813706 +0000 UTC m=+1379.534730629" Dec 03 14:32:30 crc kubenswrapper[4805]: I1203 14:32:30.858818 4805 generic.go:334] "Generic (PLEG): container finished" podID="1ab37f5e-00c9-48b9-bd52-5f28e61ebe4b" containerID="0c8a405ce922974e8020b42de906448d2a86d9092fd0c2f924f296332c72a3bb" exitCode=0 Dec 03 14:32:30 crc kubenswrapper[4805]: I1203 14:32:30.859869 4805 generic.go:334] "Generic (PLEG): container finished" podID="1ab37f5e-00c9-48b9-bd52-5f28e61ebe4b" containerID="d1ab27011808ebebb712190ef435cc83c4aff57f48b5f18c79d323ed744c8acc" exitCode=2 Dec 03 14:32:30 crc kubenswrapper[4805]: I1203 14:32:30.859979 4805 generic.go:334] "Generic (PLEG): container finished" podID="1ab37f5e-00c9-48b9-bd52-5f28e61ebe4b" containerID="f7634309afe7db68d9795f782abe922e2f805a99a8bbfbf6d4adac4ff7008304" exitCode=0 Dec 03 14:32:30 crc kubenswrapper[4805]: I1203 14:32:30.859438 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"1ab37f5e-00c9-48b9-bd52-5f28e61ebe4b","Type":"ContainerDied","Data":"0c8a405ce922974e8020b42de906448d2a86d9092fd0c2f924f296332c72a3bb"} Dec 03 14:32:30 crc kubenswrapper[4805]: I1203 14:32:30.860113 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"1ab37f5e-00c9-48b9-bd52-5f28e61ebe4b","Type":"ContainerDied","Data":"d1ab27011808ebebb712190ef435cc83c4aff57f48b5f18c79d323ed744c8acc"} Dec 03 14:32:30 crc kubenswrapper[4805]: I1203 14:32:30.860144 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"1ab37f5e-00c9-48b9-bd52-5f28e61ebe4b","Type":"ContainerDied","Data":"f7634309afe7db68d9795f782abe922e2f805a99a8bbfbf6d4adac4ff7008304"} Dec 03 14:32:31 crc kubenswrapper[4805]: I1203 14:32:31.515613 4805 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-internal-api-0" Dec 03 14:32:31 crc kubenswrapper[4805]: I1203 14:32:31.516317 4805 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Dec 03 14:32:31 crc kubenswrapper[4805]: I1203 14:32:31.583655 4805 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-internal-api-0" Dec 03 14:32:39 crc kubenswrapper[4805]: I1203 14:32:39.957144 4805 generic.go:334] "Generic (PLEG): container finished" podID="1ab37f5e-00c9-48b9-bd52-5f28e61ebe4b" containerID="8e647e4df9935ecc5cbdc26d98af67739a98402985b6ad653f7551e5b444357e" exitCode=0 Dec 03 14:32:39 crc kubenswrapper[4805]: I1203 14:32:39.957268 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"1ab37f5e-00c9-48b9-bd52-5f28e61ebe4b","Type":"ContainerDied","Data":"8e647e4df9935ecc5cbdc26d98af67739a98402985b6ad653f7551e5b444357e"} Dec 03 14:32:40 crc kubenswrapper[4805]: I1203 14:32:40.081449 4805 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 03 14:32:40 crc kubenswrapper[4805]: I1203 14:32:40.261534 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1ab37f5e-00c9-48b9-bd52-5f28e61ebe4b-scripts\") pod \"1ab37f5e-00c9-48b9-bd52-5f28e61ebe4b\" (UID: \"1ab37f5e-00c9-48b9-bd52-5f28e61ebe4b\") " Dec 03 14:32:40 crc kubenswrapper[4805]: I1203 14:32:40.261739 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1ab37f5e-00c9-48b9-bd52-5f28e61ebe4b-config-data\") pod \"1ab37f5e-00c9-48b9-bd52-5f28e61ebe4b\" (UID: \"1ab37f5e-00c9-48b9-bd52-5f28e61ebe4b\") " Dec 03 14:32:40 crc kubenswrapper[4805]: I1203 14:32:40.261777 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1ab37f5e-00c9-48b9-bd52-5f28e61ebe4b-combined-ca-bundle\") pod \"1ab37f5e-00c9-48b9-bd52-5f28e61ebe4b\" (UID: \"1ab37f5e-00c9-48b9-bd52-5f28e61ebe4b\") " Dec 03 14:32:40 crc kubenswrapper[4805]: I1203 14:32:40.261822 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-89mqv\" (UniqueName: \"kubernetes.io/projected/1ab37f5e-00c9-48b9-bd52-5f28e61ebe4b-kube-api-access-89mqv\") pod \"1ab37f5e-00c9-48b9-bd52-5f28e61ebe4b\" (UID: \"1ab37f5e-00c9-48b9-bd52-5f28e61ebe4b\") " Dec 03 14:32:40 crc kubenswrapper[4805]: I1203 14:32:40.261907 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/1ab37f5e-00c9-48b9-bd52-5f28e61ebe4b-log-httpd\") pod \"1ab37f5e-00c9-48b9-bd52-5f28e61ebe4b\" (UID: \"1ab37f5e-00c9-48b9-bd52-5f28e61ebe4b\") " Dec 03 14:32:40 crc kubenswrapper[4805]: I1203 14:32:40.261949 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/1ab37f5e-00c9-48b9-bd52-5f28e61ebe4b-sg-core-conf-yaml\") pod \"1ab37f5e-00c9-48b9-bd52-5f28e61ebe4b\" (UID: \"1ab37f5e-00c9-48b9-bd52-5f28e61ebe4b\") " Dec 03 14:32:40 crc kubenswrapper[4805]: I1203 14:32:40.261971 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/1ab37f5e-00c9-48b9-bd52-5f28e61ebe4b-run-httpd\") pod \"1ab37f5e-00c9-48b9-bd52-5f28e61ebe4b\" (UID: \"1ab37f5e-00c9-48b9-bd52-5f28e61ebe4b\") " Dec 03 14:32:40 crc kubenswrapper[4805]: I1203 14:32:40.262636 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1ab37f5e-00c9-48b9-bd52-5f28e61ebe4b-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "1ab37f5e-00c9-48b9-bd52-5f28e61ebe4b" (UID: "1ab37f5e-00c9-48b9-bd52-5f28e61ebe4b"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 14:32:40 crc kubenswrapper[4805]: I1203 14:32:40.262787 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1ab37f5e-00c9-48b9-bd52-5f28e61ebe4b-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "1ab37f5e-00c9-48b9-bd52-5f28e61ebe4b" (UID: "1ab37f5e-00c9-48b9-bd52-5f28e61ebe4b"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 14:32:40 crc kubenswrapper[4805]: I1203 14:32:40.267696 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1ab37f5e-00c9-48b9-bd52-5f28e61ebe4b-kube-api-access-89mqv" (OuterVolumeSpecName: "kube-api-access-89mqv") pod "1ab37f5e-00c9-48b9-bd52-5f28e61ebe4b" (UID: "1ab37f5e-00c9-48b9-bd52-5f28e61ebe4b"). InnerVolumeSpecName "kube-api-access-89mqv". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 14:32:40 crc kubenswrapper[4805]: I1203 14:32:40.268517 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1ab37f5e-00c9-48b9-bd52-5f28e61ebe4b-scripts" (OuterVolumeSpecName: "scripts") pod "1ab37f5e-00c9-48b9-bd52-5f28e61ebe4b" (UID: "1ab37f5e-00c9-48b9-bd52-5f28e61ebe4b"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 14:32:40 crc kubenswrapper[4805]: I1203 14:32:40.291560 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1ab37f5e-00c9-48b9-bd52-5f28e61ebe4b-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "1ab37f5e-00c9-48b9-bd52-5f28e61ebe4b" (UID: "1ab37f5e-00c9-48b9-bd52-5f28e61ebe4b"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 14:32:40 crc kubenswrapper[4805]: I1203 14:32:40.364570 4805 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-89mqv\" (UniqueName: \"kubernetes.io/projected/1ab37f5e-00c9-48b9-bd52-5f28e61ebe4b-kube-api-access-89mqv\") on node \"crc\" DevicePath \"\"" Dec 03 14:32:40 crc kubenswrapper[4805]: I1203 14:32:40.364615 4805 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/1ab37f5e-00c9-48b9-bd52-5f28e61ebe4b-log-httpd\") on node \"crc\" DevicePath \"\"" Dec 03 14:32:40 crc kubenswrapper[4805]: I1203 14:32:40.364628 4805 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/1ab37f5e-00c9-48b9-bd52-5f28e61ebe4b-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Dec 03 14:32:40 crc kubenswrapper[4805]: I1203 14:32:40.364638 4805 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/1ab37f5e-00c9-48b9-bd52-5f28e61ebe4b-run-httpd\") on node \"crc\" DevicePath \"\"" Dec 03 14:32:40 crc kubenswrapper[4805]: I1203 14:32:40.364650 4805 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1ab37f5e-00c9-48b9-bd52-5f28e61ebe4b-scripts\") on node \"crc\" DevicePath \"\"" Dec 03 14:32:40 crc kubenswrapper[4805]: I1203 14:32:40.377530 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1ab37f5e-00c9-48b9-bd52-5f28e61ebe4b-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "1ab37f5e-00c9-48b9-bd52-5f28e61ebe4b" (UID: "1ab37f5e-00c9-48b9-bd52-5f28e61ebe4b"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 14:32:40 crc kubenswrapper[4805]: I1203 14:32:40.385718 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1ab37f5e-00c9-48b9-bd52-5f28e61ebe4b-config-data" (OuterVolumeSpecName: "config-data") pod "1ab37f5e-00c9-48b9-bd52-5f28e61ebe4b" (UID: "1ab37f5e-00c9-48b9-bd52-5f28e61ebe4b"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 14:32:40 crc kubenswrapper[4805]: I1203 14:32:40.466025 4805 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1ab37f5e-00c9-48b9-bd52-5f28e61ebe4b-config-data\") on node \"crc\" DevicePath \"\"" Dec 03 14:32:40 crc kubenswrapper[4805]: I1203 14:32:40.466059 4805 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1ab37f5e-00c9-48b9-bd52-5f28e61ebe4b-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 14:32:40 crc kubenswrapper[4805]: I1203 14:32:40.970890 4805 generic.go:334] "Generic (PLEG): container finished" podID="e56c5750-e0c6-4427-ae8c-ab4af0c9b7d0" containerID="8adea027cee74f522316a8abd4ae6407b8f8c77128bc6cb459c352fdf09df968" exitCode=0 Dec 03 14:32:40 crc kubenswrapper[4805]: I1203 14:32:40.970952 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-db-sync-dxrx8" event={"ID":"e56c5750-e0c6-4427-ae8c-ab4af0c9b7d0","Type":"ContainerDied","Data":"8adea027cee74f522316a8abd4ae6407b8f8c77128bc6cb459c352fdf09df968"} Dec 03 14:32:40 crc kubenswrapper[4805]: I1203 14:32:40.977633 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"1ab37f5e-00c9-48b9-bd52-5f28e61ebe4b","Type":"ContainerDied","Data":"a3fadcacc82ebd928d9319d14e87d5f8ac6c152ece5994d53b52f36e62a876c9"} Dec 03 14:32:40 crc kubenswrapper[4805]: I1203 14:32:40.977668 4805 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 03 14:32:40 crc kubenswrapper[4805]: I1203 14:32:40.977720 4805 scope.go:117] "RemoveContainer" containerID="0c8a405ce922974e8020b42de906448d2a86d9092fd0c2f924f296332c72a3bb" Dec 03 14:32:41 crc kubenswrapper[4805]: I1203 14:32:41.004450 4805 scope.go:117] "RemoveContainer" containerID="d1ab27011808ebebb712190ef435cc83c4aff57f48b5f18c79d323ed744c8acc" Dec 03 14:32:41 crc kubenswrapper[4805]: I1203 14:32:41.012399 4805 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 03 14:32:41 crc kubenswrapper[4805]: I1203 14:32:41.020930 4805 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Dec 03 14:32:41 crc kubenswrapper[4805]: I1203 14:32:41.033604 4805 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Dec 03 14:32:41 crc kubenswrapper[4805]: E1203 14:32:41.034039 4805 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1ab37f5e-00c9-48b9-bd52-5f28e61ebe4b" containerName="proxy-httpd" Dec 03 14:32:41 crc kubenswrapper[4805]: I1203 14:32:41.034062 4805 state_mem.go:107] "Deleted CPUSet assignment" podUID="1ab37f5e-00c9-48b9-bd52-5f28e61ebe4b" containerName="proxy-httpd" Dec 03 14:32:41 crc kubenswrapper[4805]: E1203 14:32:41.034076 4805 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1ab37f5e-00c9-48b9-bd52-5f28e61ebe4b" containerName="ceilometer-central-agent" Dec 03 14:32:41 crc kubenswrapper[4805]: I1203 14:32:41.034082 4805 state_mem.go:107] "Deleted CPUSet assignment" podUID="1ab37f5e-00c9-48b9-bd52-5f28e61ebe4b" containerName="ceilometer-central-agent" Dec 03 14:32:41 crc kubenswrapper[4805]: E1203 14:32:41.034101 4805 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1ab37f5e-00c9-48b9-bd52-5f28e61ebe4b" containerName="sg-core" Dec 03 14:32:41 crc kubenswrapper[4805]: I1203 14:32:41.034107 4805 state_mem.go:107] "Deleted CPUSet assignment" podUID="1ab37f5e-00c9-48b9-bd52-5f28e61ebe4b" containerName="sg-core" Dec 03 14:32:41 crc kubenswrapper[4805]: E1203 14:32:41.034131 4805 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1ab37f5e-00c9-48b9-bd52-5f28e61ebe4b" containerName="ceilometer-notification-agent" Dec 03 14:32:41 crc kubenswrapper[4805]: I1203 14:32:41.034136 4805 state_mem.go:107] "Deleted CPUSet assignment" podUID="1ab37f5e-00c9-48b9-bd52-5f28e61ebe4b" containerName="ceilometer-notification-agent" Dec 03 14:32:41 crc kubenswrapper[4805]: I1203 14:32:41.034318 4805 memory_manager.go:354] "RemoveStaleState removing state" podUID="1ab37f5e-00c9-48b9-bd52-5f28e61ebe4b" containerName="proxy-httpd" Dec 03 14:32:41 crc kubenswrapper[4805]: I1203 14:32:41.034353 4805 memory_manager.go:354] "RemoveStaleState removing state" podUID="1ab37f5e-00c9-48b9-bd52-5f28e61ebe4b" containerName="ceilometer-notification-agent" Dec 03 14:32:41 crc kubenswrapper[4805]: I1203 14:32:41.034373 4805 memory_manager.go:354] "RemoveStaleState removing state" podUID="1ab37f5e-00c9-48b9-bd52-5f28e61ebe4b" containerName="ceilometer-central-agent" Dec 03 14:32:41 crc kubenswrapper[4805]: I1203 14:32:41.034386 4805 memory_manager.go:354] "RemoveStaleState removing state" podUID="1ab37f5e-00c9-48b9-bd52-5f28e61ebe4b" containerName="sg-core" Dec 03 14:32:41 crc kubenswrapper[4805]: I1203 14:32:41.036219 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 03 14:32:41 crc kubenswrapper[4805]: I1203 14:32:41.038751 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Dec 03 14:32:41 crc kubenswrapper[4805]: I1203 14:32:41.043286 4805 scope.go:117] "RemoveContainer" containerID="f7634309afe7db68d9795f782abe922e2f805a99a8bbfbf6d4adac4ff7008304" Dec 03 14:32:41 crc kubenswrapper[4805]: I1203 14:32:41.043454 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Dec 03 14:32:41 crc kubenswrapper[4805]: I1203 14:32:41.048658 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 03 14:32:41 crc kubenswrapper[4805]: I1203 14:32:41.082606 4805 scope.go:117] "RemoveContainer" containerID="8e647e4df9935ecc5cbdc26d98af67739a98402985b6ad653f7551e5b444357e" Dec 03 14:32:41 crc kubenswrapper[4805]: I1203 14:32:41.181700 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/35025929-9af2-4fd0-a702-76c349a6171e-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"35025929-9af2-4fd0-a702-76c349a6171e\") " pod="openstack/ceilometer-0" Dec 03 14:32:41 crc kubenswrapper[4805]: I1203 14:32:41.181757 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/35025929-9af2-4fd0-a702-76c349a6171e-config-data\") pod \"ceilometer-0\" (UID: \"35025929-9af2-4fd0-a702-76c349a6171e\") " pod="openstack/ceilometer-0" Dec 03 14:32:41 crc kubenswrapper[4805]: I1203 14:32:41.181786 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/35025929-9af2-4fd0-a702-76c349a6171e-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"35025929-9af2-4fd0-a702-76c349a6171e\") " pod="openstack/ceilometer-0" Dec 03 14:32:41 crc kubenswrapper[4805]: I1203 14:32:41.182198 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/35025929-9af2-4fd0-a702-76c349a6171e-run-httpd\") pod \"ceilometer-0\" (UID: \"35025929-9af2-4fd0-a702-76c349a6171e\") " pod="openstack/ceilometer-0" Dec 03 14:32:41 crc kubenswrapper[4805]: I1203 14:32:41.182253 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/35025929-9af2-4fd0-a702-76c349a6171e-log-httpd\") pod \"ceilometer-0\" (UID: \"35025929-9af2-4fd0-a702-76c349a6171e\") " pod="openstack/ceilometer-0" Dec 03 14:32:41 crc kubenswrapper[4805]: I1203 14:32:41.182354 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/35025929-9af2-4fd0-a702-76c349a6171e-scripts\") pod \"ceilometer-0\" (UID: \"35025929-9af2-4fd0-a702-76c349a6171e\") " pod="openstack/ceilometer-0" Dec 03 14:32:41 crc kubenswrapper[4805]: I1203 14:32:41.182423 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gsh7r\" (UniqueName: \"kubernetes.io/projected/35025929-9af2-4fd0-a702-76c349a6171e-kube-api-access-gsh7r\") pod \"ceilometer-0\" (UID: \"35025929-9af2-4fd0-a702-76c349a6171e\") " pod="openstack/ceilometer-0" Dec 03 14:32:41 crc kubenswrapper[4805]: I1203 14:32:41.283635 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/35025929-9af2-4fd0-a702-76c349a6171e-run-httpd\") pod \"ceilometer-0\" (UID: \"35025929-9af2-4fd0-a702-76c349a6171e\") " pod="openstack/ceilometer-0" Dec 03 14:32:41 crc kubenswrapper[4805]: I1203 14:32:41.283680 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/35025929-9af2-4fd0-a702-76c349a6171e-log-httpd\") pod \"ceilometer-0\" (UID: \"35025929-9af2-4fd0-a702-76c349a6171e\") " pod="openstack/ceilometer-0" Dec 03 14:32:41 crc kubenswrapper[4805]: I1203 14:32:41.283736 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/35025929-9af2-4fd0-a702-76c349a6171e-scripts\") pod \"ceilometer-0\" (UID: \"35025929-9af2-4fd0-a702-76c349a6171e\") " pod="openstack/ceilometer-0" Dec 03 14:32:41 crc kubenswrapper[4805]: I1203 14:32:41.283773 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gsh7r\" (UniqueName: \"kubernetes.io/projected/35025929-9af2-4fd0-a702-76c349a6171e-kube-api-access-gsh7r\") pod \"ceilometer-0\" (UID: \"35025929-9af2-4fd0-a702-76c349a6171e\") " pod="openstack/ceilometer-0" Dec 03 14:32:41 crc kubenswrapper[4805]: I1203 14:32:41.283850 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/35025929-9af2-4fd0-a702-76c349a6171e-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"35025929-9af2-4fd0-a702-76c349a6171e\") " pod="openstack/ceilometer-0" Dec 03 14:32:41 crc kubenswrapper[4805]: I1203 14:32:41.284305 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/35025929-9af2-4fd0-a702-76c349a6171e-log-httpd\") pod \"ceilometer-0\" (UID: \"35025929-9af2-4fd0-a702-76c349a6171e\") " pod="openstack/ceilometer-0" Dec 03 14:32:41 crc kubenswrapper[4805]: I1203 14:32:41.284305 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/35025929-9af2-4fd0-a702-76c349a6171e-run-httpd\") pod \"ceilometer-0\" (UID: \"35025929-9af2-4fd0-a702-76c349a6171e\") " pod="openstack/ceilometer-0" Dec 03 14:32:41 crc kubenswrapper[4805]: I1203 14:32:41.285101 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/35025929-9af2-4fd0-a702-76c349a6171e-config-data\") pod \"ceilometer-0\" (UID: \"35025929-9af2-4fd0-a702-76c349a6171e\") " pod="openstack/ceilometer-0" Dec 03 14:32:41 crc kubenswrapper[4805]: I1203 14:32:41.285149 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/35025929-9af2-4fd0-a702-76c349a6171e-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"35025929-9af2-4fd0-a702-76c349a6171e\") " pod="openstack/ceilometer-0" Dec 03 14:32:41 crc kubenswrapper[4805]: I1203 14:32:41.289425 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/35025929-9af2-4fd0-a702-76c349a6171e-scripts\") pod \"ceilometer-0\" (UID: \"35025929-9af2-4fd0-a702-76c349a6171e\") " pod="openstack/ceilometer-0" Dec 03 14:32:41 crc kubenswrapper[4805]: I1203 14:32:41.289654 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/35025929-9af2-4fd0-a702-76c349a6171e-config-data\") pod \"ceilometer-0\" (UID: \"35025929-9af2-4fd0-a702-76c349a6171e\") " pod="openstack/ceilometer-0" Dec 03 14:32:41 crc kubenswrapper[4805]: I1203 14:32:41.290703 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/35025929-9af2-4fd0-a702-76c349a6171e-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"35025929-9af2-4fd0-a702-76c349a6171e\") " pod="openstack/ceilometer-0" Dec 03 14:32:41 crc kubenswrapper[4805]: I1203 14:32:41.310028 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/35025929-9af2-4fd0-a702-76c349a6171e-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"35025929-9af2-4fd0-a702-76c349a6171e\") " pod="openstack/ceilometer-0" Dec 03 14:32:41 crc kubenswrapper[4805]: I1203 14:32:41.314073 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gsh7r\" (UniqueName: \"kubernetes.io/projected/35025929-9af2-4fd0-a702-76c349a6171e-kube-api-access-gsh7r\") pod \"ceilometer-0\" (UID: \"35025929-9af2-4fd0-a702-76c349a6171e\") " pod="openstack/ceilometer-0" Dec 03 14:32:41 crc kubenswrapper[4805]: I1203 14:32:41.372175 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 03 14:32:41 crc kubenswrapper[4805]: I1203 14:32:41.890291 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 03 14:32:41 crc kubenswrapper[4805]: W1203 14:32:41.891669 4805 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod35025929_9af2_4fd0_a702_76c349a6171e.slice/crio-54e46fc7bbb38965aec6ed5e313fc89c61030ebab24c17aa20f9f26ad6dd2028 WatchSource:0}: Error finding container 54e46fc7bbb38965aec6ed5e313fc89c61030ebab24c17aa20f9f26ad6dd2028: Status 404 returned error can't find the container with id 54e46fc7bbb38965aec6ed5e313fc89c61030ebab24c17aa20f9f26ad6dd2028 Dec 03 14:32:42 crc kubenswrapper[4805]: I1203 14:32:42.006191 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"35025929-9af2-4fd0-a702-76c349a6171e","Type":"ContainerStarted","Data":"54e46fc7bbb38965aec6ed5e313fc89c61030ebab24c17aa20f9f26ad6dd2028"} Dec 03 14:32:42 crc kubenswrapper[4805]: I1203 14:32:42.328972 4805 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-db-sync-dxrx8" Dec 03 14:32:42 crc kubenswrapper[4805]: I1203 14:32:42.506123 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e56c5750-e0c6-4427-ae8c-ab4af0c9b7d0-config-data\") pod \"e56c5750-e0c6-4427-ae8c-ab4af0c9b7d0\" (UID: \"e56c5750-e0c6-4427-ae8c-ab4af0c9b7d0\") " Dec 03 14:32:42 crc kubenswrapper[4805]: I1203 14:32:42.506336 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e56c5750-e0c6-4427-ae8c-ab4af0c9b7d0-combined-ca-bundle\") pod \"e56c5750-e0c6-4427-ae8c-ab4af0c9b7d0\" (UID: \"e56c5750-e0c6-4427-ae8c-ab4af0c9b7d0\") " Dec 03 14:32:42 crc kubenswrapper[4805]: I1203 14:32:42.506394 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9dr74\" (UniqueName: \"kubernetes.io/projected/e56c5750-e0c6-4427-ae8c-ab4af0c9b7d0-kube-api-access-9dr74\") pod \"e56c5750-e0c6-4427-ae8c-ab4af0c9b7d0\" (UID: \"e56c5750-e0c6-4427-ae8c-ab4af0c9b7d0\") " Dec 03 14:32:42 crc kubenswrapper[4805]: I1203 14:32:42.506506 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e56c5750-e0c6-4427-ae8c-ab4af0c9b7d0-scripts\") pod \"e56c5750-e0c6-4427-ae8c-ab4af0c9b7d0\" (UID: \"e56c5750-e0c6-4427-ae8c-ab4af0c9b7d0\") " Dec 03 14:32:42 crc kubenswrapper[4805]: I1203 14:32:42.513787 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e56c5750-e0c6-4427-ae8c-ab4af0c9b7d0-kube-api-access-9dr74" (OuterVolumeSpecName: "kube-api-access-9dr74") pod "e56c5750-e0c6-4427-ae8c-ab4af0c9b7d0" (UID: "e56c5750-e0c6-4427-ae8c-ab4af0c9b7d0"). InnerVolumeSpecName "kube-api-access-9dr74". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 14:32:42 crc kubenswrapper[4805]: I1203 14:32:42.514982 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e56c5750-e0c6-4427-ae8c-ab4af0c9b7d0-scripts" (OuterVolumeSpecName: "scripts") pod "e56c5750-e0c6-4427-ae8c-ab4af0c9b7d0" (UID: "e56c5750-e0c6-4427-ae8c-ab4af0c9b7d0"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 14:32:42 crc kubenswrapper[4805]: I1203 14:32:42.538199 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e56c5750-e0c6-4427-ae8c-ab4af0c9b7d0-config-data" (OuterVolumeSpecName: "config-data") pod "e56c5750-e0c6-4427-ae8c-ab4af0c9b7d0" (UID: "e56c5750-e0c6-4427-ae8c-ab4af0c9b7d0"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 14:32:42 crc kubenswrapper[4805]: I1203 14:32:42.539918 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e56c5750-e0c6-4427-ae8c-ab4af0c9b7d0-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "e56c5750-e0c6-4427-ae8c-ab4af0c9b7d0" (UID: "e56c5750-e0c6-4427-ae8c-ab4af0c9b7d0"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 14:32:42 crc kubenswrapper[4805]: I1203 14:32:42.608812 4805 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e56c5750-e0c6-4427-ae8c-ab4af0c9b7d0-config-data\") on node \"crc\" DevicePath \"\"" Dec 03 14:32:42 crc kubenswrapper[4805]: I1203 14:32:42.608874 4805 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e56c5750-e0c6-4427-ae8c-ab4af0c9b7d0-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 14:32:42 crc kubenswrapper[4805]: I1203 14:32:42.608889 4805 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9dr74\" (UniqueName: \"kubernetes.io/projected/e56c5750-e0c6-4427-ae8c-ab4af0c9b7d0-kube-api-access-9dr74\") on node \"crc\" DevicePath \"\"" Dec 03 14:32:42 crc kubenswrapper[4805]: I1203 14:32:42.608899 4805 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e56c5750-e0c6-4427-ae8c-ab4af0c9b7d0-scripts\") on node \"crc\" DevicePath \"\"" Dec 03 14:32:42 crc kubenswrapper[4805]: I1203 14:32:42.706717 4805 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1ab37f5e-00c9-48b9-bd52-5f28e61ebe4b" path="/var/lib/kubelet/pods/1ab37f5e-00c9-48b9-bd52-5f28e61ebe4b/volumes" Dec 03 14:32:43 crc kubenswrapper[4805]: I1203 14:32:43.016468 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-db-sync-dxrx8" event={"ID":"e56c5750-e0c6-4427-ae8c-ab4af0c9b7d0","Type":"ContainerDied","Data":"e7b3280c5b3ca26f01d3143c66f2f958b2bef41c241de233af9997ce89459f0a"} Dec 03 14:32:43 crc kubenswrapper[4805]: I1203 14:32:43.016507 4805 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="e7b3280c5b3ca26f01d3143c66f2f958b2bef41c241de233af9997ce89459f0a" Dec 03 14:32:43 crc kubenswrapper[4805]: I1203 14:32:43.016583 4805 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-db-sync-dxrx8" Dec 03 14:32:43 crc kubenswrapper[4805]: I1203 14:32:43.107780 4805 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-conductor-0"] Dec 03 14:32:43 crc kubenswrapper[4805]: E1203 14:32:43.108302 4805 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e56c5750-e0c6-4427-ae8c-ab4af0c9b7d0" containerName="nova-cell0-conductor-db-sync" Dec 03 14:32:43 crc kubenswrapper[4805]: I1203 14:32:43.108363 4805 state_mem.go:107] "Deleted CPUSet assignment" podUID="e56c5750-e0c6-4427-ae8c-ab4af0c9b7d0" containerName="nova-cell0-conductor-db-sync" Dec 03 14:32:43 crc kubenswrapper[4805]: I1203 14:32:43.108539 4805 memory_manager.go:354] "RemoveStaleState removing state" podUID="e56c5750-e0c6-4427-ae8c-ab4af0c9b7d0" containerName="nova-cell0-conductor-db-sync" Dec 03 14:32:43 crc kubenswrapper[4805]: I1203 14:32:43.109213 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-0" Dec 03 14:32:43 crc kubenswrapper[4805]: I1203 14:32:43.111433 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-nova-dockercfg-qktk6" Dec 03 14:32:43 crc kubenswrapper[4805]: I1203 14:32:43.111631 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-conductor-config-data" Dec 03 14:32:43 crc kubenswrapper[4805]: I1203 14:32:43.141395 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-0"] Dec 03 14:32:43 crc kubenswrapper[4805]: I1203 14:32:43.225627 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lqshq\" (UniqueName: \"kubernetes.io/projected/bc143d9b-fc34-4d12-a500-67de79d3c71a-kube-api-access-lqshq\") pod \"nova-cell0-conductor-0\" (UID: \"bc143d9b-fc34-4d12-a500-67de79d3c71a\") " pod="openstack/nova-cell0-conductor-0" Dec 03 14:32:43 crc kubenswrapper[4805]: I1203 14:32:43.225683 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bc143d9b-fc34-4d12-a500-67de79d3c71a-combined-ca-bundle\") pod \"nova-cell0-conductor-0\" (UID: \"bc143d9b-fc34-4d12-a500-67de79d3c71a\") " pod="openstack/nova-cell0-conductor-0" Dec 03 14:32:43 crc kubenswrapper[4805]: I1203 14:32:43.225740 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bc143d9b-fc34-4d12-a500-67de79d3c71a-config-data\") pod \"nova-cell0-conductor-0\" (UID: \"bc143d9b-fc34-4d12-a500-67de79d3c71a\") " pod="openstack/nova-cell0-conductor-0" Dec 03 14:32:43 crc kubenswrapper[4805]: I1203 14:32:43.327400 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bc143d9b-fc34-4d12-a500-67de79d3c71a-config-data\") pod \"nova-cell0-conductor-0\" (UID: \"bc143d9b-fc34-4d12-a500-67de79d3c71a\") " pod="openstack/nova-cell0-conductor-0" Dec 03 14:32:43 crc kubenswrapper[4805]: I1203 14:32:43.328156 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lqshq\" (UniqueName: \"kubernetes.io/projected/bc143d9b-fc34-4d12-a500-67de79d3c71a-kube-api-access-lqshq\") pod \"nova-cell0-conductor-0\" (UID: \"bc143d9b-fc34-4d12-a500-67de79d3c71a\") " pod="openstack/nova-cell0-conductor-0" Dec 03 14:32:43 crc kubenswrapper[4805]: I1203 14:32:43.328196 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bc143d9b-fc34-4d12-a500-67de79d3c71a-combined-ca-bundle\") pod \"nova-cell0-conductor-0\" (UID: \"bc143d9b-fc34-4d12-a500-67de79d3c71a\") " pod="openstack/nova-cell0-conductor-0" Dec 03 14:32:43 crc kubenswrapper[4805]: I1203 14:32:43.331993 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bc143d9b-fc34-4d12-a500-67de79d3c71a-combined-ca-bundle\") pod \"nova-cell0-conductor-0\" (UID: \"bc143d9b-fc34-4d12-a500-67de79d3c71a\") " pod="openstack/nova-cell0-conductor-0" Dec 03 14:32:43 crc kubenswrapper[4805]: I1203 14:32:43.332576 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bc143d9b-fc34-4d12-a500-67de79d3c71a-config-data\") pod \"nova-cell0-conductor-0\" (UID: \"bc143d9b-fc34-4d12-a500-67de79d3c71a\") " pod="openstack/nova-cell0-conductor-0" Dec 03 14:32:43 crc kubenswrapper[4805]: I1203 14:32:43.345288 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lqshq\" (UniqueName: \"kubernetes.io/projected/bc143d9b-fc34-4d12-a500-67de79d3c71a-kube-api-access-lqshq\") pod \"nova-cell0-conductor-0\" (UID: \"bc143d9b-fc34-4d12-a500-67de79d3c71a\") " pod="openstack/nova-cell0-conductor-0" Dec 03 14:32:43 crc kubenswrapper[4805]: I1203 14:32:43.558541 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-0" Dec 03 14:32:43 crc kubenswrapper[4805]: I1203 14:32:43.984267 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-0"] Dec 03 14:32:43 crc kubenswrapper[4805]: W1203 14:32:43.994204 4805 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podbc143d9b_fc34_4d12_a500_67de79d3c71a.slice/crio-9d0f22312958b860f331488550f215b5a309faf354cf119122898c38bb40633e WatchSource:0}: Error finding container 9d0f22312958b860f331488550f215b5a309faf354cf119122898c38bb40633e: Status 404 returned error can't find the container with id 9d0f22312958b860f331488550f215b5a309faf354cf119122898c38bb40633e Dec 03 14:32:44 crc kubenswrapper[4805]: I1203 14:32:44.025141 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-0" event={"ID":"bc143d9b-fc34-4d12-a500-67de79d3c71a","Type":"ContainerStarted","Data":"9d0f22312958b860f331488550f215b5a309faf354cf119122898c38bb40633e"} Dec 03 14:32:44 crc kubenswrapper[4805]: I1203 14:32:44.028325 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"35025929-9af2-4fd0-a702-76c349a6171e","Type":"ContainerStarted","Data":"dfe99b279066a8e9b93a4ec803d7bd9528727ddbf8ae2764b40f68e9fe03172b"} Dec 03 14:32:44 crc kubenswrapper[4805]: I1203 14:32:44.028363 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"35025929-9af2-4fd0-a702-76c349a6171e","Type":"ContainerStarted","Data":"1c180f0dcd29b37222c5801cb75a7ebee5cc5c54fca7eadd9e4de2fd94aa6937"} Dec 03 14:32:45 crc kubenswrapper[4805]: I1203 14:32:45.037645 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-0" event={"ID":"bc143d9b-fc34-4d12-a500-67de79d3c71a","Type":"ContainerStarted","Data":"c81ca8675195d5171b679879f92f0811427104a6c3928cf4955cf315a17edbdd"} Dec 03 14:32:46 crc kubenswrapper[4805]: I1203 14:32:46.050301 4805 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell0-conductor-0" Dec 03 14:32:46 crc kubenswrapper[4805]: I1203 14:32:46.070659 4805 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell0-conductor-0" podStartSLOduration=3.070640889 podStartE2EDuration="3.070640889s" podCreationTimestamp="2025-12-03 14:32:43 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 14:32:46.065807987 +0000 UTC m=+1395.728724910" watchObservedRunningTime="2025-12-03 14:32:46.070640889 +0000 UTC m=+1395.733557812" Dec 03 14:32:46 crc kubenswrapper[4805]: I1203 14:32:46.952741 4805 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/glance-default-internal-api-0" podUID="10877f19-dd2a-4eb2-99e0-2d0cf82d5ad5" containerName="glance-httpd" probeResult="failure" output="Get \"https://10.217.0.148:9292/healthcheck\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Dec 03 14:32:46 crc kubenswrapper[4805]: I1203 14:32:46.952980 4805 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/glance-default-internal-api-0" podUID="10877f19-dd2a-4eb2-99e0-2d0cf82d5ad5" containerName="glance-log" probeResult="failure" output="Get \"https://10.217.0.148:9292/healthcheck\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Dec 03 14:32:47 crc kubenswrapper[4805]: I1203 14:32:47.057446 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"35025929-9af2-4fd0-a702-76c349a6171e","Type":"ContainerStarted","Data":"cc7beba651205e4083e1d58cc227c37d592eb0211b10fc96c775c4dd53f8a9cb"} Dec 03 14:32:49 crc kubenswrapper[4805]: I1203 14:32:49.075564 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"35025929-9af2-4fd0-a702-76c349a6171e","Type":"ContainerStarted","Data":"5a88ff238e81b8b27e6ef5d177dbe22e894987d6855096381c3573ca12d3c345"} Dec 03 14:32:49 crc kubenswrapper[4805]: I1203 14:32:49.076145 4805 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Dec 03 14:32:49 crc kubenswrapper[4805]: I1203 14:32:49.106906 4805 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=1.9373891840000002 podStartE2EDuration="8.106883058s" podCreationTimestamp="2025-12-03 14:32:41 +0000 UTC" firstStartedPulling="2025-12-03 14:32:41.893794715 +0000 UTC m=+1391.556711638" lastFinishedPulling="2025-12-03 14:32:48.063288589 +0000 UTC m=+1397.726205512" observedRunningTime="2025-12-03 14:32:49.096521374 +0000 UTC m=+1398.759438307" watchObservedRunningTime="2025-12-03 14:32:49.106883058 +0000 UTC m=+1398.769799991" Dec 03 14:32:53 crc kubenswrapper[4805]: I1203 14:32:53.117928 4805 generic.go:334] "Generic (PLEG): container finished" podID="c4431968-a92e-4fc0-951f-f42a15942f33" containerID="0b0c768c0ff7e0144c8b1fd624455711a096a5d37b363a0e79eec761e0566c8b" exitCode=137 Dec 03 14:32:53 crc kubenswrapper[4805]: I1203 14:32:53.118504 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-api-6bd4dc884b-6qst5" event={"ID":"c4431968-a92e-4fc0-951f-f42a15942f33","Type":"ContainerDied","Data":"0b0c768c0ff7e0144c8b1fd624455711a096a5d37b363a0e79eec761e0566c8b"} Dec 03 14:32:53 crc kubenswrapper[4805]: I1203 14:32:53.119690 4805 generic.go:334] "Generic (PLEG): container finished" podID="33d42113-07b3-4db3-95b8-f43810991ce4" containerID="2bde7c7fcdbea9aa9f1f6642978c89d9478ebabcfe3db1f02634badb979b5b32" exitCode=137 Dec 03 14:32:53 crc kubenswrapper[4805]: I1203 14:32:53.119707 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-cfnapi-795fdbdfbf-8zpm7" event={"ID":"33d42113-07b3-4db3-95b8-f43810991ce4","Type":"ContainerDied","Data":"2bde7c7fcdbea9aa9f1f6642978c89d9478ebabcfe3db1f02634badb979b5b32"} Dec 03 14:32:53 crc kubenswrapper[4805]: I1203 14:32:53.239111 4805 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/heat-api-6bd4dc884b-6qst5" Dec 03 14:32:53 crc kubenswrapper[4805]: I1203 14:32:53.246955 4805 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/heat-cfnapi-795fdbdfbf-8zpm7" Dec 03 14:32:53 crc kubenswrapper[4805]: I1203 14:32:53.247338 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c4431968-a92e-4fc0-951f-f42a15942f33-combined-ca-bundle\") pod \"c4431968-a92e-4fc0-951f-f42a15942f33\" (UID: \"c4431968-a92e-4fc0-951f-f42a15942f33\") " Dec 03 14:32:53 crc kubenswrapper[4805]: I1203 14:32:53.247597 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/c4431968-a92e-4fc0-951f-f42a15942f33-config-data-custom\") pod \"c4431968-a92e-4fc0-951f-f42a15942f33\" (UID: \"c4431968-a92e-4fc0-951f-f42a15942f33\") " Dec 03 14:32:53 crc kubenswrapper[4805]: I1203 14:32:53.252989 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c4431968-a92e-4fc0-951f-f42a15942f33-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "c4431968-a92e-4fc0-951f-f42a15942f33" (UID: "c4431968-a92e-4fc0-951f-f42a15942f33"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 14:32:53 crc kubenswrapper[4805]: I1203 14:32:53.311557 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c4431968-a92e-4fc0-951f-f42a15942f33-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "c4431968-a92e-4fc0-951f-f42a15942f33" (UID: "c4431968-a92e-4fc0-951f-f42a15942f33"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 14:32:53 crc kubenswrapper[4805]: I1203 14:32:53.349594 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/33d42113-07b3-4db3-95b8-f43810991ce4-config-data-custom\") pod \"33d42113-07b3-4db3-95b8-f43810991ce4\" (UID: \"33d42113-07b3-4db3-95b8-f43810991ce4\") " Dec 03 14:32:53 crc kubenswrapper[4805]: I1203 14:32:53.349689 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4k7rf\" (UniqueName: \"kubernetes.io/projected/33d42113-07b3-4db3-95b8-f43810991ce4-kube-api-access-4k7rf\") pod \"33d42113-07b3-4db3-95b8-f43810991ce4\" (UID: \"33d42113-07b3-4db3-95b8-f43810991ce4\") " Dec 03 14:32:53 crc kubenswrapper[4805]: I1203 14:32:53.349751 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rd575\" (UniqueName: \"kubernetes.io/projected/c4431968-a92e-4fc0-951f-f42a15942f33-kube-api-access-rd575\") pod \"c4431968-a92e-4fc0-951f-f42a15942f33\" (UID: \"c4431968-a92e-4fc0-951f-f42a15942f33\") " Dec 03 14:32:53 crc kubenswrapper[4805]: I1203 14:32:53.349777 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/33d42113-07b3-4db3-95b8-f43810991ce4-config-data\") pod \"33d42113-07b3-4db3-95b8-f43810991ce4\" (UID: \"33d42113-07b3-4db3-95b8-f43810991ce4\") " Dec 03 14:32:53 crc kubenswrapper[4805]: I1203 14:32:53.349827 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/33d42113-07b3-4db3-95b8-f43810991ce4-combined-ca-bundle\") pod \"33d42113-07b3-4db3-95b8-f43810991ce4\" (UID: \"33d42113-07b3-4db3-95b8-f43810991ce4\") " Dec 03 14:32:53 crc kubenswrapper[4805]: I1203 14:32:53.349870 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c4431968-a92e-4fc0-951f-f42a15942f33-config-data\") pod \"c4431968-a92e-4fc0-951f-f42a15942f33\" (UID: \"c4431968-a92e-4fc0-951f-f42a15942f33\") " Dec 03 14:32:53 crc kubenswrapper[4805]: I1203 14:32:53.350111 4805 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c4431968-a92e-4fc0-951f-f42a15942f33-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 14:32:53 crc kubenswrapper[4805]: I1203 14:32:53.350123 4805 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/c4431968-a92e-4fc0-951f-f42a15942f33-config-data-custom\") on node \"crc\" DevicePath \"\"" Dec 03 14:32:53 crc kubenswrapper[4805]: I1203 14:32:53.353598 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/33d42113-07b3-4db3-95b8-f43810991ce4-kube-api-access-4k7rf" (OuterVolumeSpecName: "kube-api-access-4k7rf") pod "33d42113-07b3-4db3-95b8-f43810991ce4" (UID: "33d42113-07b3-4db3-95b8-f43810991ce4"). InnerVolumeSpecName "kube-api-access-4k7rf". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 14:32:53 crc kubenswrapper[4805]: I1203 14:32:53.353934 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c4431968-a92e-4fc0-951f-f42a15942f33-kube-api-access-rd575" (OuterVolumeSpecName: "kube-api-access-rd575") pod "c4431968-a92e-4fc0-951f-f42a15942f33" (UID: "c4431968-a92e-4fc0-951f-f42a15942f33"). InnerVolumeSpecName "kube-api-access-rd575". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 14:32:53 crc kubenswrapper[4805]: I1203 14:32:53.355890 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/33d42113-07b3-4db3-95b8-f43810991ce4-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "33d42113-07b3-4db3-95b8-f43810991ce4" (UID: "33d42113-07b3-4db3-95b8-f43810991ce4"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 14:32:53 crc kubenswrapper[4805]: I1203 14:32:53.380862 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/33d42113-07b3-4db3-95b8-f43810991ce4-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "33d42113-07b3-4db3-95b8-f43810991ce4" (UID: "33d42113-07b3-4db3-95b8-f43810991ce4"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 14:32:53 crc kubenswrapper[4805]: I1203 14:32:53.400495 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c4431968-a92e-4fc0-951f-f42a15942f33-config-data" (OuterVolumeSpecName: "config-data") pod "c4431968-a92e-4fc0-951f-f42a15942f33" (UID: "c4431968-a92e-4fc0-951f-f42a15942f33"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 14:32:53 crc kubenswrapper[4805]: I1203 14:32:53.415579 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/33d42113-07b3-4db3-95b8-f43810991ce4-config-data" (OuterVolumeSpecName: "config-data") pod "33d42113-07b3-4db3-95b8-f43810991ce4" (UID: "33d42113-07b3-4db3-95b8-f43810991ce4"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 14:32:53 crc kubenswrapper[4805]: I1203 14:32:53.450771 4805 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/33d42113-07b3-4db3-95b8-f43810991ce4-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 14:32:53 crc kubenswrapper[4805]: I1203 14:32:53.451010 4805 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c4431968-a92e-4fc0-951f-f42a15942f33-config-data\") on node \"crc\" DevicePath \"\"" Dec 03 14:32:53 crc kubenswrapper[4805]: I1203 14:32:53.451106 4805 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/33d42113-07b3-4db3-95b8-f43810991ce4-config-data-custom\") on node \"crc\" DevicePath \"\"" Dec 03 14:32:53 crc kubenswrapper[4805]: I1203 14:32:53.451171 4805 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4k7rf\" (UniqueName: \"kubernetes.io/projected/33d42113-07b3-4db3-95b8-f43810991ce4-kube-api-access-4k7rf\") on node \"crc\" DevicePath \"\"" Dec 03 14:32:53 crc kubenswrapper[4805]: I1203 14:32:53.451228 4805 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rd575\" (UniqueName: \"kubernetes.io/projected/c4431968-a92e-4fc0-951f-f42a15942f33-kube-api-access-rd575\") on node \"crc\" DevicePath \"\"" Dec 03 14:32:53 crc kubenswrapper[4805]: I1203 14:32:53.451288 4805 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/33d42113-07b3-4db3-95b8-f43810991ce4-config-data\") on node \"crc\" DevicePath \"\"" Dec 03 14:32:53 crc kubenswrapper[4805]: I1203 14:32:53.590932 4805 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-cell0-conductor-0" Dec 03 14:32:54 crc kubenswrapper[4805]: I1203 14:32:54.033152 4805 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-cell-mapping-bmkm9"] Dec 03 14:32:54 crc kubenswrapper[4805]: E1203 14:32:54.036996 4805 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="33d42113-07b3-4db3-95b8-f43810991ce4" containerName="heat-cfnapi" Dec 03 14:32:54 crc kubenswrapper[4805]: I1203 14:32:54.037036 4805 state_mem.go:107] "Deleted CPUSet assignment" podUID="33d42113-07b3-4db3-95b8-f43810991ce4" containerName="heat-cfnapi" Dec 03 14:32:54 crc kubenswrapper[4805]: E1203 14:32:54.037057 4805 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c4431968-a92e-4fc0-951f-f42a15942f33" containerName="heat-api" Dec 03 14:32:54 crc kubenswrapper[4805]: I1203 14:32:54.037065 4805 state_mem.go:107] "Deleted CPUSet assignment" podUID="c4431968-a92e-4fc0-951f-f42a15942f33" containerName="heat-api" Dec 03 14:32:54 crc kubenswrapper[4805]: I1203 14:32:54.037313 4805 memory_manager.go:354] "RemoveStaleState removing state" podUID="c4431968-a92e-4fc0-951f-f42a15942f33" containerName="heat-api" Dec 03 14:32:54 crc kubenswrapper[4805]: I1203 14:32:54.037327 4805 memory_manager.go:354] "RemoveStaleState removing state" podUID="33d42113-07b3-4db3-95b8-f43810991ce4" containerName="heat-cfnapi" Dec 03 14:32:54 crc kubenswrapper[4805]: I1203 14:32:54.038098 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-cell-mapping-bmkm9" Dec 03 14:32:54 crc kubenswrapper[4805]: I1203 14:32:54.040380 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-manage-scripts" Dec 03 14:32:54 crc kubenswrapper[4805]: I1203 14:32:54.041126 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-manage-config-data" Dec 03 14:32:54 crc kubenswrapper[4805]: I1203 14:32:54.062309 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-cell-mapping-bmkm9"] Dec 03 14:32:54 crc kubenswrapper[4805]: I1203 14:32:54.131096 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-api-6bd4dc884b-6qst5" event={"ID":"c4431968-a92e-4fc0-951f-f42a15942f33","Type":"ContainerDied","Data":"bf0e3e07687b91d9e6e73348065d3d6132523b5624cd87d8e1623d8a16fb1bfd"} Dec 03 14:32:54 crc kubenswrapper[4805]: I1203 14:32:54.131142 4805 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/heat-api-6bd4dc884b-6qst5" Dec 03 14:32:54 crc kubenswrapper[4805]: I1203 14:32:54.131161 4805 scope.go:117] "RemoveContainer" containerID="0b0c768c0ff7e0144c8b1fd624455711a096a5d37b363a0e79eec761e0566c8b" Dec 03 14:32:54 crc kubenswrapper[4805]: I1203 14:32:54.136084 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-cfnapi-795fdbdfbf-8zpm7" event={"ID":"33d42113-07b3-4db3-95b8-f43810991ce4","Type":"ContainerDied","Data":"782aa0b2673f8a838bacd094fec9c0de3f9f91e5da1d5fe1ec5913918509d105"} Dec 03 14:32:54 crc kubenswrapper[4805]: I1203 14:32:54.136345 4805 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/heat-cfnapi-795fdbdfbf-8zpm7" Dec 03 14:32:54 crc kubenswrapper[4805]: I1203 14:32:54.164474 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/97cf9020-6417-4d08-9cdf-f38515b63d82-config-data\") pod \"nova-cell0-cell-mapping-bmkm9\" (UID: \"97cf9020-6417-4d08-9cdf-f38515b63d82\") " pod="openstack/nova-cell0-cell-mapping-bmkm9" Dec 03 14:32:54 crc kubenswrapper[4805]: I1203 14:32:54.164570 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/97cf9020-6417-4d08-9cdf-f38515b63d82-combined-ca-bundle\") pod \"nova-cell0-cell-mapping-bmkm9\" (UID: \"97cf9020-6417-4d08-9cdf-f38515b63d82\") " pod="openstack/nova-cell0-cell-mapping-bmkm9" Dec 03 14:32:54 crc kubenswrapper[4805]: I1203 14:32:54.164637 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jsxfm\" (UniqueName: \"kubernetes.io/projected/97cf9020-6417-4d08-9cdf-f38515b63d82-kube-api-access-jsxfm\") pod \"nova-cell0-cell-mapping-bmkm9\" (UID: \"97cf9020-6417-4d08-9cdf-f38515b63d82\") " pod="openstack/nova-cell0-cell-mapping-bmkm9" Dec 03 14:32:54 crc kubenswrapper[4805]: I1203 14:32:54.164713 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/97cf9020-6417-4d08-9cdf-f38515b63d82-scripts\") pod \"nova-cell0-cell-mapping-bmkm9\" (UID: \"97cf9020-6417-4d08-9cdf-f38515b63d82\") " pod="openstack/nova-cell0-cell-mapping-bmkm9" Dec 03 14:32:54 crc kubenswrapper[4805]: I1203 14:32:54.185060 4805 scope.go:117] "RemoveContainer" containerID="2bde7c7fcdbea9aa9f1f6642978c89d9478ebabcfe3db1f02634badb979b5b32" Dec 03 14:32:54 crc kubenswrapper[4805]: I1203 14:32:54.188592 4805 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/heat-cfnapi-795fdbdfbf-8zpm7"] Dec 03 14:32:54 crc kubenswrapper[4805]: I1203 14:32:54.202697 4805 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/heat-cfnapi-795fdbdfbf-8zpm7"] Dec 03 14:32:54 crc kubenswrapper[4805]: I1203 14:32:54.227915 4805 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/heat-api-6bd4dc884b-6qst5"] Dec 03 14:32:54 crc kubenswrapper[4805]: I1203 14:32:54.246114 4805 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/heat-api-6bd4dc884b-6qst5"] Dec 03 14:32:54 crc kubenswrapper[4805]: I1203 14:32:54.262364 4805 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-0"] Dec 03 14:32:54 crc kubenswrapper[4805]: I1203 14:32:54.266787 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 03 14:32:54 crc kubenswrapper[4805]: I1203 14:32:54.270106 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jsxfm\" (UniqueName: \"kubernetes.io/projected/97cf9020-6417-4d08-9cdf-f38515b63d82-kube-api-access-jsxfm\") pod \"nova-cell0-cell-mapping-bmkm9\" (UID: \"97cf9020-6417-4d08-9cdf-f38515b63d82\") " pod="openstack/nova-cell0-cell-mapping-bmkm9" Dec 03 14:32:54 crc kubenswrapper[4805]: I1203 14:32:54.270235 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/97cf9020-6417-4d08-9cdf-f38515b63d82-scripts\") pod \"nova-cell0-cell-mapping-bmkm9\" (UID: \"97cf9020-6417-4d08-9cdf-f38515b63d82\") " pod="openstack/nova-cell0-cell-mapping-bmkm9" Dec 03 14:32:54 crc kubenswrapper[4805]: I1203 14:32:54.270291 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/97cf9020-6417-4d08-9cdf-f38515b63d82-config-data\") pod \"nova-cell0-cell-mapping-bmkm9\" (UID: \"97cf9020-6417-4d08-9cdf-f38515b63d82\") " pod="openstack/nova-cell0-cell-mapping-bmkm9" Dec 03 14:32:54 crc kubenswrapper[4805]: I1203 14:32:54.270370 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/97cf9020-6417-4d08-9cdf-f38515b63d82-combined-ca-bundle\") pod \"nova-cell0-cell-mapping-bmkm9\" (UID: \"97cf9020-6417-4d08-9cdf-f38515b63d82\") " pod="openstack/nova-cell0-cell-mapping-bmkm9" Dec 03 14:32:54 crc kubenswrapper[4805]: I1203 14:32:54.273745 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-config-data" Dec 03 14:32:54 crc kubenswrapper[4805]: I1203 14:32:54.295494 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/97cf9020-6417-4d08-9cdf-f38515b63d82-scripts\") pod \"nova-cell0-cell-mapping-bmkm9\" (UID: \"97cf9020-6417-4d08-9cdf-f38515b63d82\") " pod="openstack/nova-cell0-cell-mapping-bmkm9" Dec 03 14:32:54 crc kubenswrapper[4805]: I1203 14:32:54.298147 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Dec 03 14:32:54 crc kubenswrapper[4805]: I1203 14:32:54.311559 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/97cf9020-6417-4d08-9cdf-f38515b63d82-config-data\") pod \"nova-cell0-cell-mapping-bmkm9\" (UID: \"97cf9020-6417-4d08-9cdf-f38515b63d82\") " pod="openstack/nova-cell0-cell-mapping-bmkm9" Dec 03 14:32:54 crc kubenswrapper[4805]: I1203 14:32:54.315219 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jsxfm\" (UniqueName: \"kubernetes.io/projected/97cf9020-6417-4d08-9cdf-f38515b63d82-kube-api-access-jsxfm\") pod \"nova-cell0-cell-mapping-bmkm9\" (UID: \"97cf9020-6417-4d08-9cdf-f38515b63d82\") " pod="openstack/nova-cell0-cell-mapping-bmkm9" Dec 03 14:32:54 crc kubenswrapper[4805]: I1203 14:32:54.347655 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/97cf9020-6417-4d08-9cdf-f38515b63d82-combined-ca-bundle\") pod \"nova-cell0-cell-mapping-bmkm9\" (UID: \"97cf9020-6417-4d08-9cdf-f38515b63d82\") " pod="openstack/nova-cell0-cell-mapping-bmkm9" Dec 03 14:32:54 crc kubenswrapper[4805]: I1203 14:32:54.355149 4805 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Dec 03 14:32:54 crc kubenswrapper[4805]: I1203 14:32:54.356480 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Dec 03 14:32:54 crc kubenswrapper[4805]: I1203 14:32:54.357530 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-cell-mapping-bmkm9" Dec 03 14:32:54 crc kubenswrapper[4805]: I1203 14:32:54.358700 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-novncproxy-config-data" Dec 03 14:32:54 crc kubenswrapper[4805]: I1203 14:32:54.377932 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/08314cda-c0a3-4a26-bb3a-65cd724335a6-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"08314cda-c0a3-4a26-bb3a-65cd724335a6\") " pod="openstack/nova-api-0" Dec 03 14:32:54 crc kubenswrapper[4805]: I1203 14:32:54.377994 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/08314cda-c0a3-4a26-bb3a-65cd724335a6-logs\") pod \"nova-api-0\" (UID: \"08314cda-c0a3-4a26-bb3a-65cd724335a6\") " pod="openstack/nova-api-0" Dec 03 14:32:54 crc kubenswrapper[4805]: I1203 14:32:54.378058 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5tnl7\" (UniqueName: \"kubernetes.io/projected/08314cda-c0a3-4a26-bb3a-65cd724335a6-kube-api-access-5tnl7\") pod \"nova-api-0\" (UID: \"08314cda-c0a3-4a26-bb3a-65cd724335a6\") " pod="openstack/nova-api-0" Dec 03 14:32:54 crc kubenswrapper[4805]: I1203 14:32:54.378080 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/08314cda-c0a3-4a26-bb3a-65cd724335a6-config-data\") pod \"nova-api-0\" (UID: \"08314cda-c0a3-4a26-bb3a-65cd724335a6\") " pod="openstack/nova-api-0" Dec 03 14:32:54 crc kubenswrapper[4805]: I1203 14:32:54.401912 4805 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-scheduler-0"] Dec 03 14:32:54 crc kubenswrapper[4805]: I1203 14:32:54.403434 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Dec 03 14:32:54 crc kubenswrapper[4805]: I1203 14:32:54.414400 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-scheduler-config-data" Dec 03 14:32:54 crc kubenswrapper[4805]: I1203 14:32:54.438524 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Dec 03 14:32:54 crc kubenswrapper[4805]: I1203 14:32:54.472620 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Dec 03 14:32:54 crc kubenswrapper[4805]: I1203 14:32:54.479532 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/660cf06e-323d-478f-aa70-a928724391c3-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"660cf06e-323d-478f-aa70-a928724391c3\") " pod="openstack/nova-cell1-novncproxy-0" Dec 03 14:32:54 crc kubenswrapper[4805]: I1203 14:32:54.479573 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/08314cda-c0a3-4a26-bb3a-65cd724335a6-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"08314cda-c0a3-4a26-bb3a-65cd724335a6\") " pod="openstack/nova-api-0" Dec 03 14:32:54 crc kubenswrapper[4805]: I1203 14:32:54.479611 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-89n8v\" (UniqueName: \"kubernetes.io/projected/660cf06e-323d-478f-aa70-a928724391c3-kube-api-access-89n8v\") pod \"nova-cell1-novncproxy-0\" (UID: \"660cf06e-323d-478f-aa70-a928724391c3\") " pod="openstack/nova-cell1-novncproxy-0" Dec 03 14:32:54 crc kubenswrapper[4805]: I1203 14:32:54.479629 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/08314cda-c0a3-4a26-bb3a-65cd724335a6-logs\") pod \"nova-api-0\" (UID: \"08314cda-c0a3-4a26-bb3a-65cd724335a6\") " pod="openstack/nova-api-0" Dec 03 14:32:54 crc kubenswrapper[4805]: I1203 14:32:54.479667 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/660cf06e-323d-478f-aa70-a928724391c3-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"660cf06e-323d-478f-aa70-a928724391c3\") " pod="openstack/nova-cell1-novncproxy-0" Dec 03 14:32:54 crc kubenswrapper[4805]: I1203 14:32:54.479699 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5tnl7\" (UniqueName: \"kubernetes.io/projected/08314cda-c0a3-4a26-bb3a-65cd724335a6-kube-api-access-5tnl7\") pod \"nova-api-0\" (UID: \"08314cda-c0a3-4a26-bb3a-65cd724335a6\") " pod="openstack/nova-api-0" Dec 03 14:32:54 crc kubenswrapper[4805]: I1203 14:32:54.479720 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/08314cda-c0a3-4a26-bb3a-65cd724335a6-config-data\") pod \"nova-api-0\" (UID: \"08314cda-c0a3-4a26-bb3a-65cd724335a6\") " pod="openstack/nova-api-0" Dec 03 14:32:54 crc kubenswrapper[4805]: I1203 14:32:54.480304 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/08314cda-c0a3-4a26-bb3a-65cd724335a6-logs\") pod \"nova-api-0\" (UID: \"08314cda-c0a3-4a26-bb3a-65cd724335a6\") " pod="openstack/nova-api-0" Dec 03 14:32:54 crc kubenswrapper[4805]: I1203 14:32:54.481719 4805 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-metadata-0"] Dec 03 14:32:54 crc kubenswrapper[4805]: I1203 14:32:54.485211 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 03 14:32:54 crc kubenswrapper[4805]: I1203 14:32:54.496032 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-config-data" Dec 03 14:32:54 crc kubenswrapper[4805]: I1203 14:32:54.501577 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/08314cda-c0a3-4a26-bb3a-65cd724335a6-config-data\") pod \"nova-api-0\" (UID: \"08314cda-c0a3-4a26-bb3a-65cd724335a6\") " pod="openstack/nova-api-0" Dec 03 14:32:54 crc kubenswrapper[4805]: I1203 14:32:54.504447 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/08314cda-c0a3-4a26-bb3a-65cd724335a6-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"08314cda-c0a3-4a26-bb3a-65cd724335a6\") " pod="openstack/nova-api-0" Dec 03 14:32:54 crc kubenswrapper[4805]: I1203 14:32:54.512560 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5tnl7\" (UniqueName: \"kubernetes.io/projected/08314cda-c0a3-4a26-bb3a-65cd724335a6-kube-api-access-5tnl7\") pod \"nova-api-0\" (UID: \"08314cda-c0a3-4a26-bb3a-65cd724335a6\") " pod="openstack/nova-api-0" Dec 03 14:32:54 crc kubenswrapper[4805]: I1203 14:32:54.540895 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Dec 03 14:32:54 crc kubenswrapper[4805]: I1203 14:32:54.581315 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5c64fe61-215b-47f5-aeae-98548dce4e81-config-data\") pod \"nova-metadata-0\" (UID: \"5c64fe61-215b-47f5-aeae-98548dce4e81\") " pod="openstack/nova-metadata-0" Dec 03 14:32:54 crc kubenswrapper[4805]: I1203 14:32:54.581367 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/660cf06e-323d-478f-aa70-a928724391c3-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"660cf06e-323d-478f-aa70-a928724391c3\") " pod="openstack/nova-cell1-novncproxy-0" Dec 03 14:32:54 crc kubenswrapper[4805]: I1203 14:32:54.581412 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-89n8v\" (UniqueName: \"kubernetes.io/projected/660cf06e-323d-478f-aa70-a928724391c3-kube-api-access-89n8v\") pod \"nova-cell1-novncproxy-0\" (UID: \"660cf06e-323d-478f-aa70-a928724391c3\") " pod="openstack/nova-cell1-novncproxy-0" Dec 03 14:32:54 crc kubenswrapper[4805]: I1203 14:32:54.581434 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/5c64fe61-215b-47f5-aeae-98548dce4e81-logs\") pod \"nova-metadata-0\" (UID: \"5c64fe61-215b-47f5-aeae-98548dce4e81\") " pod="openstack/nova-metadata-0" Dec 03 14:32:54 crc kubenswrapper[4805]: I1203 14:32:54.581468 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bbp6p\" (UniqueName: \"kubernetes.io/projected/0ba4f686-dd10-475f-a65a-78041893d75d-kube-api-access-bbp6p\") pod \"nova-scheduler-0\" (UID: \"0ba4f686-dd10-475f-a65a-78041893d75d\") " pod="openstack/nova-scheduler-0" Dec 03 14:32:54 crc kubenswrapper[4805]: I1203 14:32:54.581495 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/660cf06e-323d-478f-aa70-a928724391c3-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"660cf06e-323d-478f-aa70-a928724391c3\") " pod="openstack/nova-cell1-novncproxy-0" Dec 03 14:32:54 crc kubenswrapper[4805]: I1203 14:32:54.581546 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0ba4f686-dd10-475f-a65a-78041893d75d-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"0ba4f686-dd10-475f-a65a-78041893d75d\") " pod="openstack/nova-scheduler-0" Dec 03 14:32:54 crc kubenswrapper[4805]: I1203 14:32:54.581562 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vnzwm\" (UniqueName: \"kubernetes.io/projected/5c64fe61-215b-47f5-aeae-98548dce4e81-kube-api-access-vnzwm\") pod \"nova-metadata-0\" (UID: \"5c64fe61-215b-47f5-aeae-98548dce4e81\") " pod="openstack/nova-metadata-0" Dec 03 14:32:54 crc kubenswrapper[4805]: I1203 14:32:54.581594 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0ba4f686-dd10-475f-a65a-78041893d75d-config-data\") pod \"nova-scheduler-0\" (UID: \"0ba4f686-dd10-475f-a65a-78041893d75d\") " pod="openstack/nova-scheduler-0" Dec 03 14:32:54 crc kubenswrapper[4805]: I1203 14:32:54.581632 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5c64fe61-215b-47f5-aeae-98548dce4e81-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"5c64fe61-215b-47f5-aeae-98548dce4e81\") " pod="openstack/nova-metadata-0" Dec 03 14:32:54 crc kubenswrapper[4805]: I1203 14:32:54.587260 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/660cf06e-323d-478f-aa70-a928724391c3-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"660cf06e-323d-478f-aa70-a928724391c3\") " pod="openstack/nova-cell1-novncproxy-0" Dec 03 14:32:54 crc kubenswrapper[4805]: I1203 14:32:54.587348 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/660cf06e-323d-478f-aa70-a928724391c3-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"660cf06e-323d-478f-aa70-a928724391c3\") " pod="openstack/nova-cell1-novncproxy-0" Dec 03 14:32:54 crc kubenswrapper[4805]: I1203 14:32:54.590330 4805 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-5fbc4d444f-5vhgg"] Dec 03 14:32:54 crc kubenswrapper[4805]: I1203 14:32:54.590785 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 03 14:32:54 crc kubenswrapper[4805]: I1203 14:32:54.591909 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5fbc4d444f-5vhgg" Dec 03 14:32:54 crc kubenswrapper[4805]: I1203 14:32:54.603352 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-89n8v\" (UniqueName: \"kubernetes.io/projected/660cf06e-323d-478f-aa70-a928724391c3-kube-api-access-89n8v\") pod \"nova-cell1-novncproxy-0\" (UID: \"660cf06e-323d-478f-aa70-a928724391c3\") " pod="openstack/nova-cell1-novncproxy-0" Dec 03 14:32:54 crc kubenswrapper[4805]: I1203 14:32:54.624000 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5fbc4d444f-5vhgg"] Dec 03 14:32:54 crc kubenswrapper[4805]: I1203 14:32:54.683237 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bbp6p\" (UniqueName: \"kubernetes.io/projected/0ba4f686-dd10-475f-a65a-78041893d75d-kube-api-access-bbp6p\") pod \"nova-scheduler-0\" (UID: \"0ba4f686-dd10-475f-a65a-78041893d75d\") " pod="openstack/nova-scheduler-0" Dec 03 14:32:54 crc kubenswrapper[4805]: I1203 14:32:54.683331 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e2ef61bb-5836-4986-88ae-f307ee375f0e-config\") pod \"dnsmasq-dns-5fbc4d444f-5vhgg\" (UID: \"e2ef61bb-5836-4986-88ae-f307ee375f0e\") " pod="openstack/dnsmasq-dns-5fbc4d444f-5vhgg" Dec 03 14:32:54 crc kubenswrapper[4805]: I1203 14:32:54.683367 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vnzwm\" (UniqueName: \"kubernetes.io/projected/5c64fe61-215b-47f5-aeae-98548dce4e81-kube-api-access-vnzwm\") pod \"nova-metadata-0\" (UID: \"5c64fe61-215b-47f5-aeae-98548dce4e81\") " pod="openstack/nova-metadata-0" Dec 03 14:32:54 crc kubenswrapper[4805]: I1203 14:32:54.683382 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0ba4f686-dd10-475f-a65a-78041893d75d-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"0ba4f686-dd10-475f-a65a-78041893d75d\") " pod="openstack/nova-scheduler-0" Dec 03 14:32:54 crc kubenswrapper[4805]: I1203 14:32:54.683403 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/e2ef61bb-5836-4986-88ae-f307ee375f0e-dns-swift-storage-0\") pod \"dnsmasq-dns-5fbc4d444f-5vhgg\" (UID: \"e2ef61bb-5836-4986-88ae-f307ee375f0e\") " pod="openstack/dnsmasq-dns-5fbc4d444f-5vhgg" Dec 03 14:32:54 crc kubenswrapper[4805]: I1203 14:32:54.683420 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-l9wqb\" (UniqueName: \"kubernetes.io/projected/e2ef61bb-5836-4986-88ae-f307ee375f0e-kube-api-access-l9wqb\") pod \"dnsmasq-dns-5fbc4d444f-5vhgg\" (UID: \"e2ef61bb-5836-4986-88ae-f307ee375f0e\") " pod="openstack/dnsmasq-dns-5fbc4d444f-5vhgg" Dec 03 14:32:54 crc kubenswrapper[4805]: I1203 14:32:54.683453 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0ba4f686-dd10-475f-a65a-78041893d75d-config-data\") pod \"nova-scheduler-0\" (UID: \"0ba4f686-dd10-475f-a65a-78041893d75d\") " pod="openstack/nova-scheduler-0" Dec 03 14:32:54 crc kubenswrapper[4805]: I1203 14:32:54.683478 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/e2ef61bb-5836-4986-88ae-f307ee375f0e-ovsdbserver-nb\") pod \"dnsmasq-dns-5fbc4d444f-5vhgg\" (UID: \"e2ef61bb-5836-4986-88ae-f307ee375f0e\") " pod="openstack/dnsmasq-dns-5fbc4d444f-5vhgg" Dec 03 14:32:54 crc kubenswrapper[4805]: I1203 14:32:54.683499 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5c64fe61-215b-47f5-aeae-98548dce4e81-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"5c64fe61-215b-47f5-aeae-98548dce4e81\") " pod="openstack/nova-metadata-0" Dec 03 14:32:54 crc kubenswrapper[4805]: I1203 14:32:54.683533 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5c64fe61-215b-47f5-aeae-98548dce4e81-config-data\") pod \"nova-metadata-0\" (UID: \"5c64fe61-215b-47f5-aeae-98548dce4e81\") " pod="openstack/nova-metadata-0" Dec 03 14:32:54 crc kubenswrapper[4805]: I1203 14:32:54.683567 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/e2ef61bb-5836-4986-88ae-f307ee375f0e-ovsdbserver-sb\") pod \"dnsmasq-dns-5fbc4d444f-5vhgg\" (UID: \"e2ef61bb-5836-4986-88ae-f307ee375f0e\") " pod="openstack/dnsmasq-dns-5fbc4d444f-5vhgg" Dec 03 14:32:54 crc kubenswrapper[4805]: I1203 14:32:54.683599 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/e2ef61bb-5836-4986-88ae-f307ee375f0e-dns-svc\") pod \"dnsmasq-dns-5fbc4d444f-5vhgg\" (UID: \"e2ef61bb-5836-4986-88ae-f307ee375f0e\") " pod="openstack/dnsmasq-dns-5fbc4d444f-5vhgg" Dec 03 14:32:54 crc kubenswrapper[4805]: I1203 14:32:54.683942 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/5c64fe61-215b-47f5-aeae-98548dce4e81-logs\") pod \"nova-metadata-0\" (UID: \"5c64fe61-215b-47f5-aeae-98548dce4e81\") " pod="openstack/nova-metadata-0" Dec 03 14:32:54 crc kubenswrapper[4805]: I1203 14:32:54.684407 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/5c64fe61-215b-47f5-aeae-98548dce4e81-logs\") pod \"nova-metadata-0\" (UID: \"5c64fe61-215b-47f5-aeae-98548dce4e81\") " pod="openstack/nova-metadata-0" Dec 03 14:32:54 crc kubenswrapper[4805]: I1203 14:32:54.687636 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0ba4f686-dd10-475f-a65a-78041893d75d-config-data\") pod \"nova-scheduler-0\" (UID: \"0ba4f686-dd10-475f-a65a-78041893d75d\") " pod="openstack/nova-scheduler-0" Dec 03 14:32:54 crc kubenswrapper[4805]: I1203 14:32:54.691510 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0ba4f686-dd10-475f-a65a-78041893d75d-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"0ba4f686-dd10-475f-a65a-78041893d75d\") " pod="openstack/nova-scheduler-0" Dec 03 14:32:54 crc kubenswrapper[4805]: I1203 14:32:54.692106 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5c64fe61-215b-47f5-aeae-98548dce4e81-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"5c64fe61-215b-47f5-aeae-98548dce4e81\") " pod="openstack/nova-metadata-0" Dec 03 14:32:54 crc kubenswrapper[4805]: I1203 14:32:54.695889 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5c64fe61-215b-47f5-aeae-98548dce4e81-config-data\") pod \"nova-metadata-0\" (UID: \"5c64fe61-215b-47f5-aeae-98548dce4e81\") " pod="openstack/nova-metadata-0" Dec 03 14:32:54 crc kubenswrapper[4805]: I1203 14:32:54.703211 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bbp6p\" (UniqueName: \"kubernetes.io/projected/0ba4f686-dd10-475f-a65a-78041893d75d-kube-api-access-bbp6p\") pod \"nova-scheduler-0\" (UID: \"0ba4f686-dd10-475f-a65a-78041893d75d\") " pod="openstack/nova-scheduler-0" Dec 03 14:32:54 crc kubenswrapper[4805]: I1203 14:32:54.703885 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vnzwm\" (UniqueName: \"kubernetes.io/projected/5c64fe61-215b-47f5-aeae-98548dce4e81-kube-api-access-vnzwm\") pod \"nova-metadata-0\" (UID: \"5c64fe61-215b-47f5-aeae-98548dce4e81\") " pod="openstack/nova-metadata-0" Dec 03 14:32:54 crc kubenswrapper[4805]: I1203 14:32:54.740312 4805 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="33d42113-07b3-4db3-95b8-f43810991ce4" path="/var/lib/kubelet/pods/33d42113-07b3-4db3-95b8-f43810991ce4/volumes" Dec 03 14:32:54 crc kubenswrapper[4805]: I1203 14:32:54.740977 4805 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c4431968-a92e-4fc0-951f-f42a15942f33" path="/var/lib/kubelet/pods/c4431968-a92e-4fc0-951f-f42a15942f33/volumes" Dec 03 14:32:54 crc kubenswrapper[4805]: I1203 14:32:54.785827 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/e2ef61bb-5836-4986-88ae-f307ee375f0e-ovsdbserver-sb\") pod \"dnsmasq-dns-5fbc4d444f-5vhgg\" (UID: \"e2ef61bb-5836-4986-88ae-f307ee375f0e\") " pod="openstack/dnsmasq-dns-5fbc4d444f-5vhgg" Dec 03 14:32:54 crc kubenswrapper[4805]: I1203 14:32:54.785899 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/e2ef61bb-5836-4986-88ae-f307ee375f0e-dns-svc\") pod \"dnsmasq-dns-5fbc4d444f-5vhgg\" (UID: \"e2ef61bb-5836-4986-88ae-f307ee375f0e\") " pod="openstack/dnsmasq-dns-5fbc4d444f-5vhgg" Dec 03 14:32:54 crc kubenswrapper[4805]: I1203 14:32:54.786821 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/e2ef61bb-5836-4986-88ae-f307ee375f0e-ovsdbserver-sb\") pod \"dnsmasq-dns-5fbc4d444f-5vhgg\" (UID: \"e2ef61bb-5836-4986-88ae-f307ee375f0e\") " pod="openstack/dnsmasq-dns-5fbc4d444f-5vhgg" Dec 03 14:32:54 crc kubenswrapper[4805]: I1203 14:32:54.786751 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/e2ef61bb-5836-4986-88ae-f307ee375f0e-dns-svc\") pod \"dnsmasq-dns-5fbc4d444f-5vhgg\" (UID: \"e2ef61bb-5836-4986-88ae-f307ee375f0e\") " pod="openstack/dnsmasq-dns-5fbc4d444f-5vhgg" Dec 03 14:32:54 crc kubenswrapper[4805]: I1203 14:32:54.787616 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e2ef61bb-5836-4986-88ae-f307ee375f0e-config\") pod \"dnsmasq-dns-5fbc4d444f-5vhgg\" (UID: \"e2ef61bb-5836-4986-88ae-f307ee375f0e\") " pod="openstack/dnsmasq-dns-5fbc4d444f-5vhgg" Dec 03 14:32:54 crc kubenswrapper[4805]: I1203 14:32:54.788190 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e2ef61bb-5836-4986-88ae-f307ee375f0e-config\") pod \"dnsmasq-dns-5fbc4d444f-5vhgg\" (UID: \"e2ef61bb-5836-4986-88ae-f307ee375f0e\") " pod="openstack/dnsmasq-dns-5fbc4d444f-5vhgg" Dec 03 14:32:54 crc kubenswrapper[4805]: I1203 14:32:54.788245 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/e2ef61bb-5836-4986-88ae-f307ee375f0e-dns-swift-storage-0\") pod \"dnsmasq-dns-5fbc4d444f-5vhgg\" (UID: \"e2ef61bb-5836-4986-88ae-f307ee375f0e\") " pod="openstack/dnsmasq-dns-5fbc4d444f-5vhgg" Dec 03 14:32:54 crc kubenswrapper[4805]: I1203 14:32:54.788268 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-l9wqb\" (UniqueName: \"kubernetes.io/projected/e2ef61bb-5836-4986-88ae-f307ee375f0e-kube-api-access-l9wqb\") pod \"dnsmasq-dns-5fbc4d444f-5vhgg\" (UID: \"e2ef61bb-5836-4986-88ae-f307ee375f0e\") " pod="openstack/dnsmasq-dns-5fbc4d444f-5vhgg" Dec 03 14:32:54 crc kubenswrapper[4805]: I1203 14:32:54.788330 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/e2ef61bb-5836-4986-88ae-f307ee375f0e-ovsdbserver-nb\") pod \"dnsmasq-dns-5fbc4d444f-5vhgg\" (UID: \"e2ef61bb-5836-4986-88ae-f307ee375f0e\") " pod="openstack/dnsmasq-dns-5fbc4d444f-5vhgg" Dec 03 14:32:54 crc kubenswrapper[4805]: I1203 14:32:54.789207 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/e2ef61bb-5836-4986-88ae-f307ee375f0e-dns-swift-storage-0\") pod \"dnsmasq-dns-5fbc4d444f-5vhgg\" (UID: \"e2ef61bb-5836-4986-88ae-f307ee375f0e\") " pod="openstack/dnsmasq-dns-5fbc4d444f-5vhgg" Dec 03 14:32:54 crc kubenswrapper[4805]: I1203 14:32:54.790551 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/e2ef61bb-5836-4986-88ae-f307ee375f0e-ovsdbserver-nb\") pod \"dnsmasq-dns-5fbc4d444f-5vhgg\" (UID: \"e2ef61bb-5836-4986-88ae-f307ee375f0e\") " pod="openstack/dnsmasq-dns-5fbc4d444f-5vhgg" Dec 03 14:32:54 crc kubenswrapper[4805]: I1203 14:32:54.807174 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-l9wqb\" (UniqueName: \"kubernetes.io/projected/e2ef61bb-5836-4986-88ae-f307ee375f0e-kube-api-access-l9wqb\") pod \"dnsmasq-dns-5fbc4d444f-5vhgg\" (UID: \"e2ef61bb-5836-4986-88ae-f307ee375f0e\") " pod="openstack/dnsmasq-dns-5fbc4d444f-5vhgg" Dec 03 14:32:54 crc kubenswrapper[4805]: I1203 14:32:54.842014 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Dec 03 14:32:54 crc kubenswrapper[4805]: I1203 14:32:54.862110 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Dec 03 14:32:54 crc kubenswrapper[4805]: I1203 14:32:54.871738 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 03 14:32:54 crc kubenswrapper[4805]: I1203 14:32:54.925745 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5fbc4d444f-5vhgg" Dec 03 14:32:55 crc kubenswrapper[4805]: I1203 14:32:55.030130 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-cell-mapping-bmkm9"] Dec 03 14:32:55 crc kubenswrapper[4805]: I1203 14:32:55.080663 4805 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-conductor-db-sync-42n9l"] Dec 03 14:32:55 crc kubenswrapper[4805]: I1203 14:32:55.081911 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-db-sync-42n9l" Dec 03 14:32:55 crc kubenswrapper[4805]: I1203 14:32:55.085759 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-conductor-config-data" Dec 03 14:32:55 crc kubenswrapper[4805]: I1203 14:32:55.085969 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-conductor-scripts" Dec 03 14:32:55 crc kubenswrapper[4805]: I1203 14:32:55.107012 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-db-sync-42n9l"] Dec 03 14:32:55 crc kubenswrapper[4805]: I1203 14:32:55.200999 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Dec 03 14:32:55 crc kubenswrapper[4805]: I1203 14:32:55.203689 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/527667db-e5ab-4fe2-89a5-d9110602f1d2-config-data\") pod \"nova-cell1-conductor-db-sync-42n9l\" (UID: \"527667db-e5ab-4fe2-89a5-d9110602f1d2\") " pod="openstack/nova-cell1-conductor-db-sync-42n9l" Dec 03 14:32:55 crc kubenswrapper[4805]: I1203 14:32:55.203920 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/527667db-e5ab-4fe2-89a5-d9110602f1d2-combined-ca-bundle\") pod \"nova-cell1-conductor-db-sync-42n9l\" (UID: \"527667db-e5ab-4fe2-89a5-d9110602f1d2\") " pod="openstack/nova-cell1-conductor-db-sync-42n9l" Dec 03 14:32:55 crc kubenswrapper[4805]: I1203 14:32:55.204107 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/527667db-e5ab-4fe2-89a5-d9110602f1d2-scripts\") pod \"nova-cell1-conductor-db-sync-42n9l\" (UID: \"527667db-e5ab-4fe2-89a5-d9110602f1d2\") " pod="openstack/nova-cell1-conductor-db-sync-42n9l" Dec 03 14:32:55 crc kubenswrapper[4805]: I1203 14:32:55.204312 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kmswm\" (UniqueName: \"kubernetes.io/projected/527667db-e5ab-4fe2-89a5-d9110602f1d2-kube-api-access-kmswm\") pod \"nova-cell1-conductor-db-sync-42n9l\" (UID: \"527667db-e5ab-4fe2-89a5-d9110602f1d2\") " pod="openstack/nova-cell1-conductor-db-sync-42n9l" Dec 03 14:32:55 crc kubenswrapper[4805]: I1203 14:32:55.204860 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-cell-mapping-bmkm9" event={"ID":"97cf9020-6417-4d08-9cdf-f38515b63d82","Type":"ContainerStarted","Data":"06bf9a3c07915aa35a107376c38bd43c3c95bc2c712639d4e214fe7f53144f80"} Dec 03 14:32:55 crc kubenswrapper[4805]: I1203 14:32:55.306146 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kmswm\" (UniqueName: \"kubernetes.io/projected/527667db-e5ab-4fe2-89a5-d9110602f1d2-kube-api-access-kmswm\") pod \"nova-cell1-conductor-db-sync-42n9l\" (UID: \"527667db-e5ab-4fe2-89a5-d9110602f1d2\") " pod="openstack/nova-cell1-conductor-db-sync-42n9l" Dec 03 14:32:55 crc kubenswrapper[4805]: I1203 14:32:55.306510 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/527667db-e5ab-4fe2-89a5-d9110602f1d2-config-data\") pod \"nova-cell1-conductor-db-sync-42n9l\" (UID: \"527667db-e5ab-4fe2-89a5-d9110602f1d2\") " pod="openstack/nova-cell1-conductor-db-sync-42n9l" Dec 03 14:32:55 crc kubenswrapper[4805]: I1203 14:32:55.306553 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/527667db-e5ab-4fe2-89a5-d9110602f1d2-combined-ca-bundle\") pod \"nova-cell1-conductor-db-sync-42n9l\" (UID: \"527667db-e5ab-4fe2-89a5-d9110602f1d2\") " pod="openstack/nova-cell1-conductor-db-sync-42n9l" Dec 03 14:32:55 crc kubenswrapper[4805]: I1203 14:32:55.306634 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/527667db-e5ab-4fe2-89a5-d9110602f1d2-scripts\") pod \"nova-cell1-conductor-db-sync-42n9l\" (UID: \"527667db-e5ab-4fe2-89a5-d9110602f1d2\") " pod="openstack/nova-cell1-conductor-db-sync-42n9l" Dec 03 14:32:55 crc kubenswrapper[4805]: I1203 14:32:55.313928 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/527667db-e5ab-4fe2-89a5-d9110602f1d2-scripts\") pod \"nova-cell1-conductor-db-sync-42n9l\" (UID: \"527667db-e5ab-4fe2-89a5-d9110602f1d2\") " pod="openstack/nova-cell1-conductor-db-sync-42n9l" Dec 03 14:32:55 crc kubenswrapper[4805]: I1203 14:32:55.314167 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/527667db-e5ab-4fe2-89a5-d9110602f1d2-config-data\") pod \"nova-cell1-conductor-db-sync-42n9l\" (UID: \"527667db-e5ab-4fe2-89a5-d9110602f1d2\") " pod="openstack/nova-cell1-conductor-db-sync-42n9l" Dec 03 14:32:55 crc kubenswrapper[4805]: I1203 14:32:55.316445 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/527667db-e5ab-4fe2-89a5-d9110602f1d2-combined-ca-bundle\") pod \"nova-cell1-conductor-db-sync-42n9l\" (UID: \"527667db-e5ab-4fe2-89a5-d9110602f1d2\") " pod="openstack/nova-cell1-conductor-db-sync-42n9l" Dec 03 14:32:55 crc kubenswrapper[4805]: I1203 14:32:55.329640 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kmswm\" (UniqueName: \"kubernetes.io/projected/527667db-e5ab-4fe2-89a5-d9110602f1d2-kube-api-access-kmswm\") pod \"nova-cell1-conductor-db-sync-42n9l\" (UID: \"527667db-e5ab-4fe2-89a5-d9110602f1d2\") " pod="openstack/nova-cell1-conductor-db-sync-42n9l" Dec 03 14:32:55 crc kubenswrapper[4805]: I1203 14:32:55.411331 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Dec 03 14:32:55 crc kubenswrapper[4805]: I1203 14:32:55.415595 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-db-sync-42n9l" Dec 03 14:32:55 crc kubenswrapper[4805]: I1203 14:32:55.528959 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Dec 03 14:32:55 crc kubenswrapper[4805]: W1203 14:32:55.551884 4805 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod5c64fe61_215b_47f5_aeae_98548dce4e81.slice/crio-d72b07f77a05cbd9ab86ebbef411cce283fcda627023cc7e42e1a422509f6a09 WatchSource:0}: Error finding container d72b07f77a05cbd9ab86ebbef411cce283fcda627023cc7e42e1a422509f6a09: Status 404 returned error can't find the container with id d72b07f77a05cbd9ab86ebbef411cce283fcda627023cc7e42e1a422509f6a09 Dec 03 14:32:55 crc kubenswrapper[4805]: I1203 14:32:55.578312 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Dec 03 14:32:55 crc kubenswrapper[4805]: I1203 14:32:55.676530 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5fbc4d444f-5vhgg"] Dec 03 14:32:55 crc kubenswrapper[4805]: I1203 14:32:55.947207 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-db-sync-42n9l"] Dec 03 14:32:55 crc kubenswrapper[4805]: W1203 14:32:55.971063 4805 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod527667db_e5ab_4fe2_89a5_d9110602f1d2.slice/crio-2db1823e1680280378ca4b9bb4cb18f9718ec3b957229683cd51f6789759e578 WatchSource:0}: Error finding container 2db1823e1680280378ca4b9bb4cb18f9718ec3b957229683cd51f6789759e578: Status 404 returned error can't find the container with id 2db1823e1680280378ca4b9bb4cb18f9718ec3b957229683cd51f6789759e578 Dec 03 14:32:56 crc kubenswrapper[4805]: I1203 14:32:56.220202 4805 generic.go:334] "Generic (PLEG): container finished" podID="e2ef61bb-5836-4986-88ae-f307ee375f0e" containerID="796364804b6d42eff3b0d9c02016068c2e46f8d1f2868771facca6b5efe60d07" exitCode=0 Dec 03 14:32:56 crc kubenswrapper[4805]: I1203 14:32:56.220258 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5fbc4d444f-5vhgg" event={"ID":"e2ef61bb-5836-4986-88ae-f307ee375f0e","Type":"ContainerDied","Data":"796364804b6d42eff3b0d9c02016068c2e46f8d1f2868771facca6b5efe60d07"} Dec 03 14:32:56 crc kubenswrapper[4805]: I1203 14:32:56.220494 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5fbc4d444f-5vhgg" event={"ID":"e2ef61bb-5836-4986-88ae-f307ee375f0e","Type":"ContainerStarted","Data":"6f57619b8ae32c72b2a066282ab29bc4917760e6fb6e13419cdd03cae40db7d2"} Dec 03 14:32:56 crc kubenswrapper[4805]: I1203 14:32:56.226148 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"08314cda-c0a3-4a26-bb3a-65cd724335a6","Type":"ContainerStarted","Data":"64b7f6d56431135b4076bcc6ecba0e1404bf67f2c6e6b137731012a9f7ea6dc4"} Dec 03 14:32:56 crc kubenswrapper[4805]: I1203 14:32:56.228892 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"5c64fe61-215b-47f5-aeae-98548dce4e81","Type":"ContainerStarted","Data":"d72b07f77a05cbd9ab86ebbef411cce283fcda627023cc7e42e1a422509f6a09"} Dec 03 14:32:56 crc kubenswrapper[4805]: I1203 14:32:56.231112 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-cell-mapping-bmkm9" event={"ID":"97cf9020-6417-4d08-9cdf-f38515b63d82","Type":"ContainerStarted","Data":"855ce4dd11138f5cb712a02c8377e47248650a1e2a6401e002ad634a4a558325"} Dec 03 14:32:56 crc kubenswrapper[4805]: I1203 14:32:56.236028 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"0ba4f686-dd10-475f-a65a-78041893d75d","Type":"ContainerStarted","Data":"20d8e8a0cf3cfe920107647005fd79426ba22e21aa2653799cb87d499639431d"} Dec 03 14:32:56 crc kubenswrapper[4805]: I1203 14:32:56.243471 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"660cf06e-323d-478f-aa70-a928724391c3","Type":"ContainerStarted","Data":"7d614030b5afc43de1d8596363c5c053b120ff5a827a802a64e1ff0b68c00ad5"} Dec 03 14:32:56 crc kubenswrapper[4805]: I1203 14:32:56.246960 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-db-sync-42n9l" event={"ID":"527667db-e5ab-4fe2-89a5-d9110602f1d2","Type":"ContainerStarted","Data":"3ebc3bcafd06fc724ece768dc68192d7ce17f5f76321fbde3f409727992f5ea4"} Dec 03 14:32:56 crc kubenswrapper[4805]: I1203 14:32:56.246994 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-db-sync-42n9l" event={"ID":"527667db-e5ab-4fe2-89a5-d9110602f1d2","Type":"ContainerStarted","Data":"2db1823e1680280378ca4b9bb4cb18f9718ec3b957229683cd51f6789759e578"} Dec 03 14:32:56 crc kubenswrapper[4805]: I1203 14:32:56.261733 4805 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell0-cell-mapping-bmkm9" podStartSLOduration=2.261600373 podStartE2EDuration="2.261600373s" podCreationTimestamp="2025-12-03 14:32:54 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 14:32:56.259815284 +0000 UTC m=+1405.922732207" watchObservedRunningTime="2025-12-03 14:32:56.261600373 +0000 UTC m=+1405.924517296" Dec 03 14:32:56 crc kubenswrapper[4805]: I1203 14:32:56.284450 4805 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-conductor-db-sync-42n9l" podStartSLOduration=1.284427178 podStartE2EDuration="1.284427178s" podCreationTimestamp="2025-12-03 14:32:55 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 14:32:56.27721374 +0000 UTC m=+1405.940130663" watchObservedRunningTime="2025-12-03 14:32:56.284427178 +0000 UTC m=+1405.947344101" Dec 03 14:32:58 crc kubenswrapper[4805]: I1203 14:32:58.060625 4805 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Dec 03 14:32:58 crc kubenswrapper[4805]: I1203 14:32:58.077152 4805 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Dec 03 14:32:59 crc kubenswrapper[4805]: I1203 14:32:59.286283 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"0ba4f686-dd10-475f-a65a-78041893d75d","Type":"ContainerStarted","Data":"5243aa51bbe15f29ac6238933f1d6db7ceba9c4f6359400eb45ce14c0ecf54a6"} Dec 03 14:32:59 crc kubenswrapper[4805]: I1203 14:32:59.288589 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"660cf06e-323d-478f-aa70-a928724391c3","Type":"ContainerStarted","Data":"d53e632a60c2b214a11cc34dc22d676d6f8ebe2c5767c6b29c796d3c073dff0a"} Dec 03 14:32:59 crc kubenswrapper[4805]: I1203 14:32:59.288611 4805 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-cell1-novncproxy-0" podUID="660cf06e-323d-478f-aa70-a928724391c3" containerName="nova-cell1-novncproxy-novncproxy" containerID="cri-o://d53e632a60c2b214a11cc34dc22d676d6f8ebe2c5767c6b29c796d3c073dff0a" gracePeriod=30 Dec 03 14:32:59 crc kubenswrapper[4805]: I1203 14:32:59.294446 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5fbc4d444f-5vhgg" event={"ID":"e2ef61bb-5836-4986-88ae-f307ee375f0e","Type":"ContainerStarted","Data":"a7d8422bf0b5f75574d94f6069d721089f2c5adb9c716b5ee35a9b1e43db5831"} Dec 03 14:32:59 crc kubenswrapper[4805]: I1203 14:32:59.294662 4805 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-5fbc4d444f-5vhgg" Dec 03 14:32:59 crc kubenswrapper[4805]: I1203 14:32:59.296807 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"08314cda-c0a3-4a26-bb3a-65cd724335a6","Type":"ContainerStarted","Data":"a1819f19bf4f039aeac63d0b7e68e3fb0716250eaee4e898284f30e5115c38e1"} Dec 03 14:32:59 crc kubenswrapper[4805]: I1203 14:32:59.296833 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"08314cda-c0a3-4a26-bb3a-65cd724335a6","Type":"ContainerStarted","Data":"a63cffd29ec3c44013d6c9cad06663729f09ed27ec35f6535fb35e381786dad2"} Dec 03 14:32:59 crc kubenswrapper[4805]: I1203 14:32:59.298890 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"5c64fe61-215b-47f5-aeae-98548dce4e81","Type":"ContainerStarted","Data":"2e0f517748123446673e4707811e4f69d028c5531cd4f0710a2a8ffeb838d522"} Dec 03 14:32:59 crc kubenswrapper[4805]: I1203 14:32:59.298913 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"5c64fe61-215b-47f5-aeae-98548dce4e81","Type":"ContainerStarted","Data":"11fc764facf20f5cce1f2f836536237da35bacf2ca59084754b1a7245b8d4b3a"} Dec 03 14:32:59 crc kubenswrapper[4805]: I1203 14:32:59.298998 4805 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="5c64fe61-215b-47f5-aeae-98548dce4e81" containerName="nova-metadata-log" containerID="cri-o://11fc764facf20f5cce1f2f836536237da35bacf2ca59084754b1a7245b8d4b3a" gracePeriod=30 Dec 03 14:32:59 crc kubenswrapper[4805]: I1203 14:32:59.299225 4805 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="5c64fe61-215b-47f5-aeae-98548dce4e81" containerName="nova-metadata-metadata" containerID="cri-o://2e0f517748123446673e4707811e4f69d028c5531cd4f0710a2a8ffeb838d522" gracePeriod=30 Dec 03 14:32:59 crc kubenswrapper[4805]: I1203 14:32:59.344477 4805 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-scheduler-0" podStartSLOduration=2.81238643 podStartE2EDuration="5.344456308s" podCreationTimestamp="2025-12-03 14:32:54 +0000 UTC" firstStartedPulling="2025-12-03 14:32:55.580356223 +0000 UTC m=+1405.243273136" lastFinishedPulling="2025-12-03 14:32:58.112426081 +0000 UTC m=+1407.775343014" observedRunningTime="2025-12-03 14:32:59.321395107 +0000 UTC m=+1408.984312030" watchObservedRunningTime="2025-12-03 14:32:59.344456308 +0000 UTC m=+1409.007373231" Dec 03 14:32:59 crc kubenswrapper[4805]: I1203 14:32:59.351878 4805 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-0" podStartSLOduration=2.448029436 podStartE2EDuration="5.351859141s" podCreationTimestamp="2025-12-03 14:32:54 +0000 UTC" firstStartedPulling="2025-12-03 14:32:55.206875539 +0000 UTC m=+1404.869792462" lastFinishedPulling="2025-12-03 14:32:58.110705244 +0000 UTC m=+1407.773622167" observedRunningTime="2025-12-03 14:32:59.346810053 +0000 UTC m=+1409.009726976" watchObservedRunningTime="2025-12-03 14:32:59.351859141 +0000 UTC m=+1409.014776064" Dec 03 14:32:59 crc kubenswrapper[4805]: I1203 14:32:59.367498 4805 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-metadata-0" podStartSLOduration=2.812182795 podStartE2EDuration="5.367475058s" podCreationTimestamp="2025-12-03 14:32:54 +0000 UTC" firstStartedPulling="2025-12-03 14:32:55.557097527 +0000 UTC m=+1405.220014450" lastFinishedPulling="2025-12-03 14:32:58.11238979 +0000 UTC m=+1407.775306713" observedRunningTime="2025-12-03 14:32:59.365668739 +0000 UTC m=+1409.028585662" watchObservedRunningTime="2025-12-03 14:32:59.367475058 +0000 UTC m=+1409.030391981" Dec 03 14:32:59 crc kubenswrapper[4805]: I1203 14:32:59.389464 4805 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-5fbc4d444f-5vhgg" podStartSLOduration=5.3894513 podStartE2EDuration="5.3894513s" podCreationTimestamp="2025-12-03 14:32:54 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 14:32:59.385970325 +0000 UTC m=+1409.048887238" watchObservedRunningTime="2025-12-03 14:32:59.3894513 +0000 UTC m=+1409.052368223" Dec 03 14:32:59 crc kubenswrapper[4805]: I1203 14:32:59.428054 4805 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-novncproxy-0" podStartSLOduration=2.763552543 podStartE2EDuration="5.428028326s" podCreationTimestamp="2025-12-03 14:32:54 +0000 UTC" firstStartedPulling="2025-12-03 14:32:55.447925858 +0000 UTC m=+1405.110842781" lastFinishedPulling="2025-12-03 14:32:58.112401641 +0000 UTC m=+1407.775318564" observedRunningTime="2025-12-03 14:32:59.421261391 +0000 UTC m=+1409.084178314" watchObservedRunningTime="2025-12-03 14:32:59.428028326 +0000 UTC m=+1409.090945259" Dec 03 14:32:59 crc kubenswrapper[4805]: I1203 14:32:59.842920 4805 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell1-novncproxy-0" Dec 03 14:32:59 crc kubenswrapper[4805]: I1203 14:32:59.863251 4805 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-scheduler-0" Dec 03 14:32:59 crc kubenswrapper[4805]: I1203 14:32:59.872184 4805 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Dec 03 14:32:59 crc kubenswrapper[4805]: I1203 14:32:59.872222 4805 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Dec 03 14:32:59 crc kubenswrapper[4805]: I1203 14:32:59.906896 4805 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 03 14:33:00 crc kubenswrapper[4805]: I1203 14:33:00.009101 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/5c64fe61-215b-47f5-aeae-98548dce4e81-logs\") pod \"5c64fe61-215b-47f5-aeae-98548dce4e81\" (UID: \"5c64fe61-215b-47f5-aeae-98548dce4e81\") " Dec 03 14:33:00 crc kubenswrapper[4805]: I1203 14:33:00.009192 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vnzwm\" (UniqueName: \"kubernetes.io/projected/5c64fe61-215b-47f5-aeae-98548dce4e81-kube-api-access-vnzwm\") pod \"5c64fe61-215b-47f5-aeae-98548dce4e81\" (UID: \"5c64fe61-215b-47f5-aeae-98548dce4e81\") " Dec 03 14:33:00 crc kubenswrapper[4805]: I1203 14:33:00.009277 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5c64fe61-215b-47f5-aeae-98548dce4e81-config-data\") pod \"5c64fe61-215b-47f5-aeae-98548dce4e81\" (UID: \"5c64fe61-215b-47f5-aeae-98548dce4e81\") " Dec 03 14:33:00 crc kubenswrapper[4805]: I1203 14:33:00.009382 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5c64fe61-215b-47f5-aeae-98548dce4e81-logs" (OuterVolumeSpecName: "logs") pod "5c64fe61-215b-47f5-aeae-98548dce4e81" (UID: "5c64fe61-215b-47f5-aeae-98548dce4e81"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 14:33:00 crc kubenswrapper[4805]: I1203 14:33:00.009454 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5c64fe61-215b-47f5-aeae-98548dce4e81-combined-ca-bundle\") pod \"5c64fe61-215b-47f5-aeae-98548dce4e81\" (UID: \"5c64fe61-215b-47f5-aeae-98548dce4e81\") " Dec 03 14:33:00 crc kubenswrapper[4805]: I1203 14:33:00.010015 4805 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/5c64fe61-215b-47f5-aeae-98548dce4e81-logs\") on node \"crc\" DevicePath \"\"" Dec 03 14:33:00 crc kubenswrapper[4805]: I1203 14:33:00.015069 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5c64fe61-215b-47f5-aeae-98548dce4e81-kube-api-access-vnzwm" (OuterVolumeSpecName: "kube-api-access-vnzwm") pod "5c64fe61-215b-47f5-aeae-98548dce4e81" (UID: "5c64fe61-215b-47f5-aeae-98548dce4e81"). InnerVolumeSpecName "kube-api-access-vnzwm". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 14:33:00 crc kubenswrapper[4805]: I1203 14:33:00.037166 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5c64fe61-215b-47f5-aeae-98548dce4e81-config-data" (OuterVolumeSpecName: "config-data") pod "5c64fe61-215b-47f5-aeae-98548dce4e81" (UID: "5c64fe61-215b-47f5-aeae-98548dce4e81"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 14:33:00 crc kubenswrapper[4805]: I1203 14:33:00.038542 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5c64fe61-215b-47f5-aeae-98548dce4e81-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "5c64fe61-215b-47f5-aeae-98548dce4e81" (UID: "5c64fe61-215b-47f5-aeae-98548dce4e81"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 14:33:00 crc kubenswrapper[4805]: I1203 14:33:00.111795 4805 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5c64fe61-215b-47f5-aeae-98548dce4e81-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 14:33:00 crc kubenswrapper[4805]: I1203 14:33:00.111829 4805 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vnzwm\" (UniqueName: \"kubernetes.io/projected/5c64fe61-215b-47f5-aeae-98548dce4e81-kube-api-access-vnzwm\") on node \"crc\" DevicePath \"\"" Dec 03 14:33:00 crc kubenswrapper[4805]: I1203 14:33:00.111853 4805 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5c64fe61-215b-47f5-aeae-98548dce4e81-config-data\") on node \"crc\" DevicePath \"\"" Dec 03 14:33:00 crc kubenswrapper[4805]: I1203 14:33:00.311508 4805 generic.go:334] "Generic (PLEG): container finished" podID="5c64fe61-215b-47f5-aeae-98548dce4e81" containerID="2e0f517748123446673e4707811e4f69d028c5531cd4f0710a2a8ffeb838d522" exitCode=0 Dec 03 14:33:00 crc kubenswrapper[4805]: I1203 14:33:00.311539 4805 generic.go:334] "Generic (PLEG): container finished" podID="5c64fe61-215b-47f5-aeae-98548dce4e81" containerID="11fc764facf20f5cce1f2f836536237da35bacf2ca59084754b1a7245b8d4b3a" exitCode=143 Dec 03 14:33:00 crc kubenswrapper[4805]: I1203 14:33:00.311664 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"5c64fe61-215b-47f5-aeae-98548dce4e81","Type":"ContainerDied","Data":"2e0f517748123446673e4707811e4f69d028c5531cd4f0710a2a8ffeb838d522"} Dec 03 14:33:00 crc kubenswrapper[4805]: I1203 14:33:00.311718 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"5c64fe61-215b-47f5-aeae-98548dce4e81","Type":"ContainerDied","Data":"11fc764facf20f5cce1f2f836536237da35bacf2ca59084754b1a7245b8d4b3a"} Dec 03 14:33:00 crc kubenswrapper[4805]: I1203 14:33:00.311733 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"5c64fe61-215b-47f5-aeae-98548dce4e81","Type":"ContainerDied","Data":"d72b07f77a05cbd9ab86ebbef411cce283fcda627023cc7e42e1a422509f6a09"} Dec 03 14:33:00 crc kubenswrapper[4805]: I1203 14:33:00.311754 4805 scope.go:117] "RemoveContainer" containerID="2e0f517748123446673e4707811e4f69d028c5531cd4f0710a2a8ffeb838d522" Dec 03 14:33:00 crc kubenswrapper[4805]: I1203 14:33:00.311617 4805 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 03 14:33:00 crc kubenswrapper[4805]: I1203 14:33:00.338124 4805 scope.go:117] "RemoveContainer" containerID="11fc764facf20f5cce1f2f836536237da35bacf2ca59084754b1a7245b8d4b3a" Dec 03 14:33:00 crc kubenswrapper[4805]: I1203 14:33:00.355109 4805 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Dec 03 14:33:00 crc kubenswrapper[4805]: I1203 14:33:00.362897 4805 scope.go:117] "RemoveContainer" containerID="2e0f517748123446673e4707811e4f69d028c5531cd4f0710a2a8ffeb838d522" Dec 03 14:33:00 crc kubenswrapper[4805]: E1203 14:33:00.363359 4805 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2e0f517748123446673e4707811e4f69d028c5531cd4f0710a2a8ffeb838d522\": container with ID starting with 2e0f517748123446673e4707811e4f69d028c5531cd4f0710a2a8ffeb838d522 not found: ID does not exist" containerID="2e0f517748123446673e4707811e4f69d028c5531cd4f0710a2a8ffeb838d522" Dec 03 14:33:00 crc kubenswrapper[4805]: I1203 14:33:00.363401 4805 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2e0f517748123446673e4707811e4f69d028c5531cd4f0710a2a8ffeb838d522"} err="failed to get container status \"2e0f517748123446673e4707811e4f69d028c5531cd4f0710a2a8ffeb838d522\": rpc error: code = NotFound desc = could not find container \"2e0f517748123446673e4707811e4f69d028c5531cd4f0710a2a8ffeb838d522\": container with ID starting with 2e0f517748123446673e4707811e4f69d028c5531cd4f0710a2a8ffeb838d522 not found: ID does not exist" Dec 03 14:33:00 crc kubenswrapper[4805]: I1203 14:33:00.363429 4805 scope.go:117] "RemoveContainer" containerID="11fc764facf20f5cce1f2f836536237da35bacf2ca59084754b1a7245b8d4b3a" Dec 03 14:33:00 crc kubenswrapper[4805]: E1203 14:33:00.364924 4805 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"11fc764facf20f5cce1f2f836536237da35bacf2ca59084754b1a7245b8d4b3a\": container with ID starting with 11fc764facf20f5cce1f2f836536237da35bacf2ca59084754b1a7245b8d4b3a not found: ID does not exist" containerID="11fc764facf20f5cce1f2f836536237da35bacf2ca59084754b1a7245b8d4b3a" Dec 03 14:33:00 crc kubenswrapper[4805]: I1203 14:33:00.364957 4805 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"11fc764facf20f5cce1f2f836536237da35bacf2ca59084754b1a7245b8d4b3a"} err="failed to get container status \"11fc764facf20f5cce1f2f836536237da35bacf2ca59084754b1a7245b8d4b3a\": rpc error: code = NotFound desc = could not find container \"11fc764facf20f5cce1f2f836536237da35bacf2ca59084754b1a7245b8d4b3a\": container with ID starting with 11fc764facf20f5cce1f2f836536237da35bacf2ca59084754b1a7245b8d4b3a not found: ID does not exist" Dec 03 14:33:00 crc kubenswrapper[4805]: I1203 14:33:00.364977 4805 scope.go:117] "RemoveContainer" containerID="2e0f517748123446673e4707811e4f69d028c5531cd4f0710a2a8ffeb838d522" Dec 03 14:33:00 crc kubenswrapper[4805]: I1203 14:33:00.365372 4805 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2e0f517748123446673e4707811e4f69d028c5531cd4f0710a2a8ffeb838d522"} err="failed to get container status \"2e0f517748123446673e4707811e4f69d028c5531cd4f0710a2a8ffeb838d522\": rpc error: code = NotFound desc = could not find container \"2e0f517748123446673e4707811e4f69d028c5531cd4f0710a2a8ffeb838d522\": container with ID starting with 2e0f517748123446673e4707811e4f69d028c5531cd4f0710a2a8ffeb838d522 not found: ID does not exist" Dec 03 14:33:00 crc kubenswrapper[4805]: I1203 14:33:00.365426 4805 scope.go:117] "RemoveContainer" containerID="11fc764facf20f5cce1f2f836536237da35bacf2ca59084754b1a7245b8d4b3a" Dec 03 14:33:00 crc kubenswrapper[4805]: I1203 14:33:00.365773 4805 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"11fc764facf20f5cce1f2f836536237da35bacf2ca59084754b1a7245b8d4b3a"} err="failed to get container status \"11fc764facf20f5cce1f2f836536237da35bacf2ca59084754b1a7245b8d4b3a\": rpc error: code = NotFound desc = could not find container \"11fc764facf20f5cce1f2f836536237da35bacf2ca59084754b1a7245b8d4b3a\": container with ID starting with 11fc764facf20f5cce1f2f836536237da35bacf2ca59084754b1a7245b8d4b3a not found: ID does not exist" Dec 03 14:33:00 crc kubenswrapper[4805]: I1203 14:33:00.383935 4805 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-metadata-0"] Dec 03 14:33:00 crc kubenswrapper[4805]: I1203 14:33:00.413932 4805 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-metadata-0"] Dec 03 14:33:00 crc kubenswrapper[4805]: E1203 14:33:00.414392 4805 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5c64fe61-215b-47f5-aeae-98548dce4e81" containerName="nova-metadata-log" Dec 03 14:33:00 crc kubenswrapper[4805]: I1203 14:33:00.414404 4805 state_mem.go:107] "Deleted CPUSet assignment" podUID="5c64fe61-215b-47f5-aeae-98548dce4e81" containerName="nova-metadata-log" Dec 03 14:33:00 crc kubenswrapper[4805]: E1203 14:33:00.414414 4805 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5c64fe61-215b-47f5-aeae-98548dce4e81" containerName="nova-metadata-metadata" Dec 03 14:33:00 crc kubenswrapper[4805]: I1203 14:33:00.414420 4805 state_mem.go:107] "Deleted CPUSet assignment" podUID="5c64fe61-215b-47f5-aeae-98548dce4e81" containerName="nova-metadata-metadata" Dec 03 14:33:00 crc kubenswrapper[4805]: I1203 14:33:00.414588 4805 memory_manager.go:354] "RemoveStaleState removing state" podUID="5c64fe61-215b-47f5-aeae-98548dce4e81" containerName="nova-metadata-metadata" Dec 03 14:33:00 crc kubenswrapper[4805]: I1203 14:33:00.414613 4805 memory_manager.go:354] "RemoveStaleState removing state" podUID="5c64fe61-215b-47f5-aeae-98548dce4e81" containerName="nova-metadata-log" Dec 03 14:33:00 crc kubenswrapper[4805]: I1203 14:33:00.415568 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 03 14:33:00 crc kubenswrapper[4805]: I1203 14:33:00.421090 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-config-data" Dec 03 14:33:00 crc kubenswrapper[4805]: I1203 14:33:00.421457 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-metadata-internal-svc" Dec 03 14:33:00 crc kubenswrapper[4805]: I1203 14:33:00.439785 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Dec 03 14:33:00 crc kubenswrapper[4805]: I1203 14:33:00.520739 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hdhcb\" (UniqueName: \"kubernetes.io/projected/6eb3690f-3101-448b-999c-ae7c4216d1f7-kube-api-access-hdhcb\") pod \"nova-metadata-0\" (UID: \"6eb3690f-3101-448b-999c-ae7c4216d1f7\") " pod="openstack/nova-metadata-0" Dec 03 14:33:00 crc kubenswrapper[4805]: I1203 14:33:00.520872 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/6eb3690f-3101-448b-999c-ae7c4216d1f7-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"6eb3690f-3101-448b-999c-ae7c4216d1f7\") " pod="openstack/nova-metadata-0" Dec 03 14:33:00 crc kubenswrapper[4805]: I1203 14:33:00.520927 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6eb3690f-3101-448b-999c-ae7c4216d1f7-config-data\") pod \"nova-metadata-0\" (UID: \"6eb3690f-3101-448b-999c-ae7c4216d1f7\") " pod="openstack/nova-metadata-0" Dec 03 14:33:00 crc kubenswrapper[4805]: I1203 14:33:00.520967 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6eb3690f-3101-448b-999c-ae7c4216d1f7-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"6eb3690f-3101-448b-999c-ae7c4216d1f7\") " pod="openstack/nova-metadata-0" Dec 03 14:33:00 crc kubenswrapper[4805]: I1203 14:33:00.520999 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/6eb3690f-3101-448b-999c-ae7c4216d1f7-logs\") pod \"nova-metadata-0\" (UID: \"6eb3690f-3101-448b-999c-ae7c4216d1f7\") " pod="openstack/nova-metadata-0" Dec 03 14:33:00 crc kubenswrapper[4805]: I1203 14:33:00.623590 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/6eb3690f-3101-448b-999c-ae7c4216d1f7-logs\") pod \"nova-metadata-0\" (UID: \"6eb3690f-3101-448b-999c-ae7c4216d1f7\") " pod="openstack/nova-metadata-0" Dec 03 14:33:00 crc kubenswrapper[4805]: I1203 14:33:00.623790 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hdhcb\" (UniqueName: \"kubernetes.io/projected/6eb3690f-3101-448b-999c-ae7c4216d1f7-kube-api-access-hdhcb\") pod \"nova-metadata-0\" (UID: \"6eb3690f-3101-448b-999c-ae7c4216d1f7\") " pod="openstack/nova-metadata-0" Dec 03 14:33:00 crc kubenswrapper[4805]: I1203 14:33:00.623892 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/6eb3690f-3101-448b-999c-ae7c4216d1f7-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"6eb3690f-3101-448b-999c-ae7c4216d1f7\") " pod="openstack/nova-metadata-0" Dec 03 14:33:00 crc kubenswrapper[4805]: I1203 14:33:00.623973 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/6eb3690f-3101-448b-999c-ae7c4216d1f7-logs\") pod \"nova-metadata-0\" (UID: \"6eb3690f-3101-448b-999c-ae7c4216d1f7\") " pod="openstack/nova-metadata-0" Dec 03 14:33:00 crc kubenswrapper[4805]: I1203 14:33:00.623991 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6eb3690f-3101-448b-999c-ae7c4216d1f7-config-data\") pod \"nova-metadata-0\" (UID: \"6eb3690f-3101-448b-999c-ae7c4216d1f7\") " pod="openstack/nova-metadata-0" Dec 03 14:33:00 crc kubenswrapper[4805]: I1203 14:33:00.624088 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6eb3690f-3101-448b-999c-ae7c4216d1f7-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"6eb3690f-3101-448b-999c-ae7c4216d1f7\") " pod="openstack/nova-metadata-0" Dec 03 14:33:00 crc kubenswrapper[4805]: I1203 14:33:00.633428 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6eb3690f-3101-448b-999c-ae7c4216d1f7-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"6eb3690f-3101-448b-999c-ae7c4216d1f7\") " pod="openstack/nova-metadata-0" Dec 03 14:33:00 crc kubenswrapper[4805]: I1203 14:33:00.633447 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/6eb3690f-3101-448b-999c-ae7c4216d1f7-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"6eb3690f-3101-448b-999c-ae7c4216d1f7\") " pod="openstack/nova-metadata-0" Dec 03 14:33:00 crc kubenswrapper[4805]: I1203 14:33:00.634080 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6eb3690f-3101-448b-999c-ae7c4216d1f7-config-data\") pod \"nova-metadata-0\" (UID: \"6eb3690f-3101-448b-999c-ae7c4216d1f7\") " pod="openstack/nova-metadata-0" Dec 03 14:33:00 crc kubenswrapper[4805]: I1203 14:33:00.650541 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hdhcb\" (UniqueName: \"kubernetes.io/projected/6eb3690f-3101-448b-999c-ae7c4216d1f7-kube-api-access-hdhcb\") pod \"nova-metadata-0\" (UID: \"6eb3690f-3101-448b-999c-ae7c4216d1f7\") " pod="openstack/nova-metadata-0" Dec 03 14:33:00 crc kubenswrapper[4805]: I1203 14:33:00.713506 4805 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5c64fe61-215b-47f5-aeae-98548dce4e81" path="/var/lib/kubelet/pods/5c64fe61-215b-47f5-aeae-98548dce4e81/volumes" Dec 03 14:33:00 crc kubenswrapper[4805]: I1203 14:33:00.739709 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 03 14:33:01 crc kubenswrapper[4805]: W1203 14:33:01.221243 4805 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod6eb3690f_3101_448b_999c_ae7c4216d1f7.slice/crio-59fcc66af346a6b242e1166651e94740e190fab0bb92d23d848feec9917755c3 WatchSource:0}: Error finding container 59fcc66af346a6b242e1166651e94740e190fab0bb92d23d848feec9917755c3: Status 404 returned error can't find the container with id 59fcc66af346a6b242e1166651e94740e190fab0bb92d23d848feec9917755c3 Dec 03 14:33:01 crc kubenswrapper[4805]: I1203 14:33:01.228282 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Dec 03 14:33:01 crc kubenswrapper[4805]: I1203 14:33:01.320439 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"6eb3690f-3101-448b-999c-ae7c4216d1f7","Type":"ContainerStarted","Data":"59fcc66af346a6b242e1166651e94740e190fab0bb92d23d848feec9917755c3"} Dec 03 14:33:02 crc kubenswrapper[4805]: I1203 14:33:02.334150 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"6eb3690f-3101-448b-999c-ae7c4216d1f7","Type":"ContainerStarted","Data":"f4d56fc727319480ee7c35d1447da5b62c144739673fc1cb1cfdfef2e78bd055"} Dec 03 14:33:02 crc kubenswrapper[4805]: I1203 14:33:02.334450 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"6eb3690f-3101-448b-999c-ae7c4216d1f7","Type":"ContainerStarted","Data":"45927200e3a8a365e2d0d98f280946c74d9779da0c568414671ab2c5f00ae669"} Dec 03 14:33:02 crc kubenswrapper[4805]: I1203 14:33:02.364512 4805 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-metadata-0" podStartSLOduration=2.364495693 podStartE2EDuration="2.364495693s" podCreationTimestamp="2025-12-03 14:33:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 14:33:02.363782344 +0000 UTC m=+1412.026699287" watchObservedRunningTime="2025-12-03 14:33:02.364495693 +0000 UTC m=+1412.027412616" Dec 03 14:33:03 crc kubenswrapper[4805]: I1203 14:33:03.344289 4805 generic.go:334] "Generic (PLEG): container finished" podID="97cf9020-6417-4d08-9cdf-f38515b63d82" containerID="855ce4dd11138f5cb712a02c8377e47248650a1e2a6401e002ad634a4a558325" exitCode=0 Dec 03 14:33:03 crc kubenswrapper[4805]: I1203 14:33:03.344410 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-cell-mapping-bmkm9" event={"ID":"97cf9020-6417-4d08-9cdf-f38515b63d82","Type":"ContainerDied","Data":"855ce4dd11138f5cb712a02c8377e47248650a1e2a6401e002ad634a4a558325"} Dec 03 14:33:03 crc kubenswrapper[4805]: I1203 14:33:03.346812 4805 generic.go:334] "Generic (PLEG): container finished" podID="527667db-e5ab-4fe2-89a5-d9110602f1d2" containerID="3ebc3bcafd06fc724ece768dc68192d7ce17f5f76321fbde3f409727992f5ea4" exitCode=0 Dec 03 14:33:03 crc kubenswrapper[4805]: I1203 14:33:03.347883 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-db-sync-42n9l" event={"ID":"527667db-e5ab-4fe2-89a5-d9110602f1d2","Type":"ContainerDied","Data":"3ebc3bcafd06fc724ece768dc68192d7ce17f5f76321fbde3f409727992f5ea4"} Dec 03 14:33:04 crc kubenswrapper[4805]: I1203 14:33:04.591740 4805 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Dec 03 14:33:04 crc kubenswrapper[4805]: I1203 14:33:04.592163 4805 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Dec 03 14:33:04 crc kubenswrapper[4805]: I1203 14:33:04.790997 4805 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-db-sync-42n9l" Dec 03 14:33:04 crc kubenswrapper[4805]: I1203 14:33:04.798998 4805 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-cell-mapping-bmkm9" Dec 03 14:33:04 crc kubenswrapper[4805]: I1203 14:33:04.862825 4805 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-scheduler-0" Dec 03 14:33:04 crc kubenswrapper[4805]: I1203 14:33:04.892555 4805 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-scheduler-0" Dec 03 14:33:04 crc kubenswrapper[4805]: I1203 14:33:04.904334 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/527667db-e5ab-4fe2-89a5-d9110602f1d2-combined-ca-bundle\") pod \"527667db-e5ab-4fe2-89a5-d9110602f1d2\" (UID: \"527667db-e5ab-4fe2-89a5-d9110602f1d2\") " Dec 03 14:33:04 crc kubenswrapper[4805]: I1203 14:33:04.904373 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/97cf9020-6417-4d08-9cdf-f38515b63d82-combined-ca-bundle\") pod \"97cf9020-6417-4d08-9cdf-f38515b63d82\" (UID: \"97cf9020-6417-4d08-9cdf-f38515b63d82\") " Dec 03 14:33:04 crc kubenswrapper[4805]: I1203 14:33:04.904406 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jsxfm\" (UniqueName: \"kubernetes.io/projected/97cf9020-6417-4d08-9cdf-f38515b63d82-kube-api-access-jsxfm\") pod \"97cf9020-6417-4d08-9cdf-f38515b63d82\" (UID: \"97cf9020-6417-4d08-9cdf-f38515b63d82\") " Dec 03 14:33:04 crc kubenswrapper[4805]: I1203 14:33:04.904501 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/97cf9020-6417-4d08-9cdf-f38515b63d82-scripts\") pod \"97cf9020-6417-4d08-9cdf-f38515b63d82\" (UID: \"97cf9020-6417-4d08-9cdf-f38515b63d82\") " Dec 03 14:33:04 crc kubenswrapper[4805]: I1203 14:33:04.904549 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/527667db-e5ab-4fe2-89a5-d9110602f1d2-scripts\") pod \"527667db-e5ab-4fe2-89a5-d9110602f1d2\" (UID: \"527667db-e5ab-4fe2-89a5-d9110602f1d2\") " Dec 03 14:33:04 crc kubenswrapper[4805]: I1203 14:33:04.904592 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kmswm\" (UniqueName: \"kubernetes.io/projected/527667db-e5ab-4fe2-89a5-d9110602f1d2-kube-api-access-kmswm\") pod \"527667db-e5ab-4fe2-89a5-d9110602f1d2\" (UID: \"527667db-e5ab-4fe2-89a5-d9110602f1d2\") " Dec 03 14:33:04 crc kubenswrapper[4805]: I1203 14:33:04.904657 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/97cf9020-6417-4d08-9cdf-f38515b63d82-config-data\") pod \"97cf9020-6417-4d08-9cdf-f38515b63d82\" (UID: \"97cf9020-6417-4d08-9cdf-f38515b63d82\") " Dec 03 14:33:04 crc kubenswrapper[4805]: I1203 14:33:04.904702 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/527667db-e5ab-4fe2-89a5-d9110602f1d2-config-data\") pod \"527667db-e5ab-4fe2-89a5-d9110602f1d2\" (UID: \"527667db-e5ab-4fe2-89a5-d9110602f1d2\") " Dec 03 14:33:04 crc kubenswrapper[4805]: I1203 14:33:04.920553 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/527667db-e5ab-4fe2-89a5-d9110602f1d2-kube-api-access-kmswm" (OuterVolumeSpecName: "kube-api-access-kmswm") pod "527667db-e5ab-4fe2-89a5-d9110602f1d2" (UID: "527667db-e5ab-4fe2-89a5-d9110602f1d2"). InnerVolumeSpecName "kube-api-access-kmswm". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 14:33:04 crc kubenswrapper[4805]: I1203 14:33:04.922023 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/97cf9020-6417-4d08-9cdf-f38515b63d82-kube-api-access-jsxfm" (OuterVolumeSpecName: "kube-api-access-jsxfm") pod "97cf9020-6417-4d08-9cdf-f38515b63d82" (UID: "97cf9020-6417-4d08-9cdf-f38515b63d82"). InnerVolumeSpecName "kube-api-access-jsxfm". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 14:33:04 crc kubenswrapper[4805]: I1203 14:33:04.926865 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/527667db-e5ab-4fe2-89a5-d9110602f1d2-scripts" (OuterVolumeSpecName: "scripts") pod "527667db-e5ab-4fe2-89a5-d9110602f1d2" (UID: "527667db-e5ab-4fe2-89a5-d9110602f1d2"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 14:33:04 crc kubenswrapper[4805]: I1203 14:33:04.928023 4805 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-5fbc4d444f-5vhgg" Dec 03 14:33:04 crc kubenswrapper[4805]: I1203 14:33:04.962283 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/97cf9020-6417-4d08-9cdf-f38515b63d82-scripts" (OuterVolumeSpecName: "scripts") pod "97cf9020-6417-4d08-9cdf-f38515b63d82" (UID: "97cf9020-6417-4d08-9cdf-f38515b63d82"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 14:33:05 crc kubenswrapper[4805]: I1203 14:33:05.002010 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/527667db-e5ab-4fe2-89a5-d9110602f1d2-config-data" (OuterVolumeSpecName: "config-data") pod "527667db-e5ab-4fe2-89a5-d9110602f1d2" (UID: "527667db-e5ab-4fe2-89a5-d9110602f1d2"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 14:33:05 crc kubenswrapper[4805]: I1203 14:33:05.009663 4805 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/97cf9020-6417-4d08-9cdf-f38515b63d82-scripts\") on node \"crc\" DevicePath \"\"" Dec 03 14:33:05 crc kubenswrapper[4805]: I1203 14:33:05.009804 4805 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/527667db-e5ab-4fe2-89a5-d9110602f1d2-scripts\") on node \"crc\" DevicePath \"\"" Dec 03 14:33:05 crc kubenswrapper[4805]: I1203 14:33:05.009913 4805 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kmswm\" (UniqueName: \"kubernetes.io/projected/527667db-e5ab-4fe2-89a5-d9110602f1d2-kube-api-access-kmswm\") on node \"crc\" DevicePath \"\"" Dec 03 14:33:05 crc kubenswrapper[4805]: I1203 14:33:05.010006 4805 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/527667db-e5ab-4fe2-89a5-d9110602f1d2-config-data\") on node \"crc\" DevicePath \"\"" Dec 03 14:33:05 crc kubenswrapper[4805]: I1203 14:33:05.010090 4805 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jsxfm\" (UniqueName: \"kubernetes.io/projected/97cf9020-6417-4d08-9cdf-f38515b63d82-kube-api-access-jsxfm\") on node \"crc\" DevicePath \"\"" Dec 03 14:33:05 crc kubenswrapper[4805]: I1203 14:33:05.016824 4805 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-f6bc4c6c9-x5tdg"] Dec 03 14:33:05 crc kubenswrapper[4805]: I1203 14:33:05.017076 4805 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-f6bc4c6c9-x5tdg" podUID="4994372a-ac42-4661-aed6-734070bee008" containerName="dnsmasq-dns" containerID="cri-o://fda7f3c053dd3dd390f549cc83ac2bcd36e8561e014152b04f3277f127e4ceef" gracePeriod=10 Dec 03 14:33:05 crc kubenswrapper[4805]: I1203 14:33:05.027312 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/97cf9020-6417-4d08-9cdf-f38515b63d82-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "97cf9020-6417-4d08-9cdf-f38515b63d82" (UID: "97cf9020-6417-4d08-9cdf-f38515b63d82"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 14:33:05 crc kubenswrapper[4805]: I1203 14:33:05.032051 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/97cf9020-6417-4d08-9cdf-f38515b63d82-config-data" (OuterVolumeSpecName: "config-data") pod "97cf9020-6417-4d08-9cdf-f38515b63d82" (UID: "97cf9020-6417-4d08-9cdf-f38515b63d82"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 14:33:05 crc kubenswrapper[4805]: I1203 14:33:05.051845 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/527667db-e5ab-4fe2-89a5-d9110602f1d2-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "527667db-e5ab-4fe2-89a5-d9110602f1d2" (UID: "527667db-e5ab-4fe2-89a5-d9110602f1d2"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 14:33:05 crc kubenswrapper[4805]: I1203 14:33:05.111646 4805 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/97cf9020-6417-4d08-9cdf-f38515b63d82-config-data\") on node \"crc\" DevicePath \"\"" Dec 03 14:33:05 crc kubenswrapper[4805]: I1203 14:33:05.111686 4805 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/527667db-e5ab-4fe2-89a5-d9110602f1d2-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 14:33:05 crc kubenswrapper[4805]: I1203 14:33:05.111702 4805 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/97cf9020-6417-4d08-9cdf-f38515b63d82-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 14:33:05 crc kubenswrapper[4805]: I1203 14:33:05.379315 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-cell-mapping-bmkm9" event={"ID":"97cf9020-6417-4d08-9cdf-f38515b63d82","Type":"ContainerDied","Data":"06bf9a3c07915aa35a107376c38bd43c3c95bc2c712639d4e214fe7f53144f80"} Dec 03 14:33:05 crc kubenswrapper[4805]: I1203 14:33:05.379360 4805 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="06bf9a3c07915aa35a107376c38bd43c3c95bc2c712639d4e214fe7f53144f80" Dec 03 14:33:05 crc kubenswrapper[4805]: I1203 14:33:05.379489 4805 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-cell-mapping-bmkm9" Dec 03 14:33:05 crc kubenswrapper[4805]: I1203 14:33:05.392238 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-db-sync-42n9l" event={"ID":"527667db-e5ab-4fe2-89a5-d9110602f1d2","Type":"ContainerDied","Data":"2db1823e1680280378ca4b9bb4cb18f9718ec3b957229683cd51f6789759e578"} Dec 03 14:33:05 crc kubenswrapper[4805]: I1203 14:33:05.392292 4805 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="2db1823e1680280378ca4b9bb4cb18f9718ec3b957229683cd51f6789759e578" Dec 03 14:33:05 crc kubenswrapper[4805]: I1203 14:33:05.392402 4805 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-db-sync-42n9l" Dec 03 14:33:05 crc kubenswrapper[4805]: I1203 14:33:05.449474 4805 generic.go:334] "Generic (PLEG): container finished" podID="4994372a-ac42-4661-aed6-734070bee008" containerID="fda7f3c053dd3dd390f549cc83ac2bcd36e8561e014152b04f3277f127e4ceef" exitCode=0 Dec 03 14:33:05 crc kubenswrapper[4805]: I1203 14:33:05.449963 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-f6bc4c6c9-x5tdg" event={"ID":"4994372a-ac42-4661-aed6-734070bee008","Type":"ContainerDied","Data":"fda7f3c053dd3dd390f549cc83ac2bcd36e8561e014152b04f3277f127e4ceef"} Dec 03 14:33:05 crc kubenswrapper[4805]: I1203 14:33:05.539771 4805 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-f6bc4c6c9-x5tdg" Dec 03 14:33:05 crc kubenswrapper[4805]: I1203 14:33:05.527299 4805 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-conductor-0"] Dec 03 14:33:05 crc kubenswrapper[4805]: E1203 14:33:05.553748 4805 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="97cf9020-6417-4d08-9cdf-f38515b63d82" containerName="nova-manage" Dec 03 14:33:05 crc kubenswrapper[4805]: I1203 14:33:05.553775 4805 state_mem.go:107] "Deleted CPUSet assignment" podUID="97cf9020-6417-4d08-9cdf-f38515b63d82" containerName="nova-manage" Dec 03 14:33:05 crc kubenswrapper[4805]: E1203 14:33:05.553788 4805 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4994372a-ac42-4661-aed6-734070bee008" containerName="init" Dec 03 14:33:05 crc kubenswrapper[4805]: I1203 14:33:05.553797 4805 state_mem.go:107] "Deleted CPUSet assignment" podUID="4994372a-ac42-4661-aed6-734070bee008" containerName="init" Dec 03 14:33:05 crc kubenswrapper[4805]: E1203 14:33:05.553827 4805 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="527667db-e5ab-4fe2-89a5-d9110602f1d2" containerName="nova-cell1-conductor-db-sync" Dec 03 14:33:05 crc kubenswrapper[4805]: I1203 14:33:05.553864 4805 state_mem.go:107] "Deleted CPUSet assignment" podUID="527667db-e5ab-4fe2-89a5-d9110602f1d2" containerName="nova-cell1-conductor-db-sync" Dec 03 14:33:05 crc kubenswrapper[4805]: E1203 14:33:05.553891 4805 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4994372a-ac42-4661-aed6-734070bee008" containerName="dnsmasq-dns" Dec 03 14:33:05 crc kubenswrapper[4805]: I1203 14:33:05.553899 4805 state_mem.go:107] "Deleted CPUSet assignment" podUID="4994372a-ac42-4661-aed6-734070bee008" containerName="dnsmasq-dns" Dec 03 14:33:05 crc kubenswrapper[4805]: I1203 14:33:05.554107 4805 memory_manager.go:354] "RemoveStaleState removing state" podUID="527667db-e5ab-4fe2-89a5-d9110602f1d2" containerName="nova-cell1-conductor-db-sync" Dec 03 14:33:05 crc kubenswrapper[4805]: I1203 14:33:05.554128 4805 memory_manager.go:354] "RemoveStaleState removing state" podUID="4994372a-ac42-4661-aed6-734070bee008" containerName="dnsmasq-dns" Dec 03 14:33:05 crc kubenswrapper[4805]: I1203 14:33:05.554139 4805 memory_manager.go:354] "RemoveStaleState removing state" podUID="97cf9020-6417-4d08-9cdf-f38515b63d82" containerName="nova-manage" Dec 03 14:33:05 crc kubenswrapper[4805]: I1203 14:33:05.554711 4805 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-scheduler-0" Dec 03 14:33:05 crc kubenswrapper[4805]: I1203 14:33:05.554734 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-0"] Dec 03 14:33:05 crc kubenswrapper[4805]: I1203 14:33:05.554808 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-0" Dec 03 14:33:05 crc kubenswrapper[4805]: I1203 14:33:05.563196 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-conductor-config-data" Dec 03 14:33:05 crc kubenswrapper[4805]: I1203 14:33:05.604865 4805 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Dec 03 14:33:05 crc kubenswrapper[4805]: I1203 14:33:05.605086 4805 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="08314cda-c0a3-4a26-bb3a-65cd724335a6" containerName="nova-api-log" containerID="cri-o://a63cffd29ec3c44013d6c9cad06663729f09ed27ec35f6535fb35e381786dad2" gracePeriod=30 Dec 03 14:33:05 crc kubenswrapper[4805]: I1203 14:33:05.605202 4805 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="08314cda-c0a3-4a26-bb3a-65cd724335a6" containerName="nova-api-api" containerID="cri-o://a1819f19bf4f039aeac63d0b7e68e3fb0716250eaee4e898284f30e5115c38e1" gracePeriod=30 Dec 03 14:33:05 crc kubenswrapper[4805]: I1203 14:33:05.614719 4805 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="08314cda-c0a3-4a26-bb3a-65cd724335a6" containerName="nova-api-api" probeResult="failure" output="Get \"http://10.217.0.191:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 03 14:33:05 crc kubenswrapper[4805]: I1203 14:33:05.614956 4805 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="08314cda-c0a3-4a26-bb3a-65cd724335a6" containerName="nova-api-log" probeResult="failure" output="Get \"http://10.217.0.191:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 03 14:33:05 crc kubenswrapper[4805]: I1203 14:33:05.633160 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/4994372a-ac42-4661-aed6-734070bee008-ovsdbserver-sb\") pod \"4994372a-ac42-4661-aed6-734070bee008\" (UID: \"4994372a-ac42-4661-aed6-734070bee008\") " Dec 03 14:33:05 crc kubenswrapper[4805]: I1203 14:33:05.633214 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4994372a-ac42-4661-aed6-734070bee008-config\") pod \"4994372a-ac42-4661-aed6-734070bee008\" (UID: \"4994372a-ac42-4661-aed6-734070bee008\") " Dec 03 14:33:05 crc kubenswrapper[4805]: I1203 14:33:05.633244 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/4994372a-ac42-4661-aed6-734070bee008-ovsdbserver-nb\") pod \"4994372a-ac42-4661-aed6-734070bee008\" (UID: \"4994372a-ac42-4661-aed6-734070bee008\") " Dec 03 14:33:05 crc kubenswrapper[4805]: I1203 14:33:05.633357 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/4994372a-ac42-4661-aed6-734070bee008-dns-svc\") pod \"4994372a-ac42-4661-aed6-734070bee008\" (UID: \"4994372a-ac42-4661-aed6-734070bee008\") " Dec 03 14:33:05 crc kubenswrapper[4805]: I1203 14:33:05.633475 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/4994372a-ac42-4661-aed6-734070bee008-dns-swift-storage-0\") pod \"4994372a-ac42-4661-aed6-734070bee008\" (UID: \"4994372a-ac42-4661-aed6-734070bee008\") " Dec 03 14:33:05 crc kubenswrapper[4805]: I1203 14:33:05.633532 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4dfjm\" (UniqueName: \"kubernetes.io/projected/4994372a-ac42-4661-aed6-734070bee008-kube-api-access-4dfjm\") pod \"4994372a-ac42-4661-aed6-734070bee008\" (UID: \"4994372a-ac42-4661-aed6-734070bee008\") " Dec 03 14:33:05 crc kubenswrapper[4805]: I1203 14:33:05.633817 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zq272\" (UniqueName: \"kubernetes.io/projected/b3635de5-a6a4-465c-8793-bce58d357da6-kube-api-access-zq272\") pod \"nova-cell1-conductor-0\" (UID: \"b3635de5-a6a4-465c-8793-bce58d357da6\") " pod="openstack/nova-cell1-conductor-0" Dec 03 14:33:05 crc kubenswrapper[4805]: I1203 14:33:05.633986 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b3635de5-a6a4-465c-8793-bce58d357da6-combined-ca-bundle\") pod \"nova-cell1-conductor-0\" (UID: \"b3635de5-a6a4-465c-8793-bce58d357da6\") " pod="openstack/nova-cell1-conductor-0" Dec 03 14:33:05 crc kubenswrapper[4805]: I1203 14:33:05.634022 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b3635de5-a6a4-465c-8793-bce58d357da6-config-data\") pod \"nova-cell1-conductor-0\" (UID: \"b3635de5-a6a4-465c-8793-bce58d357da6\") " pod="openstack/nova-cell1-conductor-0" Dec 03 14:33:05 crc kubenswrapper[4805]: I1203 14:33:05.651131 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4994372a-ac42-4661-aed6-734070bee008-kube-api-access-4dfjm" (OuterVolumeSpecName: "kube-api-access-4dfjm") pod "4994372a-ac42-4661-aed6-734070bee008" (UID: "4994372a-ac42-4661-aed6-734070bee008"). InnerVolumeSpecName "kube-api-access-4dfjm". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 14:33:05 crc kubenswrapper[4805]: I1203 14:33:05.682546 4805 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Dec 03 14:33:05 crc kubenswrapper[4805]: I1203 14:33:05.713315 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4994372a-ac42-4661-aed6-734070bee008-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "4994372a-ac42-4661-aed6-734070bee008" (UID: "4994372a-ac42-4661-aed6-734070bee008"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 14:33:05 crc kubenswrapper[4805]: I1203 14:33:05.730783 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4994372a-ac42-4661-aed6-734070bee008-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "4994372a-ac42-4661-aed6-734070bee008" (UID: "4994372a-ac42-4661-aed6-734070bee008"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 14:33:05 crc kubenswrapper[4805]: I1203 14:33:05.732569 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4994372a-ac42-4661-aed6-734070bee008-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "4994372a-ac42-4661-aed6-734070bee008" (UID: "4994372a-ac42-4661-aed6-734070bee008"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 14:33:05 crc kubenswrapper[4805]: I1203 14:33:05.736534 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b3635de5-a6a4-465c-8793-bce58d357da6-combined-ca-bundle\") pod \"nova-cell1-conductor-0\" (UID: \"b3635de5-a6a4-465c-8793-bce58d357da6\") " pod="openstack/nova-cell1-conductor-0" Dec 03 14:33:05 crc kubenswrapper[4805]: I1203 14:33:05.736587 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b3635de5-a6a4-465c-8793-bce58d357da6-config-data\") pod \"nova-cell1-conductor-0\" (UID: \"b3635de5-a6a4-465c-8793-bce58d357da6\") " pod="openstack/nova-cell1-conductor-0" Dec 03 14:33:05 crc kubenswrapper[4805]: I1203 14:33:05.736661 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zq272\" (UniqueName: \"kubernetes.io/projected/b3635de5-a6a4-465c-8793-bce58d357da6-kube-api-access-zq272\") pod \"nova-cell1-conductor-0\" (UID: \"b3635de5-a6a4-465c-8793-bce58d357da6\") " pod="openstack/nova-cell1-conductor-0" Dec 03 14:33:05 crc kubenswrapper[4805]: I1203 14:33:05.736759 4805 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/4994372a-ac42-4661-aed6-734070bee008-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Dec 03 14:33:05 crc kubenswrapper[4805]: I1203 14:33:05.736777 4805 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/4994372a-ac42-4661-aed6-734070bee008-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Dec 03 14:33:05 crc kubenswrapper[4805]: I1203 14:33:05.736787 4805 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/4994372a-ac42-4661-aed6-734070bee008-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 03 14:33:05 crc kubenswrapper[4805]: I1203 14:33:05.736797 4805 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4dfjm\" (UniqueName: \"kubernetes.io/projected/4994372a-ac42-4661-aed6-734070bee008-kube-api-access-4dfjm\") on node \"crc\" DevicePath \"\"" Dec 03 14:33:05 crc kubenswrapper[4805]: I1203 14:33:05.738336 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4994372a-ac42-4661-aed6-734070bee008-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "4994372a-ac42-4661-aed6-734070bee008" (UID: "4994372a-ac42-4661-aed6-734070bee008"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 14:33:05 crc kubenswrapper[4805]: I1203 14:33:05.741245 4805 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Dec 03 14:33:05 crc kubenswrapper[4805]: I1203 14:33:05.741707 4805 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Dec 03 14:33:05 crc kubenswrapper[4805]: I1203 14:33:05.742501 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b3635de5-a6a4-465c-8793-bce58d357da6-config-data\") pod \"nova-cell1-conductor-0\" (UID: \"b3635de5-a6a4-465c-8793-bce58d357da6\") " pod="openstack/nova-cell1-conductor-0" Dec 03 14:33:05 crc kubenswrapper[4805]: I1203 14:33:05.743284 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b3635de5-a6a4-465c-8793-bce58d357da6-combined-ca-bundle\") pod \"nova-cell1-conductor-0\" (UID: \"b3635de5-a6a4-465c-8793-bce58d357da6\") " pod="openstack/nova-cell1-conductor-0" Dec 03 14:33:05 crc kubenswrapper[4805]: I1203 14:33:05.751140 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4994372a-ac42-4661-aed6-734070bee008-config" (OuterVolumeSpecName: "config") pod "4994372a-ac42-4661-aed6-734070bee008" (UID: "4994372a-ac42-4661-aed6-734070bee008"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 14:33:05 crc kubenswrapper[4805]: I1203 14:33:05.760681 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zq272\" (UniqueName: \"kubernetes.io/projected/b3635de5-a6a4-465c-8793-bce58d357da6-kube-api-access-zq272\") pod \"nova-cell1-conductor-0\" (UID: \"b3635de5-a6a4-465c-8793-bce58d357da6\") " pod="openstack/nova-cell1-conductor-0" Dec 03 14:33:05 crc kubenswrapper[4805]: I1203 14:33:05.761917 4805 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Dec 03 14:33:05 crc kubenswrapper[4805]: I1203 14:33:05.839031 4805 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/4994372a-ac42-4661-aed6-734070bee008-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Dec 03 14:33:05 crc kubenswrapper[4805]: I1203 14:33:05.839061 4805 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4994372a-ac42-4661-aed6-734070bee008-config\") on node \"crc\" DevicePath \"\"" Dec 03 14:33:05 crc kubenswrapper[4805]: E1203 14:33:05.856923 4805 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod527667db_e5ab_4fe2_89a5_d9110602f1d2.slice\": RecentStats: unable to find data in memory cache]" Dec 03 14:33:05 crc kubenswrapper[4805]: I1203 14:33:05.882139 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-0" Dec 03 14:33:06 crc kubenswrapper[4805]: I1203 14:33:06.329080 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-0"] Dec 03 14:33:06 crc kubenswrapper[4805]: I1203 14:33:06.462201 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-0" event={"ID":"b3635de5-a6a4-465c-8793-bce58d357da6","Type":"ContainerStarted","Data":"4f1c9216de1e9137b935d6f44b78650d28e85a10a16b97c02bfd49bf2ef60cf1"} Dec 03 14:33:06 crc kubenswrapper[4805]: I1203 14:33:06.464275 4805 generic.go:334] "Generic (PLEG): container finished" podID="08314cda-c0a3-4a26-bb3a-65cd724335a6" containerID="a63cffd29ec3c44013d6c9cad06663729f09ed27ec35f6535fb35e381786dad2" exitCode=143 Dec 03 14:33:06 crc kubenswrapper[4805]: I1203 14:33:06.464356 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"08314cda-c0a3-4a26-bb3a-65cd724335a6","Type":"ContainerDied","Data":"a63cffd29ec3c44013d6c9cad06663729f09ed27ec35f6535fb35e381786dad2"} Dec 03 14:33:06 crc kubenswrapper[4805]: I1203 14:33:06.468994 4805 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-f6bc4c6c9-x5tdg" Dec 03 14:33:06 crc kubenswrapper[4805]: I1203 14:33:06.471976 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-f6bc4c6c9-x5tdg" event={"ID":"4994372a-ac42-4661-aed6-734070bee008","Type":"ContainerDied","Data":"57e94c3359e37674bb98418990700796598ac64a12d8b76df0724f15c0471c86"} Dec 03 14:33:06 crc kubenswrapper[4805]: I1203 14:33:06.472051 4805 scope.go:117] "RemoveContainer" containerID="fda7f3c053dd3dd390f549cc83ac2bcd36e8561e014152b04f3277f127e4ceef" Dec 03 14:33:06 crc kubenswrapper[4805]: I1203 14:33:06.496777 4805 scope.go:117] "RemoveContainer" containerID="32cefd53191c106c90ee4ec6f5b5281b248b8f25bcf06f4ae10cb2966c32886a" Dec 03 14:33:06 crc kubenswrapper[4805]: I1203 14:33:06.522237 4805 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-f6bc4c6c9-x5tdg"] Dec 03 14:33:06 crc kubenswrapper[4805]: I1203 14:33:06.529617 4805 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-f6bc4c6c9-x5tdg"] Dec 03 14:33:06 crc kubenswrapper[4805]: I1203 14:33:06.706772 4805 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4994372a-ac42-4661-aed6-734070bee008" path="/var/lib/kubelet/pods/4994372a-ac42-4661-aed6-734070bee008/volumes" Dec 03 14:33:07 crc kubenswrapper[4805]: I1203 14:33:07.478149 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-0" event={"ID":"b3635de5-a6a4-465c-8793-bce58d357da6","Type":"ContainerStarted","Data":"3a3880d7733847f6ef069cd44b31332e0a16858326e382bcdf02f8e105cac745"} Dec 03 14:33:07 crc kubenswrapper[4805]: I1203 14:33:07.478175 4805 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-scheduler-0" podUID="0ba4f686-dd10-475f-a65a-78041893d75d" containerName="nova-scheduler-scheduler" containerID="cri-o://5243aa51bbe15f29ac6238933f1d6db7ceba9c4f6359400eb45ce14c0ecf54a6" gracePeriod=30 Dec 03 14:33:07 crc kubenswrapper[4805]: I1203 14:33:07.478886 4805 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="6eb3690f-3101-448b-999c-ae7c4216d1f7" containerName="nova-metadata-log" containerID="cri-o://45927200e3a8a365e2d0d98f280946c74d9779da0c568414671ab2c5f00ae669" gracePeriod=30 Dec 03 14:33:07 crc kubenswrapper[4805]: I1203 14:33:07.479213 4805 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="6eb3690f-3101-448b-999c-ae7c4216d1f7" containerName="nova-metadata-metadata" containerID="cri-o://f4d56fc727319480ee7c35d1447da5b62c144739673fc1cb1cfdfef2e78bd055" gracePeriod=30 Dec 03 14:33:07 crc kubenswrapper[4805]: I1203 14:33:07.513377 4805 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-conductor-0" podStartSLOduration=2.513356587 podStartE2EDuration="2.513356587s" podCreationTimestamp="2025-12-03 14:33:05 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 14:33:07.509202414 +0000 UTC m=+1417.172119347" watchObservedRunningTime="2025-12-03 14:33:07.513356587 +0000 UTC m=+1417.176273510" Dec 03 14:33:08 crc kubenswrapper[4805]: I1203 14:33:08.490647 4805 generic.go:334] "Generic (PLEG): container finished" podID="6eb3690f-3101-448b-999c-ae7c4216d1f7" containerID="f4d56fc727319480ee7c35d1447da5b62c144739673fc1cb1cfdfef2e78bd055" exitCode=0 Dec 03 14:33:08 crc kubenswrapper[4805]: I1203 14:33:08.490687 4805 generic.go:334] "Generic (PLEG): container finished" podID="6eb3690f-3101-448b-999c-ae7c4216d1f7" containerID="45927200e3a8a365e2d0d98f280946c74d9779da0c568414671ab2c5f00ae669" exitCode=143 Dec 03 14:33:08 crc kubenswrapper[4805]: I1203 14:33:08.491285 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"6eb3690f-3101-448b-999c-ae7c4216d1f7","Type":"ContainerDied","Data":"f4d56fc727319480ee7c35d1447da5b62c144739673fc1cb1cfdfef2e78bd055"} Dec 03 14:33:08 crc kubenswrapper[4805]: I1203 14:33:08.491375 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"6eb3690f-3101-448b-999c-ae7c4216d1f7","Type":"ContainerDied","Data":"45927200e3a8a365e2d0d98f280946c74d9779da0c568414671ab2c5f00ae669"} Dec 03 14:33:08 crc kubenswrapper[4805]: I1203 14:33:08.491860 4805 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell1-conductor-0" Dec 03 14:33:08 crc kubenswrapper[4805]: I1203 14:33:08.879494 4805 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 03 14:33:09 crc kubenswrapper[4805]: I1203 14:33:09.007443 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6eb3690f-3101-448b-999c-ae7c4216d1f7-config-data\") pod \"6eb3690f-3101-448b-999c-ae7c4216d1f7\" (UID: \"6eb3690f-3101-448b-999c-ae7c4216d1f7\") " Dec 03 14:33:09 crc kubenswrapper[4805]: I1203 14:33:09.007609 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6eb3690f-3101-448b-999c-ae7c4216d1f7-combined-ca-bundle\") pod \"6eb3690f-3101-448b-999c-ae7c4216d1f7\" (UID: \"6eb3690f-3101-448b-999c-ae7c4216d1f7\") " Dec 03 14:33:09 crc kubenswrapper[4805]: I1203 14:33:09.007660 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/6eb3690f-3101-448b-999c-ae7c4216d1f7-nova-metadata-tls-certs\") pod \"6eb3690f-3101-448b-999c-ae7c4216d1f7\" (UID: \"6eb3690f-3101-448b-999c-ae7c4216d1f7\") " Dec 03 14:33:09 crc kubenswrapper[4805]: I1203 14:33:09.007686 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hdhcb\" (UniqueName: \"kubernetes.io/projected/6eb3690f-3101-448b-999c-ae7c4216d1f7-kube-api-access-hdhcb\") pod \"6eb3690f-3101-448b-999c-ae7c4216d1f7\" (UID: \"6eb3690f-3101-448b-999c-ae7c4216d1f7\") " Dec 03 14:33:09 crc kubenswrapper[4805]: I1203 14:33:09.007758 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/6eb3690f-3101-448b-999c-ae7c4216d1f7-logs\") pod \"6eb3690f-3101-448b-999c-ae7c4216d1f7\" (UID: \"6eb3690f-3101-448b-999c-ae7c4216d1f7\") " Dec 03 14:33:09 crc kubenswrapper[4805]: I1203 14:33:09.008282 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/6eb3690f-3101-448b-999c-ae7c4216d1f7-logs" (OuterVolumeSpecName: "logs") pod "6eb3690f-3101-448b-999c-ae7c4216d1f7" (UID: "6eb3690f-3101-448b-999c-ae7c4216d1f7"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 14:33:09 crc kubenswrapper[4805]: I1203 14:33:09.013252 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6eb3690f-3101-448b-999c-ae7c4216d1f7-kube-api-access-hdhcb" (OuterVolumeSpecName: "kube-api-access-hdhcb") pod "6eb3690f-3101-448b-999c-ae7c4216d1f7" (UID: "6eb3690f-3101-448b-999c-ae7c4216d1f7"). InnerVolumeSpecName "kube-api-access-hdhcb". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 14:33:09 crc kubenswrapper[4805]: I1203 14:33:09.036138 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6eb3690f-3101-448b-999c-ae7c4216d1f7-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "6eb3690f-3101-448b-999c-ae7c4216d1f7" (UID: "6eb3690f-3101-448b-999c-ae7c4216d1f7"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 14:33:09 crc kubenswrapper[4805]: I1203 14:33:09.040679 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6eb3690f-3101-448b-999c-ae7c4216d1f7-config-data" (OuterVolumeSpecName: "config-data") pod "6eb3690f-3101-448b-999c-ae7c4216d1f7" (UID: "6eb3690f-3101-448b-999c-ae7c4216d1f7"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 14:33:09 crc kubenswrapper[4805]: I1203 14:33:09.059084 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6eb3690f-3101-448b-999c-ae7c4216d1f7-nova-metadata-tls-certs" (OuterVolumeSpecName: "nova-metadata-tls-certs") pod "6eb3690f-3101-448b-999c-ae7c4216d1f7" (UID: "6eb3690f-3101-448b-999c-ae7c4216d1f7"). InnerVolumeSpecName "nova-metadata-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 14:33:09 crc kubenswrapper[4805]: I1203 14:33:09.110076 4805 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6eb3690f-3101-448b-999c-ae7c4216d1f7-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 14:33:09 crc kubenswrapper[4805]: I1203 14:33:09.110107 4805 reconciler_common.go:293] "Volume detached for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/6eb3690f-3101-448b-999c-ae7c4216d1f7-nova-metadata-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 03 14:33:09 crc kubenswrapper[4805]: I1203 14:33:09.110119 4805 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hdhcb\" (UniqueName: \"kubernetes.io/projected/6eb3690f-3101-448b-999c-ae7c4216d1f7-kube-api-access-hdhcb\") on node \"crc\" DevicePath \"\"" Dec 03 14:33:09 crc kubenswrapper[4805]: I1203 14:33:09.110129 4805 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/6eb3690f-3101-448b-999c-ae7c4216d1f7-logs\") on node \"crc\" DevicePath \"\"" Dec 03 14:33:09 crc kubenswrapper[4805]: I1203 14:33:09.110138 4805 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6eb3690f-3101-448b-999c-ae7c4216d1f7-config-data\") on node \"crc\" DevicePath \"\"" Dec 03 14:33:09 crc kubenswrapper[4805]: I1203 14:33:09.514669 4805 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 03 14:33:09 crc kubenswrapper[4805]: I1203 14:33:09.514878 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"6eb3690f-3101-448b-999c-ae7c4216d1f7","Type":"ContainerDied","Data":"59fcc66af346a6b242e1166651e94740e190fab0bb92d23d848feec9917755c3"} Dec 03 14:33:09 crc kubenswrapper[4805]: I1203 14:33:09.515657 4805 scope.go:117] "RemoveContainer" containerID="f4d56fc727319480ee7c35d1447da5b62c144739673fc1cb1cfdfef2e78bd055" Dec 03 14:33:09 crc kubenswrapper[4805]: I1203 14:33:09.555095 4805 scope.go:117] "RemoveContainer" containerID="45927200e3a8a365e2d0d98f280946c74d9779da0c568414671ab2c5f00ae669" Dec 03 14:33:09 crc kubenswrapper[4805]: I1203 14:33:09.586871 4805 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Dec 03 14:33:09 crc kubenswrapper[4805]: I1203 14:33:09.609803 4805 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-metadata-0"] Dec 03 14:33:09 crc kubenswrapper[4805]: I1203 14:33:09.623996 4805 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-metadata-0"] Dec 03 14:33:09 crc kubenswrapper[4805]: E1203 14:33:09.624515 4805 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6eb3690f-3101-448b-999c-ae7c4216d1f7" containerName="nova-metadata-metadata" Dec 03 14:33:09 crc kubenswrapper[4805]: I1203 14:33:09.624540 4805 state_mem.go:107] "Deleted CPUSet assignment" podUID="6eb3690f-3101-448b-999c-ae7c4216d1f7" containerName="nova-metadata-metadata" Dec 03 14:33:09 crc kubenswrapper[4805]: E1203 14:33:09.624564 4805 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6eb3690f-3101-448b-999c-ae7c4216d1f7" containerName="nova-metadata-log" Dec 03 14:33:09 crc kubenswrapper[4805]: I1203 14:33:09.624573 4805 state_mem.go:107] "Deleted CPUSet assignment" podUID="6eb3690f-3101-448b-999c-ae7c4216d1f7" containerName="nova-metadata-log" Dec 03 14:33:09 crc kubenswrapper[4805]: I1203 14:33:09.624798 4805 memory_manager.go:354] "RemoveStaleState removing state" podUID="6eb3690f-3101-448b-999c-ae7c4216d1f7" containerName="nova-metadata-metadata" Dec 03 14:33:09 crc kubenswrapper[4805]: I1203 14:33:09.624865 4805 memory_manager.go:354] "RemoveStaleState removing state" podUID="6eb3690f-3101-448b-999c-ae7c4216d1f7" containerName="nova-metadata-log" Dec 03 14:33:09 crc kubenswrapper[4805]: I1203 14:33:09.626093 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 03 14:33:09 crc kubenswrapper[4805]: I1203 14:33:09.629435 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-config-data" Dec 03 14:33:09 crc kubenswrapper[4805]: I1203 14:33:09.629812 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-metadata-internal-svc" Dec 03 14:33:09 crc kubenswrapper[4805]: I1203 14:33:09.630989 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Dec 03 14:33:09 crc kubenswrapper[4805]: I1203 14:33:09.721213 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dmqxc\" (UniqueName: \"kubernetes.io/projected/2dcd6066-cee7-40a7-ba9f-0648b6d2283b-kube-api-access-dmqxc\") pod \"nova-metadata-0\" (UID: \"2dcd6066-cee7-40a7-ba9f-0648b6d2283b\") " pod="openstack/nova-metadata-0" Dec 03 14:33:09 crc kubenswrapper[4805]: I1203 14:33:09.721256 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/2dcd6066-cee7-40a7-ba9f-0648b6d2283b-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"2dcd6066-cee7-40a7-ba9f-0648b6d2283b\") " pod="openstack/nova-metadata-0" Dec 03 14:33:09 crc kubenswrapper[4805]: I1203 14:33:09.721294 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/2dcd6066-cee7-40a7-ba9f-0648b6d2283b-logs\") pod \"nova-metadata-0\" (UID: \"2dcd6066-cee7-40a7-ba9f-0648b6d2283b\") " pod="openstack/nova-metadata-0" Dec 03 14:33:09 crc kubenswrapper[4805]: I1203 14:33:09.721314 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2dcd6066-cee7-40a7-ba9f-0648b6d2283b-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"2dcd6066-cee7-40a7-ba9f-0648b6d2283b\") " pod="openstack/nova-metadata-0" Dec 03 14:33:09 crc kubenswrapper[4805]: I1203 14:33:09.721387 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2dcd6066-cee7-40a7-ba9f-0648b6d2283b-config-data\") pod \"nova-metadata-0\" (UID: \"2dcd6066-cee7-40a7-ba9f-0648b6d2283b\") " pod="openstack/nova-metadata-0" Dec 03 14:33:09 crc kubenswrapper[4805]: I1203 14:33:09.823292 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/2dcd6066-cee7-40a7-ba9f-0648b6d2283b-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"2dcd6066-cee7-40a7-ba9f-0648b6d2283b\") " pod="openstack/nova-metadata-0" Dec 03 14:33:09 crc kubenswrapper[4805]: I1203 14:33:09.823340 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dmqxc\" (UniqueName: \"kubernetes.io/projected/2dcd6066-cee7-40a7-ba9f-0648b6d2283b-kube-api-access-dmqxc\") pod \"nova-metadata-0\" (UID: \"2dcd6066-cee7-40a7-ba9f-0648b6d2283b\") " pod="openstack/nova-metadata-0" Dec 03 14:33:09 crc kubenswrapper[4805]: I1203 14:33:09.823378 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/2dcd6066-cee7-40a7-ba9f-0648b6d2283b-logs\") pod \"nova-metadata-0\" (UID: \"2dcd6066-cee7-40a7-ba9f-0648b6d2283b\") " pod="openstack/nova-metadata-0" Dec 03 14:33:09 crc kubenswrapper[4805]: I1203 14:33:09.823395 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2dcd6066-cee7-40a7-ba9f-0648b6d2283b-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"2dcd6066-cee7-40a7-ba9f-0648b6d2283b\") " pod="openstack/nova-metadata-0" Dec 03 14:33:09 crc kubenswrapper[4805]: I1203 14:33:09.823487 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2dcd6066-cee7-40a7-ba9f-0648b6d2283b-config-data\") pod \"nova-metadata-0\" (UID: \"2dcd6066-cee7-40a7-ba9f-0648b6d2283b\") " pod="openstack/nova-metadata-0" Dec 03 14:33:09 crc kubenswrapper[4805]: I1203 14:33:09.824474 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/2dcd6066-cee7-40a7-ba9f-0648b6d2283b-logs\") pod \"nova-metadata-0\" (UID: \"2dcd6066-cee7-40a7-ba9f-0648b6d2283b\") " pod="openstack/nova-metadata-0" Dec 03 14:33:09 crc kubenswrapper[4805]: I1203 14:33:09.828415 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/2dcd6066-cee7-40a7-ba9f-0648b6d2283b-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"2dcd6066-cee7-40a7-ba9f-0648b6d2283b\") " pod="openstack/nova-metadata-0" Dec 03 14:33:09 crc kubenswrapper[4805]: I1203 14:33:09.839900 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2dcd6066-cee7-40a7-ba9f-0648b6d2283b-config-data\") pod \"nova-metadata-0\" (UID: \"2dcd6066-cee7-40a7-ba9f-0648b6d2283b\") " pod="openstack/nova-metadata-0" Dec 03 14:33:09 crc kubenswrapper[4805]: I1203 14:33:09.843438 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2dcd6066-cee7-40a7-ba9f-0648b6d2283b-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"2dcd6066-cee7-40a7-ba9f-0648b6d2283b\") " pod="openstack/nova-metadata-0" Dec 03 14:33:09 crc kubenswrapper[4805]: I1203 14:33:09.858670 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dmqxc\" (UniqueName: \"kubernetes.io/projected/2dcd6066-cee7-40a7-ba9f-0648b6d2283b-kube-api-access-dmqxc\") pod \"nova-metadata-0\" (UID: \"2dcd6066-cee7-40a7-ba9f-0648b6d2283b\") " pod="openstack/nova-metadata-0" Dec 03 14:33:09 crc kubenswrapper[4805]: E1203 14:33:09.866391 4805 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="5243aa51bbe15f29ac6238933f1d6db7ceba9c4f6359400eb45ce14c0ecf54a6" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Dec 03 14:33:09 crc kubenswrapper[4805]: E1203 14:33:09.868102 4805 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="5243aa51bbe15f29ac6238933f1d6db7ceba9c4f6359400eb45ce14c0ecf54a6" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Dec 03 14:33:09 crc kubenswrapper[4805]: E1203 14:33:09.873079 4805 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="5243aa51bbe15f29ac6238933f1d6db7ceba9c4f6359400eb45ce14c0ecf54a6" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Dec 03 14:33:09 crc kubenswrapper[4805]: E1203 14:33:09.873147 4805 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openstack/nova-scheduler-0" podUID="0ba4f686-dd10-475f-a65a-78041893d75d" containerName="nova-scheduler-scheduler" Dec 03 14:33:09 crc kubenswrapper[4805]: I1203 14:33:09.964661 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 03 14:33:10 crc kubenswrapper[4805]: I1203 14:33:10.449644 4805 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Dec 03 14:33:10 crc kubenswrapper[4805]: I1203 14:33:10.545969 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0ba4f686-dd10-475f-a65a-78041893d75d-config-data\") pod \"0ba4f686-dd10-475f-a65a-78041893d75d\" (UID: \"0ba4f686-dd10-475f-a65a-78041893d75d\") " Dec 03 14:33:10 crc kubenswrapper[4805]: I1203 14:33:10.546051 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bbp6p\" (UniqueName: \"kubernetes.io/projected/0ba4f686-dd10-475f-a65a-78041893d75d-kube-api-access-bbp6p\") pod \"0ba4f686-dd10-475f-a65a-78041893d75d\" (UID: \"0ba4f686-dd10-475f-a65a-78041893d75d\") " Dec 03 14:33:10 crc kubenswrapper[4805]: I1203 14:33:10.546120 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0ba4f686-dd10-475f-a65a-78041893d75d-combined-ca-bundle\") pod \"0ba4f686-dd10-475f-a65a-78041893d75d\" (UID: \"0ba4f686-dd10-475f-a65a-78041893d75d\") " Dec 03 14:33:10 crc kubenswrapper[4805]: I1203 14:33:10.554358 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0ba4f686-dd10-475f-a65a-78041893d75d-kube-api-access-bbp6p" (OuterVolumeSpecName: "kube-api-access-bbp6p") pod "0ba4f686-dd10-475f-a65a-78041893d75d" (UID: "0ba4f686-dd10-475f-a65a-78041893d75d"). InnerVolumeSpecName "kube-api-access-bbp6p". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 14:33:10 crc kubenswrapper[4805]: I1203 14:33:10.578330 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Dec 03 14:33:10 crc kubenswrapper[4805]: I1203 14:33:10.578388 4805 generic.go:334] "Generic (PLEG): container finished" podID="0ba4f686-dd10-475f-a65a-78041893d75d" containerID="5243aa51bbe15f29ac6238933f1d6db7ceba9c4f6359400eb45ce14c0ecf54a6" exitCode=0 Dec 03 14:33:10 crc kubenswrapper[4805]: I1203 14:33:10.578380 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"0ba4f686-dd10-475f-a65a-78041893d75d","Type":"ContainerDied","Data":"5243aa51bbe15f29ac6238933f1d6db7ceba9c4f6359400eb45ce14c0ecf54a6"} Dec 03 14:33:10 crc kubenswrapper[4805]: I1203 14:33:10.579166 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"0ba4f686-dd10-475f-a65a-78041893d75d","Type":"ContainerDied","Data":"20d8e8a0cf3cfe920107647005fd79426ba22e21aa2653799cb87d499639431d"} Dec 03 14:33:10 crc kubenswrapper[4805]: I1203 14:33:10.579200 4805 scope.go:117] "RemoveContainer" containerID="5243aa51bbe15f29ac6238933f1d6db7ceba9c4f6359400eb45ce14c0ecf54a6" Dec 03 14:33:10 crc kubenswrapper[4805]: I1203 14:33:10.579226 4805 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Dec 03 14:33:10 crc kubenswrapper[4805]: W1203 14:33:10.580621 4805 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod2dcd6066_cee7_40a7_ba9f_0648b6d2283b.slice/crio-b74eb9e10891dd812db247dd7c26b09bf2754b3c9ea7e37386e6ade1a8d6eb65 WatchSource:0}: Error finding container b74eb9e10891dd812db247dd7c26b09bf2754b3c9ea7e37386e6ade1a8d6eb65: Status 404 returned error can't find the container with id b74eb9e10891dd812db247dd7c26b09bf2754b3c9ea7e37386e6ade1a8d6eb65 Dec 03 14:33:10 crc kubenswrapper[4805]: I1203 14:33:10.589279 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0ba4f686-dd10-475f-a65a-78041893d75d-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "0ba4f686-dd10-475f-a65a-78041893d75d" (UID: "0ba4f686-dd10-475f-a65a-78041893d75d"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 14:33:10 crc kubenswrapper[4805]: I1203 14:33:10.594074 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0ba4f686-dd10-475f-a65a-78041893d75d-config-data" (OuterVolumeSpecName: "config-data") pod "0ba4f686-dd10-475f-a65a-78041893d75d" (UID: "0ba4f686-dd10-475f-a65a-78041893d75d"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 14:33:10 crc kubenswrapper[4805]: I1203 14:33:10.612661 4805 scope.go:117] "RemoveContainer" containerID="5243aa51bbe15f29ac6238933f1d6db7ceba9c4f6359400eb45ce14c0ecf54a6" Dec 03 14:33:10 crc kubenswrapper[4805]: E1203 14:33:10.616348 4805 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5243aa51bbe15f29ac6238933f1d6db7ceba9c4f6359400eb45ce14c0ecf54a6\": container with ID starting with 5243aa51bbe15f29ac6238933f1d6db7ceba9c4f6359400eb45ce14c0ecf54a6 not found: ID does not exist" containerID="5243aa51bbe15f29ac6238933f1d6db7ceba9c4f6359400eb45ce14c0ecf54a6" Dec 03 14:33:10 crc kubenswrapper[4805]: I1203 14:33:10.616388 4805 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5243aa51bbe15f29ac6238933f1d6db7ceba9c4f6359400eb45ce14c0ecf54a6"} err="failed to get container status \"5243aa51bbe15f29ac6238933f1d6db7ceba9c4f6359400eb45ce14c0ecf54a6\": rpc error: code = NotFound desc = could not find container \"5243aa51bbe15f29ac6238933f1d6db7ceba9c4f6359400eb45ce14c0ecf54a6\": container with ID starting with 5243aa51bbe15f29ac6238933f1d6db7ceba9c4f6359400eb45ce14c0ecf54a6 not found: ID does not exist" Dec 03 14:33:10 crc kubenswrapper[4805]: I1203 14:33:10.648462 4805 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0ba4f686-dd10-475f-a65a-78041893d75d-config-data\") on node \"crc\" DevicePath \"\"" Dec 03 14:33:10 crc kubenswrapper[4805]: I1203 14:33:10.648499 4805 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bbp6p\" (UniqueName: \"kubernetes.io/projected/0ba4f686-dd10-475f-a65a-78041893d75d-kube-api-access-bbp6p\") on node \"crc\" DevicePath \"\"" Dec 03 14:33:10 crc kubenswrapper[4805]: I1203 14:33:10.648509 4805 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0ba4f686-dd10-475f-a65a-78041893d75d-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 14:33:10 crc kubenswrapper[4805]: I1203 14:33:10.709175 4805 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6eb3690f-3101-448b-999c-ae7c4216d1f7" path="/var/lib/kubelet/pods/6eb3690f-3101-448b-999c-ae7c4216d1f7/volumes" Dec 03 14:33:10 crc kubenswrapper[4805]: I1203 14:33:10.961466 4805 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Dec 03 14:33:10 crc kubenswrapper[4805]: I1203 14:33:10.970784 4805 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-scheduler-0"] Dec 03 14:33:10 crc kubenswrapper[4805]: I1203 14:33:10.988548 4805 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-scheduler-0"] Dec 03 14:33:10 crc kubenswrapper[4805]: E1203 14:33:10.988947 4805 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0ba4f686-dd10-475f-a65a-78041893d75d" containerName="nova-scheduler-scheduler" Dec 03 14:33:10 crc kubenswrapper[4805]: I1203 14:33:10.988963 4805 state_mem.go:107] "Deleted CPUSet assignment" podUID="0ba4f686-dd10-475f-a65a-78041893d75d" containerName="nova-scheduler-scheduler" Dec 03 14:33:10 crc kubenswrapper[4805]: I1203 14:33:10.989219 4805 memory_manager.go:354] "RemoveStaleState removing state" podUID="0ba4f686-dd10-475f-a65a-78041893d75d" containerName="nova-scheduler-scheduler" Dec 03 14:33:10 crc kubenswrapper[4805]: I1203 14:33:10.990010 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Dec 03 14:33:10 crc kubenswrapper[4805]: I1203 14:33:10.994815 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-scheduler-config-data" Dec 03 14:33:11 crc kubenswrapper[4805]: I1203 14:33:11.013792 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Dec 03 14:33:11 crc kubenswrapper[4805]: I1203 14:33:11.054618 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c0ccb82f-2439-448d-a958-5d775682cca3-config-data\") pod \"nova-scheduler-0\" (UID: \"c0ccb82f-2439-448d-a958-5d775682cca3\") " pod="openstack/nova-scheduler-0" Dec 03 14:33:11 crc kubenswrapper[4805]: I1203 14:33:11.054688 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c0ccb82f-2439-448d-a958-5d775682cca3-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"c0ccb82f-2439-448d-a958-5d775682cca3\") " pod="openstack/nova-scheduler-0" Dec 03 14:33:11 crc kubenswrapper[4805]: I1203 14:33:11.054751 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lwh2t\" (UniqueName: \"kubernetes.io/projected/c0ccb82f-2439-448d-a958-5d775682cca3-kube-api-access-lwh2t\") pod \"nova-scheduler-0\" (UID: \"c0ccb82f-2439-448d-a958-5d775682cca3\") " pod="openstack/nova-scheduler-0" Dec 03 14:33:11 crc kubenswrapper[4805]: I1203 14:33:11.155913 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c0ccb82f-2439-448d-a958-5d775682cca3-config-data\") pod \"nova-scheduler-0\" (UID: \"c0ccb82f-2439-448d-a958-5d775682cca3\") " pod="openstack/nova-scheduler-0" Dec 03 14:33:11 crc kubenswrapper[4805]: I1203 14:33:11.155976 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c0ccb82f-2439-448d-a958-5d775682cca3-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"c0ccb82f-2439-448d-a958-5d775682cca3\") " pod="openstack/nova-scheduler-0" Dec 03 14:33:11 crc kubenswrapper[4805]: I1203 14:33:11.156017 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lwh2t\" (UniqueName: \"kubernetes.io/projected/c0ccb82f-2439-448d-a958-5d775682cca3-kube-api-access-lwh2t\") pod \"nova-scheduler-0\" (UID: \"c0ccb82f-2439-448d-a958-5d775682cca3\") " pod="openstack/nova-scheduler-0" Dec 03 14:33:11 crc kubenswrapper[4805]: I1203 14:33:11.160308 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c0ccb82f-2439-448d-a958-5d775682cca3-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"c0ccb82f-2439-448d-a958-5d775682cca3\") " pod="openstack/nova-scheduler-0" Dec 03 14:33:11 crc kubenswrapper[4805]: I1203 14:33:11.160728 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c0ccb82f-2439-448d-a958-5d775682cca3-config-data\") pod \"nova-scheduler-0\" (UID: \"c0ccb82f-2439-448d-a958-5d775682cca3\") " pod="openstack/nova-scheduler-0" Dec 03 14:33:11 crc kubenswrapper[4805]: I1203 14:33:11.176582 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lwh2t\" (UniqueName: \"kubernetes.io/projected/c0ccb82f-2439-448d-a958-5d775682cca3-kube-api-access-lwh2t\") pod \"nova-scheduler-0\" (UID: \"c0ccb82f-2439-448d-a958-5d775682cca3\") " pod="openstack/nova-scheduler-0" Dec 03 14:33:11 crc kubenswrapper[4805]: I1203 14:33:11.318022 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Dec 03 14:33:11 crc kubenswrapper[4805]: I1203 14:33:11.393129 4805 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ceilometer-0" Dec 03 14:33:11 crc kubenswrapper[4805]: I1203 14:33:11.531400 4805 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 03 14:33:11 crc kubenswrapper[4805]: I1203 14:33:11.595978 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"2dcd6066-cee7-40a7-ba9f-0648b6d2283b","Type":"ContainerStarted","Data":"59cc0d8dee55d94f88b6b089f411035edd0324469c03b366c0c24c6baa08fe63"} Dec 03 14:33:11 crc kubenswrapper[4805]: I1203 14:33:11.596033 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"2dcd6066-cee7-40a7-ba9f-0648b6d2283b","Type":"ContainerStarted","Data":"72253effa6be075609506ec28eb6dc2aa2e9f7d559f7e42103391020b725db73"} Dec 03 14:33:11 crc kubenswrapper[4805]: I1203 14:33:11.596045 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"2dcd6066-cee7-40a7-ba9f-0648b6d2283b","Type":"ContainerStarted","Data":"b74eb9e10891dd812db247dd7c26b09bf2754b3c9ea7e37386e6ade1a8d6eb65"} Dec 03 14:33:11 crc kubenswrapper[4805]: I1203 14:33:11.604306 4805 generic.go:334] "Generic (PLEG): container finished" podID="08314cda-c0a3-4a26-bb3a-65cd724335a6" containerID="a1819f19bf4f039aeac63d0b7e68e3fb0716250eaee4e898284f30e5115c38e1" exitCode=0 Dec 03 14:33:11 crc kubenswrapper[4805]: I1203 14:33:11.604377 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"08314cda-c0a3-4a26-bb3a-65cd724335a6","Type":"ContainerDied","Data":"a1819f19bf4f039aeac63d0b7e68e3fb0716250eaee4e898284f30e5115c38e1"} Dec 03 14:33:11 crc kubenswrapper[4805]: I1203 14:33:11.604422 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"08314cda-c0a3-4a26-bb3a-65cd724335a6","Type":"ContainerDied","Data":"64b7f6d56431135b4076bcc6ecba0e1404bf67f2c6e6b137731012a9f7ea6dc4"} Dec 03 14:33:11 crc kubenswrapper[4805]: I1203 14:33:11.604442 4805 scope.go:117] "RemoveContainer" containerID="a1819f19bf4f039aeac63d0b7e68e3fb0716250eaee4e898284f30e5115c38e1" Dec 03 14:33:11 crc kubenswrapper[4805]: I1203 14:33:11.604584 4805 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 03 14:33:11 crc kubenswrapper[4805]: I1203 14:33:11.627252 4805 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-metadata-0" podStartSLOduration=2.627231686 podStartE2EDuration="2.627231686s" podCreationTimestamp="2025-12-03 14:33:09 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 14:33:11.613724756 +0000 UTC m=+1421.276641679" watchObservedRunningTime="2025-12-03 14:33:11.627231686 +0000 UTC m=+1421.290148609" Dec 03 14:33:11 crc kubenswrapper[4805]: I1203 14:33:11.640296 4805 scope.go:117] "RemoveContainer" containerID="a63cffd29ec3c44013d6c9cad06663729f09ed27ec35f6535fb35e381786dad2" Dec 03 14:33:11 crc kubenswrapper[4805]: I1203 14:33:11.659100 4805 scope.go:117] "RemoveContainer" containerID="a1819f19bf4f039aeac63d0b7e68e3fb0716250eaee4e898284f30e5115c38e1" Dec 03 14:33:11 crc kubenswrapper[4805]: E1203 14:33:11.659897 4805 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a1819f19bf4f039aeac63d0b7e68e3fb0716250eaee4e898284f30e5115c38e1\": container with ID starting with a1819f19bf4f039aeac63d0b7e68e3fb0716250eaee4e898284f30e5115c38e1 not found: ID does not exist" containerID="a1819f19bf4f039aeac63d0b7e68e3fb0716250eaee4e898284f30e5115c38e1" Dec 03 14:33:11 crc kubenswrapper[4805]: I1203 14:33:11.659945 4805 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a1819f19bf4f039aeac63d0b7e68e3fb0716250eaee4e898284f30e5115c38e1"} err="failed to get container status \"a1819f19bf4f039aeac63d0b7e68e3fb0716250eaee4e898284f30e5115c38e1\": rpc error: code = NotFound desc = could not find container \"a1819f19bf4f039aeac63d0b7e68e3fb0716250eaee4e898284f30e5115c38e1\": container with ID starting with a1819f19bf4f039aeac63d0b7e68e3fb0716250eaee4e898284f30e5115c38e1 not found: ID does not exist" Dec 03 14:33:11 crc kubenswrapper[4805]: I1203 14:33:11.659966 4805 scope.go:117] "RemoveContainer" containerID="a63cffd29ec3c44013d6c9cad06663729f09ed27ec35f6535fb35e381786dad2" Dec 03 14:33:11 crc kubenswrapper[4805]: E1203 14:33:11.660596 4805 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a63cffd29ec3c44013d6c9cad06663729f09ed27ec35f6535fb35e381786dad2\": container with ID starting with a63cffd29ec3c44013d6c9cad06663729f09ed27ec35f6535fb35e381786dad2 not found: ID does not exist" containerID="a63cffd29ec3c44013d6c9cad06663729f09ed27ec35f6535fb35e381786dad2" Dec 03 14:33:11 crc kubenswrapper[4805]: I1203 14:33:11.660622 4805 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a63cffd29ec3c44013d6c9cad06663729f09ed27ec35f6535fb35e381786dad2"} err="failed to get container status \"a63cffd29ec3c44013d6c9cad06663729f09ed27ec35f6535fb35e381786dad2\": rpc error: code = NotFound desc = could not find container \"a63cffd29ec3c44013d6c9cad06663729f09ed27ec35f6535fb35e381786dad2\": container with ID starting with a63cffd29ec3c44013d6c9cad06663729f09ed27ec35f6535fb35e381786dad2 not found: ID does not exist" Dec 03 14:33:11 crc kubenswrapper[4805]: I1203 14:33:11.665068 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/08314cda-c0a3-4a26-bb3a-65cd724335a6-combined-ca-bundle\") pod \"08314cda-c0a3-4a26-bb3a-65cd724335a6\" (UID: \"08314cda-c0a3-4a26-bb3a-65cd724335a6\") " Dec 03 14:33:11 crc kubenswrapper[4805]: I1203 14:33:11.665258 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5tnl7\" (UniqueName: \"kubernetes.io/projected/08314cda-c0a3-4a26-bb3a-65cd724335a6-kube-api-access-5tnl7\") pod \"08314cda-c0a3-4a26-bb3a-65cd724335a6\" (UID: \"08314cda-c0a3-4a26-bb3a-65cd724335a6\") " Dec 03 14:33:11 crc kubenswrapper[4805]: I1203 14:33:11.665300 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/08314cda-c0a3-4a26-bb3a-65cd724335a6-config-data\") pod \"08314cda-c0a3-4a26-bb3a-65cd724335a6\" (UID: \"08314cda-c0a3-4a26-bb3a-65cd724335a6\") " Dec 03 14:33:11 crc kubenswrapper[4805]: I1203 14:33:11.665398 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/08314cda-c0a3-4a26-bb3a-65cd724335a6-logs\") pod \"08314cda-c0a3-4a26-bb3a-65cd724335a6\" (UID: \"08314cda-c0a3-4a26-bb3a-65cd724335a6\") " Dec 03 14:33:11 crc kubenswrapper[4805]: I1203 14:33:11.666254 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/08314cda-c0a3-4a26-bb3a-65cd724335a6-logs" (OuterVolumeSpecName: "logs") pod "08314cda-c0a3-4a26-bb3a-65cd724335a6" (UID: "08314cda-c0a3-4a26-bb3a-65cd724335a6"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 14:33:11 crc kubenswrapper[4805]: I1203 14:33:11.670516 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/08314cda-c0a3-4a26-bb3a-65cd724335a6-kube-api-access-5tnl7" (OuterVolumeSpecName: "kube-api-access-5tnl7") pod "08314cda-c0a3-4a26-bb3a-65cd724335a6" (UID: "08314cda-c0a3-4a26-bb3a-65cd724335a6"). InnerVolumeSpecName "kube-api-access-5tnl7". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 14:33:11 crc kubenswrapper[4805]: I1203 14:33:11.703053 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/08314cda-c0a3-4a26-bb3a-65cd724335a6-config-data" (OuterVolumeSpecName: "config-data") pod "08314cda-c0a3-4a26-bb3a-65cd724335a6" (UID: "08314cda-c0a3-4a26-bb3a-65cd724335a6"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 14:33:11 crc kubenswrapper[4805]: I1203 14:33:11.711517 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/08314cda-c0a3-4a26-bb3a-65cd724335a6-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "08314cda-c0a3-4a26-bb3a-65cd724335a6" (UID: "08314cda-c0a3-4a26-bb3a-65cd724335a6"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 14:33:11 crc kubenswrapper[4805]: I1203 14:33:11.770345 4805 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5tnl7\" (UniqueName: \"kubernetes.io/projected/08314cda-c0a3-4a26-bb3a-65cd724335a6-kube-api-access-5tnl7\") on node \"crc\" DevicePath \"\"" Dec 03 14:33:11 crc kubenswrapper[4805]: I1203 14:33:11.770389 4805 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/08314cda-c0a3-4a26-bb3a-65cd724335a6-config-data\") on node \"crc\" DevicePath \"\"" Dec 03 14:33:11 crc kubenswrapper[4805]: I1203 14:33:11.770401 4805 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/08314cda-c0a3-4a26-bb3a-65cd724335a6-logs\") on node \"crc\" DevicePath \"\"" Dec 03 14:33:11 crc kubenswrapper[4805]: I1203 14:33:11.770416 4805 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/08314cda-c0a3-4a26-bb3a-65cd724335a6-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 14:33:11 crc kubenswrapper[4805]: I1203 14:33:11.952510 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Dec 03 14:33:11 crc kubenswrapper[4805]: I1203 14:33:11.960622 4805 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Dec 03 14:33:11 crc kubenswrapper[4805]: I1203 14:33:11.969649 4805 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-0"] Dec 03 14:33:12 crc kubenswrapper[4805]: I1203 14:33:12.004337 4805 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-0"] Dec 03 14:33:12 crc kubenswrapper[4805]: E1203 14:33:12.004751 4805 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="08314cda-c0a3-4a26-bb3a-65cd724335a6" containerName="nova-api-log" Dec 03 14:33:12 crc kubenswrapper[4805]: I1203 14:33:12.004764 4805 state_mem.go:107] "Deleted CPUSet assignment" podUID="08314cda-c0a3-4a26-bb3a-65cd724335a6" containerName="nova-api-log" Dec 03 14:33:12 crc kubenswrapper[4805]: E1203 14:33:12.004800 4805 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="08314cda-c0a3-4a26-bb3a-65cd724335a6" containerName="nova-api-api" Dec 03 14:33:12 crc kubenswrapper[4805]: I1203 14:33:12.004806 4805 state_mem.go:107] "Deleted CPUSet assignment" podUID="08314cda-c0a3-4a26-bb3a-65cd724335a6" containerName="nova-api-api" Dec 03 14:33:12 crc kubenswrapper[4805]: I1203 14:33:12.004994 4805 memory_manager.go:354] "RemoveStaleState removing state" podUID="08314cda-c0a3-4a26-bb3a-65cd724335a6" containerName="nova-api-log" Dec 03 14:33:12 crc kubenswrapper[4805]: I1203 14:33:12.005008 4805 memory_manager.go:354] "RemoveStaleState removing state" podUID="08314cda-c0a3-4a26-bb3a-65cd724335a6" containerName="nova-api-api" Dec 03 14:33:12 crc kubenswrapper[4805]: I1203 14:33:12.005974 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 03 14:33:12 crc kubenswrapper[4805]: I1203 14:33:12.011858 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-config-data" Dec 03 14:33:12 crc kubenswrapper[4805]: I1203 14:33:12.019161 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Dec 03 14:33:12 crc kubenswrapper[4805]: I1203 14:33:12.076387 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f6f6151a-8cbd-431e-b448-bf70fce67cde-logs\") pod \"nova-api-0\" (UID: \"f6f6151a-8cbd-431e-b448-bf70fce67cde\") " pod="openstack/nova-api-0" Dec 03 14:33:12 crc kubenswrapper[4805]: I1203 14:33:12.076490 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vdklr\" (UniqueName: \"kubernetes.io/projected/f6f6151a-8cbd-431e-b448-bf70fce67cde-kube-api-access-vdklr\") pod \"nova-api-0\" (UID: \"f6f6151a-8cbd-431e-b448-bf70fce67cde\") " pod="openstack/nova-api-0" Dec 03 14:33:12 crc kubenswrapper[4805]: I1203 14:33:12.076551 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f6f6151a-8cbd-431e-b448-bf70fce67cde-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"f6f6151a-8cbd-431e-b448-bf70fce67cde\") " pod="openstack/nova-api-0" Dec 03 14:33:12 crc kubenswrapper[4805]: I1203 14:33:12.076569 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f6f6151a-8cbd-431e-b448-bf70fce67cde-config-data\") pod \"nova-api-0\" (UID: \"f6f6151a-8cbd-431e-b448-bf70fce67cde\") " pod="openstack/nova-api-0" Dec 03 14:33:12 crc kubenswrapper[4805]: I1203 14:33:12.180921 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f6f6151a-8cbd-431e-b448-bf70fce67cde-logs\") pod \"nova-api-0\" (UID: \"f6f6151a-8cbd-431e-b448-bf70fce67cde\") " pod="openstack/nova-api-0" Dec 03 14:33:12 crc kubenswrapper[4805]: I1203 14:33:12.181045 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vdklr\" (UniqueName: \"kubernetes.io/projected/f6f6151a-8cbd-431e-b448-bf70fce67cde-kube-api-access-vdklr\") pod \"nova-api-0\" (UID: \"f6f6151a-8cbd-431e-b448-bf70fce67cde\") " pod="openstack/nova-api-0" Dec 03 14:33:12 crc kubenswrapper[4805]: I1203 14:33:12.181083 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f6f6151a-8cbd-431e-b448-bf70fce67cde-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"f6f6151a-8cbd-431e-b448-bf70fce67cde\") " pod="openstack/nova-api-0" Dec 03 14:33:12 crc kubenswrapper[4805]: I1203 14:33:12.181106 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f6f6151a-8cbd-431e-b448-bf70fce67cde-config-data\") pod \"nova-api-0\" (UID: \"f6f6151a-8cbd-431e-b448-bf70fce67cde\") " pod="openstack/nova-api-0" Dec 03 14:33:12 crc kubenswrapper[4805]: I1203 14:33:12.182231 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f6f6151a-8cbd-431e-b448-bf70fce67cde-logs\") pod \"nova-api-0\" (UID: \"f6f6151a-8cbd-431e-b448-bf70fce67cde\") " pod="openstack/nova-api-0" Dec 03 14:33:12 crc kubenswrapper[4805]: I1203 14:33:12.185748 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f6f6151a-8cbd-431e-b448-bf70fce67cde-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"f6f6151a-8cbd-431e-b448-bf70fce67cde\") " pod="openstack/nova-api-0" Dec 03 14:33:12 crc kubenswrapper[4805]: I1203 14:33:12.186527 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f6f6151a-8cbd-431e-b448-bf70fce67cde-config-data\") pod \"nova-api-0\" (UID: \"f6f6151a-8cbd-431e-b448-bf70fce67cde\") " pod="openstack/nova-api-0" Dec 03 14:33:12 crc kubenswrapper[4805]: I1203 14:33:12.198523 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vdklr\" (UniqueName: \"kubernetes.io/projected/f6f6151a-8cbd-431e-b448-bf70fce67cde-kube-api-access-vdklr\") pod \"nova-api-0\" (UID: \"f6f6151a-8cbd-431e-b448-bf70fce67cde\") " pod="openstack/nova-api-0" Dec 03 14:33:12 crc kubenswrapper[4805]: I1203 14:33:12.364286 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 03 14:33:12 crc kubenswrapper[4805]: I1203 14:33:12.614135 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"c0ccb82f-2439-448d-a958-5d775682cca3","Type":"ContainerStarted","Data":"47fa01711b4e60668d53aa93490a39d0905127c17fbc1ca51fd562ed6c516ebc"} Dec 03 14:33:12 crc kubenswrapper[4805]: I1203 14:33:12.614172 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"c0ccb82f-2439-448d-a958-5d775682cca3","Type":"ContainerStarted","Data":"a738f147993d9d90919dd4b17115af79a515cd7403dea206122ef203e5e01b6c"} Dec 03 14:33:12 crc kubenswrapper[4805]: I1203 14:33:12.652524 4805 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-scheduler-0" podStartSLOduration=2.652490603 podStartE2EDuration="2.652490603s" podCreationTimestamp="2025-12-03 14:33:10 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 14:33:12.648224186 +0000 UTC m=+1422.311141109" watchObservedRunningTime="2025-12-03 14:33:12.652490603 +0000 UTC m=+1422.315407546" Dec 03 14:33:12 crc kubenswrapper[4805]: I1203 14:33:12.705741 4805 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="08314cda-c0a3-4a26-bb3a-65cd724335a6" path="/var/lib/kubelet/pods/08314cda-c0a3-4a26-bb3a-65cd724335a6/volumes" Dec 03 14:33:12 crc kubenswrapper[4805]: I1203 14:33:12.706641 4805 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0ba4f686-dd10-475f-a65a-78041893d75d" path="/var/lib/kubelet/pods/0ba4f686-dd10-475f-a65a-78041893d75d/volumes" Dec 03 14:33:12 crc kubenswrapper[4805]: W1203 14:33:12.821750 4805 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podf6f6151a_8cbd_431e_b448_bf70fce67cde.slice/crio-8c7a9e480902e2ee368ec28411b1d380197f9c3dc31d7645669ec6a80eb4892a WatchSource:0}: Error finding container 8c7a9e480902e2ee368ec28411b1d380197f9c3dc31d7645669ec6a80eb4892a: Status 404 returned error can't find the container with id 8c7a9e480902e2ee368ec28411b1d380197f9c3dc31d7645669ec6a80eb4892a Dec 03 14:33:12 crc kubenswrapper[4805]: I1203 14:33:12.823435 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Dec 03 14:33:13 crc kubenswrapper[4805]: I1203 14:33:13.628904 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"f6f6151a-8cbd-431e-b448-bf70fce67cde","Type":"ContainerStarted","Data":"27fd08f618dd7cc317a2f311da01536f7fdd4aae45c1d3cdb98d85e29257f3a7"} Dec 03 14:33:13 crc kubenswrapper[4805]: I1203 14:33:13.629278 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"f6f6151a-8cbd-431e-b448-bf70fce67cde","Type":"ContainerStarted","Data":"1d2296f16c415ddf1a330122ee2e86cbcf674b0148eae3c574415d0228e09439"} Dec 03 14:33:13 crc kubenswrapper[4805]: I1203 14:33:13.629296 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"f6f6151a-8cbd-431e-b448-bf70fce67cde","Type":"ContainerStarted","Data":"8c7a9e480902e2ee368ec28411b1d380197f9c3dc31d7645669ec6a80eb4892a"} Dec 03 14:33:13 crc kubenswrapper[4805]: I1203 14:33:13.649326 4805 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-0" podStartSLOduration=2.649306991 podStartE2EDuration="2.649306991s" podCreationTimestamp="2025-12-03 14:33:11 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 14:33:13.644727206 +0000 UTC m=+1423.307644129" watchObservedRunningTime="2025-12-03 14:33:13.649306991 +0000 UTC m=+1423.312223914" Dec 03 14:33:13 crc kubenswrapper[4805]: I1203 14:33:13.917526 4805 patch_prober.go:28] interesting pod/machine-config-daemon-gskh4 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 03 14:33:13 crc kubenswrapper[4805]: I1203 14:33:13.917587 4805 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-gskh4" podUID="6098937f-e3f6-45e8-a647-4994a79cd711" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 03 14:33:14 crc kubenswrapper[4805]: I1203 14:33:14.965520 4805 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Dec 03 14:33:14 crc kubenswrapper[4805]: I1203 14:33:14.965921 4805 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Dec 03 14:33:15 crc kubenswrapper[4805]: I1203 14:33:15.619547 4805 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/kube-state-metrics-0"] Dec 03 14:33:15 crc kubenswrapper[4805]: I1203 14:33:15.619787 4805 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/kube-state-metrics-0" podUID="978876d6-1a61-47c1-a0af-eaec298dbe47" containerName="kube-state-metrics" containerID="cri-o://21beea7642ced06d70a91581568a623fd5e9cda04779e8724086621a25095b2f" gracePeriod=30 Dec 03 14:33:15 crc kubenswrapper[4805]: I1203 14:33:15.920306 4805 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-cell1-conductor-0" Dec 03 14:33:16 crc kubenswrapper[4805]: I1203 14:33:16.159994 4805 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Dec 03 14:33:16 crc kubenswrapper[4805]: I1203 14:33:16.260858 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mmj8n\" (UniqueName: \"kubernetes.io/projected/978876d6-1a61-47c1-a0af-eaec298dbe47-kube-api-access-mmj8n\") pod \"978876d6-1a61-47c1-a0af-eaec298dbe47\" (UID: \"978876d6-1a61-47c1-a0af-eaec298dbe47\") " Dec 03 14:33:16 crc kubenswrapper[4805]: I1203 14:33:16.267883 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/978876d6-1a61-47c1-a0af-eaec298dbe47-kube-api-access-mmj8n" (OuterVolumeSpecName: "kube-api-access-mmj8n") pod "978876d6-1a61-47c1-a0af-eaec298dbe47" (UID: "978876d6-1a61-47c1-a0af-eaec298dbe47"). InnerVolumeSpecName "kube-api-access-mmj8n". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 14:33:16 crc kubenswrapper[4805]: I1203 14:33:16.318987 4805 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-scheduler-0" Dec 03 14:33:16 crc kubenswrapper[4805]: I1203 14:33:16.363499 4805 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mmj8n\" (UniqueName: \"kubernetes.io/projected/978876d6-1a61-47c1-a0af-eaec298dbe47-kube-api-access-mmj8n\") on node \"crc\" DevicePath \"\"" Dec 03 14:33:16 crc kubenswrapper[4805]: I1203 14:33:16.685338 4805 generic.go:334] "Generic (PLEG): container finished" podID="978876d6-1a61-47c1-a0af-eaec298dbe47" containerID="21beea7642ced06d70a91581568a623fd5e9cda04779e8724086621a25095b2f" exitCode=2 Dec 03 14:33:16 crc kubenswrapper[4805]: I1203 14:33:16.685393 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"978876d6-1a61-47c1-a0af-eaec298dbe47","Type":"ContainerDied","Data":"21beea7642ced06d70a91581568a623fd5e9cda04779e8724086621a25095b2f"} Dec 03 14:33:16 crc kubenswrapper[4805]: I1203 14:33:16.685425 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"978876d6-1a61-47c1-a0af-eaec298dbe47","Type":"ContainerDied","Data":"4019ce98731b0c3fb180a1b5df437792cc076f25c91ddb833823160217cc22f5"} Dec 03 14:33:16 crc kubenswrapper[4805]: I1203 14:33:16.685444 4805 scope.go:117] "RemoveContainer" containerID="21beea7642ced06d70a91581568a623fd5e9cda04779e8724086621a25095b2f" Dec 03 14:33:16 crc kubenswrapper[4805]: I1203 14:33:16.685599 4805 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Dec 03 14:33:16 crc kubenswrapper[4805]: I1203 14:33:16.722969 4805 scope.go:117] "RemoveContainer" containerID="21beea7642ced06d70a91581568a623fd5e9cda04779e8724086621a25095b2f" Dec 03 14:33:16 crc kubenswrapper[4805]: E1203 14:33:16.723433 4805 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"21beea7642ced06d70a91581568a623fd5e9cda04779e8724086621a25095b2f\": container with ID starting with 21beea7642ced06d70a91581568a623fd5e9cda04779e8724086621a25095b2f not found: ID does not exist" containerID="21beea7642ced06d70a91581568a623fd5e9cda04779e8724086621a25095b2f" Dec 03 14:33:16 crc kubenswrapper[4805]: I1203 14:33:16.723481 4805 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"21beea7642ced06d70a91581568a623fd5e9cda04779e8724086621a25095b2f"} err="failed to get container status \"21beea7642ced06d70a91581568a623fd5e9cda04779e8724086621a25095b2f\": rpc error: code = NotFound desc = could not find container \"21beea7642ced06d70a91581568a623fd5e9cda04779e8724086621a25095b2f\": container with ID starting with 21beea7642ced06d70a91581568a623fd5e9cda04779e8724086621a25095b2f not found: ID does not exist" Dec 03 14:33:16 crc kubenswrapper[4805]: I1203 14:33:16.726209 4805 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/kube-state-metrics-0"] Dec 03 14:33:16 crc kubenswrapper[4805]: I1203 14:33:16.739741 4805 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/kube-state-metrics-0"] Dec 03 14:33:16 crc kubenswrapper[4805]: I1203 14:33:16.761936 4805 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/kube-state-metrics-0"] Dec 03 14:33:16 crc kubenswrapper[4805]: E1203 14:33:16.762371 4805 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="978876d6-1a61-47c1-a0af-eaec298dbe47" containerName="kube-state-metrics" Dec 03 14:33:16 crc kubenswrapper[4805]: I1203 14:33:16.762393 4805 state_mem.go:107] "Deleted CPUSet assignment" podUID="978876d6-1a61-47c1-a0af-eaec298dbe47" containerName="kube-state-metrics" Dec 03 14:33:16 crc kubenswrapper[4805]: I1203 14:33:16.762579 4805 memory_manager.go:354] "RemoveStaleState removing state" podUID="978876d6-1a61-47c1-a0af-eaec298dbe47" containerName="kube-state-metrics" Dec 03 14:33:16 crc kubenswrapper[4805]: I1203 14:33:16.763233 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Dec 03 14:33:16 crc kubenswrapper[4805]: I1203 14:33:16.766543 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"kube-state-metrics-tls-config" Dec 03 14:33:16 crc kubenswrapper[4805]: I1203 14:33:16.766599 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-kube-state-metrics-svc" Dec 03 14:33:16 crc kubenswrapper[4805]: I1203 14:33:16.768247 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/kube-state-metrics-0"] Dec 03 14:33:16 crc kubenswrapper[4805]: I1203 14:33:16.871648 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-state-metrics-tls-certs\" (UniqueName: \"kubernetes.io/secret/91f65152-9858-4a6f-ba45-9e0556490322-kube-state-metrics-tls-certs\") pod \"kube-state-metrics-0\" (UID: \"91f65152-9858-4a6f-ba45-9e0556490322\") " pod="openstack/kube-state-metrics-0" Dec 03 14:33:16 crc kubenswrapper[4805]: I1203 14:33:16.872158 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qrbpn\" (UniqueName: \"kubernetes.io/projected/91f65152-9858-4a6f-ba45-9e0556490322-kube-api-access-qrbpn\") pod \"kube-state-metrics-0\" (UID: \"91f65152-9858-4a6f-ba45-9e0556490322\") " pod="openstack/kube-state-metrics-0" Dec 03 14:33:16 crc kubenswrapper[4805]: I1203 14:33:16.872282 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/91f65152-9858-4a6f-ba45-9e0556490322-combined-ca-bundle\") pod \"kube-state-metrics-0\" (UID: \"91f65152-9858-4a6f-ba45-9e0556490322\") " pod="openstack/kube-state-metrics-0" Dec 03 14:33:16 crc kubenswrapper[4805]: I1203 14:33:16.872322 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-state-metrics-tls-config\" (UniqueName: \"kubernetes.io/secret/91f65152-9858-4a6f-ba45-9e0556490322-kube-state-metrics-tls-config\") pod \"kube-state-metrics-0\" (UID: \"91f65152-9858-4a6f-ba45-9e0556490322\") " pod="openstack/kube-state-metrics-0" Dec 03 14:33:16 crc kubenswrapper[4805]: I1203 14:33:16.974670 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qrbpn\" (UniqueName: \"kubernetes.io/projected/91f65152-9858-4a6f-ba45-9e0556490322-kube-api-access-qrbpn\") pod \"kube-state-metrics-0\" (UID: \"91f65152-9858-4a6f-ba45-9e0556490322\") " pod="openstack/kube-state-metrics-0" Dec 03 14:33:16 crc kubenswrapper[4805]: I1203 14:33:16.974780 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/91f65152-9858-4a6f-ba45-9e0556490322-combined-ca-bundle\") pod \"kube-state-metrics-0\" (UID: \"91f65152-9858-4a6f-ba45-9e0556490322\") " pod="openstack/kube-state-metrics-0" Dec 03 14:33:16 crc kubenswrapper[4805]: I1203 14:33:16.974810 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-state-metrics-tls-config\" (UniqueName: \"kubernetes.io/secret/91f65152-9858-4a6f-ba45-9e0556490322-kube-state-metrics-tls-config\") pod \"kube-state-metrics-0\" (UID: \"91f65152-9858-4a6f-ba45-9e0556490322\") " pod="openstack/kube-state-metrics-0" Dec 03 14:33:16 crc kubenswrapper[4805]: I1203 14:33:16.974900 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-state-metrics-tls-certs\" (UniqueName: \"kubernetes.io/secret/91f65152-9858-4a6f-ba45-9e0556490322-kube-state-metrics-tls-certs\") pod \"kube-state-metrics-0\" (UID: \"91f65152-9858-4a6f-ba45-9e0556490322\") " pod="openstack/kube-state-metrics-0" Dec 03 14:33:16 crc kubenswrapper[4805]: I1203 14:33:16.980537 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-state-metrics-tls-config\" (UniqueName: \"kubernetes.io/secret/91f65152-9858-4a6f-ba45-9e0556490322-kube-state-metrics-tls-config\") pod \"kube-state-metrics-0\" (UID: \"91f65152-9858-4a6f-ba45-9e0556490322\") " pod="openstack/kube-state-metrics-0" Dec 03 14:33:16 crc kubenswrapper[4805]: I1203 14:33:16.983740 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/91f65152-9858-4a6f-ba45-9e0556490322-combined-ca-bundle\") pod \"kube-state-metrics-0\" (UID: \"91f65152-9858-4a6f-ba45-9e0556490322\") " pod="openstack/kube-state-metrics-0" Dec 03 14:33:16 crc kubenswrapper[4805]: I1203 14:33:16.987770 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-state-metrics-tls-certs\" (UniqueName: \"kubernetes.io/secret/91f65152-9858-4a6f-ba45-9e0556490322-kube-state-metrics-tls-certs\") pod \"kube-state-metrics-0\" (UID: \"91f65152-9858-4a6f-ba45-9e0556490322\") " pod="openstack/kube-state-metrics-0" Dec 03 14:33:17 crc kubenswrapper[4805]: I1203 14:33:17.000060 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qrbpn\" (UniqueName: \"kubernetes.io/projected/91f65152-9858-4a6f-ba45-9e0556490322-kube-api-access-qrbpn\") pod \"kube-state-metrics-0\" (UID: \"91f65152-9858-4a6f-ba45-9e0556490322\") " pod="openstack/kube-state-metrics-0" Dec 03 14:33:17 crc kubenswrapper[4805]: I1203 14:33:17.095234 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Dec 03 14:33:17 crc kubenswrapper[4805]: I1203 14:33:17.530365 4805 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 03 14:33:17 crc kubenswrapper[4805]: I1203 14:33:17.530629 4805 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="35025929-9af2-4fd0-a702-76c349a6171e" containerName="ceilometer-central-agent" containerID="cri-o://1c180f0dcd29b37222c5801cb75a7ebee5cc5c54fca7eadd9e4de2fd94aa6937" gracePeriod=30 Dec 03 14:33:17 crc kubenswrapper[4805]: I1203 14:33:17.530728 4805 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="35025929-9af2-4fd0-a702-76c349a6171e" containerName="proxy-httpd" containerID="cri-o://5a88ff238e81b8b27e6ef5d177dbe22e894987d6855096381c3573ca12d3c345" gracePeriod=30 Dec 03 14:33:17 crc kubenswrapper[4805]: I1203 14:33:17.530741 4805 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="35025929-9af2-4fd0-a702-76c349a6171e" containerName="sg-core" containerID="cri-o://cc7beba651205e4083e1d58cc227c37d592eb0211b10fc96c775c4dd53f8a9cb" gracePeriod=30 Dec 03 14:33:17 crc kubenswrapper[4805]: I1203 14:33:17.530758 4805 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="35025929-9af2-4fd0-a702-76c349a6171e" containerName="ceilometer-notification-agent" containerID="cri-o://dfe99b279066a8e9b93a4ec803d7bd9528727ddbf8ae2764b40f68e9fe03172b" gracePeriod=30 Dec 03 14:33:17 crc kubenswrapper[4805]: I1203 14:33:17.639893 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/kube-state-metrics-0"] Dec 03 14:33:17 crc kubenswrapper[4805]: I1203 14:33:17.698650 4805 generic.go:334] "Generic (PLEG): container finished" podID="35025929-9af2-4fd0-a702-76c349a6171e" containerID="5a88ff238e81b8b27e6ef5d177dbe22e894987d6855096381c3573ca12d3c345" exitCode=0 Dec 03 14:33:17 crc kubenswrapper[4805]: I1203 14:33:17.698689 4805 generic.go:334] "Generic (PLEG): container finished" podID="35025929-9af2-4fd0-a702-76c349a6171e" containerID="cc7beba651205e4083e1d58cc227c37d592eb0211b10fc96c775c4dd53f8a9cb" exitCode=2 Dec 03 14:33:17 crc kubenswrapper[4805]: I1203 14:33:17.698717 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"35025929-9af2-4fd0-a702-76c349a6171e","Type":"ContainerDied","Data":"5a88ff238e81b8b27e6ef5d177dbe22e894987d6855096381c3573ca12d3c345"} Dec 03 14:33:17 crc kubenswrapper[4805]: I1203 14:33:17.699168 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"35025929-9af2-4fd0-a702-76c349a6171e","Type":"ContainerDied","Data":"cc7beba651205e4083e1d58cc227c37d592eb0211b10fc96c775c4dd53f8a9cb"} Dec 03 14:33:17 crc kubenswrapper[4805]: I1203 14:33:17.700477 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"91f65152-9858-4a6f-ba45-9e0556490322","Type":"ContainerStarted","Data":"72e9f29f611a0cc161520517119c8f1f472dcf530d4d99b3265474b547a58d05"} Dec 03 14:33:18 crc kubenswrapper[4805]: I1203 14:33:18.712829 4805 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="978876d6-1a61-47c1-a0af-eaec298dbe47" path="/var/lib/kubelet/pods/978876d6-1a61-47c1-a0af-eaec298dbe47/volumes" Dec 03 14:33:18 crc kubenswrapper[4805]: I1203 14:33:18.716779 4805 generic.go:334] "Generic (PLEG): container finished" podID="35025929-9af2-4fd0-a702-76c349a6171e" containerID="1c180f0dcd29b37222c5801cb75a7ebee5cc5c54fca7eadd9e4de2fd94aa6937" exitCode=0 Dec 03 14:33:18 crc kubenswrapper[4805]: I1203 14:33:18.716879 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"35025929-9af2-4fd0-a702-76c349a6171e","Type":"ContainerDied","Data":"1c180f0dcd29b37222c5801cb75a7ebee5cc5c54fca7eadd9e4de2fd94aa6937"} Dec 03 14:33:18 crc kubenswrapper[4805]: I1203 14:33:18.719388 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"91f65152-9858-4a6f-ba45-9e0556490322","Type":"ContainerStarted","Data":"0630fee1a45729f68d0800d01c1c0073753aefb8ffe3aaba1de3ec1f19750d97"} Dec 03 14:33:18 crc kubenswrapper[4805]: I1203 14:33:18.721026 4805 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/kube-state-metrics-0" Dec 03 14:33:18 crc kubenswrapper[4805]: I1203 14:33:18.745817 4805 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/kube-state-metrics-0" podStartSLOduration=2.4031497809999998 podStartE2EDuration="2.745799561s" podCreationTimestamp="2025-12-03 14:33:16 +0000 UTC" firstStartedPulling="2025-12-03 14:33:17.645668874 +0000 UTC m=+1427.308585797" lastFinishedPulling="2025-12-03 14:33:17.988318654 +0000 UTC m=+1427.651235577" observedRunningTime="2025-12-03 14:33:18.742515771 +0000 UTC m=+1428.405432704" watchObservedRunningTime="2025-12-03 14:33:18.745799561 +0000 UTC m=+1428.408716484" Dec 03 14:33:19 crc kubenswrapper[4805]: I1203 14:33:19.966431 4805 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Dec 03 14:33:19 crc kubenswrapper[4805]: I1203 14:33:19.966782 4805 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Dec 03 14:33:20 crc kubenswrapper[4805]: I1203 14:33:20.742648 4805 generic.go:334] "Generic (PLEG): container finished" podID="35025929-9af2-4fd0-a702-76c349a6171e" containerID="dfe99b279066a8e9b93a4ec803d7bd9528727ddbf8ae2764b40f68e9fe03172b" exitCode=0 Dec 03 14:33:20 crc kubenswrapper[4805]: I1203 14:33:20.742743 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"35025929-9af2-4fd0-a702-76c349a6171e","Type":"ContainerDied","Data":"dfe99b279066a8e9b93a4ec803d7bd9528727ddbf8ae2764b40f68e9fe03172b"} Dec 03 14:33:20 crc kubenswrapper[4805]: I1203 14:33:20.983109 4805 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="2dcd6066-cee7-40a7-ba9f-0648b6d2283b" containerName="nova-metadata-metadata" probeResult="failure" output="Get \"https://10.217.0.199:8775/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Dec 03 14:33:20 crc kubenswrapper[4805]: I1203 14:33:20.983186 4805 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="2dcd6066-cee7-40a7-ba9f-0648b6d2283b" containerName="nova-metadata-log" probeResult="failure" output="Get \"https://10.217.0.199:8775/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Dec 03 14:33:21 crc kubenswrapper[4805]: I1203 14:33:21.299189 4805 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 03 14:33:21 crc kubenswrapper[4805]: I1203 14:33:21.318832 4805 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-scheduler-0" Dec 03 14:33:21 crc kubenswrapper[4805]: I1203 14:33:21.369359 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/35025929-9af2-4fd0-a702-76c349a6171e-scripts\") pod \"35025929-9af2-4fd0-a702-76c349a6171e\" (UID: \"35025929-9af2-4fd0-a702-76c349a6171e\") " Dec 03 14:33:21 crc kubenswrapper[4805]: I1203 14:33:21.369415 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/35025929-9af2-4fd0-a702-76c349a6171e-sg-core-conf-yaml\") pod \"35025929-9af2-4fd0-a702-76c349a6171e\" (UID: \"35025929-9af2-4fd0-a702-76c349a6171e\") " Dec 03 14:33:21 crc kubenswrapper[4805]: I1203 14:33:21.369464 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/35025929-9af2-4fd0-a702-76c349a6171e-log-httpd\") pod \"35025929-9af2-4fd0-a702-76c349a6171e\" (UID: \"35025929-9af2-4fd0-a702-76c349a6171e\") " Dec 03 14:33:21 crc kubenswrapper[4805]: I1203 14:33:21.369515 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gsh7r\" (UniqueName: \"kubernetes.io/projected/35025929-9af2-4fd0-a702-76c349a6171e-kube-api-access-gsh7r\") pod \"35025929-9af2-4fd0-a702-76c349a6171e\" (UID: \"35025929-9af2-4fd0-a702-76c349a6171e\") " Dec 03 14:33:21 crc kubenswrapper[4805]: I1203 14:33:21.369571 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/35025929-9af2-4fd0-a702-76c349a6171e-run-httpd\") pod \"35025929-9af2-4fd0-a702-76c349a6171e\" (UID: \"35025929-9af2-4fd0-a702-76c349a6171e\") " Dec 03 14:33:21 crc kubenswrapper[4805]: I1203 14:33:21.369700 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/35025929-9af2-4fd0-a702-76c349a6171e-config-data\") pod \"35025929-9af2-4fd0-a702-76c349a6171e\" (UID: \"35025929-9af2-4fd0-a702-76c349a6171e\") " Dec 03 14:33:21 crc kubenswrapper[4805]: I1203 14:33:21.369750 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/35025929-9af2-4fd0-a702-76c349a6171e-combined-ca-bundle\") pod \"35025929-9af2-4fd0-a702-76c349a6171e\" (UID: \"35025929-9af2-4fd0-a702-76c349a6171e\") " Dec 03 14:33:21 crc kubenswrapper[4805]: I1203 14:33:21.370168 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/35025929-9af2-4fd0-a702-76c349a6171e-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "35025929-9af2-4fd0-a702-76c349a6171e" (UID: "35025929-9af2-4fd0-a702-76c349a6171e"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 14:33:21 crc kubenswrapper[4805]: I1203 14:33:21.370634 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/35025929-9af2-4fd0-a702-76c349a6171e-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "35025929-9af2-4fd0-a702-76c349a6171e" (UID: "35025929-9af2-4fd0-a702-76c349a6171e"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 14:33:21 crc kubenswrapper[4805]: I1203 14:33:21.371717 4805 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-scheduler-0" Dec 03 14:33:21 crc kubenswrapper[4805]: I1203 14:33:21.376286 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/35025929-9af2-4fd0-a702-76c349a6171e-scripts" (OuterVolumeSpecName: "scripts") pod "35025929-9af2-4fd0-a702-76c349a6171e" (UID: "35025929-9af2-4fd0-a702-76c349a6171e"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 14:33:21 crc kubenswrapper[4805]: I1203 14:33:21.377249 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/35025929-9af2-4fd0-a702-76c349a6171e-kube-api-access-gsh7r" (OuterVolumeSpecName: "kube-api-access-gsh7r") pod "35025929-9af2-4fd0-a702-76c349a6171e" (UID: "35025929-9af2-4fd0-a702-76c349a6171e"). InnerVolumeSpecName "kube-api-access-gsh7r". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 14:33:21 crc kubenswrapper[4805]: I1203 14:33:21.441345 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/35025929-9af2-4fd0-a702-76c349a6171e-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "35025929-9af2-4fd0-a702-76c349a6171e" (UID: "35025929-9af2-4fd0-a702-76c349a6171e"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 14:33:21 crc kubenswrapper[4805]: I1203 14:33:21.462465 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/35025929-9af2-4fd0-a702-76c349a6171e-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "35025929-9af2-4fd0-a702-76c349a6171e" (UID: "35025929-9af2-4fd0-a702-76c349a6171e"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 14:33:21 crc kubenswrapper[4805]: I1203 14:33:21.472040 4805 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/35025929-9af2-4fd0-a702-76c349a6171e-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 14:33:21 crc kubenswrapper[4805]: I1203 14:33:21.472070 4805 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/35025929-9af2-4fd0-a702-76c349a6171e-scripts\") on node \"crc\" DevicePath \"\"" Dec 03 14:33:21 crc kubenswrapper[4805]: I1203 14:33:21.472113 4805 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/35025929-9af2-4fd0-a702-76c349a6171e-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Dec 03 14:33:21 crc kubenswrapper[4805]: I1203 14:33:21.472122 4805 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/35025929-9af2-4fd0-a702-76c349a6171e-log-httpd\") on node \"crc\" DevicePath \"\"" Dec 03 14:33:21 crc kubenswrapper[4805]: I1203 14:33:21.472130 4805 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gsh7r\" (UniqueName: \"kubernetes.io/projected/35025929-9af2-4fd0-a702-76c349a6171e-kube-api-access-gsh7r\") on node \"crc\" DevicePath \"\"" Dec 03 14:33:21 crc kubenswrapper[4805]: I1203 14:33:21.472139 4805 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/35025929-9af2-4fd0-a702-76c349a6171e-run-httpd\") on node \"crc\" DevicePath \"\"" Dec 03 14:33:21 crc kubenswrapper[4805]: I1203 14:33:21.504408 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/35025929-9af2-4fd0-a702-76c349a6171e-config-data" (OuterVolumeSpecName: "config-data") pod "35025929-9af2-4fd0-a702-76c349a6171e" (UID: "35025929-9af2-4fd0-a702-76c349a6171e"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 14:33:21 crc kubenswrapper[4805]: I1203 14:33:21.574414 4805 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/35025929-9af2-4fd0-a702-76c349a6171e-config-data\") on node \"crc\" DevicePath \"\"" Dec 03 14:33:21 crc kubenswrapper[4805]: I1203 14:33:21.755131 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"35025929-9af2-4fd0-a702-76c349a6171e","Type":"ContainerDied","Data":"54e46fc7bbb38965aec6ed5e313fc89c61030ebab24c17aa20f9f26ad6dd2028"} Dec 03 14:33:21 crc kubenswrapper[4805]: I1203 14:33:21.755201 4805 scope.go:117] "RemoveContainer" containerID="5a88ff238e81b8b27e6ef5d177dbe22e894987d6855096381c3573ca12d3c345" Dec 03 14:33:21 crc kubenswrapper[4805]: I1203 14:33:21.755151 4805 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 03 14:33:21 crc kubenswrapper[4805]: I1203 14:33:21.779459 4805 scope.go:117] "RemoveContainer" containerID="cc7beba651205e4083e1d58cc227c37d592eb0211b10fc96c775c4dd53f8a9cb" Dec 03 14:33:21 crc kubenswrapper[4805]: I1203 14:33:21.797890 4805 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 03 14:33:21 crc kubenswrapper[4805]: I1203 14:33:21.803041 4805 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-scheduler-0" Dec 03 14:33:21 crc kubenswrapper[4805]: I1203 14:33:21.815078 4805 scope.go:117] "RemoveContainer" containerID="dfe99b279066a8e9b93a4ec803d7bd9528727ddbf8ae2764b40f68e9fe03172b" Dec 03 14:33:21 crc kubenswrapper[4805]: I1203 14:33:21.822007 4805 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Dec 03 14:33:21 crc kubenswrapper[4805]: I1203 14:33:21.832543 4805 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Dec 03 14:33:21 crc kubenswrapper[4805]: E1203 14:33:21.833125 4805 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="35025929-9af2-4fd0-a702-76c349a6171e" containerName="ceilometer-notification-agent" Dec 03 14:33:21 crc kubenswrapper[4805]: I1203 14:33:21.833148 4805 state_mem.go:107] "Deleted CPUSet assignment" podUID="35025929-9af2-4fd0-a702-76c349a6171e" containerName="ceilometer-notification-agent" Dec 03 14:33:21 crc kubenswrapper[4805]: E1203 14:33:21.833173 4805 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="35025929-9af2-4fd0-a702-76c349a6171e" containerName="proxy-httpd" Dec 03 14:33:21 crc kubenswrapper[4805]: I1203 14:33:21.833181 4805 state_mem.go:107] "Deleted CPUSet assignment" podUID="35025929-9af2-4fd0-a702-76c349a6171e" containerName="proxy-httpd" Dec 03 14:33:21 crc kubenswrapper[4805]: E1203 14:33:21.833198 4805 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="35025929-9af2-4fd0-a702-76c349a6171e" containerName="ceilometer-central-agent" Dec 03 14:33:21 crc kubenswrapper[4805]: I1203 14:33:21.833206 4805 state_mem.go:107] "Deleted CPUSet assignment" podUID="35025929-9af2-4fd0-a702-76c349a6171e" containerName="ceilometer-central-agent" Dec 03 14:33:21 crc kubenswrapper[4805]: E1203 14:33:21.833245 4805 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="35025929-9af2-4fd0-a702-76c349a6171e" containerName="sg-core" Dec 03 14:33:21 crc kubenswrapper[4805]: I1203 14:33:21.833253 4805 state_mem.go:107] "Deleted CPUSet assignment" podUID="35025929-9af2-4fd0-a702-76c349a6171e" containerName="sg-core" Dec 03 14:33:21 crc kubenswrapper[4805]: I1203 14:33:21.833472 4805 memory_manager.go:354] "RemoveStaleState removing state" podUID="35025929-9af2-4fd0-a702-76c349a6171e" containerName="ceilometer-notification-agent" Dec 03 14:33:21 crc kubenswrapper[4805]: I1203 14:33:21.833496 4805 memory_manager.go:354] "RemoveStaleState removing state" podUID="35025929-9af2-4fd0-a702-76c349a6171e" containerName="proxy-httpd" Dec 03 14:33:21 crc kubenswrapper[4805]: I1203 14:33:21.833513 4805 memory_manager.go:354] "RemoveStaleState removing state" podUID="35025929-9af2-4fd0-a702-76c349a6171e" containerName="sg-core" Dec 03 14:33:21 crc kubenswrapper[4805]: I1203 14:33:21.833525 4805 memory_manager.go:354] "RemoveStaleState removing state" podUID="35025929-9af2-4fd0-a702-76c349a6171e" containerName="ceilometer-central-agent" Dec 03 14:33:21 crc kubenswrapper[4805]: I1203 14:33:21.835791 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 03 14:33:21 crc kubenswrapper[4805]: I1203 14:33:21.842423 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Dec 03 14:33:21 crc kubenswrapper[4805]: I1203 14:33:21.842613 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ceilometer-internal-svc" Dec 03 14:33:21 crc kubenswrapper[4805]: I1203 14:33:21.842758 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Dec 03 14:33:21 crc kubenswrapper[4805]: I1203 14:33:21.847181 4805 scope.go:117] "RemoveContainer" containerID="1c180f0dcd29b37222c5801cb75a7ebee5cc5c54fca7eadd9e4de2fd94aa6937" Dec 03 14:33:21 crc kubenswrapper[4805]: I1203 14:33:21.847742 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 03 14:33:21 crc kubenswrapper[4805]: I1203 14:33:21.882220 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/0b2edb05-f8c1-4573-aadc-a19a90ec6575-log-httpd\") pod \"ceilometer-0\" (UID: \"0b2edb05-f8c1-4573-aadc-a19a90ec6575\") " pod="openstack/ceilometer-0" Dec 03 14:33:21 crc kubenswrapper[4805]: I1203 14:33:21.882317 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0b2edb05-f8c1-4573-aadc-a19a90ec6575-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"0b2edb05-f8c1-4573-aadc-a19a90ec6575\") " pod="openstack/ceilometer-0" Dec 03 14:33:21 crc kubenswrapper[4805]: I1203 14:33:21.882408 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0b2edb05-f8c1-4573-aadc-a19a90ec6575-config-data\") pod \"ceilometer-0\" (UID: \"0b2edb05-f8c1-4573-aadc-a19a90ec6575\") " pod="openstack/ceilometer-0" Dec 03 14:33:21 crc kubenswrapper[4805]: I1203 14:33:21.882429 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0b2edb05-f8c1-4573-aadc-a19a90ec6575-scripts\") pod \"ceilometer-0\" (UID: \"0b2edb05-f8c1-4573-aadc-a19a90ec6575\") " pod="openstack/ceilometer-0" Dec 03 14:33:21 crc kubenswrapper[4805]: I1203 14:33:21.882573 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/0b2edb05-f8c1-4573-aadc-a19a90ec6575-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"0b2edb05-f8c1-4573-aadc-a19a90ec6575\") " pod="openstack/ceilometer-0" Dec 03 14:33:21 crc kubenswrapper[4805]: I1203 14:33:21.882676 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/0b2edb05-f8c1-4573-aadc-a19a90ec6575-run-httpd\") pod \"ceilometer-0\" (UID: \"0b2edb05-f8c1-4573-aadc-a19a90ec6575\") " pod="openstack/ceilometer-0" Dec 03 14:33:21 crc kubenswrapper[4805]: I1203 14:33:21.882857 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/0b2edb05-f8c1-4573-aadc-a19a90ec6575-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"0b2edb05-f8c1-4573-aadc-a19a90ec6575\") " pod="openstack/ceilometer-0" Dec 03 14:33:21 crc kubenswrapper[4805]: I1203 14:33:21.882901 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4qgpd\" (UniqueName: \"kubernetes.io/projected/0b2edb05-f8c1-4573-aadc-a19a90ec6575-kube-api-access-4qgpd\") pod \"ceilometer-0\" (UID: \"0b2edb05-f8c1-4573-aadc-a19a90ec6575\") " pod="openstack/ceilometer-0" Dec 03 14:33:21 crc kubenswrapper[4805]: I1203 14:33:21.984410 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/0b2edb05-f8c1-4573-aadc-a19a90ec6575-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"0b2edb05-f8c1-4573-aadc-a19a90ec6575\") " pod="openstack/ceilometer-0" Dec 03 14:33:21 crc kubenswrapper[4805]: I1203 14:33:21.984465 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4qgpd\" (UniqueName: \"kubernetes.io/projected/0b2edb05-f8c1-4573-aadc-a19a90ec6575-kube-api-access-4qgpd\") pod \"ceilometer-0\" (UID: \"0b2edb05-f8c1-4573-aadc-a19a90ec6575\") " pod="openstack/ceilometer-0" Dec 03 14:33:21 crc kubenswrapper[4805]: I1203 14:33:21.984523 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/0b2edb05-f8c1-4573-aadc-a19a90ec6575-log-httpd\") pod \"ceilometer-0\" (UID: \"0b2edb05-f8c1-4573-aadc-a19a90ec6575\") " pod="openstack/ceilometer-0" Dec 03 14:33:21 crc kubenswrapper[4805]: I1203 14:33:21.984591 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0b2edb05-f8c1-4573-aadc-a19a90ec6575-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"0b2edb05-f8c1-4573-aadc-a19a90ec6575\") " pod="openstack/ceilometer-0" Dec 03 14:33:21 crc kubenswrapper[4805]: I1203 14:33:21.984635 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0b2edb05-f8c1-4573-aadc-a19a90ec6575-config-data\") pod \"ceilometer-0\" (UID: \"0b2edb05-f8c1-4573-aadc-a19a90ec6575\") " pod="openstack/ceilometer-0" Dec 03 14:33:21 crc kubenswrapper[4805]: I1203 14:33:21.984656 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0b2edb05-f8c1-4573-aadc-a19a90ec6575-scripts\") pod \"ceilometer-0\" (UID: \"0b2edb05-f8c1-4573-aadc-a19a90ec6575\") " pod="openstack/ceilometer-0" Dec 03 14:33:21 crc kubenswrapper[4805]: I1203 14:33:21.984720 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/0b2edb05-f8c1-4573-aadc-a19a90ec6575-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"0b2edb05-f8c1-4573-aadc-a19a90ec6575\") " pod="openstack/ceilometer-0" Dec 03 14:33:21 crc kubenswrapper[4805]: I1203 14:33:21.984748 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/0b2edb05-f8c1-4573-aadc-a19a90ec6575-run-httpd\") pod \"ceilometer-0\" (UID: \"0b2edb05-f8c1-4573-aadc-a19a90ec6575\") " pod="openstack/ceilometer-0" Dec 03 14:33:21 crc kubenswrapper[4805]: I1203 14:33:21.985283 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/0b2edb05-f8c1-4573-aadc-a19a90ec6575-run-httpd\") pod \"ceilometer-0\" (UID: \"0b2edb05-f8c1-4573-aadc-a19a90ec6575\") " pod="openstack/ceilometer-0" Dec 03 14:33:21 crc kubenswrapper[4805]: I1203 14:33:21.985774 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/0b2edb05-f8c1-4573-aadc-a19a90ec6575-log-httpd\") pod \"ceilometer-0\" (UID: \"0b2edb05-f8c1-4573-aadc-a19a90ec6575\") " pod="openstack/ceilometer-0" Dec 03 14:33:21 crc kubenswrapper[4805]: I1203 14:33:21.989341 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/0b2edb05-f8c1-4573-aadc-a19a90ec6575-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"0b2edb05-f8c1-4573-aadc-a19a90ec6575\") " pod="openstack/ceilometer-0" Dec 03 14:33:21 crc kubenswrapper[4805]: I1203 14:33:21.991343 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/0b2edb05-f8c1-4573-aadc-a19a90ec6575-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"0b2edb05-f8c1-4573-aadc-a19a90ec6575\") " pod="openstack/ceilometer-0" Dec 03 14:33:21 crc kubenswrapper[4805]: I1203 14:33:21.991747 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0b2edb05-f8c1-4573-aadc-a19a90ec6575-config-data\") pod \"ceilometer-0\" (UID: \"0b2edb05-f8c1-4573-aadc-a19a90ec6575\") " pod="openstack/ceilometer-0" Dec 03 14:33:21 crc kubenswrapper[4805]: I1203 14:33:21.992434 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0b2edb05-f8c1-4573-aadc-a19a90ec6575-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"0b2edb05-f8c1-4573-aadc-a19a90ec6575\") " pod="openstack/ceilometer-0" Dec 03 14:33:21 crc kubenswrapper[4805]: I1203 14:33:21.993166 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0b2edb05-f8c1-4573-aadc-a19a90ec6575-scripts\") pod \"ceilometer-0\" (UID: \"0b2edb05-f8c1-4573-aadc-a19a90ec6575\") " pod="openstack/ceilometer-0" Dec 03 14:33:22 crc kubenswrapper[4805]: I1203 14:33:22.006737 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4qgpd\" (UniqueName: \"kubernetes.io/projected/0b2edb05-f8c1-4573-aadc-a19a90ec6575-kube-api-access-4qgpd\") pod \"ceilometer-0\" (UID: \"0b2edb05-f8c1-4573-aadc-a19a90ec6575\") " pod="openstack/ceilometer-0" Dec 03 14:33:22 crc kubenswrapper[4805]: I1203 14:33:22.171327 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 03 14:33:22 crc kubenswrapper[4805]: I1203 14:33:22.365071 4805 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Dec 03 14:33:22 crc kubenswrapper[4805]: I1203 14:33:22.365578 4805 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Dec 03 14:33:22 crc kubenswrapper[4805]: I1203 14:33:22.613167 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 03 14:33:22 crc kubenswrapper[4805]: W1203 14:33:22.618304 4805 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod0b2edb05_f8c1_4573_aadc_a19a90ec6575.slice/crio-3997d71ff0ec1f77bfd6b4586b7b83c680e25b72c75ddb6bf5f6779946c60be6 WatchSource:0}: Error finding container 3997d71ff0ec1f77bfd6b4586b7b83c680e25b72c75ddb6bf5f6779946c60be6: Status 404 returned error can't find the container with id 3997d71ff0ec1f77bfd6b4586b7b83c680e25b72c75ddb6bf5f6779946c60be6 Dec 03 14:33:22 crc kubenswrapper[4805]: I1203 14:33:22.710009 4805 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="35025929-9af2-4fd0-a702-76c349a6171e" path="/var/lib/kubelet/pods/35025929-9af2-4fd0-a702-76c349a6171e/volumes" Dec 03 14:33:22 crc kubenswrapper[4805]: I1203 14:33:22.769292 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"0b2edb05-f8c1-4573-aadc-a19a90ec6575","Type":"ContainerStarted","Data":"3997d71ff0ec1f77bfd6b4586b7b83c680e25b72c75ddb6bf5f6779946c60be6"} Dec 03 14:33:23 crc kubenswrapper[4805]: I1203 14:33:23.447064 4805 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="f6f6151a-8cbd-431e-b448-bf70fce67cde" containerName="nova-api-log" probeResult="failure" output="Get \"http://10.217.0.201:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 03 14:33:23 crc kubenswrapper[4805]: I1203 14:33:23.447086 4805 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="f6f6151a-8cbd-431e-b448-bf70fce67cde" containerName="nova-api-api" probeResult="failure" output="Get \"http://10.217.0.201:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 03 14:33:23 crc kubenswrapper[4805]: I1203 14:33:23.781039 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"0b2edb05-f8c1-4573-aadc-a19a90ec6575","Type":"ContainerStarted","Data":"1a04e270afd803e2f0ae6e27199ac82282063350f51f59add0e60f560f536649"} Dec 03 14:33:24 crc kubenswrapper[4805]: I1203 14:33:24.794677 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"0b2edb05-f8c1-4573-aadc-a19a90ec6575","Type":"ContainerStarted","Data":"ecfff5277ed2e0011128d2830b165a01ac0550b1d6a0c7a6dbbaf9a4ad39bfe4"} Dec 03 14:33:25 crc kubenswrapper[4805]: I1203 14:33:25.121198 4805 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-7scbf"] Dec 03 14:33:25 crc kubenswrapper[4805]: I1203 14:33:25.123171 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-7scbf" Dec 03 14:33:25 crc kubenswrapper[4805]: I1203 14:33:25.151607 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-7scbf"] Dec 03 14:33:25 crc kubenswrapper[4805]: I1203 14:33:25.155772 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/94555924-f465-4fb9-aaf1-95afdd1b3a66-catalog-content\") pod \"redhat-marketplace-7scbf\" (UID: \"94555924-f465-4fb9-aaf1-95afdd1b3a66\") " pod="openshift-marketplace/redhat-marketplace-7scbf" Dec 03 14:33:25 crc kubenswrapper[4805]: I1203 14:33:25.155918 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pftll\" (UniqueName: \"kubernetes.io/projected/94555924-f465-4fb9-aaf1-95afdd1b3a66-kube-api-access-pftll\") pod \"redhat-marketplace-7scbf\" (UID: \"94555924-f465-4fb9-aaf1-95afdd1b3a66\") " pod="openshift-marketplace/redhat-marketplace-7scbf" Dec 03 14:33:25 crc kubenswrapper[4805]: I1203 14:33:25.156058 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/94555924-f465-4fb9-aaf1-95afdd1b3a66-utilities\") pod \"redhat-marketplace-7scbf\" (UID: \"94555924-f465-4fb9-aaf1-95afdd1b3a66\") " pod="openshift-marketplace/redhat-marketplace-7scbf" Dec 03 14:33:25 crc kubenswrapper[4805]: I1203 14:33:25.258931 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/94555924-f465-4fb9-aaf1-95afdd1b3a66-utilities\") pod \"redhat-marketplace-7scbf\" (UID: \"94555924-f465-4fb9-aaf1-95afdd1b3a66\") " pod="openshift-marketplace/redhat-marketplace-7scbf" Dec 03 14:33:25 crc kubenswrapper[4805]: I1203 14:33:25.259037 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/94555924-f465-4fb9-aaf1-95afdd1b3a66-catalog-content\") pod \"redhat-marketplace-7scbf\" (UID: \"94555924-f465-4fb9-aaf1-95afdd1b3a66\") " pod="openshift-marketplace/redhat-marketplace-7scbf" Dec 03 14:33:25 crc kubenswrapper[4805]: I1203 14:33:25.259088 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pftll\" (UniqueName: \"kubernetes.io/projected/94555924-f465-4fb9-aaf1-95afdd1b3a66-kube-api-access-pftll\") pod \"redhat-marketplace-7scbf\" (UID: \"94555924-f465-4fb9-aaf1-95afdd1b3a66\") " pod="openshift-marketplace/redhat-marketplace-7scbf" Dec 03 14:33:25 crc kubenswrapper[4805]: I1203 14:33:25.260025 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/94555924-f465-4fb9-aaf1-95afdd1b3a66-utilities\") pod \"redhat-marketplace-7scbf\" (UID: \"94555924-f465-4fb9-aaf1-95afdd1b3a66\") " pod="openshift-marketplace/redhat-marketplace-7scbf" Dec 03 14:33:25 crc kubenswrapper[4805]: I1203 14:33:25.260279 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/94555924-f465-4fb9-aaf1-95afdd1b3a66-catalog-content\") pod \"redhat-marketplace-7scbf\" (UID: \"94555924-f465-4fb9-aaf1-95afdd1b3a66\") " pod="openshift-marketplace/redhat-marketplace-7scbf" Dec 03 14:33:25 crc kubenswrapper[4805]: I1203 14:33:25.281771 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pftll\" (UniqueName: \"kubernetes.io/projected/94555924-f465-4fb9-aaf1-95afdd1b3a66-kube-api-access-pftll\") pod \"redhat-marketplace-7scbf\" (UID: \"94555924-f465-4fb9-aaf1-95afdd1b3a66\") " pod="openshift-marketplace/redhat-marketplace-7scbf" Dec 03 14:33:25 crc kubenswrapper[4805]: I1203 14:33:25.466337 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-7scbf" Dec 03 14:33:25 crc kubenswrapper[4805]: I1203 14:33:25.807489 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"0b2edb05-f8c1-4573-aadc-a19a90ec6575","Type":"ContainerStarted","Data":"4f421ab97684b172723215aa66ba4b646542ef506fa51035c997453656f031f7"} Dec 03 14:33:26 crc kubenswrapper[4805]: I1203 14:33:26.039075 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-7scbf"] Dec 03 14:33:26 crc kubenswrapper[4805]: W1203 14:33:26.039593 4805 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod94555924_f465_4fb9_aaf1_95afdd1b3a66.slice/crio-946fc62a8fd4fd311c3394f51ad5d1dc5f87a57b876303716a5c0f4d89cfbd30 WatchSource:0}: Error finding container 946fc62a8fd4fd311c3394f51ad5d1dc5f87a57b876303716a5c0f4d89cfbd30: Status 404 returned error can't find the container with id 946fc62a8fd4fd311c3394f51ad5d1dc5f87a57b876303716a5c0f4d89cfbd30 Dec 03 14:33:26 crc kubenswrapper[4805]: E1203 14:33:26.565416 4805 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod94555924_f465_4fb9_aaf1_95afdd1b3a66.slice/crio-conmon-c76f027685120547acaa3c7dc4c9af2a48664e21451d91ffac1b0db630b99891.scope\": RecentStats: unable to find data in memory cache]" Dec 03 14:33:26 crc kubenswrapper[4805]: I1203 14:33:26.819049 4805 generic.go:334] "Generic (PLEG): container finished" podID="94555924-f465-4fb9-aaf1-95afdd1b3a66" containerID="c76f027685120547acaa3c7dc4c9af2a48664e21451d91ffac1b0db630b99891" exitCode=0 Dec 03 14:33:26 crc kubenswrapper[4805]: I1203 14:33:26.819133 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-7scbf" event={"ID":"94555924-f465-4fb9-aaf1-95afdd1b3a66","Type":"ContainerDied","Data":"c76f027685120547acaa3c7dc4c9af2a48664e21451d91ffac1b0db630b99891"} Dec 03 14:33:26 crc kubenswrapper[4805]: I1203 14:33:26.821286 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-7scbf" event={"ID":"94555924-f465-4fb9-aaf1-95afdd1b3a66","Type":"ContainerStarted","Data":"946fc62a8fd4fd311c3394f51ad5d1dc5f87a57b876303716a5c0f4d89cfbd30"} Dec 03 14:33:26 crc kubenswrapper[4805]: I1203 14:33:26.826091 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"0b2edb05-f8c1-4573-aadc-a19a90ec6575","Type":"ContainerStarted","Data":"2796d8e7ced38b7980ca7effec53d78ecc6fc7774b9505bbb31b085653f72b91"} Dec 03 14:33:26 crc kubenswrapper[4805]: I1203 14:33:26.826374 4805 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Dec 03 14:33:26 crc kubenswrapper[4805]: I1203 14:33:26.868471 4805 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=2.217675146 podStartE2EDuration="5.868450011s" podCreationTimestamp="2025-12-03 14:33:21 +0000 UTC" firstStartedPulling="2025-12-03 14:33:22.620676222 +0000 UTC m=+1432.283593155" lastFinishedPulling="2025-12-03 14:33:26.271451097 +0000 UTC m=+1435.934368020" observedRunningTime="2025-12-03 14:33:26.863133086 +0000 UTC m=+1436.526050009" watchObservedRunningTime="2025-12-03 14:33:26.868450011 +0000 UTC m=+1436.531366934" Dec 03 14:33:27 crc kubenswrapper[4805]: I1203 14:33:27.106664 4805 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/kube-state-metrics-0" Dec 03 14:33:27 crc kubenswrapper[4805]: I1203 14:33:27.837701 4805 generic.go:334] "Generic (PLEG): container finished" podID="94555924-f465-4fb9-aaf1-95afdd1b3a66" containerID="8a1675db1b86630d89429bb261bfe6f98a7818dab56bb9753962937b40fb4690" exitCode=0 Dec 03 14:33:27 crc kubenswrapper[4805]: I1203 14:33:27.837807 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-7scbf" event={"ID":"94555924-f465-4fb9-aaf1-95afdd1b3a66","Type":"ContainerDied","Data":"8a1675db1b86630d89429bb261bfe6f98a7818dab56bb9753962937b40fb4690"} Dec 03 14:33:28 crc kubenswrapper[4805]: I1203 14:33:28.848888 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-7scbf" event={"ID":"94555924-f465-4fb9-aaf1-95afdd1b3a66","Type":"ContainerStarted","Data":"25d99f4b23894a9130fc530916f9b7235391aa390814fcee160a2ae205de87bc"} Dec 03 14:33:29 crc kubenswrapper[4805]: I1203 14:33:29.713753 4805 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Dec 03 14:33:29 crc kubenswrapper[4805]: I1203 14:33:29.734120 4805 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-7scbf" podStartSLOduration=3.370924503 podStartE2EDuration="4.734099788s" podCreationTimestamp="2025-12-03 14:33:25 +0000 UTC" firstStartedPulling="2025-12-03 14:33:26.822396667 +0000 UTC m=+1436.485313590" lastFinishedPulling="2025-12-03 14:33:28.185571952 +0000 UTC m=+1437.848488875" observedRunningTime="2025-12-03 14:33:28.878126839 +0000 UTC m=+1438.541043762" watchObservedRunningTime="2025-12-03 14:33:29.734099788 +0000 UTC m=+1439.397016711" Dec 03 14:33:29 crc kubenswrapper[4805]: I1203 14:33:29.852139 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-89n8v\" (UniqueName: \"kubernetes.io/projected/660cf06e-323d-478f-aa70-a928724391c3-kube-api-access-89n8v\") pod \"660cf06e-323d-478f-aa70-a928724391c3\" (UID: \"660cf06e-323d-478f-aa70-a928724391c3\") " Dec 03 14:33:29 crc kubenswrapper[4805]: I1203 14:33:29.852211 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/660cf06e-323d-478f-aa70-a928724391c3-combined-ca-bundle\") pod \"660cf06e-323d-478f-aa70-a928724391c3\" (UID: \"660cf06e-323d-478f-aa70-a928724391c3\") " Dec 03 14:33:29 crc kubenswrapper[4805]: I1203 14:33:29.852328 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/660cf06e-323d-478f-aa70-a928724391c3-config-data\") pod \"660cf06e-323d-478f-aa70-a928724391c3\" (UID: \"660cf06e-323d-478f-aa70-a928724391c3\") " Dec 03 14:33:29 crc kubenswrapper[4805]: I1203 14:33:29.858383 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/660cf06e-323d-478f-aa70-a928724391c3-kube-api-access-89n8v" (OuterVolumeSpecName: "kube-api-access-89n8v") pod "660cf06e-323d-478f-aa70-a928724391c3" (UID: "660cf06e-323d-478f-aa70-a928724391c3"). InnerVolumeSpecName "kube-api-access-89n8v". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 14:33:29 crc kubenswrapper[4805]: I1203 14:33:29.859790 4805 generic.go:334] "Generic (PLEG): container finished" podID="660cf06e-323d-478f-aa70-a928724391c3" containerID="d53e632a60c2b214a11cc34dc22d676d6f8ebe2c5767c6b29c796d3c073dff0a" exitCode=137 Dec 03 14:33:29 crc kubenswrapper[4805]: I1203 14:33:29.860718 4805 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Dec 03 14:33:29 crc kubenswrapper[4805]: I1203 14:33:29.861352 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"660cf06e-323d-478f-aa70-a928724391c3","Type":"ContainerDied","Data":"d53e632a60c2b214a11cc34dc22d676d6f8ebe2c5767c6b29c796d3c073dff0a"} Dec 03 14:33:29 crc kubenswrapper[4805]: I1203 14:33:29.861387 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"660cf06e-323d-478f-aa70-a928724391c3","Type":"ContainerDied","Data":"7d614030b5afc43de1d8596363c5c053b120ff5a827a802a64e1ff0b68c00ad5"} Dec 03 14:33:29 crc kubenswrapper[4805]: I1203 14:33:29.861406 4805 scope.go:117] "RemoveContainer" containerID="d53e632a60c2b214a11cc34dc22d676d6f8ebe2c5767c6b29c796d3c073dff0a" Dec 03 14:33:29 crc kubenswrapper[4805]: I1203 14:33:29.882449 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/660cf06e-323d-478f-aa70-a928724391c3-config-data" (OuterVolumeSpecName: "config-data") pod "660cf06e-323d-478f-aa70-a928724391c3" (UID: "660cf06e-323d-478f-aa70-a928724391c3"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 14:33:29 crc kubenswrapper[4805]: I1203 14:33:29.894777 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/660cf06e-323d-478f-aa70-a928724391c3-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "660cf06e-323d-478f-aa70-a928724391c3" (UID: "660cf06e-323d-478f-aa70-a928724391c3"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 14:33:29 crc kubenswrapper[4805]: I1203 14:33:29.954743 4805 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-89n8v\" (UniqueName: \"kubernetes.io/projected/660cf06e-323d-478f-aa70-a928724391c3-kube-api-access-89n8v\") on node \"crc\" DevicePath \"\"" Dec 03 14:33:29 crc kubenswrapper[4805]: I1203 14:33:29.954790 4805 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/660cf06e-323d-478f-aa70-a928724391c3-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 14:33:29 crc kubenswrapper[4805]: I1203 14:33:29.954803 4805 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/660cf06e-323d-478f-aa70-a928724391c3-config-data\") on node \"crc\" DevicePath \"\"" Dec 03 14:33:29 crc kubenswrapper[4805]: I1203 14:33:29.965681 4805 scope.go:117] "RemoveContainer" containerID="d53e632a60c2b214a11cc34dc22d676d6f8ebe2c5767c6b29c796d3c073dff0a" Dec 03 14:33:29 crc kubenswrapper[4805]: E1203 14:33:29.966234 4805 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d53e632a60c2b214a11cc34dc22d676d6f8ebe2c5767c6b29c796d3c073dff0a\": container with ID starting with d53e632a60c2b214a11cc34dc22d676d6f8ebe2c5767c6b29c796d3c073dff0a not found: ID does not exist" containerID="d53e632a60c2b214a11cc34dc22d676d6f8ebe2c5767c6b29c796d3c073dff0a" Dec 03 14:33:29 crc kubenswrapper[4805]: I1203 14:33:29.966338 4805 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d53e632a60c2b214a11cc34dc22d676d6f8ebe2c5767c6b29c796d3c073dff0a"} err="failed to get container status \"d53e632a60c2b214a11cc34dc22d676d6f8ebe2c5767c6b29c796d3c073dff0a\": rpc error: code = NotFound desc = could not find container \"d53e632a60c2b214a11cc34dc22d676d6f8ebe2c5767c6b29c796d3c073dff0a\": container with ID starting with d53e632a60c2b214a11cc34dc22d676d6f8ebe2c5767c6b29c796d3c073dff0a not found: ID does not exist" Dec 03 14:33:29 crc kubenswrapper[4805]: I1203 14:33:29.973369 4805 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-metadata-0" Dec 03 14:33:29 crc kubenswrapper[4805]: I1203 14:33:29.974316 4805 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-metadata-0" Dec 03 14:33:29 crc kubenswrapper[4805]: I1203 14:33:29.978722 4805 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-metadata-0" Dec 03 14:33:30 crc kubenswrapper[4805]: I1203 14:33:30.195400 4805 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Dec 03 14:33:30 crc kubenswrapper[4805]: I1203 14:33:30.205784 4805 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Dec 03 14:33:30 crc kubenswrapper[4805]: I1203 14:33:30.221647 4805 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Dec 03 14:33:30 crc kubenswrapper[4805]: E1203 14:33:30.222150 4805 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="660cf06e-323d-478f-aa70-a928724391c3" containerName="nova-cell1-novncproxy-novncproxy" Dec 03 14:33:30 crc kubenswrapper[4805]: I1203 14:33:30.222173 4805 state_mem.go:107] "Deleted CPUSet assignment" podUID="660cf06e-323d-478f-aa70-a928724391c3" containerName="nova-cell1-novncproxy-novncproxy" Dec 03 14:33:30 crc kubenswrapper[4805]: I1203 14:33:30.222409 4805 memory_manager.go:354] "RemoveStaleState removing state" podUID="660cf06e-323d-478f-aa70-a928724391c3" containerName="nova-cell1-novncproxy-novncproxy" Dec 03 14:33:30 crc kubenswrapper[4805]: I1203 14:33:30.223219 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Dec 03 14:33:30 crc kubenswrapper[4805]: I1203 14:33:30.226590 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-novncproxy-cell1-public-svc" Dec 03 14:33:30 crc kubenswrapper[4805]: I1203 14:33:30.226679 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-novncproxy-cell1-vencrypt" Dec 03 14:33:30 crc kubenswrapper[4805]: I1203 14:33:30.226938 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-novncproxy-config-data" Dec 03 14:33:30 crc kubenswrapper[4805]: I1203 14:33:30.238259 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Dec 03 14:33:30 crc kubenswrapper[4805]: I1203 14:33:30.361753 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4bpcv\" (UniqueName: \"kubernetes.io/projected/d386b865-1a97-468b-ba0e-7733abb94034-kube-api-access-4bpcv\") pod \"nova-cell1-novncproxy-0\" (UID: \"d386b865-1a97-468b-ba0e-7733abb94034\") " pod="openstack/nova-cell1-novncproxy-0" Dec 03 14:33:30 crc kubenswrapper[4805]: I1203 14:33:30.361853 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"vencrypt-tls-certs\" (UniqueName: \"kubernetes.io/secret/d386b865-1a97-468b-ba0e-7733abb94034-vencrypt-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"d386b865-1a97-468b-ba0e-7733abb94034\") " pod="openstack/nova-cell1-novncproxy-0" Dec 03 14:33:30 crc kubenswrapper[4805]: I1203 14:33:30.361935 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-novncproxy-tls-certs\" (UniqueName: \"kubernetes.io/secret/d386b865-1a97-468b-ba0e-7733abb94034-nova-novncproxy-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"d386b865-1a97-468b-ba0e-7733abb94034\") " pod="openstack/nova-cell1-novncproxy-0" Dec 03 14:33:30 crc kubenswrapper[4805]: I1203 14:33:30.362048 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d386b865-1a97-468b-ba0e-7733abb94034-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"d386b865-1a97-468b-ba0e-7733abb94034\") " pod="openstack/nova-cell1-novncproxy-0" Dec 03 14:33:30 crc kubenswrapper[4805]: I1203 14:33:30.362126 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d386b865-1a97-468b-ba0e-7733abb94034-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"d386b865-1a97-468b-ba0e-7733abb94034\") " pod="openstack/nova-cell1-novncproxy-0" Dec 03 14:33:30 crc kubenswrapper[4805]: I1203 14:33:30.464112 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"vencrypt-tls-certs\" (UniqueName: \"kubernetes.io/secret/d386b865-1a97-468b-ba0e-7733abb94034-vencrypt-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"d386b865-1a97-468b-ba0e-7733abb94034\") " pod="openstack/nova-cell1-novncproxy-0" Dec 03 14:33:30 crc kubenswrapper[4805]: I1203 14:33:30.464220 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-novncproxy-tls-certs\" (UniqueName: \"kubernetes.io/secret/d386b865-1a97-468b-ba0e-7733abb94034-nova-novncproxy-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"d386b865-1a97-468b-ba0e-7733abb94034\") " pod="openstack/nova-cell1-novncproxy-0" Dec 03 14:33:30 crc kubenswrapper[4805]: I1203 14:33:30.464258 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d386b865-1a97-468b-ba0e-7733abb94034-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"d386b865-1a97-468b-ba0e-7733abb94034\") " pod="openstack/nova-cell1-novncproxy-0" Dec 03 14:33:30 crc kubenswrapper[4805]: I1203 14:33:30.464308 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d386b865-1a97-468b-ba0e-7733abb94034-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"d386b865-1a97-468b-ba0e-7733abb94034\") " pod="openstack/nova-cell1-novncproxy-0" Dec 03 14:33:30 crc kubenswrapper[4805]: I1203 14:33:30.464426 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4bpcv\" (UniqueName: \"kubernetes.io/projected/d386b865-1a97-468b-ba0e-7733abb94034-kube-api-access-4bpcv\") pod \"nova-cell1-novncproxy-0\" (UID: \"d386b865-1a97-468b-ba0e-7733abb94034\") " pod="openstack/nova-cell1-novncproxy-0" Dec 03 14:33:30 crc kubenswrapper[4805]: I1203 14:33:30.469310 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d386b865-1a97-468b-ba0e-7733abb94034-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"d386b865-1a97-468b-ba0e-7733abb94034\") " pod="openstack/nova-cell1-novncproxy-0" Dec 03 14:33:30 crc kubenswrapper[4805]: I1203 14:33:30.471038 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d386b865-1a97-468b-ba0e-7733abb94034-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"d386b865-1a97-468b-ba0e-7733abb94034\") " pod="openstack/nova-cell1-novncproxy-0" Dec 03 14:33:30 crc kubenswrapper[4805]: I1203 14:33:30.471722 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"vencrypt-tls-certs\" (UniqueName: \"kubernetes.io/secret/d386b865-1a97-468b-ba0e-7733abb94034-vencrypt-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"d386b865-1a97-468b-ba0e-7733abb94034\") " pod="openstack/nova-cell1-novncproxy-0" Dec 03 14:33:30 crc kubenswrapper[4805]: I1203 14:33:30.477541 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-novncproxy-tls-certs\" (UniqueName: \"kubernetes.io/secret/d386b865-1a97-468b-ba0e-7733abb94034-nova-novncproxy-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"d386b865-1a97-468b-ba0e-7733abb94034\") " pod="openstack/nova-cell1-novncproxy-0" Dec 03 14:33:30 crc kubenswrapper[4805]: I1203 14:33:30.503395 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4bpcv\" (UniqueName: \"kubernetes.io/projected/d386b865-1a97-468b-ba0e-7733abb94034-kube-api-access-4bpcv\") pod \"nova-cell1-novncproxy-0\" (UID: \"d386b865-1a97-468b-ba0e-7733abb94034\") " pod="openstack/nova-cell1-novncproxy-0" Dec 03 14:33:30 crc kubenswrapper[4805]: I1203 14:33:30.539534 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Dec 03 14:33:30 crc kubenswrapper[4805]: I1203 14:33:30.707719 4805 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="660cf06e-323d-478f-aa70-a928724391c3" path="/var/lib/kubelet/pods/660cf06e-323d-478f-aa70-a928724391c3/volumes" Dec 03 14:33:30 crc kubenswrapper[4805]: I1203 14:33:30.874085 4805 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-metadata-0" Dec 03 14:33:31 crc kubenswrapper[4805]: I1203 14:33:31.033568 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Dec 03 14:33:31 crc kubenswrapper[4805]: I1203 14:33:31.882026 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"d386b865-1a97-468b-ba0e-7733abb94034","Type":"ContainerStarted","Data":"09d1cc5da07a9fe7ef974de8183d4a2ebf6d5c6b3e324056ee06e66a25507bb1"} Dec 03 14:33:31 crc kubenswrapper[4805]: I1203 14:33:31.884909 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"d386b865-1a97-468b-ba0e-7733abb94034","Type":"ContainerStarted","Data":"50622ea8f2d81f9675a553b87106415d999615fc429f02fcdc8f7be07ddb0c22"} Dec 03 14:33:31 crc kubenswrapper[4805]: I1203 14:33:31.902956 4805 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-novncproxy-0" podStartSLOduration=1.902909321 podStartE2EDuration="1.902909321s" podCreationTimestamp="2025-12-03 14:33:30 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 14:33:31.899885929 +0000 UTC m=+1441.562802862" watchObservedRunningTime="2025-12-03 14:33:31.902909321 +0000 UTC m=+1441.565826254" Dec 03 14:33:32 crc kubenswrapper[4805]: I1203 14:33:32.370900 4805 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Dec 03 14:33:32 crc kubenswrapper[4805]: I1203 14:33:32.371301 4805 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Dec 03 14:33:32 crc kubenswrapper[4805]: I1203 14:33:32.373261 4805 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Dec 03 14:33:32 crc kubenswrapper[4805]: I1203 14:33:32.376155 4805 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Dec 03 14:33:32 crc kubenswrapper[4805]: I1203 14:33:32.896163 4805 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Dec 03 14:33:32 crc kubenswrapper[4805]: I1203 14:33:32.903090 4805 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Dec 03 14:33:33 crc kubenswrapper[4805]: I1203 14:33:33.083785 4805 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-79b5d74c8c-qmdpk"] Dec 03 14:33:33 crc kubenswrapper[4805]: I1203 14:33:33.091590 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-79b5d74c8c-qmdpk" Dec 03 14:33:33 crc kubenswrapper[4805]: I1203 14:33:33.131745 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-79b5d74c8c-qmdpk"] Dec 03 14:33:33 crc kubenswrapper[4805]: I1203 14:33:33.226466 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/73f56cb0-f79a-4e60-ae70-f86f1595a8c9-config\") pod \"dnsmasq-dns-79b5d74c8c-qmdpk\" (UID: \"73f56cb0-f79a-4e60-ae70-f86f1595a8c9\") " pod="openstack/dnsmasq-dns-79b5d74c8c-qmdpk" Dec 03 14:33:33 crc kubenswrapper[4805]: I1203 14:33:33.226517 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cbqwb\" (UniqueName: \"kubernetes.io/projected/73f56cb0-f79a-4e60-ae70-f86f1595a8c9-kube-api-access-cbqwb\") pod \"dnsmasq-dns-79b5d74c8c-qmdpk\" (UID: \"73f56cb0-f79a-4e60-ae70-f86f1595a8c9\") " pod="openstack/dnsmasq-dns-79b5d74c8c-qmdpk" Dec 03 14:33:33 crc kubenswrapper[4805]: I1203 14:33:33.226592 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/73f56cb0-f79a-4e60-ae70-f86f1595a8c9-ovsdbserver-sb\") pod \"dnsmasq-dns-79b5d74c8c-qmdpk\" (UID: \"73f56cb0-f79a-4e60-ae70-f86f1595a8c9\") " pod="openstack/dnsmasq-dns-79b5d74c8c-qmdpk" Dec 03 14:33:33 crc kubenswrapper[4805]: I1203 14:33:33.226615 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/73f56cb0-f79a-4e60-ae70-f86f1595a8c9-dns-svc\") pod \"dnsmasq-dns-79b5d74c8c-qmdpk\" (UID: \"73f56cb0-f79a-4e60-ae70-f86f1595a8c9\") " pod="openstack/dnsmasq-dns-79b5d74c8c-qmdpk" Dec 03 14:33:33 crc kubenswrapper[4805]: I1203 14:33:33.226715 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/73f56cb0-f79a-4e60-ae70-f86f1595a8c9-dns-swift-storage-0\") pod \"dnsmasq-dns-79b5d74c8c-qmdpk\" (UID: \"73f56cb0-f79a-4e60-ae70-f86f1595a8c9\") " pod="openstack/dnsmasq-dns-79b5d74c8c-qmdpk" Dec 03 14:33:33 crc kubenswrapper[4805]: I1203 14:33:33.226766 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/73f56cb0-f79a-4e60-ae70-f86f1595a8c9-ovsdbserver-nb\") pod \"dnsmasq-dns-79b5d74c8c-qmdpk\" (UID: \"73f56cb0-f79a-4e60-ae70-f86f1595a8c9\") " pod="openstack/dnsmasq-dns-79b5d74c8c-qmdpk" Dec 03 14:33:33 crc kubenswrapper[4805]: I1203 14:33:33.329077 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/73f56cb0-f79a-4e60-ae70-f86f1595a8c9-dns-swift-storage-0\") pod \"dnsmasq-dns-79b5d74c8c-qmdpk\" (UID: \"73f56cb0-f79a-4e60-ae70-f86f1595a8c9\") " pod="openstack/dnsmasq-dns-79b5d74c8c-qmdpk" Dec 03 14:33:33 crc kubenswrapper[4805]: I1203 14:33:33.329131 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/73f56cb0-f79a-4e60-ae70-f86f1595a8c9-ovsdbserver-nb\") pod \"dnsmasq-dns-79b5d74c8c-qmdpk\" (UID: \"73f56cb0-f79a-4e60-ae70-f86f1595a8c9\") " pod="openstack/dnsmasq-dns-79b5d74c8c-qmdpk" Dec 03 14:33:33 crc kubenswrapper[4805]: I1203 14:33:33.329238 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/73f56cb0-f79a-4e60-ae70-f86f1595a8c9-config\") pod \"dnsmasq-dns-79b5d74c8c-qmdpk\" (UID: \"73f56cb0-f79a-4e60-ae70-f86f1595a8c9\") " pod="openstack/dnsmasq-dns-79b5d74c8c-qmdpk" Dec 03 14:33:33 crc kubenswrapper[4805]: I1203 14:33:33.329260 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cbqwb\" (UniqueName: \"kubernetes.io/projected/73f56cb0-f79a-4e60-ae70-f86f1595a8c9-kube-api-access-cbqwb\") pod \"dnsmasq-dns-79b5d74c8c-qmdpk\" (UID: \"73f56cb0-f79a-4e60-ae70-f86f1595a8c9\") " pod="openstack/dnsmasq-dns-79b5d74c8c-qmdpk" Dec 03 14:33:33 crc kubenswrapper[4805]: I1203 14:33:33.329290 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/73f56cb0-f79a-4e60-ae70-f86f1595a8c9-dns-svc\") pod \"dnsmasq-dns-79b5d74c8c-qmdpk\" (UID: \"73f56cb0-f79a-4e60-ae70-f86f1595a8c9\") " pod="openstack/dnsmasq-dns-79b5d74c8c-qmdpk" Dec 03 14:33:33 crc kubenswrapper[4805]: I1203 14:33:33.329305 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/73f56cb0-f79a-4e60-ae70-f86f1595a8c9-ovsdbserver-sb\") pod \"dnsmasq-dns-79b5d74c8c-qmdpk\" (UID: \"73f56cb0-f79a-4e60-ae70-f86f1595a8c9\") " pod="openstack/dnsmasq-dns-79b5d74c8c-qmdpk" Dec 03 14:33:33 crc kubenswrapper[4805]: I1203 14:33:33.330411 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/73f56cb0-f79a-4e60-ae70-f86f1595a8c9-ovsdbserver-sb\") pod \"dnsmasq-dns-79b5d74c8c-qmdpk\" (UID: \"73f56cb0-f79a-4e60-ae70-f86f1595a8c9\") " pod="openstack/dnsmasq-dns-79b5d74c8c-qmdpk" Dec 03 14:33:33 crc kubenswrapper[4805]: I1203 14:33:33.330420 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/73f56cb0-f79a-4e60-ae70-f86f1595a8c9-dns-swift-storage-0\") pod \"dnsmasq-dns-79b5d74c8c-qmdpk\" (UID: \"73f56cb0-f79a-4e60-ae70-f86f1595a8c9\") " pod="openstack/dnsmasq-dns-79b5d74c8c-qmdpk" Dec 03 14:33:33 crc kubenswrapper[4805]: I1203 14:33:33.330420 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/73f56cb0-f79a-4e60-ae70-f86f1595a8c9-ovsdbserver-nb\") pod \"dnsmasq-dns-79b5d74c8c-qmdpk\" (UID: \"73f56cb0-f79a-4e60-ae70-f86f1595a8c9\") " pod="openstack/dnsmasq-dns-79b5d74c8c-qmdpk" Dec 03 14:33:33 crc kubenswrapper[4805]: I1203 14:33:33.330464 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/73f56cb0-f79a-4e60-ae70-f86f1595a8c9-dns-svc\") pod \"dnsmasq-dns-79b5d74c8c-qmdpk\" (UID: \"73f56cb0-f79a-4e60-ae70-f86f1595a8c9\") " pod="openstack/dnsmasq-dns-79b5d74c8c-qmdpk" Dec 03 14:33:33 crc kubenswrapper[4805]: I1203 14:33:33.331105 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/73f56cb0-f79a-4e60-ae70-f86f1595a8c9-config\") pod \"dnsmasq-dns-79b5d74c8c-qmdpk\" (UID: \"73f56cb0-f79a-4e60-ae70-f86f1595a8c9\") " pod="openstack/dnsmasq-dns-79b5d74c8c-qmdpk" Dec 03 14:33:33 crc kubenswrapper[4805]: I1203 14:33:33.358678 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cbqwb\" (UniqueName: \"kubernetes.io/projected/73f56cb0-f79a-4e60-ae70-f86f1595a8c9-kube-api-access-cbqwb\") pod \"dnsmasq-dns-79b5d74c8c-qmdpk\" (UID: \"73f56cb0-f79a-4e60-ae70-f86f1595a8c9\") " pod="openstack/dnsmasq-dns-79b5d74c8c-qmdpk" Dec 03 14:33:33 crc kubenswrapper[4805]: I1203 14:33:33.421285 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-79b5d74c8c-qmdpk" Dec 03 14:33:33 crc kubenswrapper[4805]: W1203 14:33:33.905196 4805 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod73f56cb0_f79a_4e60_ae70_f86f1595a8c9.slice/crio-d1c6c41b5239625860972078b347eb1f56398f99e7eb6345fb420006a803c4c0 WatchSource:0}: Error finding container d1c6c41b5239625860972078b347eb1f56398f99e7eb6345fb420006a803c4c0: Status 404 returned error can't find the container with id d1c6c41b5239625860972078b347eb1f56398f99e7eb6345fb420006a803c4c0 Dec 03 14:33:33 crc kubenswrapper[4805]: I1203 14:33:33.909532 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-79b5d74c8c-qmdpk"] Dec 03 14:33:34 crc kubenswrapper[4805]: I1203 14:33:34.923090 4805 generic.go:334] "Generic (PLEG): container finished" podID="73f56cb0-f79a-4e60-ae70-f86f1595a8c9" containerID="8d5a86cad745b1c5d43012a05e252adf40a5ecc7c85e407457caa86d1b0e9dc4" exitCode=0 Dec 03 14:33:34 crc kubenswrapper[4805]: I1203 14:33:34.923152 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-79b5d74c8c-qmdpk" event={"ID":"73f56cb0-f79a-4e60-ae70-f86f1595a8c9","Type":"ContainerDied","Data":"8d5a86cad745b1c5d43012a05e252adf40a5ecc7c85e407457caa86d1b0e9dc4"} Dec 03 14:33:34 crc kubenswrapper[4805]: I1203 14:33:34.923693 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-79b5d74c8c-qmdpk" event={"ID":"73f56cb0-f79a-4e60-ae70-f86f1595a8c9","Type":"ContainerStarted","Data":"d1c6c41b5239625860972078b347eb1f56398f99e7eb6345fb420006a803c4c0"} Dec 03 14:33:35 crc kubenswrapper[4805]: I1203 14:33:35.350265 4805 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Dec 03 14:33:35 crc kubenswrapper[4805]: I1203 14:33:35.466824 4805 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-7scbf" Dec 03 14:33:35 crc kubenswrapper[4805]: I1203 14:33:35.466887 4805 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-7scbf" Dec 03 14:33:35 crc kubenswrapper[4805]: I1203 14:33:35.525434 4805 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-7scbf" Dec 03 14:33:35 crc kubenswrapper[4805]: I1203 14:33:35.539735 4805 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell1-novncproxy-0" Dec 03 14:33:35 crc kubenswrapper[4805]: I1203 14:33:35.938676 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-79b5d74c8c-qmdpk" event={"ID":"73f56cb0-f79a-4e60-ae70-f86f1595a8c9","Type":"ContainerStarted","Data":"0b641002c33e06eb4a00fef3b57217534b7ea1fc136c66dcc4bf637adfa49dee"} Dec 03 14:33:35 crc kubenswrapper[4805]: I1203 14:33:35.938876 4805 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="f6f6151a-8cbd-431e-b448-bf70fce67cde" containerName="nova-api-log" containerID="cri-o://1d2296f16c415ddf1a330122ee2e86cbcf674b0148eae3c574415d0228e09439" gracePeriod=30 Dec 03 14:33:35 crc kubenswrapper[4805]: I1203 14:33:35.938932 4805 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="f6f6151a-8cbd-431e-b448-bf70fce67cde" containerName="nova-api-api" containerID="cri-o://27fd08f618dd7cc317a2f311da01536f7fdd4aae45c1d3cdb98d85e29257f3a7" gracePeriod=30 Dec 03 14:33:35 crc kubenswrapper[4805]: I1203 14:33:35.969100 4805 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-79b5d74c8c-qmdpk" podStartSLOduration=2.969080493 podStartE2EDuration="2.969080493s" podCreationTimestamp="2025-12-03 14:33:33 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 14:33:35.963605824 +0000 UTC m=+1445.626522747" watchObservedRunningTime="2025-12-03 14:33:35.969080493 +0000 UTC m=+1445.631997416" Dec 03 14:33:35 crc kubenswrapper[4805]: I1203 14:33:35.998277 4805 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-7scbf" Dec 03 14:33:36 crc kubenswrapper[4805]: I1203 14:33:36.046991 4805 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-7scbf"] Dec 03 14:33:36 crc kubenswrapper[4805]: I1203 14:33:36.474217 4805 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 03 14:33:36 crc kubenswrapper[4805]: I1203 14:33:36.474766 4805 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="0b2edb05-f8c1-4573-aadc-a19a90ec6575" containerName="ceilometer-central-agent" containerID="cri-o://1a04e270afd803e2f0ae6e27199ac82282063350f51f59add0e60f560f536649" gracePeriod=30 Dec 03 14:33:36 crc kubenswrapper[4805]: I1203 14:33:36.474925 4805 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="0b2edb05-f8c1-4573-aadc-a19a90ec6575" containerName="proxy-httpd" containerID="cri-o://2796d8e7ced38b7980ca7effec53d78ecc6fc7774b9505bbb31b085653f72b91" gracePeriod=30 Dec 03 14:33:36 crc kubenswrapper[4805]: I1203 14:33:36.475042 4805 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="0b2edb05-f8c1-4573-aadc-a19a90ec6575" containerName="sg-core" containerID="cri-o://4f421ab97684b172723215aa66ba4b646542ef506fa51035c997453656f031f7" gracePeriod=30 Dec 03 14:33:36 crc kubenswrapper[4805]: I1203 14:33:36.475017 4805 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="0b2edb05-f8c1-4573-aadc-a19a90ec6575" containerName="ceilometer-notification-agent" containerID="cri-o://ecfff5277ed2e0011128d2830b165a01ac0550b1d6a0c7a6dbbaf9a4ad39bfe4" gracePeriod=30 Dec 03 14:33:36 crc kubenswrapper[4805]: I1203 14:33:36.489152 4805 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/ceilometer-0" podUID="0b2edb05-f8c1-4573-aadc-a19a90ec6575" containerName="proxy-httpd" probeResult="failure" output="Get \"https://10.217.0.203:3000/\": EOF" Dec 03 14:33:36 crc kubenswrapper[4805]: E1203 14:33:36.827292 4805 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod0b2edb05_f8c1_4573_aadc_a19a90ec6575.slice/crio-conmon-2796d8e7ced38b7980ca7effec53d78ecc6fc7774b9505bbb31b085653f72b91.scope\": RecentStats: unable to find data in memory cache]" Dec 03 14:33:36 crc kubenswrapper[4805]: I1203 14:33:36.952140 4805 generic.go:334] "Generic (PLEG): container finished" podID="0b2edb05-f8c1-4573-aadc-a19a90ec6575" containerID="2796d8e7ced38b7980ca7effec53d78ecc6fc7774b9505bbb31b085653f72b91" exitCode=0 Dec 03 14:33:36 crc kubenswrapper[4805]: I1203 14:33:36.952178 4805 generic.go:334] "Generic (PLEG): container finished" podID="0b2edb05-f8c1-4573-aadc-a19a90ec6575" containerID="4f421ab97684b172723215aa66ba4b646542ef506fa51035c997453656f031f7" exitCode=2 Dec 03 14:33:36 crc kubenswrapper[4805]: I1203 14:33:36.952186 4805 generic.go:334] "Generic (PLEG): container finished" podID="0b2edb05-f8c1-4573-aadc-a19a90ec6575" containerID="1a04e270afd803e2f0ae6e27199ac82282063350f51f59add0e60f560f536649" exitCode=0 Dec 03 14:33:36 crc kubenswrapper[4805]: I1203 14:33:36.952186 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"0b2edb05-f8c1-4573-aadc-a19a90ec6575","Type":"ContainerDied","Data":"2796d8e7ced38b7980ca7effec53d78ecc6fc7774b9505bbb31b085653f72b91"} Dec 03 14:33:36 crc kubenswrapper[4805]: I1203 14:33:36.952236 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"0b2edb05-f8c1-4573-aadc-a19a90ec6575","Type":"ContainerDied","Data":"4f421ab97684b172723215aa66ba4b646542ef506fa51035c997453656f031f7"} Dec 03 14:33:36 crc kubenswrapper[4805]: I1203 14:33:36.952249 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"0b2edb05-f8c1-4573-aadc-a19a90ec6575","Type":"ContainerDied","Data":"1a04e270afd803e2f0ae6e27199ac82282063350f51f59add0e60f560f536649"} Dec 03 14:33:36 crc kubenswrapper[4805]: I1203 14:33:36.954293 4805 generic.go:334] "Generic (PLEG): container finished" podID="f6f6151a-8cbd-431e-b448-bf70fce67cde" containerID="1d2296f16c415ddf1a330122ee2e86cbcf674b0148eae3c574415d0228e09439" exitCode=143 Dec 03 14:33:36 crc kubenswrapper[4805]: I1203 14:33:36.954329 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"f6f6151a-8cbd-431e-b448-bf70fce67cde","Type":"ContainerDied","Data":"1d2296f16c415ddf1a330122ee2e86cbcf674b0148eae3c574415d0228e09439"} Dec 03 14:33:36 crc kubenswrapper[4805]: I1203 14:33:36.954644 4805 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-79b5d74c8c-qmdpk" Dec 03 14:33:37 crc kubenswrapper[4805]: I1203 14:33:37.708797 4805 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 03 14:33:37 crc kubenswrapper[4805]: I1203 14:33:37.819156 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/0b2edb05-f8c1-4573-aadc-a19a90ec6575-run-httpd\") pod \"0b2edb05-f8c1-4573-aadc-a19a90ec6575\" (UID: \"0b2edb05-f8c1-4573-aadc-a19a90ec6575\") " Dec 03 14:33:37 crc kubenswrapper[4805]: I1203 14:33:37.819233 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0b2edb05-f8c1-4573-aadc-a19a90ec6575-scripts\") pod \"0b2edb05-f8c1-4573-aadc-a19a90ec6575\" (UID: \"0b2edb05-f8c1-4573-aadc-a19a90ec6575\") " Dec 03 14:33:37 crc kubenswrapper[4805]: I1203 14:33:37.819254 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0b2edb05-f8c1-4573-aadc-a19a90ec6575-config-data\") pod \"0b2edb05-f8c1-4573-aadc-a19a90ec6575\" (UID: \"0b2edb05-f8c1-4573-aadc-a19a90ec6575\") " Dec 03 14:33:37 crc kubenswrapper[4805]: I1203 14:33:37.819356 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4qgpd\" (UniqueName: \"kubernetes.io/projected/0b2edb05-f8c1-4573-aadc-a19a90ec6575-kube-api-access-4qgpd\") pod \"0b2edb05-f8c1-4573-aadc-a19a90ec6575\" (UID: \"0b2edb05-f8c1-4573-aadc-a19a90ec6575\") " Dec 03 14:33:37 crc kubenswrapper[4805]: I1203 14:33:37.819469 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/0b2edb05-f8c1-4573-aadc-a19a90ec6575-log-httpd\") pod \"0b2edb05-f8c1-4573-aadc-a19a90ec6575\" (UID: \"0b2edb05-f8c1-4573-aadc-a19a90ec6575\") " Dec 03 14:33:37 crc kubenswrapper[4805]: I1203 14:33:37.819492 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/0b2edb05-f8c1-4573-aadc-a19a90ec6575-sg-core-conf-yaml\") pod \"0b2edb05-f8c1-4573-aadc-a19a90ec6575\" (UID: \"0b2edb05-f8c1-4573-aadc-a19a90ec6575\") " Dec 03 14:33:37 crc kubenswrapper[4805]: I1203 14:33:37.819530 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/0b2edb05-f8c1-4573-aadc-a19a90ec6575-ceilometer-tls-certs\") pod \"0b2edb05-f8c1-4573-aadc-a19a90ec6575\" (UID: \"0b2edb05-f8c1-4573-aadc-a19a90ec6575\") " Dec 03 14:33:37 crc kubenswrapper[4805]: I1203 14:33:37.819556 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0b2edb05-f8c1-4573-aadc-a19a90ec6575-combined-ca-bundle\") pod \"0b2edb05-f8c1-4573-aadc-a19a90ec6575\" (UID: \"0b2edb05-f8c1-4573-aadc-a19a90ec6575\") " Dec 03 14:33:37 crc kubenswrapper[4805]: I1203 14:33:37.819872 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0b2edb05-f8c1-4573-aadc-a19a90ec6575-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "0b2edb05-f8c1-4573-aadc-a19a90ec6575" (UID: "0b2edb05-f8c1-4573-aadc-a19a90ec6575"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 14:33:37 crc kubenswrapper[4805]: I1203 14:33:37.820368 4805 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/0b2edb05-f8c1-4573-aadc-a19a90ec6575-log-httpd\") on node \"crc\" DevicePath \"\"" Dec 03 14:33:37 crc kubenswrapper[4805]: I1203 14:33:37.820734 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0b2edb05-f8c1-4573-aadc-a19a90ec6575-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "0b2edb05-f8c1-4573-aadc-a19a90ec6575" (UID: "0b2edb05-f8c1-4573-aadc-a19a90ec6575"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 14:33:37 crc kubenswrapper[4805]: I1203 14:33:37.825419 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0b2edb05-f8c1-4573-aadc-a19a90ec6575-scripts" (OuterVolumeSpecName: "scripts") pod "0b2edb05-f8c1-4573-aadc-a19a90ec6575" (UID: "0b2edb05-f8c1-4573-aadc-a19a90ec6575"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 14:33:37 crc kubenswrapper[4805]: I1203 14:33:37.825849 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0b2edb05-f8c1-4573-aadc-a19a90ec6575-kube-api-access-4qgpd" (OuterVolumeSpecName: "kube-api-access-4qgpd") pod "0b2edb05-f8c1-4573-aadc-a19a90ec6575" (UID: "0b2edb05-f8c1-4573-aadc-a19a90ec6575"). InnerVolumeSpecName "kube-api-access-4qgpd". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 14:33:37 crc kubenswrapper[4805]: I1203 14:33:37.857493 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0b2edb05-f8c1-4573-aadc-a19a90ec6575-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "0b2edb05-f8c1-4573-aadc-a19a90ec6575" (UID: "0b2edb05-f8c1-4573-aadc-a19a90ec6575"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 14:33:37 crc kubenswrapper[4805]: I1203 14:33:37.883146 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0b2edb05-f8c1-4573-aadc-a19a90ec6575-ceilometer-tls-certs" (OuterVolumeSpecName: "ceilometer-tls-certs") pod "0b2edb05-f8c1-4573-aadc-a19a90ec6575" (UID: "0b2edb05-f8c1-4573-aadc-a19a90ec6575"). InnerVolumeSpecName "ceilometer-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 14:33:37 crc kubenswrapper[4805]: I1203 14:33:37.915814 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0b2edb05-f8c1-4573-aadc-a19a90ec6575-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "0b2edb05-f8c1-4573-aadc-a19a90ec6575" (UID: "0b2edb05-f8c1-4573-aadc-a19a90ec6575"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 14:33:37 crc kubenswrapper[4805]: I1203 14:33:37.922346 4805 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/0b2edb05-f8c1-4573-aadc-a19a90ec6575-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Dec 03 14:33:37 crc kubenswrapper[4805]: I1203 14:33:37.922697 4805 reconciler_common.go:293] "Volume detached for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/0b2edb05-f8c1-4573-aadc-a19a90ec6575-ceilometer-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 03 14:33:37 crc kubenswrapper[4805]: I1203 14:33:37.922711 4805 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0b2edb05-f8c1-4573-aadc-a19a90ec6575-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 14:33:37 crc kubenswrapper[4805]: I1203 14:33:37.922722 4805 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/0b2edb05-f8c1-4573-aadc-a19a90ec6575-run-httpd\") on node \"crc\" DevicePath \"\"" Dec 03 14:33:37 crc kubenswrapper[4805]: I1203 14:33:37.922736 4805 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0b2edb05-f8c1-4573-aadc-a19a90ec6575-scripts\") on node \"crc\" DevicePath \"\"" Dec 03 14:33:37 crc kubenswrapper[4805]: I1203 14:33:37.922747 4805 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4qgpd\" (UniqueName: \"kubernetes.io/projected/0b2edb05-f8c1-4573-aadc-a19a90ec6575-kube-api-access-4qgpd\") on node \"crc\" DevicePath \"\"" Dec 03 14:33:37 crc kubenswrapper[4805]: I1203 14:33:37.930820 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0b2edb05-f8c1-4573-aadc-a19a90ec6575-config-data" (OuterVolumeSpecName: "config-data") pod "0b2edb05-f8c1-4573-aadc-a19a90ec6575" (UID: "0b2edb05-f8c1-4573-aadc-a19a90ec6575"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 14:33:37 crc kubenswrapper[4805]: I1203 14:33:37.966961 4805 generic.go:334] "Generic (PLEG): container finished" podID="0b2edb05-f8c1-4573-aadc-a19a90ec6575" containerID="ecfff5277ed2e0011128d2830b165a01ac0550b1d6a0c7a6dbbaf9a4ad39bfe4" exitCode=0 Dec 03 14:33:37 crc kubenswrapper[4805]: I1203 14:33:37.967048 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"0b2edb05-f8c1-4573-aadc-a19a90ec6575","Type":"ContainerDied","Data":"ecfff5277ed2e0011128d2830b165a01ac0550b1d6a0c7a6dbbaf9a4ad39bfe4"} Dec 03 14:33:37 crc kubenswrapper[4805]: I1203 14:33:37.967103 4805 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 03 14:33:37 crc kubenswrapper[4805]: I1203 14:33:37.968246 4805 scope.go:117] "RemoveContainer" containerID="2796d8e7ced38b7980ca7effec53d78ecc6fc7774b9505bbb31b085653f72b91" Dec 03 14:33:37 crc kubenswrapper[4805]: I1203 14:33:37.968811 4805 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-7scbf" podUID="94555924-f465-4fb9-aaf1-95afdd1b3a66" containerName="registry-server" containerID="cri-o://25d99f4b23894a9130fc530916f9b7235391aa390814fcee160a2ae205de87bc" gracePeriod=2 Dec 03 14:33:37 crc kubenswrapper[4805]: I1203 14:33:37.972908 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"0b2edb05-f8c1-4573-aadc-a19a90ec6575","Type":"ContainerDied","Data":"3997d71ff0ec1f77bfd6b4586b7b83c680e25b72c75ddb6bf5f6779946c60be6"} Dec 03 14:33:37 crc kubenswrapper[4805]: I1203 14:33:37.999122 4805 scope.go:117] "RemoveContainer" containerID="4f421ab97684b172723215aa66ba4b646542ef506fa51035c997453656f031f7" Dec 03 14:33:38 crc kubenswrapper[4805]: I1203 14:33:38.016863 4805 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 03 14:33:38 crc kubenswrapper[4805]: I1203 14:33:38.023740 4805 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0b2edb05-f8c1-4573-aadc-a19a90ec6575-config-data\") on node \"crc\" DevicePath \"\"" Dec 03 14:33:38 crc kubenswrapper[4805]: I1203 14:33:38.030245 4805 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Dec 03 14:33:38 crc kubenswrapper[4805]: I1203 14:33:38.042910 4805 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Dec 03 14:33:38 crc kubenswrapper[4805]: E1203 14:33:38.043486 4805 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0b2edb05-f8c1-4573-aadc-a19a90ec6575" containerName="sg-core" Dec 03 14:33:38 crc kubenswrapper[4805]: I1203 14:33:38.043503 4805 state_mem.go:107] "Deleted CPUSet assignment" podUID="0b2edb05-f8c1-4573-aadc-a19a90ec6575" containerName="sg-core" Dec 03 14:33:38 crc kubenswrapper[4805]: E1203 14:33:38.043517 4805 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0b2edb05-f8c1-4573-aadc-a19a90ec6575" containerName="ceilometer-central-agent" Dec 03 14:33:38 crc kubenswrapper[4805]: I1203 14:33:38.043524 4805 state_mem.go:107] "Deleted CPUSet assignment" podUID="0b2edb05-f8c1-4573-aadc-a19a90ec6575" containerName="ceilometer-central-agent" Dec 03 14:33:38 crc kubenswrapper[4805]: E1203 14:33:38.043537 4805 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0b2edb05-f8c1-4573-aadc-a19a90ec6575" containerName="ceilometer-notification-agent" Dec 03 14:33:38 crc kubenswrapper[4805]: I1203 14:33:38.043544 4805 state_mem.go:107] "Deleted CPUSet assignment" podUID="0b2edb05-f8c1-4573-aadc-a19a90ec6575" containerName="ceilometer-notification-agent" Dec 03 14:33:38 crc kubenswrapper[4805]: E1203 14:33:38.043564 4805 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0b2edb05-f8c1-4573-aadc-a19a90ec6575" containerName="proxy-httpd" Dec 03 14:33:38 crc kubenswrapper[4805]: I1203 14:33:38.043571 4805 state_mem.go:107] "Deleted CPUSet assignment" podUID="0b2edb05-f8c1-4573-aadc-a19a90ec6575" containerName="proxy-httpd" Dec 03 14:33:38 crc kubenswrapper[4805]: I1203 14:33:38.043783 4805 memory_manager.go:354] "RemoveStaleState removing state" podUID="0b2edb05-f8c1-4573-aadc-a19a90ec6575" containerName="sg-core" Dec 03 14:33:38 crc kubenswrapper[4805]: I1203 14:33:38.043797 4805 memory_manager.go:354] "RemoveStaleState removing state" podUID="0b2edb05-f8c1-4573-aadc-a19a90ec6575" containerName="ceilometer-central-agent" Dec 03 14:33:38 crc kubenswrapper[4805]: I1203 14:33:38.043820 4805 memory_manager.go:354] "RemoveStaleState removing state" podUID="0b2edb05-f8c1-4573-aadc-a19a90ec6575" containerName="ceilometer-notification-agent" Dec 03 14:33:38 crc kubenswrapper[4805]: I1203 14:33:38.043831 4805 memory_manager.go:354] "RemoveStaleState removing state" podUID="0b2edb05-f8c1-4573-aadc-a19a90ec6575" containerName="proxy-httpd" Dec 03 14:33:38 crc kubenswrapper[4805]: I1203 14:33:38.045968 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 03 14:33:38 crc kubenswrapper[4805]: I1203 14:33:38.049506 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Dec 03 14:33:38 crc kubenswrapper[4805]: I1203 14:33:38.049753 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Dec 03 14:33:38 crc kubenswrapper[4805]: I1203 14:33:38.049907 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ceilometer-internal-svc" Dec 03 14:33:38 crc kubenswrapper[4805]: I1203 14:33:38.051715 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 03 14:33:38 crc kubenswrapper[4805]: I1203 14:33:38.071960 4805 scope.go:117] "RemoveContainer" containerID="ecfff5277ed2e0011128d2830b165a01ac0550b1d6a0c7a6dbbaf9a4ad39bfe4" Dec 03 14:33:38 crc kubenswrapper[4805]: I1203 14:33:38.101487 4805 scope.go:117] "RemoveContainer" containerID="1a04e270afd803e2f0ae6e27199ac82282063350f51f59add0e60f560f536649" Dec 03 14:33:38 crc kubenswrapper[4805]: I1203 14:33:38.126316 4805 scope.go:117] "RemoveContainer" containerID="2796d8e7ced38b7980ca7effec53d78ecc6fc7774b9505bbb31b085653f72b91" Dec 03 14:33:38 crc kubenswrapper[4805]: E1203 14:33:38.127021 4805 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2796d8e7ced38b7980ca7effec53d78ecc6fc7774b9505bbb31b085653f72b91\": container with ID starting with 2796d8e7ced38b7980ca7effec53d78ecc6fc7774b9505bbb31b085653f72b91 not found: ID does not exist" containerID="2796d8e7ced38b7980ca7effec53d78ecc6fc7774b9505bbb31b085653f72b91" Dec 03 14:33:38 crc kubenswrapper[4805]: I1203 14:33:38.127085 4805 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2796d8e7ced38b7980ca7effec53d78ecc6fc7774b9505bbb31b085653f72b91"} err="failed to get container status \"2796d8e7ced38b7980ca7effec53d78ecc6fc7774b9505bbb31b085653f72b91\": rpc error: code = NotFound desc = could not find container \"2796d8e7ced38b7980ca7effec53d78ecc6fc7774b9505bbb31b085653f72b91\": container with ID starting with 2796d8e7ced38b7980ca7effec53d78ecc6fc7774b9505bbb31b085653f72b91 not found: ID does not exist" Dec 03 14:33:38 crc kubenswrapper[4805]: I1203 14:33:38.127124 4805 scope.go:117] "RemoveContainer" containerID="4f421ab97684b172723215aa66ba4b646542ef506fa51035c997453656f031f7" Dec 03 14:33:38 crc kubenswrapper[4805]: E1203 14:33:38.127647 4805 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4f421ab97684b172723215aa66ba4b646542ef506fa51035c997453656f031f7\": container with ID starting with 4f421ab97684b172723215aa66ba4b646542ef506fa51035c997453656f031f7 not found: ID does not exist" containerID="4f421ab97684b172723215aa66ba4b646542ef506fa51035c997453656f031f7" Dec 03 14:33:38 crc kubenswrapper[4805]: I1203 14:33:38.127711 4805 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4f421ab97684b172723215aa66ba4b646542ef506fa51035c997453656f031f7"} err="failed to get container status \"4f421ab97684b172723215aa66ba4b646542ef506fa51035c997453656f031f7\": rpc error: code = NotFound desc = could not find container \"4f421ab97684b172723215aa66ba4b646542ef506fa51035c997453656f031f7\": container with ID starting with 4f421ab97684b172723215aa66ba4b646542ef506fa51035c997453656f031f7 not found: ID does not exist" Dec 03 14:33:38 crc kubenswrapper[4805]: I1203 14:33:38.127748 4805 scope.go:117] "RemoveContainer" containerID="ecfff5277ed2e0011128d2830b165a01ac0550b1d6a0c7a6dbbaf9a4ad39bfe4" Dec 03 14:33:38 crc kubenswrapper[4805]: E1203 14:33:38.128060 4805 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ecfff5277ed2e0011128d2830b165a01ac0550b1d6a0c7a6dbbaf9a4ad39bfe4\": container with ID starting with ecfff5277ed2e0011128d2830b165a01ac0550b1d6a0c7a6dbbaf9a4ad39bfe4 not found: ID does not exist" containerID="ecfff5277ed2e0011128d2830b165a01ac0550b1d6a0c7a6dbbaf9a4ad39bfe4" Dec 03 14:33:38 crc kubenswrapper[4805]: I1203 14:33:38.128087 4805 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ecfff5277ed2e0011128d2830b165a01ac0550b1d6a0c7a6dbbaf9a4ad39bfe4"} err="failed to get container status \"ecfff5277ed2e0011128d2830b165a01ac0550b1d6a0c7a6dbbaf9a4ad39bfe4\": rpc error: code = NotFound desc = could not find container \"ecfff5277ed2e0011128d2830b165a01ac0550b1d6a0c7a6dbbaf9a4ad39bfe4\": container with ID starting with ecfff5277ed2e0011128d2830b165a01ac0550b1d6a0c7a6dbbaf9a4ad39bfe4 not found: ID does not exist" Dec 03 14:33:38 crc kubenswrapper[4805]: I1203 14:33:38.128104 4805 scope.go:117] "RemoveContainer" containerID="1a04e270afd803e2f0ae6e27199ac82282063350f51f59add0e60f560f536649" Dec 03 14:33:38 crc kubenswrapper[4805]: E1203 14:33:38.128397 4805 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1a04e270afd803e2f0ae6e27199ac82282063350f51f59add0e60f560f536649\": container with ID starting with 1a04e270afd803e2f0ae6e27199ac82282063350f51f59add0e60f560f536649 not found: ID does not exist" containerID="1a04e270afd803e2f0ae6e27199ac82282063350f51f59add0e60f560f536649" Dec 03 14:33:38 crc kubenswrapper[4805]: I1203 14:33:38.128463 4805 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1a04e270afd803e2f0ae6e27199ac82282063350f51f59add0e60f560f536649"} err="failed to get container status \"1a04e270afd803e2f0ae6e27199ac82282063350f51f59add0e60f560f536649\": rpc error: code = NotFound desc = could not find container \"1a04e270afd803e2f0ae6e27199ac82282063350f51f59add0e60f560f536649\": container with ID starting with 1a04e270afd803e2f0ae6e27199ac82282063350f51f59add0e60f560f536649 not found: ID does not exist" Dec 03 14:33:38 crc kubenswrapper[4805]: I1203 14:33:38.227199 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/8622d91c-e48b-46a9-a3e7-ba9b9f2dd257-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"8622d91c-e48b-46a9-a3e7-ba9b9f2dd257\") " pod="openstack/ceilometer-0" Dec 03 14:33:38 crc kubenswrapper[4805]: I1203 14:33:38.227263 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/8622d91c-e48b-46a9-a3e7-ba9b9f2dd257-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"8622d91c-e48b-46a9-a3e7-ba9b9f2dd257\") " pod="openstack/ceilometer-0" Dec 03 14:33:38 crc kubenswrapper[4805]: I1203 14:33:38.227379 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9scm6\" (UniqueName: \"kubernetes.io/projected/8622d91c-e48b-46a9-a3e7-ba9b9f2dd257-kube-api-access-9scm6\") pod \"ceilometer-0\" (UID: \"8622d91c-e48b-46a9-a3e7-ba9b9f2dd257\") " pod="openstack/ceilometer-0" Dec 03 14:33:38 crc kubenswrapper[4805]: I1203 14:33:38.227604 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8622d91c-e48b-46a9-a3e7-ba9b9f2dd257-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"8622d91c-e48b-46a9-a3e7-ba9b9f2dd257\") " pod="openstack/ceilometer-0" Dec 03 14:33:38 crc kubenswrapper[4805]: I1203 14:33:38.227686 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8622d91c-e48b-46a9-a3e7-ba9b9f2dd257-config-data\") pod \"ceilometer-0\" (UID: \"8622d91c-e48b-46a9-a3e7-ba9b9f2dd257\") " pod="openstack/ceilometer-0" Dec 03 14:33:38 crc kubenswrapper[4805]: I1203 14:33:38.227712 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/8622d91c-e48b-46a9-a3e7-ba9b9f2dd257-run-httpd\") pod \"ceilometer-0\" (UID: \"8622d91c-e48b-46a9-a3e7-ba9b9f2dd257\") " pod="openstack/ceilometer-0" Dec 03 14:33:38 crc kubenswrapper[4805]: I1203 14:33:38.227741 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8622d91c-e48b-46a9-a3e7-ba9b9f2dd257-scripts\") pod \"ceilometer-0\" (UID: \"8622d91c-e48b-46a9-a3e7-ba9b9f2dd257\") " pod="openstack/ceilometer-0" Dec 03 14:33:38 crc kubenswrapper[4805]: I1203 14:33:38.227991 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/8622d91c-e48b-46a9-a3e7-ba9b9f2dd257-log-httpd\") pod \"ceilometer-0\" (UID: \"8622d91c-e48b-46a9-a3e7-ba9b9f2dd257\") " pod="openstack/ceilometer-0" Dec 03 14:33:38 crc kubenswrapper[4805]: I1203 14:33:38.329486 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9scm6\" (UniqueName: \"kubernetes.io/projected/8622d91c-e48b-46a9-a3e7-ba9b9f2dd257-kube-api-access-9scm6\") pod \"ceilometer-0\" (UID: \"8622d91c-e48b-46a9-a3e7-ba9b9f2dd257\") " pod="openstack/ceilometer-0" Dec 03 14:33:38 crc kubenswrapper[4805]: I1203 14:33:38.329548 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8622d91c-e48b-46a9-a3e7-ba9b9f2dd257-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"8622d91c-e48b-46a9-a3e7-ba9b9f2dd257\") " pod="openstack/ceilometer-0" Dec 03 14:33:38 crc kubenswrapper[4805]: I1203 14:33:38.329577 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/8622d91c-e48b-46a9-a3e7-ba9b9f2dd257-run-httpd\") pod \"ceilometer-0\" (UID: \"8622d91c-e48b-46a9-a3e7-ba9b9f2dd257\") " pod="openstack/ceilometer-0" Dec 03 14:33:38 crc kubenswrapper[4805]: I1203 14:33:38.329591 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8622d91c-e48b-46a9-a3e7-ba9b9f2dd257-config-data\") pod \"ceilometer-0\" (UID: \"8622d91c-e48b-46a9-a3e7-ba9b9f2dd257\") " pod="openstack/ceilometer-0" Dec 03 14:33:38 crc kubenswrapper[4805]: I1203 14:33:38.329609 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8622d91c-e48b-46a9-a3e7-ba9b9f2dd257-scripts\") pod \"ceilometer-0\" (UID: \"8622d91c-e48b-46a9-a3e7-ba9b9f2dd257\") " pod="openstack/ceilometer-0" Dec 03 14:33:38 crc kubenswrapper[4805]: I1203 14:33:38.329669 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/8622d91c-e48b-46a9-a3e7-ba9b9f2dd257-log-httpd\") pod \"ceilometer-0\" (UID: \"8622d91c-e48b-46a9-a3e7-ba9b9f2dd257\") " pod="openstack/ceilometer-0" Dec 03 14:33:38 crc kubenswrapper[4805]: I1203 14:33:38.329729 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/8622d91c-e48b-46a9-a3e7-ba9b9f2dd257-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"8622d91c-e48b-46a9-a3e7-ba9b9f2dd257\") " pod="openstack/ceilometer-0" Dec 03 14:33:38 crc kubenswrapper[4805]: I1203 14:33:38.329771 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/8622d91c-e48b-46a9-a3e7-ba9b9f2dd257-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"8622d91c-e48b-46a9-a3e7-ba9b9f2dd257\") " pod="openstack/ceilometer-0" Dec 03 14:33:38 crc kubenswrapper[4805]: I1203 14:33:38.330272 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/8622d91c-e48b-46a9-a3e7-ba9b9f2dd257-log-httpd\") pod \"ceilometer-0\" (UID: \"8622d91c-e48b-46a9-a3e7-ba9b9f2dd257\") " pod="openstack/ceilometer-0" Dec 03 14:33:38 crc kubenswrapper[4805]: I1203 14:33:38.330436 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/8622d91c-e48b-46a9-a3e7-ba9b9f2dd257-run-httpd\") pod \"ceilometer-0\" (UID: \"8622d91c-e48b-46a9-a3e7-ba9b9f2dd257\") " pod="openstack/ceilometer-0" Dec 03 14:33:38 crc kubenswrapper[4805]: I1203 14:33:38.333743 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/8622d91c-e48b-46a9-a3e7-ba9b9f2dd257-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"8622d91c-e48b-46a9-a3e7-ba9b9f2dd257\") " pod="openstack/ceilometer-0" Dec 03 14:33:38 crc kubenswrapper[4805]: I1203 14:33:38.334278 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8622d91c-e48b-46a9-a3e7-ba9b9f2dd257-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"8622d91c-e48b-46a9-a3e7-ba9b9f2dd257\") " pod="openstack/ceilometer-0" Dec 03 14:33:38 crc kubenswrapper[4805]: I1203 14:33:38.334718 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8622d91c-e48b-46a9-a3e7-ba9b9f2dd257-config-data\") pod \"ceilometer-0\" (UID: \"8622d91c-e48b-46a9-a3e7-ba9b9f2dd257\") " pod="openstack/ceilometer-0" Dec 03 14:33:38 crc kubenswrapper[4805]: I1203 14:33:38.334974 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/8622d91c-e48b-46a9-a3e7-ba9b9f2dd257-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"8622d91c-e48b-46a9-a3e7-ba9b9f2dd257\") " pod="openstack/ceilometer-0" Dec 03 14:33:38 crc kubenswrapper[4805]: I1203 14:33:38.342602 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8622d91c-e48b-46a9-a3e7-ba9b9f2dd257-scripts\") pod \"ceilometer-0\" (UID: \"8622d91c-e48b-46a9-a3e7-ba9b9f2dd257\") " pod="openstack/ceilometer-0" Dec 03 14:33:38 crc kubenswrapper[4805]: I1203 14:33:38.350855 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9scm6\" (UniqueName: \"kubernetes.io/projected/8622d91c-e48b-46a9-a3e7-ba9b9f2dd257-kube-api-access-9scm6\") pod \"ceilometer-0\" (UID: \"8622d91c-e48b-46a9-a3e7-ba9b9f2dd257\") " pod="openstack/ceilometer-0" Dec 03 14:33:38 crc kubenswrapper[4805]: I1203 14:33:38.383479 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 03 14:33:38 crc kubenswrapper[4805]: I1203 14:33:38.396626 4805 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 03 14:33:38 crc kubenswrapper[4805]: I1203 14:33:38.704993 4805 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0b2edb05-f8c1-4573-aadc-a19a90ec6575" path="/var/lib/kubelet/pods/0b2edb05-f8c1-4573-aadc-a19a90ec6575/volumes" Dec 03 14:33:38 crc kubenswrapper[4805]: W1203 14:33:38.868023 4805 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod8622d91c_e48b_46a9_a3e7_ba9b9f2dd257.slice/crio-11e06a37d4e01ebbe21956ff796cbe0c8579599266d9cf9a9c109fe3597c9d5f WatchSource:0}: Error finding container 11e06a37d4e01ebbe21956ff796cbe0c8579599266d9cf9a9c109fe3597c9d5f: Status 404 returned error can't find the container with id 11e06a37d4e01ebbe21956ff796cbe0c8579599266d9cf9a9c109fe3597c9d5f Dec 03 14:33:38 crc kubenswrapper[4805]: I1203 14:33:38.890477 4805 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 03 14:33:38 crc kubenswrapper[4805]: I1203 14:33:38.982707 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"8622d91c-e48b-46a9-a3e7-ba9b9f2dd257","Type":"ContainerStarted","Data":"11e06a37d4e01ebbe21956ff796cbe0c8579599266d9cf9a9c109fe3597c9d5f"} Dec 03 14:33:38 crc kubenswrapper[4805]: I1203 14:33:38.986400 4805 generic.go:334] "Generic (PLEG): container finished" podID="94555924-f465-4fb9-aaf1-95afdd1b3a66" containerID="25d99f4b23894a9130fc530916f9b7235391aa390814fcee160a2ae205de87bc" exitCode=0 Dec 03 14:33:38 crc kubenswrapper[4805]: I1203 14:33:38.986434 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-7scbf" event={"ID":"94555924-f465-4fb9-aaf1-95afdd1b3a66","Type":"ContainerDied","Data":"25d99f4b23894a9130fc530916f9b7235391aa390814fcee160a2ae205de87bc"} Dec 03 14:33:39 crc kubenswrapper[4805]: I1203 14:33:39.652028 4805 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-7scbf" Dec 03 14:33:39 crc kubenswrapper[4805]: I1203 14:33:39.658961 4805 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 03 14:33:39 crc kubenswrapper[4805]: I1203 14:33:39.754288 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f6f6151a-8cbd-431e-b448-bf70fce67cde-combined-ca-bundle\") pod \"f6f6151a-8cbd-431e-b448-bf70fce67cde\" (UID: \"f6f6151a-8cbd-431e-b448-bf70fce67cde\") " Dec 03 14:33:39 crc kubenswrapper[4805]: I1203 14:33:39.754335 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f6f6151a-8cbd-431e-b448-bf70fce67cde-config-data\") pod \"f6f6151a-8cbd-431e-b448-bf70fce67cde\" (UID: \"f6f6151a-8cbd-431e-b448-bf70fce67cde\") " Dec 03 14:33:39 crc kubenswrapper[4805]: I1203 14:33:39.754434 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vdklr\" (UniqueName: \"kubernetes.io/projected/f6f6151a-8cbd-431e-b448-bf70fce67cde-kube-api-access-vdklr\") pod \"f6f6151a-8cbd-431e-b448-bf70fce67cde\" (UID: \"f6f6151a-8cbd-431e-b448-bf70fce67cde\") " Dec 03 14:33:39 crc kubenswrapper[4805]: I1203 14:33:39.754461 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f6f6151a-8cbd-431e-b448-bf70fce67cde-logs\") pod \"f6f6151a-8cbd-431e-b448-bf70fce67cde\" (UID: \"f6f6151a-8cbd-431e-b448-bf70fce67cde\") " Dec 03 14:33:39 crc kubenswrapper[4805]: I1203 14:33:39.754491 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/94555924-f465-4fb9-aaf1-95afdd1b3a66-catalog-content\") pod \"94555924-f465-4fb9-aaf1-95afdd1b3a66\" (UID: \"94555924-f465-4fb9-aaf1-95afdd1b3a66\") " Dec 03 14:33:39 crc kubenswrapper[4805]: I1203 14:33:39.754527 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pftll\" (UniqueName: \"kubernetes.io/projected/94555924-f465-4fb9-aaf1-95afdd1b3a66-kube-api-access-pftll\") pod \"94555924-f465-4fb9-aaf1-95afdd1b3a66\" (UID: \"94555924-f465-4fb9-aaf1-95afdd1b3a66\") " Dec 03 14:33:39 crc kubenswrapper[4805]: I1203 14:33:39.754585 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/94555924-f465-4fb9-aaf1-95afdd1b3a66-utilities\") pod \"94555924-f465-4fb9-aaf1-95afdd1b3a66\" (UID: \"94555924-f465-4fb9-aaf1-95afdd1b3a66\") " Dec 03 14:33:39 crc kubenswrapper[4805]: I1203 14:33:39.755747 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f6f6151a-8cbd-431e-b448-bf70fce67cde-logs" (OuterVolumeSpecName: "logs") pod "f6f6151a-8cbd-431e-b448-bf70fce67cde" (UID: "f6f6151a-8cbd-431e-b448-bf70fce67cde"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 14:33:39 crc kubenswrapper[4805]: I1203 14:33:39.755754 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/94555924-f465-4fb9-aaf1-95afdd1b3a66-utilities" (OuterVolumeSpecName: "utilities") pod "94555924-f465-4fb9-aaf1-95afdd1b3a66" (UID: "94555924-f465-4fb9-aaf1-95afdd1b3a66"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 14:33:39 crc kubenswrapper[4805]: I1203 14:33:39.763123 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f6f6151a-8cbd-431e-b448-bf70fce67cde-kube-api-access-vdklr" (OuterVolumeSpecName: "kube-api-access-vdklr") pod "f6f6151a-8cbd-431e-b448-bf70fce67cde" (UID: "f6f6151a-8cbd-431e-b448-bf70fce67cde"). InnerVolumeSpecName "kube-api-access-vdklr". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 14:33:39 crc kubenswrapper[4805]: I1203 14:33:39.763193 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/94555924-f465-4fb9-aaf1-95afdd1b3a66-kube-api-access-pftll" (OuterVolumeSpecName: "kube-api-access-pftll") pod "94555924-f465-4fb9-aaf1-95afdd1b3a66" (UID: "94555924-f465-4fb9-aaf1-95afdd1b3a66"). InnerVolumeSpecName "kube-api-access-pftll". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 14:33:39 crc kubenswrapper[4805]: I1203 14:33:39.802097 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f6f6151a-8cbd-431e-b448-bf70fce67cde-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "f6f6151a-8cbd-431e-b448-bf70fce67cde" (UID: "f6f6151a-8cbd-431e-b448-bf70fce67cde"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 14:33:39 crc kubenswrapper[4805]: I1203 14:33:39.823334 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f6f6151a-8cbd-431e-b448-bf70fce67cde-config-data" (OuterVolumeSpecName: "config-data") pod "f6f6151a-8cbd-431e-b448-bf70fce67cde" (UID: "f6f6151a-8cbd-431e-b448-bf70fce67cde"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 14:33:39 crc kubenswrapper[4805]: I1203 14:33:39.856834 4805 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f6f6151a-8cbd-431e-b448-bf70fce67cde-config-data\") on node \"crc\" DevicePath \"\"" Dec 03 14:33:39 crc kubenswrapper[4805]: I1203 14:33:39.856894 4805 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vdklr\" (UniqueName: \"kubernetes.io/projected/f6f6151a-8cbd-431e-b448-bf70fce67cde-kube-api-access-vdklr\") on node \"crc\" DevicePath \"\"" Dec 03 14:33:39 crc kubenswrapper[4805]: I1203 14:33:39.856909 4805 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f6f6151a-8cbd-431e-b448-bf70fce67cde-logs\") on node \"crc\" DevicePath \"\"" Dec 03 14:33:39 crc kubenswrapper[4805]: I1203 14:33:39.856921 4805 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pftll\" (UniqueName: \"kubernetes.io/projected/94555924-f465-4fb9-aaf1-95afdd1b3a66-kube-api-access-pftll\") on node \"crc\" DevicePath \"\"" Dec 03 14:33:39 crc kubenswrapper[4805]: I1203 14:33:39.856933 4805 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/94555924-f465-4fb9-aaf1-95afdd1b3a66-utilities\") on node \"crc\" DevicePath \"\"" Dec 03 14:33:39 crc kubenswrapper[4805]: I1203 14:33:39.856947 4805 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f6f6151a-8cbd-431e-b448-bf70fce67cde-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 14:33:39 crc kubenswrapper[4805]: I1203 14:33:39.997304 4805 generic.go:334] "Generic (PLEG): container finished" podID="f6f6151a-8cbd-431e-b448-bf70fce67cde" containerID="27fd08f618dd7cc317a2f311da01536f7fdd4aae45c1d3cdb98d85e29257f3a7" exitCode=0 Dec 03 14:33:39 crc kubenswrapper[4805]: I1203 14:33:39.997391 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"f6f6151a-8cbd-431e-b448-bf70fce67cde","Type":"ContainerDied","Data":"27fd08f618dd7cc317a2f311da01536f7fdd4aae45c1d3cdb98d85e29257f3a7"} Dec 03 14:33:39 crc kubenswrapper[4805]: I1203 14:33:39.997419 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"f6f6151a-8cbd-431e-b448-bf70fce67cde","Type":"ContainerDied","Data":"8c7a9e480902e2ee368ec28411b1d380197f9c3dc31d7645669ec6a80eb4892a"} Dec 03 14:33:39 crc kubenswrapper[4805]: I1203 14:33:39.997435 4805 scope.go:117] "RemoveContainer" containerID="27fd08f618dd7cc317a2f311da01536f7fdd4aae45c1d3cdb98d85e29257f3a7" Dec 03 14:33:39 crc kubenswrapper[4805]: I1203 14:33:39.997555 4805 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 03 14:33:40 crc kubenswrapper[4805]: I1203 14:33:40.003447 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-7scbf" event={"ID":"94555924-f465-4fb9-aaf1-95afdd1b3a66","Type":"ContainerDied","Data":"946fc62a8fd4fd311c3394f51ad5d1dc5f87a57b876303716a5c0f4d89cfbd30"} Dec 03 14:33:40 crc kubenswrapper[4805]: I1203 14:33:40.003502 4805 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-7scbf" Dec 03 14:33:40 crc kubenswrapper[4805]: I1203 14:33:40.035729 4805 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Dec 03 14:33:40 crc kubenswrapper[4805]: I1203 14:33:40.035971 4805 scope.go:117] "RemoveContainer" containerID="1d2296f16c415ddf1a330122ee2e86cbcf674b0148eae3c574415d0228e09439" Dec 03 14:33:40 crc kubenswrapper[4805]: I1203 14:33:40.044761 4805 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-0"] Dec 03 14:33:40 crc kubenswrapper[4805]: I1203 14:33:40.055385 4805 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-0"] Dec 03 14:33:40 crc kubenswrapper[4805]: E1203 14:33:40.055831 4805 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="94555924-f465-4fb9-aaf1-95afdd1b3a66" containerName="extract-content" Dec 03 14:33:40 crc kubenswrapper[4805]: I1203 14:33:40.055944 4805 state_mem.go:107] "Deleted CPUSet assignment" podUID="94555924-f465-4fb9-aaf1-95afdd1b3a66" containerName="extract-content" Dec 03 14:33:40 crc kubenswrapper[4805]: E1203 14:33:40.055965 4805 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="94555924-f465-4fb9-aaf1-95afdd1b3a66" containerName="registry-server" Dec 03 14:33:40 crc kubenswrapper[4805]: I1203 14:33:40.055974 4805 state_mem.go:107] "Deleted CPUSet assignment" podUID="94555924-f465-4fb9-aaf1-95afdd1b3a66" containerName="registry-server" Dec 03 14:33:40 crc kubenswrapper[4805]: E1203 14:33:40.055998 4805 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="94555924-f465-4fb9-aaf1-95afdd1b3a66" containerName="extract-utilities" Dec 03 14:33:40 crc kubenswrapper[4805]: I1203 14:33:40.056006 4805 state_mem.go:107] "Deleted CPUSet assignment" podUID="94555924-f465-4fb9-aaf1-95afdd1b3a66" containerName="extract-utilities" Dec 03 14:33:40 crc kubenswrapper[4805]: E1203 14:33:40.056023 4805 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f6f6151a-8cbd-431e-b448-bf70fce67cde" containerName="nova-api-api" Dec 03 14:33:40 crc kubenswrapper[4805]: I1203 14:33:40.056058 4805 state_mem.go:107] "Deleted CPUSet assignment" podUID="f6f6151a-8cbd-431e-b448-bf70fce67cde" containerName="nova-api-api" Dec 03 14:33:40 crc kubenswrapper[4805]: E1203 14:33:40.056075 4805 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f6f6151a-8cbd-431e-b448-bf70fce67cde" containerName="nova-api-log" Dec 03 14:33:40 crc kubenswrapper[4805]: I1203 14:33:40.056083 4805 state_mem.go:107] "Deleted CPUSet assignment" podUID="f6f6151a-8cbd-431e-b448-bf70fce67cde" containerName="nova-api-log" Dec 03 14:33:40 crc kubenswrapper[4805]: I1203 14:33:40.056310 4805 memory_manager.go:354] "RemoveStaleState removing state" podUID="94555924-f465-4fb9-aaf1-95afdd1b3a66" containerName="registry-server" Dec 03 14:33:40 crc kubenswrapper[4805]: I1203 14:33:40.056340 4805 memory_manager.go:354] "RemoveStaleState removing state" podUID="f6f6151a-8cbd-431e-b448-bf70fce67cde" containerName="nova-api-log" Dec 03 14:33:40 crc kubenswrapper[4805]: I1203 14:33:40.056368 4805 memory_manager.go:354] "RemoveStaleState removing state" podUID="f6f6151a-8cbd-431e-b448-bf70fce67cde" containerName="nova-api-api" Dec 03 14:33:40 crc kubenswrapper[4805]: I1203 14:33:40.057563 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 03 14:33:40 crc kubenswrapper[4805]: I1203 14:33:40.058201 4805 scope.go:117] "RemoveContainer" containerID="27fd08f618dd7cc317a2f311da01536f7fdd4aae45c1d3cdb98d85e29257f3a7" Dec 03 14:33:40 crc kubenswrapper[4805]: E1203 14:33:40.058638 4805 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"27fd08f618dd7cc317a2f311da01536f7fdd4aae45c1d3cdb98d85e29257f3a7\": container with ID starting with 27fd08f618dd7cc317a2f311da01536f7fdd4aae45c1d3cdb98d85e29257f3a7 not found: ID does not exist" containerID="27fd08f618dd7cc317a2f311da01536f7fdd4aae45c1d3cdb98d85e29257f3a7" Dec 03 14:33:40 crc kubenswrapper[4805]: I1203 14:33:40.058673 4805 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"27fd08f618dd7cc317a2f311da01536f7fdd4aae45c1d3cdb98d85e29257f3a7"} err="failed to get container status \"27fd08f618dd7cc317a2f311da01536f7fdd4aae45c1d3cdb98d85e29257f3a7\": rpc error: code = NotFound desc = could not find container \"27fd08f618dd7cc317a2f311da01536f7fdd4aae45c1d3cdb98d85e29257f3a7\": container with ID starting with 27fd08f618dd7cc317a2f311da01536f7fdd4aae45c1d3cdb98d85e29257f3a7 not found: ID does not exist" Dec 03 14:33:40 crc kubenswrapper[4805]: I1203 14:33:40.058699 4805 scope.go:117] "RemoveContainer" containerID="1d2296f16c415ddf1a330122ee2e86cbcf674b0148eae3c574415d0228e09439" Dec 03 14:33:40 crc kubenswrapper[4805]: E1203 14:33:40.060818 4805 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1d2296f16c415ddf1a330122ee2e86cbcf674b0148eae3c574415d0228e09439\": container with ID starting with 1d2296f16c415ddf1a330122ee2e86cbcf674b0148eae3c574415d0228e09439 not found: ID does not exist" containerID="1d2296f16c415ddf1a330122ee2e86cbcf674b0148eae3c574415d0228e09439" Dec 03 14:33:40 crc kubenswrapper[4805]: I1203 14:33:40.060895 4805 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1d2296f16c415ddf1a330122ee2e86cbcf674b0148eae3c574415d0228e09439"} err="failed to get container status \"1d2296f16c415ddf1a330122ee2e86cbcf674b0148eae3c574415d0228e09439\": rpc error: code = NotFound desc = could not find container \"1d2296f16c415ddf1a330122ee2e86cbcf674b0148eae3c574415d0228e09439\": container with ID starting with 1d2296f16c415ddf1a330122ee2e86cbcf674b0148eae3c574415d0228e09439 not found: ID does not exist" Dec 03 14:33:40 crc kubenswrapper[4805]: I1203 14:33:40.060935 4805 scope.go:117] "RemoveContainer" containerID="25d99f4b23894a9130fc530916f9b7235391aa390814fcee160a2ae205de87bc" Dec 03 14:33:40 crc kubenswrapper[4805]: I1203 14:33:40.060979 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-config-data" Dec 03 14:33:40 crc kubenswrapper[4805]: I1203 14:33:40.061108 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-public-svc" Dec 03 14:33:40 crc kubenswrapper[4805]: I1203 14:33:40.061137 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-internal-svc" Dec 03 14:33:40 crc kubenswrapper[4805]: I1203 14:33:40.069787 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Dec 03 14:33:40 crc kubenswrapper[4805]: I1203 14:33:40.084137 4805 scope.go:117] "RemoveContainer" containerID="8a1675db1b86630d89429bb261bfe6f98a7818dab56bb9753962937b40fb4690" Dec 03 14:33:40 crc kubenswrapper[4805]: I1203 14:33:40.111196 4805 scope.go:117] "RemoveContainer" containerID="c76f027685120547acaa3c7dc4c9af2a48664e21451d91ffac1b0db630b99891" Dec 03 14:33:40 crc kubenswrapper[4805]: I1203 14:33:40.164703 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pgxts\" (UniqueName: \"kubernetes.io/projected/68636027-732d-46fa-aba1-7f89927292ac-kube-api-access-pgxts\") pod \"nova-api-0\" (UID: \"68636027-732d-46fa-aba1-7f89927292ac\") " pod="openstack/nova-api-0" Dec 03 14:33:40 crc kubenswrapper[4805]: I1203 14:33:40.164748 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/68636027-732d-46fa-aba1-7f89927292ac-internal-tls-certs\") pod \"nova-api-0\" (UID: \"68636027-732d-46fa-aba1-7f89927292ac\") " pod="openstack/nova-api-0" Dec 03 14:33:40 crc kubenswrapper[4805]: I1203 14:33:40.165045 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/68636027-732d-46fa-aba1-7f89927292ac-logs\") pod \"nova-api-0\" (UID: \"68636027-732d-46fa-aba1-7f89927292ac\") " pod="openstack/nova-api-0" Dec 03 14:33:40 crc kubenswrapper[4805]: I1203 14:33:40.165151 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/68636027-732d-46fa-aba1-7f89927292ac-config-data\") pod \"nova-api-0\" (UID: \"68636027-732d-46fa-aba1-7f89927292ac\") " pod="openstack/nova-api-0" Dec 03 14:33:40 crc kubenswrapper[4805]: I1203 14:33:40.165173 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/68636027-732d-46fa-aba1-7f89927292ac-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"68636027-732d-46fa-aba1-7f89927292ac\") " pod="openstack/nova-api-0" Dec 03 14:33:40 crc kubenswrapper[4805]: I1203 14:33:40.165191 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/68636027-732d-46fa-aba1-7f89927292ac-public-tls-certs\") pod \"nova-api-0\" (UID: \"68636027-732d-46fa-aba1-7f89927292ac\") " pod="openstack/nova-api-0" Dec 03 14:33:40 crc kubenswrapper[4805]: I1203 14:33:40.266994 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/68636027-732d-46fa-aba1-7f89927292ac-logs\") pod \"nova-api-0\" (UID: \"68636027-732d-46fa-aba1-7f89927292ac\") " pod="openstack/nova-api-0" Dec 03 14:33:40 crc kubenswrapper[4805]: I1203 14:33:40.267422 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/68636027-732d-46fa-aba1-7f89927292ac-config-data\") pod \"nova-api-0\" (UID: \"68636027-732d-46fa-aba1-7f89927292ac\") " pod="openstack/nova-api-0" Dec 03 14:33:40 crc kubenswrapper[4805]: I1203 14:33:40.267506 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/68636027-732d-46fa-aba1-7f89927292ac-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"68636027-732d-46fa-aba1-7f89927292ac\") " pod="openstack/nova-api-0" Dec 03 14:33:40 crc kubenswrapper[4805]: I1203 14:33:40.267574 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/68636027-732d-46fa-aba1-7f89927292ac-public-tls-certs\") pod \"nova-api-0\" (UID: \"68636027-732d-46fa-aba1-7f89927292ac\") " pod="openstack/nova-api-0" Dec 03 14:33:40 crc kubenswrapper[4805]: I1203 14:33:40.267532 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/68636027-732d-46fa-aba1-7f89927292ac-logs\") pod \"nova-api-0\" (UID: \"68636027-732d-46fa-aba1-7f89927292ac\") " pod="openstack/nova-api-0" Dec 03 14:33:40 crc kubenswrapper[4805]: I1203 14:33:40.267878 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pgxts\" (UniqueName: \"kubernetes.io/projected/68636027-732d-46fa-aba1-7f89927292ac-kube-api-access-pgxts\") pod \"nova-api-0\" (UID: \"68636027-732d-46fa-aba1-7f89927292ac\") " pod="openstack/nova-api-0" Dec 03 14:33:40 crc kubenswrapper[4805]: I1203 14:33:40.267976 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/68636027-732d-46fa-aba1-7f89927292ac-internal-tls-certs\") pod \"nova-api-0\" (UID: \"68636027-732d-46fa-aba1-7f89927292ac\") " pod="openstack/nova-api-0" Dec 03 14:33:40 crc kubenswrapper[4805]: I1203 14:33:40.277218 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/68636027-732d-46fa-aba1-7f89927292ac-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"68636027-732d-46fa-aba1-7f89927292ac\") " pod="openstack/nova-api-0" Dec 03 14:33:40 crc kubenswrapper[4805]: I1203 14:33:40.277231 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/68636027-732d-46fa-aba1-7f89927292ac-config-data\") pod \"nova-api-0\" (UID: \"68636027-732d-46fa-aba1-7f89927292ac\") " pod="openstack/nova-api-0" Dec 03 14:33:40 crc kubenswrapper[4805]: I1203 14:33:40.277350 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/68636027-732d-46fa-aba1-7f89927292ac-public-tls-certs\") pod \"nova-api-0\" (UID: \"68636027-732d-46fa-aba1-7f89927292ac\") " pod="openstack/nova-api-0" Dec 03 14:33:40 crc kubenswrapper[4805]: I1203 14:33:40.282287 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/68636027-732d-46fa-aba1-7f89927292ac-internal-tls-certs\") pod \"nova-api-0\" (UID: \"68636027-732d-46fa-aba1-7f89927292ac\") " pod="openstack/nova-api-0" Dec 03 14:33:40 crc kubenswrapper[4805]: I1203 14:33:40.297441 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pgxts\" (UniqueName: \"kubernetes.io/projected/68636027-732d-46fa-aba1-7f89927292ac-kube-api-access-pgxts\") pod \"nova-api-0\" (UID: \"68636027-732d-46fa-aba1-7f89927292ac\") " pod="openstack/nova-api-0" Dec 03 14:33:40 crc kubenswrapper[4805]: I1203 14:33:40.381416 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 03 14:33:40 crc kubenswrapper[4805]: I1203 14:33:40.540348 4805 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-cell1-novncproxy-0" Dec 03 14:33:40 crc kubenswrapper[4805]: I1203 14:33:40.572643 4805 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-cell1-novncproxy-0" Dec 03 14:33:40 crc kubenswrapper[4805]: I1203 14:33:40.707880 4805 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f6f6151a-8cbd-431e-b448-bf70fce67cde" path="/var/lib/kubelet/pods/f6f6151a-8cbd-431e-b448-bf70fce67cde/volumes" Dec 03 14:33:40 crc kubenswrapper[4805]: I1203 14:33:40.751687 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/94555924-f465-4fb9-aaf1-95afdd1b3a66-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "94555924-f465-4fb9-aaf1-95afdd1b3a66" (UID: "94555924-f465-4fb9-aaf1-95afdd1b3a66"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 14:33:40 crc kubenswrapper[4805]: I1203 14:33:40.777678 4805 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/94555924-f465-4fb9-aaf1-95afdd1b3a66-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 03 14:33:40 crc kubenswrapper[4805]: W1203 14:33:40.837621 4805 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod68636027_732d_46fa_aba1_7f89927292ac.slice/crio-25a2d3afcc061c8c20cf82112ab18f83427acb4f96f81bae1375681ed8e7c0d8 WatchSource:0}: Error finding container 25a2d3afcc061c8c20cf82112ab18f83427acb4f96f81bae1375681ed8e7c0d8: Status 404 returned error can't find the container with id 25a2d3afcc061c8c20cf82112ab18f83427acb4f96f81bae1375681ed8e7c0d8 Dec 03 14:33:40 crc kubenswrapper[4805]: I1203 14:33:40.838155 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Dec 03 14:33:41 crc kubenswrapper[4805]: I1203 14:33:41.012499 4805 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-7scbf"] Dec 03 14:33:41 crc kubenswrapper[4805]: I1203 14:33:41.037967 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"68636027-732d-46fa-aba1-7f89927292ac","Type":"ContainerStarted","Data":"25a2d3afcc061c8c20cf82112ab18f83427acb4f96f81bae1375681ed8e7c0d8"} Dec 03 14:33:41 crc kubenswrapper[4805]: I1203 14:33:41.038692 4805 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-7scbf"] Dec 03 14:33:41 crc kubenswrapper[4805]: I1203 14:33:41.058438 4805 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-cell1-novncproxy-0" Dec 03 14:33:41 crc kubenswrapper[4805]: I1203 14:33:41.252439 4805 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-cell-mapping-pkj2h"] Dec 03 14:33:41 crc kubenswrapper[4805]: I1203 14:33:41.254202 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-cell-mapping-pkj2h" Dec 03 14:33:41 crc kubenswrapper[4805]: I1203 14:33:41.256258 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-manage-scripts" Dec 03 14:33:41 crc kubenswrapper[4805]: I1203 14:33:41.259437 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-manage-config-data" Dec 03 14:33:41 crc kubenswrapper[4805]: I1203 14:33:41.272722 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-cell-mapping-pkj2h"] Dec 03 14:33:41 crc kubenswrapper[4805]: I1203 14:33:41.387078 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cae0a70b-e47c-40ff-9b40-40a2aa570097-config-data\") pod \"nova-cell1-cell-mapping-pkj2h\" (UID: \"cae0a70b-e47c-40ff-9b40-40a2aa570097\") " pod="openstack/nova-cell1-cell-mapping-pkj2h" Dec 03 14:33:41 crc kubenswrapper[4805]: I1203 14:33:41.387126 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/cae0a70b-e47c-40ff-9b40-40a2aa570097-scripts\") pod \"nova-cell1-cell-mapping-pkj2h\" (UID: \"cae0a70b-e47c-40ff-9b40-40a2aa570097\") " pod="openstack/nova-cell1-cell-mapping-pkj2h" Dec 03 14:33:41 crc kubenswrapper[4805]: I1203 14:33:41.387181 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cae0a70b-e47c-40ff-9b40-40a2aa570097-combined-ca-bundle\") pod \"nova-cell1-cell-mapping-pkj2h\" (UID: \"cae0a70b-e47c-40ff-9b40-40a2aa570097\") " pod="openstack/nova-cell1-cell-mapping-pkj2h" Dec 03 14:33:41 crc kubenswrapper[4805]: I1203 14:33:41.387496 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hmpbx\" (UniqueName: \"kubernetes.io/projected/cae0a70b-e47c-40ff-9b40-40a2aa570097-kube-api-access-hmpbx\") pod \"nova-cell1-cell-mapping-pkj2h\" (UID: \"cae0a70b-e47c-40ff-9b40-40a2aa570097\") " pod="openstack/nova-cell1-cell-mapping-pkj2h" Dec 03 14:33:41 crc kubenswrapper[4805]: I1203 14:33:41.491113 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hmpbx\" (UniqueName: \"kubernetes.io/projected/cae0a70b-e47c-40ff-9b40-40a2aa570097-kube-api-access-hmpbx\") pod \"nova-cell1-cell-mapping-pkj2h\" (UID: \"cae0a70b-e47c-40ff-9b40-40a2aa570097\") " pod="openstack/nova-cell1-cell-mapping-pkj2h" Dec 03 14:33:41 crc kubenswrapper[4805]: I1203 14:33:41.491300 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cae0a70b-e47c-40ff-9b40-40a2aa570097-config-data\") pod \"nova-cell1-cell-mapping-pkj2h\" (UID: \"cae0a70b-e47c-40ff-9b40-40a2aa570097\") " pod="openstack/nova-cell1-cell-mapping-pkj2h" Dec 03 14:33:41 crc kubenswrapper[4805]: I1203 14:33:41.491529 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/cae0a70b-e47c-40ff-9b40-40a2aa570097-scripts\") pod \"nova-cell1-cell-mapping-pkj2h\" (UID: \"cae0a70b-e47c-40ff-9b40-40a2aa570097\") " pod="openstack/nova-cell1-cell-mapping-pkj2h" Dec 03 14:33:41 crc kubenswrapper[4805]: I1203 14:33:41.491637 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cae0a70b-e47c-40ff-9b40-40a2aa570097-combined-ca-bundle\") pod \"nova-cell1-cell-mapping-pkj2h\" (UID: \"cae0a70b-e47c-40ff-9b40-40a2aa570097\") " pod="openstack/nova-cell1-cell-mapping-pkj2h" Dec 03 14:33:41 crc kubenswrapper[4805]: I1203 14:33:41.497456 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cae0a70b-e47c-40ff-9b40-40a2aa570097-combined-ca-bundle\") pod \"nova-cell1-cell-mapping-pkj2h\" (UID: \"cae0a70b-e47c-40ff-9b40-40a2aa570097\") " pod="openstack/nova-cell1-cell-mapping-pkj2h" Dec 03 14:33:41 crc kubenswrapper[4805]: I1203 14:33:41.501409 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/cae0a70b-e47c-40ff-9b40-40a2aa570097-scripts\") pod \"nova-cell1-cell-mapping-pkj2h\" (UID: \"cae0a70b-e47c-40ff-9b40-40a2aa570097\") " pod="openstack/nova-cell1-cell-mapping-pkj2h" Dec 03 14:33:41 crc kubenswrapper[4805]: I1203 14:33:41.505071 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cae0a70b-e47c-40ff-9b40-40a2aa570097-config-data\") pod \"nova-cell1-cell-mapping-pkj2h\" (UID: \"cae0a70b-e47c-40ff-9b40-40a2aa570097\") " pod="openstack/nova-cell1-cell-mapping-pkj2h" Dec 03 14:33:41 crc kubenswrapper[4805]: I1203 14:33:41.512907 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hmpbx\" (UniqueName: \"kubernetes.io/projected/cae0a70b-e47c-40ff-9b40-40a2aa570097-kube-api-access-hmpbx\") pod \"nova-cell1-cell-mapping-pkj2h\" (UID: \"cae0a70b-e47c-40ff-9b40-40a2aa570097\") " pod="openstack/nova-cell1-cell-mapping-pkj2h" Dec 03 14:33:41 crc kubenswrapper[4805]: I1203 14:33:41.573449 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-cell-mapping-pkj2h" Dec 03 14:33:42 crc kubenswrapper[4805]: I1203 14:33:42.052574 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"8622d91c-e48b-46a9-a3e7-ba9b9f2dd257","Type":"ContainerStarted","Data":"8abb17bdb2ad683fc76656be33369b0da58f730c86bd5999b64e778d928b2cfb"} Dec 03 14:33:42 crc kubenswrapper[4805]: I1203 14:33:42.056420 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"68636027-732d-46fa-aba1-7f89927292ac","Type":"ContainerStarted","Data":"d3db209ff6c299ff5989f450f491b40fe35f9073b078b22a33a3a15f6a882663"} Dec 03 14:33:42 crc kubenswrapper[4805]: I1203 14:33:42.057858 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"68636027-732d-46fa-aba1-7f89927292ac","Type":"ContainerStarted","Data":"2508ae57d37d01ca93f1023c81c6c09acbbbbf95890d14a4951ae7cdfb8ddbc2"} Dec 03 14:33:42 crc kubenswrapper[4805]: I1203 14:33:42.092440 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-cell-mapping-pkj2h"] Dec 03 14:33:42 crc kubenswrapper[4805]: I1203 14:33:42.101154 4805 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-0" podStartSLOduration=2.101140105 podStartE2EDuration="2.101140105s" podCreationTimestamp="2025-12-03 14:33:40 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 14:33:42.07344964 +0000 UTC m=+1451.736366563" watchObservedRunningTime="2025-12-03 14:33:42.101140105 +0000 UTC m=+1451.764057028" Dec 03 14:33:42 crc kubenswrapper[4805]: I1203 14:33:42.704935 4805 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="94555924-f465-4fb9-aaf1-95afdd1b3a66" path="/var/lib/kubelet/pods/94555924-f465-4fb9-aaf1-95afdd1b3a66/volumes" Dec 03 14:33:43 crc kubenswrapper[4805]: I1203 14:33:43.066550 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"8622d91c-e48b-46a9-a3e7-ba9b9f2dd257","Type":"ContainerStarted","Data":"e1b9fd988130f8d3338ae8849b1514d22bc3e2dde1c18a30f2f1007209da0368"} Dec 03 14:33:43 crc kubenswrapper[4805]: I1203 14:33:43.069352 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-cell-mapping-pkj2h" event={"ID":"cae0a70b-e47c-40ff-9b40-40a2aa570097","Type":"ContainerStarted","Data":"d384c8b48d1a97e7c0c06910a1049b12545dc37afe379f1e5d0a1b926448aa6b"} Dec 03 14:33:43 crc kubenswrapper[4805]: I1203 14:33:43.069417 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-cell-mapping-pkj2h" event={"ID":"cae0a70b-e47c-40ff-9b40-40a2aa570097","Type":"ContainerStarted","Data":"eb74f826e57fdb74da5e57b7dfc2b92730485025ca3d05c13c452b7d09004cc4"} Dec 03 14:33:43 crc kubenswrapper[4805]: I1203 14:33:43.094438 4805 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-cell-mapping-pkj2h" podStartSLOduration=2.094418924 podStartE2EDuration="2.094418924s" podCreationTimestamp="2025-12-03 14:33:41 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 14:33:43.090632571 +0000 UTC m=+1452.753549504" watchObservedRunningTime="2025-12-03 14:33:43.094418924 +0000 UTC m=+1452.757335847" Dec 03 14:33:43 crc kubenswrapper[4805]: I1203 14:33:43.424103 4805 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-79b5d74c8c-qmdpk" Dec 03 14:33:43 crc kubenswrapper[4805]: I1203 14:33:43.506300 4805 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5fbc4d444f-5vhgg"] Dec 03 14:33:43 crc kubenswrapper[4805]: I1203 14:33:43.506525 4805 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-5fbc4d444f-5vhgg" podUID="e2ef61bb-5836-4986-88ae-f307ee375f0e" containerName="dnsmasq-dns" containerID="cri-o://a7d8422bf0b5f75574d94f6069d721089f2c5adb9c716b5ee35a9b1e43db5831" gracePeriod=10 Dec 03 14:33:43 crc kubenswrapper[4805]: I1203 14:33:43.917091 4805 patch_prober.go:28] interesting pod/machine-config-daemon-gskh4 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 03 14:33:43 crc kubenswrapper[4805]: I1203 14:33:43.917158 4805 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-gskh4" podUID="6098937f-e3f6-45e8-a647-4994a79cd711" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 03 14:33:44 crc kubenswrapper[4805]: I1203 14:33:44.080948 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"8622d91c-e48b-46a9-a3e7-ba9b9f2dd257","Type":"ContainerStarted","Data":"cdfe7cc1361c48b38876bd8b0e963bcd5f0e6bd7b096019ecfd6d8523b35b5e5"} Dec 03 14:33:44 crc kubenswrapper[4805]: I1203 14:33:44.083159 4805 generic.go:334] "Generic (PLEG): container finished" podID="e2ef61bb-5836-4986-88ae-f307ee375f0e" containerID="a7d8422bf0b5f75574d94f6069d721089f2c5adb9c716b5ee35a9b1e43db5831" exitCode=0 Dec 03 14:33:44 crc kubenswrapper[4805]: I1203 14:33:44.083252 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5fbc4d444f-5vhgg" event={"ID":"e2ef61bb-5836-4986-88ae-f307ee375f0e","Type":"ContainerDied","Data":"a7d8422bf0b5f75574d94f6069d721089f2c5adb9c716b5ee35a9b1e43db5831"} Dec 03 14:33:44 crc kubenswrapper[4805]: I1203 14:33:44.569227 4805 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5fbc4d444f-5vhgg" Dec 03 14:33:44 crc kubenswrapper[4805]: I1203 14:33:44.657252 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/e2ef61bb-5836-4986-88ae-f307ee375f0e-ovsdbserver-sb\") pod \"e2ef61bb-5836-4986-88ae-f307ee375f0e\" (UID: \"e2ef61bb-5836-4986-88ae-f307ee375f0e\") " Dec 03 14:33:44 crc kubenswrapper[4805]: I1203 14:33:44.657333 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e2ef61bb-5836-4986-88ae-f307ee375f0e-config\") pod \"e2ef61bb-5836-4986-88ae-f307ee375f0e\" (UID: \"e2ef61bb-5836-4986-88ae-f307ee375f0e\") " Dec 03 14:33:44 crc kubenswrapper[4805]: I1203 14:33:44.657453 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/e2ef61bb-5836-4986-88ae-f307ee375f0e-dns-swift-storage-0\") pod \"e2ef61bb-5836-4986-88ae-f307ee375f0e\" (UID: \"e2ef61bb-5836-4986-88ae-f307ee375f0e\") " Dec 03 14:33:44 crc kubenswrapper[4805]: I1203 14:33:44.657610 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/e2ef61bb-5836-4986-88ae-f307ee375f0e-dns-svc\") pod \"e2ef61bb-5836-4986-88ae-f307ee375f0e\" (UID: \"e2ef61bb-5836-4986-88ae-f307ee375f0e\") " Dec 03 14:33:44 crc kubenswrapper[4805]: I1203 14:33:44.657672 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-l9wqb\" (UniqueName: \"kubernetes.io/projected/e2ef61bb-5836-4986-88ae-f307ee375f0e-kube-api-access-l9wqb\") pod \"e2ef61bb-5836-4986-88ae-f307ee375f0e\" (UID: \"e2ef61bb-5836-4986-88ae-f307ee375f0e\") " Dec 03 14:33:44 crc kubenswrapper[4805]: I1203 14:33:44.657715 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/e2ef61bb-5836-4986-88ae-f307ee375f0e-ovsdbserver-nb\") pod \"e2ef61bb-5836-4986-88ae-f307ee375f0e\" (UID: \"e2ef61bb-5836-4986-88ae-f307ee375f0e\") " Dec 03 14:33:44 crc kubenswrapper[4805]: I1203 14:33:44.666147 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e2ef61bb-5836-4986-88ae-f307ee375f0e-kube-api-access-l9wqb" (OuterVolumeSpecName: "kube-api-access-l9wqb") pod "e2ef61bb-5836-4986-88ae-f307ee375f0e" (UID: "e2ef61bb-5836-4986-88ae-f307ee375f0e"). InnerVolumeSpecName "kube-api-access-l9wqb". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 14:33:44 crc kubenswrapper[4805]: I1203 14:33:44.727631 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e2ef61bb-5836-4986-88ae-f307ee375f0e-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "e2ef61bb-5836-4986-88ae-f307ee375f0e" (UID: "e2ef61bb-5836-4986-88ae-f307ee375f0e"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 14:33:44 crc kubenswrapper[4805]: I1203 14:33:44.728698 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e2ef61bb-5836-4986-88ae-f307ee375f0e-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "e2ef61bb-5836-4986-88ae-f307ee375f0e" (UID: "e2ef61bb-5836-4986-88ae-f307ee375f0e"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 14:33:44 crc kubenswrapper[4805]: I1203 14:33:44.735379 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e2ef61bb-5836-4986-88ae-f307ee375f0e-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "e2ef61bb-5836-4986-88ae-f307ee375f0e" (UID: "e2ef61bb-5836-4986-88ae-f307ee375f0e"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 14:33:44 crc kubenswrapper[4805]: I1203 14:33:44.737175 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e2ef61bb-5836-4986-88ae-f307ee375f0e-config" (OuterVolumeSpecName: "config") pod "e2ef61bb-5836-4986-88ae-f307ee375f0e" (UID: "e2ef61bb-5836-4986-88ae-f307ee375f0e"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 14:33:44 crc kubenswrapper[4805]: I1203 14:33:44.737622 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e2ef61bb-5836-4986-88ae-f307ee375f0e-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "e2ef61bb-5836-4986-88ae-f307ee375f0e" (UID: "e2ef61bb-5836-4986-88ae-f307ee375f0e"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 14:33:44 crc kubenswrapper[4805]: I1203 14:33:44.762425 4805 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/e2ef61bb-5836-4986-88ae-f307ee375f0e-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Dec 03 14:33:44 crc kubenswrapper[4805]: I1203 14:33:44.762466 4805 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e2ef61bb-5836-4986-88ae-f307ee375f0e-config\") on node \"crc\" DevicePath \"\"" Dec 03 14:33:44 crc kubenswrapper[4805]: I1203 14:33:44.762480 4805 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/e2ef61bb-5836-4986-88ae-f307ee375f0e-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Dec 03 14:33:44 crc kubenswrapper[4805]: I1203 14:33:44.762492 4805 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/e2ef61bb-5836-4986-88ae-f307ee375f0e-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 03 14:33:44 crc kubenswrapper[4805]: I1203 14:33:44.762503 4805 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-l9wqb\" (UniqueName: \"kubernetes.io/projected/e2ef61bb-5836-4986-88ae-f307ee375f0e-kube-api-access-l9wqb\") on node \"crc\" DevicePath \"\"" Dec 03 14:33:44 crc kubenswrapper[4805]: I1203 14:33:44.762519 4805 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/e2ef61bb-5836-4986-88ae-f307ee375f0e-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Dec 03 14:33:45 crc kubenswrapper[4805]: I1203 14:33:45.095164 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5fbc4d444f-5vhgg" event={"ID":"e2ef61bb-5836-4986-88ae-f307ee375f0e","Type":"ContainerDied","Data":"6f57619b8ae32c72b2a066282ab29bc4917760e6fb6e13419cdd03cae40db7d2"} Dec 03 14:33:45 crc kubenswrapper[4805]: I1203 14:33:45.096403 4805 scope.go:117] "RemoveContainer" containerID="a7d8422bf0b5f75574d94f6069d721089f2c5adb9c716b5ee35a9b1e43db5831" Dec 03 14:33:45 crc kubenswrapper[4805]: I1203 14:33:45.095209 4805 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5fbc4d444f-5vhgg" Dec 03 14:33:45 crc kubenswrapper[4805]: I1203 14:33:45.119783 4805 scope.go:117] "RemoveContainer" containerID="796364804b6d42eff3b0d9c02016068c2e46f8d1f2868771facca6b5efe60d07" Dec 03 14:33:45 crc kubenswrapper[4805]: I1203 14:33:45.133774 4805 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5fbc4d444f-5vhgg"] Dec 03 14:33:45 crc kubenswrapper[4805]: I1203 14:33:45.144985 4805 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-5fbc4d444f-5vhgg"] Dec 03 14:33:46 crc kubenswrapper[4805]: I1203 14:33:46.110746 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"8622d91c-e48b-46a9-a3e7-ba9b9f2dd257","Type":"ContainerStarted","Data":"bf767c08ed5f4dc072b01a0f6bd513d3a6c2d7098a2aebfcd86c041599dc20ef"} Dec 03 14:33:46 crc kubenswrapper[4805]: I1203 14:33:46.111206 4805 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Dec 03 14:33:46 crc kubenswrapper[4805]: I1203 14:33:46.111210 4805 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="8622d91c-e48b-46a9-a3e7-ba9b9f2dd257" containerName="sg-core" containerID="cri-o://cdfe7cc1361c48b38876bd8b0e963bcd5f0e6bd7b096019ecfd6d8523b35b5e5" gracePeriod=30 Dec 03 14:33:46 crc kubenswrapper[4805]: I1203 14:33:46.111261 4805 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="8622d91c-e48b-46a9-a3e7-ba9b9f2dd257" containerName="ceilometer-notification-agent" containerID="cri-o://e1b9fd988130f8d3338ae8849b1514d22bc3e2dde1c18a30f2f1007209da0368" gracePeriod=30 Dec 03 14:33:46 crc kubenswrapper[4805]: I1203 14:33:46.111259 4805 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="8622d91c-e48b-46a9-a3e7-ba9b9f2dd257" containerName="proxy-httpd" containerID="cri-o://bf767c08ed5f4dc072b01a0f6bd513d3a6c2d7098a2aebfcd86c041599dc20ef" gracePeriod=30 Dec 03 14:33:46 crc kubenswrapper[4805]: I1203 14:33:46.110911 4805 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="8622d91c-e48b-46a9-a3e7-ba9b9f2dd257" containerName="ceilometer-central-agent" containerID="cri-o://8abb17bdb2ad683fc76656be33369b0da58f730c86bd5999b64e778d928b2cfb" gracePeriod=30 Dec 03 14:33:46 crc kubenswrapper[4805]: I1203 14:33:46.149958 4805 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=1.78914466 podStartE2EDuration="8.149933863s" podCreationTimestamp="2025-12-03 14:33:38 +0000 UTC" firstStartedPulling="2025-12-03 14:33:38.869559669 +0000 UTC m=+1448.532476602" lastFinishedPulling="2025-12-03 14:33:45.230348882 +0000 UTC m=+1454.893265805" observedRunningTime="2025-12-03 14:33:46.147376684 +0000 UTC m=+1455.810293607" watchObservedRunningTime="2025-12-03 14:33:46.149933863 +0000 UTC m=+1455.812850796" Dec 03 14:33:46 crc kubenswrapper[4805]: I1203 14:33:46.705850 4805 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e2ef61bb-5836-4986-88ae-f307ee375f0e" path="/var/lib/kubelet/pods/e2ef61bb-5836-4986-88ae-f307ee375f0e/volumes" Dec 03 14:33:47 crc kubenswrapper[4805]: I1203 14:33:47.123492 4805 generic.go:334] "Generic (PLEG): container finished" podID="8622d91c-e48b-46a9-a3e7-ba9b9f2dd257" containerID="bf767c08ed5f4dc072b01a0f6bd513d3a6c2d7098a2aebfcd86c041599dc20ef" exitCode=0 Dec 03 14:33:47 crc kubenswrapper[4805]: I1203 14:33:47.123623 4805 generic.go:334] "Generic (PLEG): container finished" podID="8622d91c-e48b-46a9-a3e7-ba9b9f2dd257" containerID="cdfe7cc1361c48b38876bd8b0e963bcd5f0e6bd7b096019ecfd6d8523b35b5e5" exitCode=2 Dec 03 14:33:47 crc kubenswrapper[4805]: I1203 14:33:47.123539 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"8622d91c-e48b-46a9-a3e7-ba9b9f2dd257","Type":"ContainerDied","Data":"bf767c08ed5f4dc072b01a0f6bd513d3a6c2d7098a2aebfcd86c041599dc20ef"} Dec 03 14:33:47 crc kubenswrapper[4805]: I1203 14:33:47.123670 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"8622d91c-e48b-46a9-a3e7-ba9b9f2dd257","Type":"ContainerDied","Data":"cdfe7cc1361c48b38876bd8b0e963bcd5f0e6bd7b096019ecfd6d8523b35b5e5"} Dec 03 14:33:47 crc kubenswrapper[4805]: I1203 14:33:47.123687 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"8622d91c-e48b-46a9-a3e7-ba9b9f2dd257","Type":"ContainerDied","Data":"e1b9fd988130f8d3338ae8849b1514d22bc3e2dde1c18a30f2f1007209da0368"} Dec 03 14:33:47 crc kubenswrapper[4805]: I1203 14:33:47.123639 4805 generic.go:334] "Generic (PLEG): container finished" podID="8622d91c-e48b-46a9-a3e7-ba9b9f2dd257" containerID="e1b9fd988130f8d3338ae8849b1514d22bc3e2dde1c18a30f2f1007209da0368" exitCode=0 Dec 03 14:33:48 crc kubenswrapper[4805]: I1203 14:33:48.139961 4805 generic.go:334] "Generic (PLEG): container finished" podID="cae0a70b-e47c-40ff-9b40-40a2aa570097" containerID="d384c8b48d1a97e7c0c06910a1049b12545dc37afe379f1e5d0a1b926448aa6b" exitCode=0 Dec 03 14:33:48 crc kubenswrapper[4805]: I1203 14:33:48.140190 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-cell-mapping-pkj2h" event={"ID":"cae0a70b-e47c-40ff-9b40-40a2aa570097","Type":"ContainerDied","Data":"d384c8b48d1a97e7c0c06910a1049b12545dc37afe379f1e5d0a1b926448aa6b"} Dec 03 14:33:48 crc kubenswrapper[4805]: I1203 14:33:48.995984 4805 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 03 14:33:49 crc kubenswrapper[4805]: I1203 14:33:49.044524 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/8622d91c-e48b-46a9-a3e7-ba9b9f2dd257-log-httpd\") pod \"8622d91c-e48b-46a9-a3e7-ba9b9f2dd257\" (UID: \"8622d91c-e48b-46a9-a3e7-ba9b9f2dd257\") " Dec 03 14:33:49 crc kubenswrapper[4805]: I1203 14:33:49.044709 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8622d91c-e48b-46a9-a3e7-ba9b9f2dd257-config-data\") pod \"8622d91c-e48b-46a9-a3e7-ba9b9f2dd257\" (UID: \"8622d91c-e48b-46a9-a3e7-ba9b9f2dd257\") " Dec 03 14:33:49 crc kubenswrapper[4805]: I1203 14:33:49.044783 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/8622d91c-e48b-46a9-a3e7-ba9b9f2dd257-sg-core-conf-yaml\") pod \"8622d91c-e48b-46a9-a3e7-ba9b9f2dd257\" (UID: \"8622d91c-e48b-46a9-a3e7-ba9b9f2dd257\") " Dec 03 14:33:49 crc kubenswrapper[4805]: I1203 14:33:49.044900 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8622d91c-e48b-46a9-a3e7-ba9b9f2dd257-combined-ca-bundle\") pod \"8622d91c-e48b-46a9-a3e7-ba9b9f2dd257\" (UID: \"8622d91c-e48b-46a9-a3e7-ba9b9f2dd257\") " Dec 03 14:33:49 crc kubenswrapper[4805]: I1203 14:33:49.044937 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/8622d91c-e48b-46a9-a3e7-ba9b9f2dd257-run-httpd\") pod \"8622d91c-e48b-46a9-a3e7-ba9b9f2dd257\" (UID: \"8622d91c-e48b-46a9-a3e7-ba9b9f2dd257\") " Dec 03 14:33:49 crc kubenswrapper[4805]: I1203 14:33:49.044972 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9scm6\" (UniqueName: \"kubernetes.io/projected/8622d91c-e48b-46a9-a3e7-ba9b9f2dd257-kube-api-access-9scm6\") pod \"8622d91c-e48b-46a9-a3e7-ba9b9f2dd257\" (UID: \"8622d91c-e48b-46a9-a3e7-ba9b9f2dd257\") " Dec 03 14:33:49 crc kubenswrapper[4805]: I1203 14:33:49.045027 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/8622d91c-e48b-46a9-a3e7-ba9b9f2dd257-ceilometer-tls-certs\") pod \"8622d91c-e48b-46a9-a3e7-ba9b9f2dd257\" (UID: \"8622d91c-e48b-46a9-a3e7-ba9b9f2dd257\") " Dec 03 14:33:49 crc kubenswrapper[4805]: I1203 14:33:49.045071 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8622d91c-e48b-46a9-a3e7-ba9b9f2dd257-scripts\") pod \"8622d91c-e48b-46a9-a3e7-ba9b9f2dd257\" (UID: \"8622d91c-e48b-46a9-a3e7-ba9b9f2dd257\") " Dec 03 14:33:49 crc kubenswrapper[4805]: I1203 14:33:49.045181 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8622d91c-e48b-46a9-a3e7-ba9b9f2dd257-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "8622d91c-e48b-46a9-a3e7-ba9b9f2dd257" (UID: "8622d91c-e48b-46a9-a3e7-ba9b9f2dd257"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 14:33:49 crc kubenswrapper[4805]: I1203 14:33:49.045597 4805 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/8622d91c-e48b-46a9-a3e7-ba9b9f2dd257-run-httpd\") on node \"crc\" DevicePath \"\"" Dec 03 14:33:49 crc kubenswrapper[4805]: I1203 14:33:49.045793 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8622d91c-e48b-46a9-a3e7-ba9b9f2dd257-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "8622d91c-e48b-46a9-a3e7-ba9b9f2dd257" (UID: "8622d91c-e48b-46a9-a3e7-ba9b9f2dd257"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 14:33:49 crc kubenswrapper[4805]: I1203 14:33:49.062942 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8622d91c-e48b-46a9-a3e7-ba9b9f2dd257-scripts" (OuterVolumeSpecName: "scripts") pod "8622d91c-e48b-46a9-a3e7-ba9b9f2dd257" (UID: "8622d91c-e48b-46a9-a3e7-ba9b9f2dd257"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 14:33:49 crc kubenswrapper[4805]: I1203 14:33:49.063055 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8622d91c-e48b-46a9-a3e7-ba9b9f2dd257-kube-api-access-9scm6" (OuterVolumeSpecName: "kube-api-access-9scm6") pod "8622d91c-e48b-46a9-a3e7-ba9b9f2dd257" (UID: "8622d91c-e48b-46a9-a3e7-ba9b9f2dd257"). InnerVolumeSpecName "kube-api-access-9scm6". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 14:33:49 crc kubenswrapper[4805]: I1203 14:33:49.081301 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8622d91c-e48b-46a9-a3e7-ba9b9f2dd257-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "8622d91c-e48b-46a9-a3e7-ba9b9f2dd257" (UID: "8622d91c-e48b-46a9-a3e7-ba9b9f2dd257"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 14:33:49 crc kubenswrapper[4805]: I1203 14:33:49.099230 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8622d91c-e48b-46a9-a3e7-ba9b9f2dd257-ceilometer-tls-certs" (OuterVolumeSpecName: "ceilometer-tls-certs") pod "8622d91c-e48b-46a9-a3e7-ba9b9f2dd257" (UID: "8622d91c-e48b-46a9-a3e7-ba9b9f2dd257"). InnerVolumeSpecName "ceilometer-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 14:33:49 crc kubenswrapper[4805]: I1203 14:33:49.119054 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8622d91c-e48b-46a9-a3e7-ba9b9f2dd257-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "8622d91c-e48b-46a9-a3e7-ba9b9f2dd257" (UID: "8622d91c-e48b-46a9-a3e7-ba9b9f2dd257"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 14:33:49 crc kubenswrapper[4805]: I1203 14:33:49.145195 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8622d91c-e48b-46a9-a3e7-ba9b9f2dd257-config-data" (OuterVolumeSpecName: "config-data") pod "8622d91c-e48b-46a9-a3e7-ba9b9f2dd257" (UID: "8622d91c-e48b-46a9-a3e7-ba9b9f2dd257"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 14:33:49 crc kubenswrapper[4805]: I1203 14:33:49.147493 4805 reconciler_common.go:293] "Volume detached for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/8622d91c-e48b-46a9-a3e7-ba9b9f2dd257-ceilometer-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 03 14:33:49 crc kubenswrapper[4805]: I1203 14:33:49.147524 4805 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8622d91c-e48b-46a9-a3e7-ba9b9f2dd257-scripts\") on node \"crc\" DevicePath \"\"" Dec 03 14:33:49 crc kubenswrapper[4805]: I1203 14:33:49.147535 4805 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/8622d91c-e48b-46a9-a3e7-ba9b9f2dd257-log-httpd\") on node \"crc\" DevicePath \"\"" Dec 03 14:33:49 crc kubenswrapper[4805]: I1203 14:33:49.147545 4805 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8622d91c-e48b-46a9-a3e7-ba9b9f2dd257-config-data\") on node \"crc\" DevicePath \"\"" Dec 03 14:33:49 crc kubenswrapper[4805]: I1203 14:33:49.147554 4805 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/8622d91c-e48b-46a9-a3e7-ba9b9f2dd257-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Dec 03 14:33:49 crc kubenswrapper[4805]: I1203 14:33:49.147562 4805 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8622d91c-e48b-46a9-a3e7-ba9b9f2dd257-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 14:33:49 crc kubenswrapper[4805]: I1203 14:33:49.147570 4805 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9scm6\" (UniqueName: \"kubernetes.io/projected/8622d91c-e48b-46a9-a3e7-ba9b9f2dd257-kube-api-access-9scm6\") on node \"crc\" DevicePath \"\"" Dec 03 14:33:49 crc kubenswrapper[4805]: I1203 14:33:49.153109 4805 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 03 14:33:49 crc kubenswrapper[4805]: I1203 14:33:49.153105 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"8622d91c-e48b-46a9-a3e7-ba9b9f2dd257","Type":"ContainerDied","Data":"8abb17bdb2ad683fc76656be33369b0da58f730c86bd5999b64e778d928b2cfb"} Dec 03 14:33:49 crc kubenswrapper[4805]: I1203 14:33:49.153207 4805 scope.go:117] "RemoveContainer" containerID="bf767c08ed5f4dc072b01a0f6bd513d3a6c2d7098a2aebfcd86c041599dc20ef" Dec 03 14:33:49 crc kubenswrapper[4805]: I1203 14:33:49.153354 4805 generic.go:334] "Generic (PLEG): container finished" podID="8622d91c-e48b-46a9-a3e7-ba9b9f2dd257" containerID="8abb17bdb2ad683fc76656be33369b0da58f730c86bd5999b64e778d928b2cfb" exitCode=0 Dec 03 14:33:49 crc kubenswrapper[4805]: I1203 14:33:49.153488 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"8622d91c-e48b-46a9-a3e7-ba9b9f2dd257","Type":"ContainerDied","Data":"11e06a37d4e01ebbe21956ff796cbe0c8579599266d9cf9a9c109fe3597c9d5f"} Dec 03 14:33:49 crc kubenswrapper[4805]: I1203 14:33:49.236750 4805 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 03 14:33:49 crc kubenswrapper[4805]: I1203 14:33:49.250497 4805 scope.go:117] "RemoveContainer" containerID="cdfe7cc1361c48b38876bd8b0e963bcd5f0e6bd7b096019ecfd6d8523b35b5e5" Dec 03 14:33:49 crc kubenswrapper[4805]: I1203 14:33:49.256475 4805 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Dec 03 14:33:49 crc kubenswrapper[4805]: I1203 14:33:49.272707 4805 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Dec 03 14:33:49 crc kubenswrapper[4805]: E1203 14:33:49.273211 4805 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8622d91c-e48b-46a9-a3e7-ba9b9f2dd257" containerName="sg-core" Dec 03 14:33:49 crc kubenswrapper[4805]: I1203 14:33:49.273226 4805 state_mem.go:107] "Deleted CPUSet assignment" podUID="8622d91c-e48b-46a9-a3e7-ba9b9f2dd257" containerName="sg-core" Dec 03 14:33:49 crc kubenswrapper[4805]: E1203 14:33:49.273264 4805 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8622d91c-e48b-46a9-a3e7-ba9b9f2dd257" containerName="ceilometer-central-agent" Dec 03 14:33:49 crc kubenswrapper[4805]: I1203 14:33:49.273275 4805 state_mem.go:107] "Deleted CPUSet assignment" podUID="8622d91c-e48b-46a9-a3e7-ba9b9f2dd257" containerName="ceilometer-central-agent" Dec 03 14:33:49 crc kubenswrapper[4805]: E1203 14:33:49.273291 4805 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e2ef61bb-5836-4986-88ae-f307ee375f0e" containerName="init" Dec 03 14:33:49 crc kubenswrapper[4805]: I1203 14:33:49.273299 4805 state_mem.go:107] "Deleted CPUSet assignment" podUID="e2ef61bb-5836-4986-88ae-f307ee375f0e" containerName="init" Dec 03 14:33:49 crc kubenswrapper[4805]: E1203 14:33:49.273315 4805 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8622d91c-e48b-46a9-a3e7-ba9b9f2dd257" containerName="ceilometer-notification-agent" Dec 03 14:33:49 crc kubenswrapper[4805]: I1203 14:33:49.273323 4805 state_mem.go:107] "Deleted CPUSet assignment" podUID="8622d91c-e48b-46a9-a3e7-ba9b9f2dd257" containerName="ceilometer-notification-agent" Dec 03 14:33:49 crc kubenswrapper[4805]: E1203 14:33:49.273341 4805 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e2ef61bb-5836-4986-88ae-f307ee375f0e" containerName="dnsmasq-dns" Dec 03 14:33:49 crc kubenswrapper[4805]: I1203 14:33:49.273350 4805 state_mem.go:107] "Deleted CPUSet assignment" podUID="e2ef61bb-5836-4986-88ae-f307ee375f0e" containerName="dnsmasq-dns" Dec 03 14:33:49 crc kubenswrapper[4805]: E1203 14:33:49.273363 4805 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8622d91c-e48b-46a9-a3e7-ba9b9f2dd257" containerName="proxy-httpd" Dec 03 14:33:49 crc kubenswrapper[4805]: I1203 14:33:49.273371 4805 state_mem.go:107] "Deleted CPUSet assignment" podUID="8622d91c-e48b-46a9-a3e7-ba9b9f2dd257" containerName="proxy-httpd" Dec 03 14:33:49 crc kubenswrapper[4805]: I1203 14:33:49.273682 4805 memory_manager.go:354] "RemoveStaleState removing state" podUID="8622d91c-e48b-46a9-a3e7-ba9b9f2dd257" containerName="ceilometer-notification-agent" Dec 03 14:33:49 crc kubenswrapper[4805]: I1203 14:33:49.273699 4805 memory_manager.go:354] "RemoveStaleState removing state" podUID="8622d91c-e48b-46a9-a3e7-ba9b9f2dd257" containerName="ceilometer-central-agent" Dec 03 14:33:49 crc kubenswrapper[4805]: I1203 14:33:49.273719 4805 memory_manager.go:354] "RemoveStaleState removing state" podUID="8622d91c-e48b-46a9-a3e7-ba9b9f2dd257" containerName="sg-core" Dec 03 14:33:49 crc kubenswrapper[4805]: I1203 14:33:49.273728 4805 memory_manager.go:354] "RemoveStaleState removing state" podUID="8622d91c-e48b-46a9-a3e7-ba9b9f2dd257" containerName="proxy-httpd" Dec 03 14:33:49 crc kubenswrapper[4805]: I1203 14:33:49.273745 4805 memory_manager.go:354] "RemoveStaleState removing state" podUID="e2ef61bb-5836-4986-88ae-f307ee375f0e" containerName="dnsmasq-dns" Dec 03 14:33:49 crc kubenswrapper[4805]: I1203 14:33:49.275526 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 03 14:33:49 crc kubenswrapper[4805]: I1203 14:33:49.277474 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Dec 03 14:33:49 crc kubenswrapper[4805]: I1203 14:33:49.277662 4805 scope.go:117] "RemoveContainer" containerID="e1b9fd988130f8d3338ae8849b1514d22bc3e2dde1c18a30f2f1007209da0368" Dec 03 14:33:49 crc kubenswrapper[4805]: I1203 14:33:49.277879 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Dec 03 14:33:49 crc kubenswrapper[4805]: I1203 14:33:49.278932 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ceilometer-internal-svc" Dec 03 14:33:49 crc kubenswrapper[4805]: I1203 14:33:49.281430 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 03 14:33:49 crc kubenswrapper[4805]: I1203 14:33:49.301522 4805 scope.go:117] "RemoveContainer" containerID="8abb17bdb2ad683fc76656be33369b0da58f730c86bd5999b64e778d928b2cfb" Dec 03 14:33:49 crc kubenswrapper[4805]: I1203 14:33:49.322727 4805 scope.go:117] "RemoveContainer" containerID="bf767c08ed5f4dc072b01a0f6bd513d3a6c2d7098a2aebfcd86c041599dc20ef" Dec 03 14:33:49 crc kubenswrapper[4805]: E1203 14:33:49.323113 4805 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"bf767c08ed5f4dc072b01a0f6bd513d3a6c2d7098a2aebfcd86c041599dc20ef\": container with ID starting with bf767c08ed5f4dc072b01a0f6bd513d3a6c2d7098a2aebfcd86c041599dc20ef not found: ID does not exist" containerID="bf767c08ed5f4dc072b01a0f6bd513d3a6c2d7098a2aebfcd86c041599dc20ef" Dec 03 14:33:49 crc kubenswrapper[4805]: I1203 14:33:49.323146 4805 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"bf767c08ed5f4dc072b01a0f6bd513d3a6c2d7098a2aebfcd86c041599dc20ef"} err="failed to get container status \"bf767c08ed5f4dc072b01a0f6bd513d3a6c2d7098a2aebfcd86c041599dc20ef\": rpc error: code = NotFound desc = could not find container \"bf767c08ed5f4dc072b01a0f6bd513d3a6c2d7098a2aebfcd86c041599dc20ef\": container with ID starting with bf767c08ed5f4dc072b01a0f6bd513d3a6c2d7098a2aebfcd86c041599dc20ef not found: ID does not exist" Dec 03 14:33:49 crc kubenswrapper[4805]: I1203 14:33:49.323170 4805 scope.go:117] "RemoveContainer" containerID="cdfe7cc1361c48b38876bd8b0e963bcd5f0e6bd7b096019ecfd6d8523b35b5e5" Dec 03 14:33:49 crc kubenswrapper[4805]: E1203 14:33:49.323384 4805 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"cdfe7cc1361c48b38876bd8b0e963bcd5f0e6bd7b096019ecfd6d8523b35b5e5\": container with ID starting with cdfe7cc1361c48b38876bd8b0e963bcd5f0e6bd7b096019ecfd6d8523b35b5e5 not found: ID does not exist" containerID="cdfe7cc1361c48b38876bd8b0e963bcd5f0e6bd7b096019ecfd6d8523b35b5e5" Dec 03 14:33:49 crc kubenswrapper[4805]: I1203 14:33:49.323407 4805 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"cdfe7cc1361c48b38876bd8b0e963bcd5f0e6bd7b096019ecfd6d8523b35b5e5"} err="failed to get container status \"cdfe7cc1361c48b38876bd8b0e963bcd5f0e6bd7b096019ecfd6d8523b35b5e5\": rpc error: code = NotFound desc = could not find container \"cdfe7cc1361c48b38876bd8b0e963bcd5f0e6bd7b096019ecfd6d8523b35b5e5\": container with ID starting with cdfe7cc1361c48b38876bd8b0e963bcd5f0e6bd7b096019ecfd6d8523b35b5e5 not found: ID does not exist" Dec 03 14:33:49 crc kubenswrapper[4805]: I1203 14:33:49.323424 4805 scope.go:117] "RemoveContainer" containerID="e1b9fd988130f8d3338ae8849b1514d22bc3e2dde1c18a30f2f1007209da0368" Dec 03 14:33:49 crc kubenswrapper[4805]: E1203 14:33:49.323662 4805 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e1b9fd988130f8d3338ae8849b1514d22bc3e2dde1c18a30f2f1007209da0368\": container with ID starting with e1b9fd988130f8d3338ae8849b1514d22bc3e2dde1c18a30f2f1007209da0368 not found: ID does not exist" containerID="e1b9fd988130f8d3338ae8849b1514d22bc3e2dde1c18a30f2f1007209da0368" Dec 03 14:33:49 crc kubenswrapper[4805]: I1203 14:33:49.323690 4805 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e1b9fd988130f8d3338ae8849b1514d22bc3e2dde1c18a30f2f1007209da0368"} err="failed to get container status \"e1b9fd988130f8d3338ae8849b1514d22bc3e2dde1c18a30f2f1007209da0368\": rpc error: code = NotFound desc = could not find container \"e1b9fd988130f8d3338ae8849b1514d22bc3e2dde1c18a30f2f1007209da0368\": container with ID starting with e1b9fd988130f8d3338ae8849b1514d22bc3e2dde1c18a30f2f1007209da0368 not found: ID does not exist" Dec 03 14:33:49 crc kubenswrapper[4805]: I1203 14:33:49.323709 4805 scope.go:117] "RemoveContainer" containerID="8abb17bdb2ad683fc76656be33369b0da58f730c86bd5999b64e778d928b2cfb" Dec 03 14:33:49 crc kubenswrapper[4805]: E1203 14:33:49.324032 4805 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8abb17bdb2ad683fc76656be33369b0da58f730c86bd5999b64e778d928b2cfb\": container with ID starting with 8abb17bdb2ad683fc76656be33369b0da58f730c86bd5999b64e778d928b2cfb not found: ID does not exist" containerID="8abb17bdb2ad683fc76656be33369b0da58f730c86bd5999b64e778d928b2cfb" Dec 03 14:33:49 crc kubenswrapper[4805]: I1203 14:33:49.324053 4805 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8abb17bdb2ad683fc76656be33369b0da58f730c86bd5999b64e778d928b2cfb"} err="failed to get container status \"8abb17bdb2ad683fc76656be33369b0da58f730c86bd5999b64e778d928b2cfb\": rpc error: code = NotFound desc = could not find container \"8abb17bdb2ad683fc76656be33369b0da58f730c86bd5999b64e778d928b2cfb\": container with ID starting with 8abb17bdb2ad683fc76656be33369b0da58f730c86bd5999b64e778d928b2cfb not found: ID does not exist" Dec 03 14:33:49 crc kubenswrapper[4805]: I1203 14:33:49.353762 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hdg4f\" (UniqueName: \"kubernetes.io/projected/aee26d6d-7304-4001-90f8-c0ed886a3cda-kube-api-access-hdg4f\") pod \"ceilometer-0\" (UID: \"aee26d6d-7304-4001-90f8-c0ed886a3cda\") " pod="openstack/ceilometer-0" Dec 03 14:33:49 crc kubenswrapper[4805]: I1203 14:33:49.353821 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/aee26d6d-7304-4001-90f8-c0ed886a3cda-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"aee26d6d-7304-4001-90f8-c0ed886a3cda\") " pod="openstack/ceilometer-0" Dec 03 14:33:49 crc kubenswrapper[4805]: I1203 14:33:49.353909 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/aee26d6d-7304-4001-90f8-c0ed886a3cda-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"aee26d6d-7304-4001-90f8-c0ed886a3cda\") " pod="openstack/ceilometer-0" Dec 03 14:33:49 crc kubenswrapper[4805]: I1203 14:33:49.353965 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/aee26d6d-7304-4001-90f8-c0ed886a3cda-config-data\") pod \"ceilometer-0\" (UID: \"aee26d6d-7304-4001-90f8-c0ed886a3cda\") " pod="openstack/ceilometer-0" Dec 03 14:33:49 crc kubenswrapper[4805]: I1203 14:33:49.354005 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/aee26d6d-7304-4001-90f8-c0ed886a3cda-scripts\") pod \"ceilometer-0\" (UID: \"aee26d6d-7304-4001-90f8-c0ed886a3cda\") " pod="openstack/ceilometer-0" Dec 03 14:33:49 crc kubenswrapper[4805]: I1203 14:33:49.354246 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/aee26d6d-7304-4001-90f8-c0ed886a3cda-log-httpd\") pod \"ceilometer-0\" (UID: \"aee26d6d-7304-4001-90f8-c0ed886a3cda\") " pod="openstack/ceilometer-0" Dec 03 14:33:49 crc kubenswrapper[4805]: I1203 14:33:49.354462 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/aee26d6d-7304-4001-90f8-c0ed886a3cda-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"aee26d6d-7304-4001-90f8-c0ed886a3cda\") " pod="openstack/ceilometer-0" Dec 03 14:33:49 crc kubenswrapper[4805]: I1203 14:33:49.354507 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/aee26d6d-7304-4001-90f8-c0ed886a3cda-run-httpd\") pod \"ceilometer-0\" (UID: \"aee26d6d-7304-4001-90f8-c0ed886a3cda\") " pod="openstack/ceilometer-0" Dec 03 14:33:49 crc kubenswrapper[4805]: I1203 14:33:49.456251 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/aee26d6d-7304-4001-90f8-c0ed886a3cda-log-httpd\") pod \"ceilometer-0\" (UID: \"aee26d6d-7304-4001-90f8-c0ed886a3cda\") " pod="openstack/ceilometer-0" Dec 03 14:33:49 crc kubenswrapper[4805]: I1203 14:33:49.456362 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/aee26d6d-7304-4001-90f8-c0ed886a3cda-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"aee26d6d-7304-4001-90f8-c0ed886a3cda\") " pod="openstack/ceilometer-0" Dec 03 14:33:49 crc kubenswrapper[4805]: I1203 14:33:49.456396 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/aee26d6d-7304-4001-90f8-c0ed886a3cda-run-httpd\") pod \"ceilometer-0\" (UID: \"aee26d6d-7304-4001-90f8-c0ed886a3cda\") " pod="openstack/ceilometer-0" Dec 03 14:33:49 crc kubenswrapper[4805]: I1203 14:33:49.456429 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hdg4f\" (UniqueName: \"kubernetes.io/projected/aee26d6d-7304-4001-90f8-c0ed886a3cda-kube-api-access-hdg4f\") pod \"ceilometer-0\" (UID: \"aee26d6d-7304-4001-90f8-c0ed886a3cda\") " pod="openstack/ceilometer-0" Dec 03 14:33:49 crc kubenswrapper[4805]: I1203 14:33:49.456454 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/aee26d6d-7304-4001-90f8-c0ed886a3cda-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"aee26d6d-7304-4001-90f8-c0ed886a3cda\") " pod="openstack/ceilometer-0" Dec 03 14:33:49 crc kubenswrapper[4805]: I1203 14:33:49.456479 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/aee26d6d-7304-4001-90f8-c0ed886a3cda-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"aee26d6d-7304-4001-90f8-c0ed886a3cda\") " pod="openstack/ceilometer-0" Dec 03 14:33:49 crc kubenswrapper[4805]: I1203 14:33:49.456891 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/aee26d6d-7304-4001-90f8-c0ed886a3cda-log-httpd\") pod \"ceilometer-0\" (UID: \"aee26d6d-7304-4001-90f8-c0ed886a3cda\") " pod="openstack/ceilometer-0" Dec 03 14:33:49 crc kubenswrapper[4805]: I1203 14:33:49.456536 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/aee26d6d-7304-4001-90f8-c0ed886a3cda-config-data\") pod \"ceilometer-0\" (UID: \"aee26d6d-7304-4001-90f8-c0ed886a3cda\") " pod="openstack/ceilometer-0" Dec 03 14:33:49 crc kubenswrapper[4805]: I1203 14:33:49.456976 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/aee26d6d-7304-4001-90f8-c0ed886a3cda-scripts\") pod \"ceilometer-0\" (UID: \"aee26d6d-7304-4001-90f8-c0ed886a3cda\") " pod="openstack/ceilometer-0" Dec 03 14:33:49 crc kubenswrapper[4805]: I1203 14:33:49.459231 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/aee26d6d-7304-4001-90f8-c0ed886a3cda-run-httpd\") pod \"ceilometer-0\" (UID: \"aee26d6d-7304-4001-90f8-c0ed886a3cda\") " pod="openstack/ceilometer-0" Dec 03 14:33:49 crc kubenswrapper[4805]: I1203 14:33:49.476749 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/aee26d6d-7304-4001-90f8-c0ed886a3cda-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"aee26d6d-7304-4001-90f8-c0ed886a3cda\") " pod="openstack/ceilometer-0" Dec 03 14:33:49 crc kubenswrapper[4805]: I1203 14:33:49.476829 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/aee26d6d-7304-4001-90f8-c0ed886a3cda-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"aee26d6d-7304-4001-90f8-c0ed886a3cda\") " pod="openstack/ceilometer-0" Dec 03 14:33:49 crc kubenswrapper[4805]: I1203 14:33:49.476897 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/aee26d6d-7304-4001-90f8-c0ed886a3cda-scripts\") pod \"ceilometer-0\" (UID: \"aee26d6d-7304-4001-90f8-c0ed886a3cda\") " pod="openstack/ceilometer-0" Dec 03 14:33:49 crc kubenswrapper[4805]: I1203 14:33:49.476934 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/aee26d6d-7304-4001-90f8-c0ed886a3cda-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"aee26d6d-7304-4001-90f8-c0ed886a3cda\") " pod="openstack/ceilometer-0" Dec 03 14:33:49 crc kubenswrapper[4805]: I1203 14:33:49.477608 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/aee26d6d-7304-4001-90f8-c0ed886a3cda-config-data\") pod \"ceilometer-0\" (UID: \"aee26d6d-7304-4001-90f8-c0ed886a3cda\") " pod="openstack/ceilometer-0" Dec 03 14:33:49 crc kubenswrapper[4805]: I1203 14:33:49.479721 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hdg4f\" (UniqueName: \"kubernetes.io/projected/aee26d6d-7304-4001-90f8-c0ed886a3cda-kube-api-access-hdg4f\") pod \"ceilometer-0\" (UID: \"aee26d6d-7304-4001-90f8-c0ed886a3cda\") " pod="openstack/ceilometer-0" Dec 03 14:33:49 crc kubenswrapper[4805]: I1203 14:33:49.526340 4805 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-cell-mapping-pkj2h" Dec 03 14:33:49 crc kubenswrapper[4805]: I1203 14:33:49.599239 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 03 14:33:49 crc kubenswrapper[4805]: I1203 14:33:49.672060 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hmpbx\" (UniqueName: \"kubernetes.io/projected/cae0a70b-e47c-40ff-9b40-40a2aa570097-kube-api-access-hmpbx\") pod \"cae0a70b-e47c-40ff-9b40-40a2aa570097\" (UID: \"cae0a70b-e47c-40ff-9b40-40a2aa570097\") " Dec 03 14:33:49 crc kubenswrapper[4805]: I1203 14:33:49.672140 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cae0a70b-e47c-40ff-9b40-40a2aa570097-combined-ca-bundle\") pod \"cae0a70b-e47c-40ff-9b40-40a2aa570097\" (UID: \"cae0a70b-e47c-40ff-9b40-40a2aa570097\") " Dec 03 14:33:49 crc kubenswrapper[4805]: I1203 14:33:49.672293 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cae0a70b-e47c-40ff-9b40-40a2aa570097-config-data\") pod \"cae0a70b-e47c-40ff-9b40-40a2aa570097\" (UID: \"cae0a70b-e47c-40ff-9b40-40a2aa570097\") " Dec 03 14:33:49 crc kubenswrapper[4805]: I1203 14:33:49.672347 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/cae0a70b-e47c-40ff-9b40-40a2aa570097-scripts\") pod \"cae0a70b-e47c-40ff-9b40-40a2aa570097\" (UID: \"cae0a70b-e47c-40ff-9b40-40a2aa570097\") " Dec 03 14:33:49 crc kubenswrapper[4805]: I1203 14:33:49.676570 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/cae0a70b-e47c-40ff-9b40-40a2aa570097-scripts" (OuterVolumeSpecName: "scripts") pod "cae0a70b-e47c-40ff-9b40-40a2aa570097" (UID: "cae0a70b-e47c-40ff-9b40-40a2aa570097"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 14:33:49 crc kubenswrapper[4805]: I1203 14:33:49.678773 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cae0a70b-e47c-40ff-9b40-40a2aa570097-kube-api-access-hmpbx" (OuterVolumeSpecName: "kube-api-access-hmpbx") pod "cae0a70b-e47c-40ff-9b40-40a2aa570097" (UID: "cae0a70b-e47c-40ff-9b40-40a2aa570097"). InnerVolumeSpecName "kube-api-access-hmpbx". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 14:33:49 crc kubenswrapper[4805]: I1203 14:33:49.704741 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/cae0a70b-e47c-40ff-9b40-40a2aa570097-config-data" (OuterVolumeSpecName: "config-data") pod "cae0a70b-e47c-40ff-9b40-40a2aa570097" (UID: "cae0a70b-e47c-40ff-9b40-40a2aa570097"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 14:33:49 crc kubenswrapper[4805]: I1203 14:33:49.708115 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/cae0a70b-e47c-40ff-9b40-40a2aa570097-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "cae0a70b-e47c-40ff-9b40-40a2aa570097" (UID: "cae0a70b-e47c-40ff-9b40-40a2aa570097"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 14:33:49 crc kubenswrapper[4805]: I1203 14:33:49.775076 4805 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hmpbx\" (UniqueName: \"kubernetes.io/projected/cae0a70b-e47c-40ff-9b40-40a2aa570097-kube-api-access-hmpbx\") on node \"crc\" DevicePath \"\"" Dec 03 14:33:49 crc kubenswrapper[4805]: I1203 14:33:49.775462 4805 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cae0a70b-e47c-40ff-9b40-40a2aa570097-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 14:33:49 crc kubenswrapper[4805]: I1203 14:33:49.775472 4805 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cae0a70b-e47c-40ff-9b40-40a2aa570097-config-data\") on node \"crc\" DevicePath \"\"" Dec 03 14:33:49 crc kubenswrapper[4805]: I1203 14:33:49.775480 4805 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/cae0a70b-e47c-40ff-9b40-40a2aa570097-scripts\") on node \"crc\" DevicePath \"\"" Dec 03 14:33:50 crc kubenswrapper[4805]: I1203 14:33:50.048123 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 03 14:33:50 crc kubenswrapper[4805]: W1203 14:33:50.052092 4805 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podaee26d6d_7304_4001_90f8_c0ed886a3cda.slice/crio-b44eea74232b42f321d7fb0358f8f4f1fcb9b6397d3e8156b964488f587b25dc WatchSource:0}: Error finding container b44eea74232b42f321d7fb0358f8f4f1fcb9b6397d3e8156b964488f587b25dc: Status 404 returned error can't find the container with id b44eea74232b42f321d7fb0358f8f4f1fcb9b6397d3e8156b964488f587b25dc Dec 03 14:33:50 crc kubenswrapper[4805]: I1203 14:33:50.163944 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-cell-mapping-pkj2h" event={"ID":"cae0a70b-e47c-40ff-9b40-40a2aa570097","Type":"ContainerDied","Data":"eb74f826e57fdb74da5e57b7dfc2b92730485025ca3d05c13c452b7d09004cc4"} Dec 03 14:33:50 crc kubenswrapper[4805]: I1203 14:33:50.163993 4805 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="eb74f826e57fdb74da5e57b7dfc2b92730485025ca3d05c13c452b7d09004cc4" Dec 03 14:33:50 crc kubenswrapper[4805]: I1203 14:33:50.164007 4805 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-cell-mapping-pkj2h" Dec 03 14:33:50 crc kubenswrapper[4805]: I1203 14:33:50.165373 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"aee26d6d-7304-4001-90f8-c0ed886a3cda","Type":"ContainerStarted","Data":"b44eea74232b42f321d7fb0358f8f4f1fcb9b6397d3e8156b964488f587b25dc"} Dec 03 14:33:50 crc kubenswrapper[4805]: I1203 14:33:50.350308 4805 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Dec 03 14:33:50 crc kubenswrapper[4805]: I1203 14:33:50.351098 4805 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-scheduler-0" podUID="c0ccb82f-2439-448d-a958-5d775682cca3" containerName="nova-scheduler-scheduler" containerID="cri-o://47fa01711b4e60668d53aa93490a39d0905127c17fbc1ca51fd562ed6c516ebc" gracePeriod=30 Dec 03 14:33:50 crc kubenswrapper[4805]: I1203 14:33:50.360969 4805 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Dec 03 14:33:50 crc kubenswrapper[4805]: I1203 14:33:50.361262 4805 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="68636027-732d-46fa-aba1-7f89927292ac" containerName="nova-api-log" containerID="cri-o://2508ae57d37d01ca93f1023c81c6c09acbbbbf95890d14a4951ae7cdfb8ddbc2" gracePeriod=30 Dec 03 14:33:50 crc kubenswrapper[4805]: I1203 14:33:50.361439 4805 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="68636027-732d-46fa-aba1-7f89927292ac" containerName="nova-api-api" containerID="cri-o://d3db209ff6c299ff5989f450f491b40fe35f9073b078b22a33a3a15f6a882663" gracePeriod=30 Dec 03 14:33:50 crc kubenswrapper[4805]: I1203 14:33:50.438191 4805 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Dec 03 14:33:50 crc kubenswrapper[4805]: I1203 14:33:50.438758 4805 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="2dcd6066-cee7-40a7-ba9f-0648b6d2283b" containerName="nova-metadata-log" containerID="cri-o://72253effa6be075609506ec28eb6dc2aa2e9f7d559f7e42103391020b725db73" gracePeriod=30 Dec 03 14:33:50 crc kubenswrapper[4805]: I1203 14:33:50.438963 4805 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="2dcd6066-cee7-40a7-ba9f-0648b6d2283b" containerName="nova-metadata-metadata" containerID="cri-o://59cc0d8dee55d94f88b6b089f411035edd0324469c03b366c0c24c6baa08fe63" gracePeriod=30 Dec 03 14:33:50 crc kubenswrapper[4805]: I1203 14:33:50.707832 4805 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8622d91c-e48b-46a9-a3e7-ba9b9f2dd257" path="/var/lib/kubelet/pods/8622d91c-e48b-46a9-a3e7-ba9b9f2dd257/volumes" Dec 03 14:33:50 crc kubenswrapper[4805]: I1203 14:33:50.867262 4805 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 03 14:33:51 crc kubenswrapper[4805]: I1203 14:33:51.001797 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/68636027-732d-46fa-aba1-7f89927292ac-combined-ca-bundle\") pod \"68636027-732d-46fa-aba1-7f89927292ac\" (UID: \"68636027-732d-46fa-aba1-7f89927292ac\") " Dec 03 14:33:51 crc kubenswrapper[4805]: I1203 14:33:51.001977 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/68636027-732d-46fa-aba1-7f89927292ac-public-tls-certs\") pod \"68636027-732d-46fa-aba1-7f89927292ac\" (UID: \"68636027-732d-46fa-aba1-7f89927292ac\") " Dec 03 14:33:51 crc kubenswrapper[4805]: I1203 14:33:51.002072 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/68636027-732d-46fa-aba1-7f89927292ac-logs\") pod \"68636027-732d-46fa-aba1-7f89927292ac\" (UID: \"68636027-732d-46fa-aba1-7f89927292ac\") " Dec 03 14:33:51 crc kubenswrapper[4805]: I1203 14:33:51.002102 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/68636027-732d-46fa-aba1-7f89927292ac-internal-tls-certs\") pod \"68636027-732d-46fa-aba1-7f89927292ac\" (UID: \"68636027-732d-46fa-aba1-7f89927292ac\") " Dec 03 14:33:51 crc kubenswrapper[4805]: I1203 14:33:51.002177 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pgxts\" (UniqueName: \"kubernetes.io/projected/68636027-732d-46fa-aba1-7f89927292ac-kube-api-access-pgxts\") pod \"68636027-732d-46fa-aba1-7f89927292ac\" (UID: \"68636027-732d-46fa-aba1-7f89927292ac\") " Dec 03 14:33:51 crc kubenswrapper[4805]: I1203 14:33:51.002282 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/68636027-732d-46fa-aba1-7f89927292ac-config-data\") pod \"68636027-732d-46fa-aba1-7f89927292ac\" (UID: \"68636027-732d-46fa-aba1-7f89927292ac\") " Dec 03 14:33:51 crc kubenswrapper[4805]: I1203 14:33:51.002821 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/68636027-732d-46fa-aba1-7f89927292ac-logs" (OuterVolumeSpecName: "logs") pod "68636027-732d-46fa-aba1-7f89927292ac" (UID: "68636027-732d-46fa-aba1-7f89927292ac"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 14:33:51 crc kubenswrapper[4805]: I1203 14:33:51.007065 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/68636027-732d-46fa-aba1-7f89927292ac-kube-api-access-pgxts" (OuterVolumeSpecName: "kube-api-access-pgxts") pod "68636027-732d-46fa-aba1-7f89927292ac" (UID: "68636027-732d-46fa-aba1-7f89927292ac"). InnerVolumeSpecName "kube-api-access-pgxts". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 14:33:51 crc kubenswrapper[4805]: I1203 14:33:51.031442 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/68636027-732d-46fa-aba1-7f89927292ac-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "68636027-732d-46fa-aba1-7f89927292ac" (UID: "68636027-732d-46fa-aba1-7f89927292ac"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 14:33:51 crc kubenswrapper[4805]: I1203 14:33:51.041585 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/68636027-732d-46fa-aba1-7f89927292ac-config-data" (OuterVolumeSpecName: "config-data") pod "68636027-732d-46fa-aba1-7f89927292ac" (UID: "68636027-732d-46fa-aba1-7f89927292ac"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 14:33:51 crc kubenswrapper[4805]: I1203 14:33:51.053721 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/68636027-732d-46fa-aba1-7f89927292ac-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "68636027-732d-46fa-aba1-7f89927292ac" (UID: "68636027-732d-46fa-aba1-7f89927292ac"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 14:33:51 crc kubenswrapper[4805]: I1203 14:33:51.063434 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/68636027-732d-46fa-aba1-7f89927292ac-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "68636027-732d-46fa-aba1-7f89927292ac" (UID: "68636027-732d-46fa-aba1-7f89927292ac"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 14:33:51 crc kubenswrapper[4805]: I1203 14:33:51.104444 4805 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pgxts\" (UniqueName: \"kubernetes.io/projected/68636027-732d-46fa-aba1-7f89927292ac-kube-api-access-pgxts\") on node \"crc\" DevicePath \"\"" Dec 03 14:33:51 crc kubenswrapper[4805]: I1203 14:33:51.104489 4805 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/68636027-732d-46fa-aba1-7f89927292ac-config-data\") on node \"crc\" DevicePath \"\"" Dec 03 14:33:51 crc kubenswrapper[4805]: I1203 14:33:51.104503 4805 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/68636027-732d-46fa-aba1-7f89927292ac-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 14:33:51 crc kubenswrapper[4805]: I1203 14:33:51.104515 4805 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/68636027-732d-46fa-aba1-7f89927292ac-public-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 03 14:33:51 crc kubenswrapper[4805]: I1203 14:33:51.104528 4805 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/68636027-732d-46fa-aba1-7f89927292ac-logs\") on node \"crc\" DevicePath \"\"" Dec 03 14:33:51 crc kubenswrapper[4805]: I1203 14:33:51.104541 4805 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/68636027-732d-46fa-aba1-7f89927292ac-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 03 14:33:51 crc kubenswrapper[4805]: I1203 14:33:51.179010 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"aee26d6d-7304-4001-90f8-c0ed886a3cda","Type":"ContainerStarted","Data":"73941e0d6546b90dd2830c242d337c821bfefb112e9c263e904f7c35bfea49e4"} Dec 03 14:33:51 crc kubenswrapper[4805]: I1203 14:33:51.184959 4805 generic.go:334] "Generic (PLEG): container finished" podID="2dcd6066-cee7-40a7-ba9f-0648b6d2283b" containerID="72253effa6be075609506ec28eb6dc2aa2e9f7d559f7e42103391020b725db73" exitCode=143 Dec 03 14:33:51 crc kubenswrapper[4805]: I1203 14:33:51.185019 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"2dcd6066-cee7-40a7-ba9f-0648b6d2283b","Type":"ContainerDied","Data":"72253effa6be075609506ec28eb6dc2aa2e9f7d559f7e42103391020b725db73"} Dec 03 14:33:51 crc kubenswrapper[4805]: I1203 14:33:51.187072 4805 generic.go:334] "Generic (PLEG): container finished" podID="68636027-732d-46fa-aba1-7f89927292ac" containerID="d3db209ff6c299ff5989f450f491b40fe35f9073b078b22a33a3a15f6a882663" exitCode=0 Dec 03 14:33:51 crc kubenswrapper[4805]: I1203 14:33:51.187114 4805 generic.go:334] "Generic (PLEG): container finished" podID="68636027-732d-46fa-aba1-7f89927292ac" containerID="2508ae57d37d01ca93f1023c81c6c09acbbbbf95890d14a4951ae7cdfb8ddbc2" exitCode=143 Dec 03 14:33:51 crc kubenswrapper[4805]: I1203 14:33:51.187129 4805 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 03 14:33:51 crc kubenswrapper[4805]: I1203 14:33:51.187141 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"68636027-732d-46fa-aba1-7f89927292ac","Type":"ContainerDied","Data":"d3db209ff6c299ff5989f450f491b40fe35f9073b078b22a33a3a15f6a882663"} Dec 03 14:33:51 crc kubenswrapper[4805]: I1203 14:33:51.187174 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"68636027-732d-46fa-aba1-7f89927292ac","Type":"ContainerDied","Data":"2508ae57d37d01ca93f1023c81c6c09acbbbbf95890d14a4951ae7cdfb8ddbc2"} Dec 03 14:33:51 crc kubenswrapper[4805]: I1203 14:33:51.187187 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"68636027-732d-46fa-aba1-7f89927292ac","Type":"ContainerDied","Data":"25a2d3afcc061c8c20cf82112ab18f83427acb4f96f81bae1375681ed8e7c0d8"} Dec 03 14:33:51 crc kubenswrapper[4805]: I1203 14:33:51.187209 4805 scope.go:117] "RemoveContainer" containerID="d3db209ff6c299ff5989f450f491b40fe35f9073b078b22a33a3a15f6a882663" Dec 03 14:33:51 crc kubenswrapper[4805]: I1203 14:33:51.219080 4805 scope.go:117] "RemoveContainer" containerID="2508ae57d37d01ca93f1023c81c6c09acbbbbf95890d14a4951ae7cdfb8ddbc2" Dec 03 14:33:51 crc kubenswrapper[4805]: I1203 14:33:51.235972 4805 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Dec 03 14:33:51 crc kubenswrapper[4805]: I1203 14:33:51.244435 4805 scope.go:117] "RemoveContainer" containerID="d3db209ff6c299ff5989f450f491b40fe35f9073b078b22a33a3a15f6a882663" Dec 03 14:33:51 crc kubenswrapper[4805]: E1203 14:33:51.249126 4805 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d3db209ff6c299ff5989f450f491b40fe35f9073b078b22a33a3a15f6a882663\": container with ID starting with d3db209ff6c299ff5989f450f491b40fe35f9073b078b22a33a3a15f6a882663 not found: ID does not exist" containerID="d3db209ff6c299ff5989f450f491b40fe35f9073b078b22a33a3a15f6a882663" Dec 03 14:33:51 crc kubenswrapper[4805]: I1203 14:33:51.249190 4805 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d3db209ff6c299ff5989f450f491b40fe35f9073b078b22a33a3a15f6a882663"} err="failed to get container status \"d3db209ff6c299ff5989f450f491b40fe35f9073b078b22a33a3a15f6a882663\": rpc error: code = NotFound desc = could not find container \"d3db209ff6c299ff5989f450f491b40fe35f9073b078b22a33a3a15f6a882663\": container with ID starting with d3db209ff6c299ff5989f450f491b40fe35f9073b078b22a33a3a15f6a882663 not found: ID does not exist" Dec 03 14:33:51 crc kubenswrapper[4805]: I1203 14:33:51.249222 4805 scope.go:117] "RemoveContainer" containerID="2508ae57d37d01ca93f1023c81c6c09acbbbbf95890d14a4951ae7cdfb8ddbc2" Dec 03 14:33:51 crc kubenswrapper[4805]: E1203 14:33:51.250139 4805 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2508ae57d37d01ca93f1023c81c6c09acbbbbf95890d14a4951ae7cdfb8ddbc2\": container with ID starting with 2508ae57d37d01ca93f1023c81c6c09acbbbbf95890d14a4951ae7cdfb8ddbc2 not found: ID does not exist" containerID="2508ae57d37d01ca93f1023c81c6c09acbbbbf95890d14a4951ae7cdfb8ddbc2" Dec 03 14:33:51 crc kubenswrapper[4805]: I1203 14:33:51.250190 4805 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2508ae57d37d01ca93f1023c81c6c09acbbbbf95890d14a4951ae7cdfb8ddbc2"} err="failed to get container status \"2508ae57d37d01ca93f1023c81c6c09acbbbbf95890d14a4951ae7cdfb8ddbc2\": rpc error: code = NotFound desc = could not find container \"2508ae57d37d01ca93f1023c81c6c09acbbbbf95890d14a4951ae7cdfb8ddbc2\": container with ID starting with 2508ae57d37d01ca93f1023c81c6c09acbbbbf95890d14a4951ae7cdfb8ddbc2 not found: ID does not exist" Dec 03 14:33:51 crc kubenswrapper[4805]: I1203 14:33:51.250221 4805 scope.go:117] "RemoveContainer" containerID="d3db209ff6c299ff5989f450f491b40fe35f9073b078b22a33a3a15f6a882663" Dec 03 14:33:51 crc kubenswrapper[4805]: I1203 14:33:51.253295 4805 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d3db209ff6c299ff5989f450f491b40fe35f9073b078b22a33a3a15f6a882663"} err="failed to get container status \"d3db209ff6c299ff5989f450f491b40fe35f9073b078b22a33a3a15f6a882663\": rpc error: code = NotFound desc = could not find container \"d3db209ff6c299ff5989f450f491b40fe35f9073b078b22a33a3a15f6a882663\": container with ID starting with d3db209ff6c299ff5989f450f491b40fe35f9073b078b22a33a3a15f6a882663 not found: ID does not exist" Dec 03 14:33:51 crc kubenswrapper[4805]: I1203 14:33:51.253343 4805 scope.go:117] "RemoveContainer" containerID="2508ae57d37d01ca93f1023c81c6c09acbbbbf95890d14a4951ae7cdfb8ddbc2" Dec 03 14:33:51 crc kubenswrapper[4805]: I1203 14:33:51.253447 4805 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-0"] Dec 03 14:33:51 crc kubenswrapper[4805]: I1203 14:33:51.256689 4805 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2508ae57d37d01ca93f1023c81c6c09acbbbbf95890d14a4951ae7cdfb8ddbc2"} err="failed to get container status \"2508ae57d37d01ca93f1023c81c6c09acbbbbf95890d14a4951ae7cdfb8ddbc2\": rpc error: code = NotFound desc = could not find container \"2508ae57d37d01ca93f1023c81c6c09acbbbbf95890d14a4951ae7cdfb8ddbc2\": container with ID starting with 2508ae57d37d01ca93f1023c81c6c09acbbbbf95890d14a4951ae7cdfb8ddbc2 not found: ID does not exist" Dec 03 14:33:51 crc kubenswrapper[4805]: I1203 14:33:51.264729 4805 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-0"] Dec 03 14:33:51 crc kubenswrapper[4805]: E1203 14:33:51.265286 4805 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="68636027-732d-46fa-aba1-7f89927292ac" containerName="nova-api-log" Dec 03 14:33:51 crc kubenswrapper[4805]: I1203 14:33:51.265311 4805 state_mem.go:107] "Deleted CPUSet assignment" podUID="68636027-732d-46fa-aba1-7f89927292ac" containerName="nova-api-log" Dec 03 14:33:51 crc kubenswrapper[4805]: E1203 14:33:51.265326 4805 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="68636027-732d-46fa-aba1-7f89927292ac" containerName="nova-api-api" Dec 03 14:33:51 crc kubenswrapper[4805]: I1203 14:33:51.265337 4805 state_mem.go:107] "Deleted CPUSet assignment" podUID="68636027-732d-46fa-aba1-7f89927292ac" containerName="nova-api-api" Dec 03 14:33:51 crc kubenswrapper[4805]: E1203 14:33:51.265349 4805 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cae0a70b-e47c-40ff-9b40-40a2aa570097" containerName="nova-manage" Dec 03 14:33:51 crc kubenswrapper[4805]: I1203 14:33:51.265360 4805 state_mem.go:107] "Deleted CPUSet assignment" podUID="cae0a70b-e47c-40ff-9b40-40a2aa570097" containerName="nova-manage" Dec 03 14:33:51 crc kubenswrapper[4805]: I1203 14:33:51.265576 4805 memory_manager.go:354] "RemoveStaleState removing state" podUID="cae0a70b-e47c-40ff-9b40-40a2aa570097" containerName="nova-manage" Dec 03 14:33:51 crc kubenswrapper[4805]: I1203 14:33:51.265599 4805 memory_manager.go:354] "RemoveStaleState removing state" podUID="68636027-732d-46fa-aba1-7f89927292ac" containerName="nova-api-log" Dec 03 14:33:51 crc kubenswrapper[4805]: I1203 14:33:51.265628 4805 memory_manager.go:354] "RemoveStaleState removing state" podUID="68636027-732d-46fa-aba1-7f89927292ac" containerName="nova-api-api" Dec 03 14:33:51 crc kubenswrapper[4805]: I1203 14:33:51.266930 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 03 14:33:51 crc kubenswrapper[4805]: I1203 14:33:51.275670 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Dec 03 14:33:51 crc kubenswrapper[4805]: I1203 14:33:51.276383 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-config-data" Dec 03 14:33:51 crc kubenswrapper[4805]: I1203 14:33:51.276614 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-public-svc" Dec 03 14:33:51 crc kubenswrapper[4805]: I1203 14:33:51.278988 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-internal-svc" Dec 03 14:33:51 crc kubenswrapper[4805]: E1203 14:33:51.326354 4805 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="47fa01711b4e60668d53aa93490a39d0905127c17fbc1ca51fd562ed6c516ebc" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Dec 03 14:33:51 crc kubenswrapper[4805]: E1203 14:33:51.327967 4805 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="47fa01711b4e60668d53aa93490a39d0905127c17fbc1ca51fd562ed6c516ebc" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Dec 03 14:33:51 crc kubenswrapper[4805]: E1203 14:33:51.329856 4805 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="47fa01711b4e60668d53aa93490a39d0905127c17fbc1ca51fd562ed6c516ebc" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Dec 03 14:33:51 crc kubenswrapper[4805]: E1203 14:33:51.329894 4805 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openstack/nova-scheduler-0" podUID="c0ccb82f-2439-448d-a958-5d775682cca3" containerName="nova-scheduler-scheduler" Dec 03 14:33:51 crc kubenswrapper[4805]: I1203 14:33:51.415922 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/65386594-9cac-40ef-b3a8-2f3ee225bfb6-internal-tls-certs\") pod \"nova-api-0\" (UID: \"65386594-9cac-40ef-b3a8-2f3ee225bfb6\") " pod="openstack/nova-api-0" Dec 03 14:33:51 crc kubenswrapper[4805]: I1203 14:33:51.416323 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/65386594-9cac-40ef-b3a8-2f3ee225bfb6-public-tls-certs\") pod \"nova-api-0\" (UID: \"65386594-9cac-40ef-b3a8-2f3ee225bfb6\") " pod="openstack/nova-api-0" Dec 03 14:33:51 crc kubenswrapper[4805]: I1203 14:33:51.416354 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/65386594-9cac-40ef-b3a8-2f3ee225bfb6-config-data\") pod \"nova-api-0\" (UID: \"65386594-9cac-40ef-b3a8-2f3ee225bfb6\") " pod="openstack/nova-api-0" Dec 03 14:33:51 crc kubenswrapper[4805]: I1203 14:33:51.416374 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/65386594-9cac-40ef-b3a8-2f3ee225bfb6-logs\") pod \"nova-api-0\" (UID: \"65386594-9cac-40ef-b3a8-2f3ee225bfb6\") " pod="openstack/nova-api-0" Dec 03 14:33:51 crc kubenswrapper[4805]: I1203 14:33:51.416633 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9kgqr\" (UniqueName: \"kubernetes.io/projected/65386594-9cac-40ef-b3a8-2f3ee225bfb6-kube-api-access-9kgqr\") pod \"nova-api-0\" (UID: \"65386594-9cac-40ef-b3a8-2f3ee225bfb6\") " pod="openstack/nova-api-0" Dec 03 14:33:51 crc kubenswrapper[4805]: I1203 14:33:51.416703 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/65386594-9cac-40ef-b3a8-2f3ee225bfb6-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"65386594-9cac-40ef-b3a8-2f3ee225bfb6\") " pod="openstack/nova-api-0" Dec 03 14:33:51 crc kubenswrapper[4805]: I1203 14:33:51.518023 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9kgqr\" (UniqueName: \"kubernetes.io/projected/65386594-9cac-40ef-b3a8-2f3ee225bfb6-kube-api-access-9kgqr\") pod \"nova-api-0\" (UID: \"65386594-9cac-40ef-b3a8-2f3ee225bfb6\") " pod="openstack/nova-api-0" Dec 03 14:33:51 crc kubenswrapper[4805]: I1203 14:33:51.518072 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/65386594-9cac-40ef-b3a8-2f3ee225bfb6-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"65386594-9cac-40ef-b3a8-2f3ee225bfb6\") " pod="openstack/nova-api-0" Dec 03 14:33:51 crc kubenswrapper[4805]: I1203 14:33:51.518104 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/65386594-9cac-40ef-b3a8-2f3ee225bfb6-internal-tls-certs\") pod \"nova-api-0\" (UID: \"65386594-9cac-40ef-b3a8-2f3ee225bfb6\") " pod="openstack/nova-api-0" Dec 03 14:33:51 crc kubenswrapper[4805]: I1203 14:33:51.518161 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/65386594-9cac-40ef-b3a8-2f3ee225bfb6-public-tls-certs\") pod \"nova-api-0\" (UID: \"65386594-9cac-40ef-b3a8-2f3ee225bfb6\") " pod="openstack/nova-api-0" Dec 03 14:33:51 crc kubenswrapper[4805]: I1203 14:33:51.518189 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/65386594-9cac-40ef-b3a8-2f3ee225bfb6-config-data\") pod \"nova-api-0\" (UID: \"65386594-9cac-40ef-b3a8-2f3ee225bfb6\") " pod="openstack/nova-api-0" Dec 03 14:33:51 crc kubenswrapper[4805]: I1203 14:33:51.518211 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/65386594-9cac-40ef-b3a8-2f3ee225bfb6-logs\") pod \"nova-api-0\" (UID: \"65386594-9cac-40ef-b3a8-2f3ee225bfb6\") " pod="openstack/nova-api-0" Dec 03 14:33:51 crc kubenswrapper[4805]: I1203 14:33:51.518762 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/65386594-9cac-40ef-b3a8-2f3ee225bfb6-logs\") pod \"nova-api-0\" (UID: \"65386594-9cac-40ef-b3a8-2f3ee225bfb6\") " pod="openstack/nova-api-0" Dec 03 14:33:51 crc kubenswrapper[4805]: I1203 14:33:51.522210 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/65386594-9cac-40ef-b3a8-2f3ee225bfb6-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"65386594-9cac-40ef-b3a8-2f3ee225bfb6\") " pod="openstack/nova-api-0" Dec 03 14:33:51 crc kubenswrapper[4805]: I1203 14:33:51.522329 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/65386594-9cac-40ef-b3a8-2f3ee225bfb6-public-tls-certs\") pod \"nova-api-0\" (UID: \"65386594-9cac-40ef-b3a8-2f3ee225bfb6\") " pod="openstack/nova-api-0" Dec 03 14:33:51 crc kubenswrapper[4805]: I1203 14:33:51.523015 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/65386594-9cac-40ef-b3a8-2f3ee225bfb6-internal-tls-certs\") pod \"nova-api-0\" (UID: \"65386594-9cac-40ef-b3a8-2f3ee225bfb6\") " pod="openstack/nova-api-0" Dec 03 14:33:51 crc kubenswrapper[4805]: I1203 14:33:51.523400 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/65386594-9cac-40ef-b3a8-2f3ee225bfb6-config-data\") pod \"nova-api-0\" (UID: \"65386594-9cac-40ef-b3a8-2f3ee225bfb6\") " pod="openstack/nova-api-0" Dec 03 14:33:51 crc kubenswrapper[4805]: I1203 14:33:51.539464 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9kgqr\" (UniqueName: \"kubernetes.io/projected/65386594-9cac-40ef-b3a8-2f3ee225bfb6-kube-api-access-9kgqr\") pod \"nova-api-0\" (UID: \"65386594-9cac-40ef-b3a8-2f3ee225bfb6\") " pod="openstack/nova-api-0" Dec 03 14:33:51 crc kubenswrapper[4805]: I1203 14:33:51.595710 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 03 14:33:52 crc kubenswrapper[4805]: I1203 14:33:52.059912 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Dec 03 14:33:52 crc kubenswrapper[4805]: I1203 14:33:52.205436 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"65386594-9cac-40ef-b3a8-2f3ee225bfb6","Type":"ContainerStarted","Data":"100a4d3849853b9e3eaa6293a0ac0d5dc9b96c950065e2ea9850cd0cd827405f"} Dec 03 14:33:52 crc kubenswrapper[4805]: I1203 14:33:52.212612 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"aee26d6d-7304-4001-90f8-c0ed886a3cda","Type":"ContainerStarted","Data":"c6e36473c9af1c0d702a0744c34ad51f28c604114568f3a35c2d7eaf0de6e461"} Dec 03 14:33:52 crc kubenswrapper[4805]: I1203 14:33:52.704551 4805 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="68636027-732d-46fa-aba1-7f89927292ac" path="/var/lib/kubelet/pods/68636027-732d-46fa-aba1-7f89927292ac/volumes" Dec 03 14:33:53 crc kubenswrapper[4805]: I1203 14:33:53.223479 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"65386594-9cac-40ef-b3a8-2f3ee225bfb6","Type":"ContainerStarted","Data":"4da654a0af3df609ab1e3c4bb88e2ef3d9cfd6430ec2f58f6ff3bb3a56410212"} Dec 03 14:33:53 crc kubenswrapper[4805]: I1203 14:33:53.223828 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"65386594-9cac-40ef-b3a8-2f3ee225bfb6","Type":"ContainerStarted","Data":"d0c0b35ab970b9e12b5336b6f9f24114229c6ed42b6105da38bee2c1d59af43c"} Dec 03 14:33:53 crc kubenswrapper[4805]: I1203 14:33:53.226981 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"aee26d6d-7304-4001-90f8-c0ed886a3cda","Type":"ContainerStarted","Data":"f62e2a86bdda5a487b4221e167d77f2dbb49598c46ba7c129af938f19abfce52"} Dec 03 14:33:53 crc kubenswrapper[4805]: I1203 14:33:53.259743 4805 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-0" podStartSLOduration=2.2597204 podStartE2EDuration="2.2597204s" podCreationTimestamp="2025-12-03 14:33:51 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 14:33:53.249420889 +0000 UTC m=+1462.912337812" watchObservedRunningTime="2025-12-03 14:33:53.2597204 +0000 UTC m=+1462.922637323" Dec 03 14:33:54 crc kubenswrapper[4805]: I1203 14:33:54.237973 4805 generic.go:334] "Generic (PLEG): container finished" podID="2dcd6066-cee7-40a7-ba9f-0648b6d2283b" containerID="59cc0d8dee55d94f88b6b089f411035edd0324469c03b366c0c24c6baa08fe63" exitCode=0 Dec 03 14:33:54 crc kubenswrapper[4805]: I1203 14:33:54.238039 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"2dcd6066-cee7-40a7-ba9f-0648b6d2283b","Type":"ContainerDied","Data":"59cc0d8dee55d94f88b6b089f411035edd0324469c03b366c0c24c6baa08fe63"} Dec 03 14:33:54 crc kubenswrapper[4805]: I1203 14:33:54.664569 4805 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 03 14:33:54 crc kubenswrapper[4805]: I1203 14:33:54.789694 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/2dcd6066-cee7-40a7-ba9f-0648b6d2283b-logs\") pod \"2dcd6066-cee7-40a7-ba9f-0648b6d2283b\" (UID: \"2dcd6066-cee7-40a7-ba9f-0648b6d2283b\") " Dec 03 14:33:54 crc kubenswrapper[4805]: I1203 14:33:54.789940 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/2dcd6066-cee7-40a7-ba9f-0648b6d2283b-nova-metadata-tls-certs\") pod \"2dcd6066-cee7-40a7-ba9f-0648b6d2283b\" (UID: \"2dcd6066-cee7-40a7-ba9f-0648b6d2283b\") " Dec 03 14:33:54 crc kubenswrapper[4805]: I1203 14:33:54.789981 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2dcd6066-cee7-40a7-ba9f-0648b6d2283b-combined-ca-bundle\") pod \"2dcd6066-cee7-40a7-ba9f-0648b6d2283b\" (UID: \"2dcd6066-cee7-40a7-ba9f-0648b6d2283b\") " Dec 03 14:33:54 crc kubenswrapper[4805]: I1203 14:33:54.790020 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dmqxc\" (UniqueName: \"kubernetes.io/projected/2dcd6066-cee7-40a7-ba9f-0648b6d2283b-kube-api-access-dmqxc\") pod \"2dcd6066-cee7-40a7-ba9f-0648b6d2283b\" (UID: \"2dcd6066-cee7-40a7-ba9f-0648b6d2283b\") " Dec 03 14:33:54 crc kubenswrapper[4805]: I1203 14:33:54.790039 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2dcd6066-cee7-40a7-ba9f-0648b6d2283b-config-data\") pod \"2dcd6066-cee7-40a7-ba9f-0648b6d2283b\" (UID: \"2dcd6066-cee7-40a7-ba9f-0648b6d2283b\") " Dec 03 14:33:54 crc kubenswrapper[4805]: I1203 14:33:54.790367 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2dcd6066-cee7-40a7-ba9f-0648b6d2283b-logs" (OuterVolumeSpecName: "logs") pod "2dcd6066-cee7-40a7-ba9f-0648b6d2283b" (UID: "2dcd6066-cee7-40a7-ba9f-0648b6d2283b"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 14:33:54 crc kubenswrapper[4805]: I1203 14:33:54.791892 4805 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/2dcd6066-cee7-40a7-ba9f-0648b6d2283b-logs\") on node \"crc\" DevicePath \"\"" Dec 03 14:33:54 crc kubenswrapper[4805]: I1203 14:33:54.799503 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2dcd6066-cee7-40a7-ba9f-0648b6d2283b-kube-api-access-dmqxc" (OuterVolumeSpecName: "kube-api-access-dmqxc") pod "2dcd6066-cee7-40a7-ba9f-0648b6d2283b" (UID: "2dcd6066-cee7-40a7-ba9f-0648b6d2283b"). InnerVolumeSpecName "kube-api-access-dmqxc". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 14:33:54 crc kubenswrapper[4805]: I1203 14:33:54.822299 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2dcd6066-cee7-40a7-ba9f-0648b6d2283b-config-data" (OuterVolumeSpecName: "config-data") pod "2dcd6066-cee7-40a7-ba9f-0648b6d2283b" (UID: "2dcd6066-cee7-40a7-ba9f-0648b6d2283b"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 14:33:54 crc kubenswrapper[4805]: I1203 14:33:54.824745 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2dcd6066-cee7-40a7-ba9f-0648b6d2283b-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "2dcd6066-cee7-40a7-ba9f-0648b6d2283b" (UID: "2dcd6066-cee7-40a7-ba9f-0648b6d2283b"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 14:33:54 crc kubenswrapper[4805]: I1203 14:33:54.864611 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2dcd6066-cee7-40a7-ba9f-0648b6d2283b-nova-metadata-tls-certs" (OuterVolumeSpecName: "nova-metadata-tls-certs") pod "2dcd6066-cee7-40a7-ba9f-0648b6d2283b" (UID: "2dcd6066-cee7-40a7-ba9f-0648b6d2283b"). InnerVolumeSpecName "nova-metadata-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 14:33:54 crc kubenswrapper[4805]: I1203 14:33:54.893695 4805 reconciler_common.go:293] "Volume detached for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/2dcd6066-cee7-40a7-ba9f-0648b6d2283b-nova-metadata-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 03 14:33:54 crc kubenswrapper[4805]: I1203 14:33:54.893741 4805 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2dcd6066-cee7-40a7-ba9f-0648b6d2283b-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 14:33:54 crc kubenswrapper[4805]: I1203 14:33:54.893753 4805 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dmqxc\" (UniqueName: \"kubernetes.io/projected/2dcd6066-cee7-40a7-ba9f-0648b6d2283b-kube-api-access-dmqxc\") on node \"crc\" DevicePath \"\"" Dec 03 14:33:54 crc kubenswrapper[4805]: I1203 14:33:54.893764 4805 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2dcd6066-cee7-40a7-ba9f-0648b6d2283b-config-data\") on node \"crc\" DevicePath \"\"" Dec 03 14:33:55 crc kubenswrapper[4805]: I1203 14:33:55.250850 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"2dcd6066-cee7-40a7-ba9f-0648b6d2283b","Type":"ContainerDied","Data":"b74eb9e10891dd812db247dd7c26b09bf2754b3c9ea7e37386e6ade1a8d6eb65"} Dec 03 14:33:55 crc kubenswrapper[4805]: I1203 14:33:55.250898 4805 scope.go:117] "RemoveContainer" containerID="59cc0d8dee55d94f88b6b089f411035edd0324469c03b366c0c24c6baa08fe63" Dec 03 14:33:55 crc kubenswrapper[4805]: I1203 14:33:55.251005 4805 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 03 14:33:55 crc kubenswrapper[4805]: I1203 14:33:55.254118 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"aee26d6d-7304-4001-90f8-c0ed886a3cda","Type":"ContainerStarted","Data":"2cf776a26e533c761ac54e168d3c4825a871fbc428f01ab40780a9f315b80d05"} Dec 03 14:33:55 crc kubenswrapper[4805]: I1203 14:33:55.254452 4805 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Dec 03 14:33:55 crc kubenswrapper[4805]: I1203 14:33:55.282615 4805 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=2.272052043 podStartE2EDuration="6.282595749s" podCreationTimestamp="2025-12-03 14:33:49 +0000 UTC" firstStartedPulling="2025-12-03 14:33:50.055427988 +0000 UTC m=+1459.718344911" lastFinishedPulling="2025-12-03 14:33:54.065971694 +0000 UTC m=+1463.728888617" observedRunningTime="2025-12-03 14:33:55.277360416 +0000 UTC m=+1464.940277339" watchObservedRunningTime="2025-12-03 14:33:55.282595749 +0000 UTC m=+1464.945512672" Dec 03 14:33:55 crc kubenswrapper[4805]: I1203 14:33:55.298953 4805 scope.go:117] "RemoveContainer" containerID="72253effa6be075609506ec28eb6dc2aa2e9f7d559f7e42103391020b725db73" Dec 03 14:33:55 crc kubenswrapper[4805]: I1203 14:33:55.333703 4805 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Dec 03 14:33:55 crc kubenswrapper[4805]: I1203 14:33:55.359433 4805 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-metadata-0"] Dec 03 14:33:55 crc kubenswrapper[4805]: I1203 14:33:55.369528 4805 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-metadata-0"] Dec 03 14:33:55 crc kubenswrapper[4805]: E1203 14:33:55.370001 4805 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2dcd6066-cee7-40a7-ba9f-0648b6d2283b" containerName="nova-metadata-log" Dec 03 14:33:55 crc kubenswrapper[4805]: I1203 14:33:55.370023 4805 state_mem.go:107] "Deleted CPUSet assignment" podUID="2dcd6066-cee7-40a7-ba9f-0648b6d2283b" containerName="nova-metadata-log" Dec 03 14:33:55 crc kubenswrapper[4805]: E1203 14:33:55.370045 4805 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2dcd6066-cee7-40a7-ba9f-0648b6d2283b" containerName="nova-metadata-metadata" Dec 03 14:33:55 crc kubenswrapper[4805]: I1203 14:33:55.370059 4805 state_mem.go:107] "Deleted CPUSet assignment" podUID="2dcd6066-cee7-40a7-ba9f-0648b6d2283b" containerName="nova-metadata-metadata" Dec 03 14:33:55 crc kubenswrapper[4805]: I1203 14:33:55.370257 4805 memory_manager.go:354] "RemoveStaleState removing state" podUID="2dcd6066-cee7-40a7-ba9f-0648b6d2283b" containerName="nova-metadata-log" Dec 03 14:33:55 crc kubenswrapper[4805]: I1203 14:33:55.370275 4805 memory_manager.go:354] "RemoveStaleState removing state" podUID="2dcd6066-cee7-40a7-ba9f-0648b6d2283b" containerName="nova-metadata-metadata" Dec 03 14:33:55 crc kubenswrapper[4805]: I1203 14:33:55.371408 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 03 14:33:55 crc kubenswrapper[4805]: I1203 14:33:55.374417 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-config-data" Dec 03 14:33:55 crc kubenswrapper[4805]: I1203 14:33:55.374953 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-metadata-internal-svc" Dec 03 14:33:55 crc kubenswrapper[4805]: I1203 14:33:55.382252 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Dec 03 14:33:55 crc kubenswrapper[4805]: I1203 14:33:55.507692 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7dvhr\" (UniqueName: \"kubernetes.io/projected/9acb704f-0e77-458e-b31e-7b32cf8619e5-kube-api-access-7dvhr\") pod \"nova-metadata-0\" (UID: \"9acb704f-0e77-458e-b31e-7b32cf8619e5\") " pod="openstack/nova-metadata-0" Dec 03 14:33:55 crc kubenswrapper[4805]: I1203 14:33:55.509498 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/9acb704f-0e77-458e-b31e-7b32cf8619e5-logs\") pod \"nova-metadata-0\" (UID: \"9acb704f-0e77-458e-b31e-7b32cf8619e5\") " pod="openstack/nova-metadata-0" Dec 03 14:33:55 crc kubenswrapper[4805]: I1203 14:33:55.509556 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9acb704f-0e77-458e-b31e-7b32cf8619e5-config-data\") pod \"nova-metadata-0\" (UID: \"9acb704f-0e77-458e-b31e-7b32cf8619e5\") " pod="openstack/nova-metadata-0" Dec 03 14:33:55 crc kubenswrapper[4805]: I1203 14:33:55.509595 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9acb704f-0e77-458e-b31e-7b32cf8619e5-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"9acb704f-0e77-458e-b31e-7b32cf8619e5\") " pod="openstack/nova-metadata-0" Dec 03 14:33:55 crc kubenswrapper[4805]: I1203 14:33:55.509704 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/9acb704f-0e77-458e-b31e-7b32cf8619e5-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"9acb704f-0e77-458e-b31e-7b32cf8619e5\") " pod="openstack/nova-metadata-0" Dec 03 14:33:55 crc kubenswrapper[4805]: I1203 14:33:55.611127 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7dvhr\" (UniqueName: \"kubernetes.io/projected/9acb704f-0e77-458e-b31e-7b32cf8619e5-kube-api-access-7dvhr\") pod \"nova-metadata-0\" (UID: \"9acb704f-0e77-458e-b31e-7b32cf8619e5\") " pod="openstack/nova-metadata-0" Dec 03 14:33:55 crc kubenswrapper[4805]: I1203 14:33:55.611185 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/9acb704f-0e77-458e-b31e-7b32cf8619e5-logs\") pod \"nova-metadata-0\" (UID: \"9acb704f-0e77-458e-b31e-7b32cf8619e5\") " pod="openstack/nova-metadata-0" Dec 03 14:33:55 crc kubenswrapper[4805]: I1203 14:33:55.611220 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9acb704f-0e77-458e-b31e-7b32cf8619e5-config-data\") pod \"nova-metadata-0\" (UID: \"9acb704f-0e77-458e-b31e-7b32cf8619e5\") " pod="openstack/nova-metadata-0" Dec 03 14:33:55 crc kubenswrapper[4805]: I1203 14:33:55.611252 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9acb704f-0e77-458e-b31e-7b32cf8619e5-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"9acb704f-0e77-458e-b31e-7b32cf8619e5\") " pod="openstack/nova-metadata-0" Dec 03 14:33:55 crc kubenswrapper[4805]: I1203 14:33:55.611317 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/9acb704f-0e77-458e-b31e-7b32cf8619e5-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"9acb704f-0e77-458e-b31e-7b32cf8619e5\") " pod="openstack/nova-metadata-0" Dec 03 14:33:55 crc kubenswrapper[4805]: I1203 14:33:55.612486 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/9acb704f-0e77-458e-b31e-7b32cf8619e5-logs\") pod \"nova-metadata-0\" (UID: \"9acb704f-0e77-458e-b31e-7b32cf8619e5\") " pod="openstack/nova-metadata-0" Dec 03 14:33:55 crc kubenswrapper[4805]: I1203 14:33:55.616485 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9acb704f-0e77-458e-b31e-7b32cf8619e5-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"9acb704f-0e77-458e-b31e-7b32cf8619e5\") " pod="openstack/nova-metadata-0" Dec 03 14:33:55 crc kubenswrapper[4805]: I1203 14:33:55.616590 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/9acb704f-0e77-458e-b31e-7b32cf8619e5-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"9acb704f-0e77-458e-b31e-7b32cf8619e5\") " pod="openstack/nova-metadata-0" Dec 03 14:33:55 crc kubenswrapper[4805]: I1203 14:33:55.617190 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9acb704f-0e77-458e-b31e-7b32cf8619e5-config-data\") pod \"nova-metadata-0\" (UID: \"9acb704f-0e77-458e-b31e-7b32cf8619e5\") " pod="openstack/nova-metadata-0" Dec 03 14:33:55 crc kubenswrapper[4805]: I1203 14:33:55.631218 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7dvhr\" (UniqueName: \"kubernetes.io/projected/9acb704f-0e77-458e-b31e-7b32cf8619e5-kube-api-access-7dvhr\") pod \"nova-metadata-0\" (UID: \"9acb704f-0e77-458e-b31e-7b32cf8619e5\") " pod="openstack/nova-metadata-0" Dec 03 14:33:55 crc kubenswrapper[4805]: I1203 14:33:55.696154 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 03 14:33:55 crc kubenswrapper[4805]: I1203 14:33:55.711854 4805 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Dec 03 14:33:55 crc kubenswrapper[4805]: I1203 14:33:55.814463 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c0ccb82f-2439-448d-a958-5d775682cca3-config-data\") pod \"c0ccb82f-2439-448d-a958-5d775682cca3\" (UID: \"c0ccb82f-2439-448d-a958-5d775682cca3\") " Dec 03 14:33:55 crc kubenswrapper[4805]: I1203 14:33:55.814853 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lwh2t\" (UniqueName: \"kubernetes.io/projected/c0ccb82f-2439-448d-a958-5d775682cca3-kube-api-access-lwh2t\") pod \"c0ccb82f-2439-448d-a958-5d775682cca3\" (UID: \"c0ccb82f-2439-448d-a958-5d775682cca3\") " Dec 03 14:33:55 crc kubenswrapper[4805]: I1203 14:33:55.815078 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c0ccb82f-2439-448d-a958-5d775682cca3-combined-ca-bundle\") pod \"c0ccb82f-2439-448d-a958-5d775682cca3\" (UID: \"c0ccb82f-2439-448d-a958-5d775682cca3\") " Dec 03 14:33:55 crc kubenswrapper[4805]: I1203 14:33:55.822046 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c0ccb82f-2439-448d-a958-5d775682cca3-kube-api-access-lwh2t" (OuterVolumeSpecName: "kube-api-access-lwh2t") pod "c0ccb82f-2439-448d-a958-5d775682cca3" (UID: "c0ccb82f-2439-448d-a958-5d775682cca3"). InnerVolumeSpecName "kube-api-access-lwh2t". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 14:33:55 crc kubenswrapper[4805]: I1203 14:33:55.849272 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c0ccb82f-2439-448d-a958-5d775682cca3-config-data" (OuterVolumeSpecName: "config-data") pod "c0ccb82f-2439-448d-a958-5d775682cca3" (UID: "c0ccb82f-2439-448d-a958-5d775682cca3"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 14:33:55 crc kubenswrapper[4805]: I1203 14:33:55.856110 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c0ccb82f-2439-448d-a958-5d775682cca3-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "c0ccb82f-2439-448d-a958-5d775682cca3" (UID: "c0ccb82f-2439-448d-a958-5d775682cca3"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 14:33:55 crc kubenswrapper[4805]: I1203 14:33:55.917985 4805 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c0ccb82f-2439-448d-a958-5d775682cca3-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 14:33:55 crc kubenswrapper[4805]: I1203 14:33:55.918456 4805 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c0ccb82f-2439-448d-a958-5d775682cca3-config-data\") on node \"crc\" DevicePath \"\"" Dec 03 14:33:55 crc kubenswrapper[4805]: I1203 14:33:55.918585 4805 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lwh2t\" (UniqueName: \"kubernetes.io/projected/c0ccb82f-2439-448d-a958-5d775682cca3-kube-api-access-lwh2t\") on node \"crc\" DevicePath \"\"" Dec 03 14:33:56 crc kubenswrapper[4805]: I1203 14:33:56.219918 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Dec 03 14:33:56 crc kubenswrapper[4805]: I1203 14:33:56.265295 4805 generic.go:334] "Generic (PLEG): container finished" podID="c0ccb82f-2439-448d-a958-5d775682cca3" containerID="47fa01711b4e60668d53aa93490a39d0905127c17fbc1ca51fd562ed6c516ebc" exitCode=0 Dec 03 14:33:56 crc kubenswrapper[4805]: I1203 14:33:56.265367 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"c0ccb82f-2439-448d-a958-5d775682cca3","Type":"ContainerDied","Data":"47fa01711b4e60668d53aa93490a39d0905127c17fbc1ca51fd562ed6c516ebc"} Dec 03 14:33:56 crc kubenswrapper[4805]: I1203 14:33:56.265376 4805 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Dec 03 14:33:56 crc kubenswrapper[4805]: I1203 14:33:56.265397 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"c0ccb82f-2439-448d-a958-5d775682cca3","Type":"ContainerDied","Data":"a738f147993d9d90919dd4b17115af79a515cd7403dea206122ef203e5e01b6c"} Dec 03 14:33:56 crc kubenswrapper[4805]: I1203 14:33:56.265417 4805 scope.go:117] "RemoveContainer" containerID="47fa01711b4e60668d53aa93490a39d0905127c17fbc1ca51fd562ed6c516ebc" Dec 03 14:33:56 crc kubenswrapper[4805]: I1203 14:33:56.272101 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"9acb704f-0e77-458e-b31e-7b32cf8619e5","Type":"ContainerStarted","Data":"be99f03f414aba962b30240d72af81dfe383609018acf26bd5ad09ff4c81dbe9"} Dec 03 14:33:56 crc kubenswrapper[4805]: I1203 14:33:56.333186 4805 scope.go:117] "RemoveContainer" containerID="47fa01711b4e60668d53aa93490a39d0905127c17fbc1ca51fd562ed6c516ebc" Dec 03 14:33:56 crc kubenswrapper[4805]: E1203 14:33:56.333690 4805 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"47fa01711b4e60668d53aa93490a39d0905127c17fbc1ca51fd562ed6c516ebc\": container with ID starting with 47fa01711b4e60668d53aa93490a39d0905127c17fbc1ca51fd562ed6c516ebc not found: ID does not exist" containerID="47fa01711b4e60668d53aa93490a39d0905127c17fbc1ca51fd562ed6c516ebc" Dec 03 14:33:56 crc kubenswrapper[4805]: I1203 14:33:56.333736 4805 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"47fa01711b4e60668d53aa93490a39d0905127c17fbc1ca51fd562ed6c516ebc"} err="failed to get container status \"47fa01711b4e60668d53aa93490a39d0905127c17fbc1ca51fd562ed6c516ebc\": rpc error: code = NotFound desc = could not find container \"47fa01711b4e60668d53aa93490a39d0905127c17fbc1ca51fd562ed6c516ebc\": container with ID starting with 47fa01711b4e60668d53aa93490a39d0905127c17fbc1ca51fd562ed6c516ebc not found: ID does not exist" Dec 03 14:33:56 crc kubenswrapper[4805]: I1203 14:33:56.371478 4805 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Dec 03 14:33:56 crc kubenswrapper[4805]: I1203 14:33:56.391906 4805 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-scheduler-0"] Dec 03 14:33:56 crc kubenswrapper[4805]: I1203 14:33:56.400929 4805 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-scheduler-0"] Dec 03 14:33:56 crc kubenswrapper[4805]: E1203 14:33:56.401383 4805 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c0ccb82f-2439-448d-a958-5d775682cca3" containerName="nova-scheduler-scheduler" Dec 03 14:33:56 crc kubenswrapper[4805]: I1203 14:33:56.401402 4805 state_mem.go:107] "Deleted CPUSet assignment" podUID="c0ccb82f-2439-448d-a958-5d775682cca3" containerName="nova-scheduler-scheduler" Dec 03 14:33:56 crc kubenswrapper[4805]: I1203 14:33:56.401658 4805 memory_manager.go:354] "RemoveStaleState removing state" podUID="c0ccb82f-2439-448d-a958-5d775682cca3" containerName="nova-scheduler-scheduler" Dec 03 14:33:56 crc kubenswrapper[4805]: I1203 14:33:56.402345 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Dec 03 14:33:56 crc kubenswrapper[4805]: I1203 14:33:56.408109 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-scheduler-config-data" Dec 03 14:33:56 crc kubenswrapper[4805]: I1203 14:33:56.410385 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Dec 03 14:33:56 crc kubenswrapper[4805]: I1203 14:33:56.531070 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e30e8472-484d-47a9-bb8f-839d41ec04be-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"e30e8472-484d-47a9-bb8f-839d41ec04be\") " pod="openstack/nova-scheduler-0" Dec 03 14:33:56 crc kubenswrapper[4805]: I1203 14:33:56.531221 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e30e8472-484d-47a9-bb8f-839d41ec04be-config-data\") pod \"nova-scheduler-0\" (UID: \"e30e8472-484d-47a9-bb8f-839d41ec04be\") " pod="openstack/nova-scheduler-0" Dec 03 14:33:56 crc kubenswrapper[4805]: I1203 14:33:56.531258 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5gf4w\" (UniqueName: \"kubernetes.io/projected/e30e8472-484d-47a9-bb8f-839d41ec04be-kube-api-access-5gf4w\") pod \"nova-scheduler-0\" (UID: \"e30e8472-484d-47a9-bb8f-839d41ec04be\") " pod="openstack/nova-scheduler-0" Dec 03 14:33:56 crc kubenswrapper[4805]: I1203 14:33:56.632469 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e30e8472-484d-47a9-bb8f-839d41ec04be-config-data\") pod \"nova-scheduler-0\" (UID: \"e30e8472-484d-47a9-bb8f-839d41ec04be\") " pod="openstack/nova-scheduler-0" Dec 03 14:33:56 crc kubenswrapper[4805]: I1203 14:33:56.632511 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5gf4w\" (UniqueName: \"kubernetes.io/projected/e30e8472-484d-47a9-bb8f-839d41ec04be-kube-api-access-5gf4w\") pod \"nova-scheduler-0\" (UID: \"e30e8472-484d-47a9-bb8f-839d41ec04be\") " pod="openstack/nova-scheduler-0" Dec 03 14:33:56 crc kubenswrapper[4805]: I1203 14:33:56.632623 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e30e8472-484d-47a9-bb8f-839d41ec04be-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"e30e8472-484d-47a9-bb8f-839d41ec04be\") " pod="openstack/nova-scheduler-0" Dec 03 14:33:56 crc kubenswrapper[4805]: I1203 14:33:56.636205 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e30e8472-484d-47a9-bb8f-839d41ec04be-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"e30e8472-484d-47a9-bb8f-839d41ec04be\") " pod="openstack/nova-scheduler-0" Dec 03 14:33:56 crc kubenswrapper[4805]: I1203 14:33:56.636230 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e30e8472-484d-47a9-bb8f-839d41ec04be-config-data\") pod \"nova-scheduler-0\" (UID: \"e30e8472-484d-47a9-bb8f-839d41ec04be\") " pod="openstack/nova-scheduler-0" Dec 03 14:33:56 crc kubenswrapper[4805]: I1203 14:33:56.649348 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5gf4w\" (UniqueName: \"kubernetes.io/projected/e30e8472-484d-47a9-bb8f-839d41ec04be-kube-api-access-5gf4w\") pod \"nova-scheduler-0\" (UID: \"e30e8472-484d-47a9-bb8f-839d41ec04be\") " pod="openstack/nova-scheduler-0" Dec 03 14:33:56 crc kubenswrapper[4805]: I1203 14:33:56.705132 4805 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2dcd6066-cee7-40a7-ba9f-0648b6d2283b" path="/var/lib/kubelet/pods/2dcd6066-cee7-40a7-ba9f-0648b6d2283b/volumes" Dec 03 14:33:56 crc kubenswrapper[4805]: I1203 14:33:56.706021 4805 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c0ccb82f-2439-448d-a958-5d775682cca3" path="/var/lib/kubelet/pods/c0ccb82f-2439-448d-a958-5d775682cca3/volumes" Dec 03 14:33:56 crc kubenswrapper[4805]: I1203 14:33:56.730603 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Dec 03 14:33:57 crc kubenswrapper[4805]: I1203 14:33:57.184722 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Dec 03 14:33:57 crc kubenswrapper[4805]: I1203 14:33:57.289670 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"9acb704f-0e77-458e-b31e-7b32cf8619e5","Type":"ContainerStarted","Data":"db81cad7f23cfe75e13cb04fc0f3c5a9245351a60e5a174f013529a9b0f97329"} Dec 03 14:33:57 crc kubenswrapper[4805]: I1203 14:33:57.289734 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"9acb704f-0e77-458e-b31e-7b32cf8619e5","Type":"ContainerStarted","Data":"5e5ffbdfaaee289d04bcc3f069a6d13347f5ec41eb0ba7c8e17c7620cc46d35a"} Dec 03 14:33:57 crc kubenswrapper[4805]: I1203 14:33:57.294082 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"e30e8472-484d-47a9-bb8f-839d41ec04be","Type":"ContainerStarted","Data":"ec6d8e844206c2d6b929c1734d7a30097fc579c177b437eae3b688fc51315f6b"} Dec 03 14:33:58 crc kubenswrapper[4805]: I1203 14:33:58.307233 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"e30e8472-484d-47a9-bb8f-839d41ec04be","Type":"ContainerStarted","Data":"7dddfccb35f273b850609af3f3759167c37a8fefff88a6fbdf0f30555b49d4de"} Dec 03 14:33:58 crc kubenswrapper[4805]: I1203 14:33:58.324375 4805 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-metadata-0" podStartSLOduration=3.324350223 podStartE2EDuration="3.324350223s" podCreationTimestamp="2025-12-03 14:33:55 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 14:33:57.318793949 +0000 UTC m=+1466.981710872" watchObservedRunningTime="2025-12-03 14:33:58.324350223 +0000 UTC m=+1467.987267146" Dec 03 14:33:58 crc kubenswrapper[4805]: I1203 14:33:58.327860 4805 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-scheduler-0" podStartSLOduration=2.327846638 podStartE2EDuration="2.327846638s" podCreationTimestamp="2025-12-03 14:33:56 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 14:33:58.319150821 +0000 UTC m=+1467.982067764" watchObservedRunningTime="2025-12-03 14:33:58.327846638 +0000 UTC m=+1467.990763561" Dec 03 14:34:00 crc kubenswrapper[4805]: I1203 14:34:00.711110 4805 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Dec 03 14:34:00 crc kubenswrapper[4805]: I1203 14:34:00.711533 4805 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Dec 03 14:34:01 crc kubenswrapper[4805]: I1203 14:34:01.596369 4805 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Dec 03 14:34:01 crc kubenswrapper[4805]: I1203 14:34:01.596807 4805 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Dec 03 14:34:01 crc kubenswrapper[4805]: I1203 14:34:01.731643 4805 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-scheduler-0" Dec 03 14:34:02 crc kubenswrapper[4805]: I1203 14:34:02.614093 4805 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="65386594-9cac-40ef-b3a8-2f3ee225bfb6" containerName="nova-api-api" probeResult="failure" output="Get \"https://10.217.0.211:8774/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Dec 03 14:34:02 crc kubenswrapper[4805]: I1203 14:34:02.614119 4805 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="65386594-9cac-40ef-b3a8-2f3ee225bfb6" containerName="nova-api-log" probeResult="failure" output="Get \"https://10.217.0.211:8774/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Dec 03 14:34:05 crc kubenswrapper[4805]: I1203 14:34:05.696580 4805 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Dec 03 14:34:05 crc kubenswrapper[4805]: I1203 14:34:05.697161 4805 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Dec 03 14:34:06 crc kubenswrapper[4805]: I1203 14:34:06.721065 4805 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="9acb704f-0e77-458e-b31e-7b32cf8619e5" containerName="nova-metadata-metadata" probeResult="failure" output="Get \"https://10.217.0.212:8775/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Dec 03 14:34:06 crc kubenswrapper[4805]: I1203 14:34:06.721120 4805 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="9acb704f-0e77-458e-b31e-7b32cf8619e5" containerName="nova-metadata-log" probeResult="failure" output="Get \"https://10.217.0.212:8775/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Dec 03 14:34:06 crc kubenswrapper[4805]: I1203 14:34:06.731201 4805 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-scheduler-0" Dec 03 14:34:06 crc kubenswrapper[4805]: I1203 14:34:06.757582 4805 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-scheduler-0" Dec 03 14:34:07 crc kubenswrapper[4805]: I1203 14:34:07.451338 4805 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-scheduler-0" Dec 03 14:34:11 crc kubenswrapper[4805]: I1203 14:34:11.603620 4805 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Dec 03 14:34:11 crc kubenswrapper[4805]: I1203 14:34:11.604240 4805 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Dec 03 14:34:11 crc kubenswrapper[4805]: I1203 14:34:11.604863 4805 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Dec 03 14:34:11 crc kubenswrapper[4805]: I1203 14:34:11.605343 4805 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Dec 03 14:34:11 crc kubenswrapper[4805]: I1203 14:34:11.608822 4805 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Dec 03 14:34:11 crc kubenswrapper[4805]: I1203 14:34:11.611121 4805 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Dec 03 14:34:13 crc kubenswrapper[4805]: I1203 14:34:13.917449 4805 patch_prober.go:28] interesting pod/machine-config-daemon-gskh4 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 03 14:34:13 crc kubenswrapper[4805]: I1203 14:34:13.917709 4805 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-gskh4" podUID="6098937f-e3f6-45e8-a647-4994a79cd711" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 03 14:34:13 crc kubenswrapper[4805]: I1203 14:34:13.917751 4805 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-gskh4" Dec 03 14:34:14 crc kubenswrapper[4805]: I1203 14:34:13.918448 4805 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"a8c52725eecea0a7c49a5c7b85e82a2cd4d97ad1c78cf047a931df09ab5530e8"} pod="openshift-machine-config-operator/machine-config-daemon-gskh4" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 03 14:34:14 crc kubenswrapper[4805]: I1203 14:34:13.918501 4805 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-gskh4" podUID="6098937f-e3f6-45e8-a647-4994a79cd711" containerName="machine-config-daemon" containerID="cri-o://a8c52725eecea0a7c49a5c7b85e82a2cd4d97ad1c78cf047a931df09ab5530e8" gracePeriod=600 Dec 03 14:34:14 crc kubenswrapper[4805]: I1203 14:34:14.489415 4805 generic.go:334] "Generic (PLEG): container finished" podID="6098937f-e3f6-45e8-a647-4994a79cd711" containerID="a8c52725eecea0a7c49a5c7b85e82a2cd4d97ad1c78cf047a931df09ab5530e8" exitCode=0 Dec 03 14:34:14 crc kubenswrapper[4805]: I1203 14:34:14.489496 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-gskh4" event={"ID":"6098937f-e3f6-45e8-a647-4994a79cd711","Type":"ContainerDied","Data":"a8c52725eecea0a7c49a5c7b85e82a2cd4d97ad1c78cf047a931df09ab5530e8"} Dec 03 14:34:14 crc kubenswrapper[4805]: I1203 14:34:14.489755 4805 scope.go:117] "RemoveContainer" containerID="b84d108663e877514e58fb0f23591399246f6b24b642394275299320cb0e52d9" Dec 03 14:34:15 crc kubenswrapper[4805]: I1203 14:34:15.500991 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-gskh4" event={"ID":"6098937f-e3f6-45e8-a647-4994a79cd711","Type":"ContainerStarted","Data":"748e11b7efe02ff9a381aeee0aa4ddad32272e1564bc028d779c442a7b7f2149"} Dec 03 14:34:15 crc kubenswrapper[4805]: I1203 14:34:15.702785 4805 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-metadata-0" Dec 03 14:34:15 crc kubenswrapper[4805]: I1203 14:34:15.703304 4805 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-metadata-0" Dec 03 14:34:15 crc kubenswrapper[4805]: I1203 14:34:15.709036 4805 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-metadata-0" Dec 03 14:34:15 crc kubenswrapper[4805]: I1203 14:34:15.709319 4805 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-metadata-0" Dec 03 14:34:19 crc kubenswrapper[4805]: I1203 14:34:19.608143 4805 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ceilometer-0" Dec 03 14:34:30 crc kubenswrapper[4805]: I1203 14:34:30.395645 4805 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/rabbitmq-server-0"] Dec 03 14:34:31 crc kubenswrapper[4805]: I1203 14:34:31.255032 4805 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Dec 03 14:34:34 crc kubenswrapper[4805]: I1203 14:34:34.634196 4805 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/rabbitmq-server-0" podUID="1b34fa5c-2867-4dbe-9285-80c327699864" containerName="rabbitmq" containerID="cri-o://b834218684c741c79335521d4745edfa20b11d70a0f812104cdab4f0a3846643" gracePeriod=604796 Dec 03 14:34:35 crc kubenswrapper[4805]: I1203 14:34:35.453944 4805 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/rabbitmq-cell1-server-0" podUID="656f691c-1188-47b9-b8dd-b393a230602a" containerName="rabbitmq" containerID="cri-o://770e52863be49086afe2bea1b59e31fac9f82d73b30ec6fe23a2c1d8d92fe910" gracePeriod=604796 Dec 03 14:34:41 crc kubenswrapper[4805]: I1203 14:34:41.594732 4805 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Dec 03 14:34:41 crc kubenswrapper[4805]: I1203 14:34:41.716638 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/1b34fa5c-2867-4dbe-9285-80c327699864-rabbitmq-plugins\") pod \"1b34fa5c-2867-4dbe-9285-80c327699864\" (UID: \"1b34fa5c-2867-4dbe-9285-80c327699864\") " Dec 03 14:34:41 crc kubenswrapper[4805]: I1203 14:34:41.716720 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/1b34fa5c-2867-4dbe-9285-80c327699864-rabbitmq-erlang-cookie\") pod \"1b34fa5c-2867-4dbe-9285-80c327699864\" (UID: \"1b34fa5c-2867-4dbe-9285-80c327699864\") " Dec 03 14:34:41 crc kubenswrapper[4805]: I1203 14:34:41.717382 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/1b34fa5c-2867-4dbe-9285-80c327699864-server-conf\") pod \"1b34fa5c-2867-4dbe-9285-80c327699864\" (UID: \"1b34fa5c-2867-4dbe-9285-80c327699864\") " Dec 03 14:34:41 crc kubenswrapper[4805]: I1203 14:34:41.717424 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/1b34fa5c-2867-4dbe-9285-80c327699864-plugins-conf\") pod \"1b34fa5c-2867-4dbe-9285-80c327699864\" (UID: \"1b34fa5c-2867-4dbe-9285-80c327699864\") " Dec 03 14:34:41 crc kubenswrapper[4805]: I1203 14:34:41.717448 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/1b34fa5c-2867-4dbe-9285-80c327699864-rabbitmq-tls\") pod \"1b34fa5c-2867-4dbe-9285-80c327699864\" (UID: \"1b34fa5c-2867-4dbe-9285-80c327699864\") " Dec 03 14:34:41 crc kubenswrapper[4805]: I1203 14:34:41.717485 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/1b34fa5c-2867-4dbe-9285-80c327699864-pod-info\") pod \"1b34fa5c-2867-4dbe-9285-80c327699864\" (UID: \"1b34fa5c-2867-4dbe-9285-80c327699864\") " Dec 03 14:34:41 crc kubenswrapper[4805]: I1203 14:34:41.717507 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tbx6c\" (UniqueName: \"kubernetes.io/projected/1b34fa5c-2867-4dbe-9285-80c327699864-kube-api-access-tbx6c\") pod \"1b34fa5c-2867-4dbe-9285-80c327699864\" (UID: \"1b34fa5c-2867-4dbe-9285-80c327699864\") " Dec 03 14:34:41 crc kubenswrapper[4805]: I1203 14:34:41.717546 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"persistence\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"1b34fa5c-2867-4dbe-9285-80c327699864\" (UID: \"1b34fa5c-2867-4dbe-9285-80c327699864\") " Dec 03 14:34:41 crc kubenswrapper[4805]: I1203 14:34:41.717572 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/1b34fa5c-2867-4dbe-9285-80c327699864-erlang-cookie-secret\") pod \"1b34fa5c-2867-4dbe-9285-80c327699864\" (UID: \"1b34fa5c-2867-4dbe-9285-80c327699864\") " Dec 03 14:34:41 crc kubenswrapper[4805]: I1203 14:34:41.717598 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/1b34fa5c-2867-4dbe-9285-80c327699864-config-data\") pod \"1b34fa5c-2867-4dbe-9285-80c327699864\" (UID: \"1b34fa5c-2867-4dbe-9285-80c327699864\") " Dec 03 14:34:41 crc kubenswrapper[4805]: I1203 14:34:41.717631 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/1b34fa5c-2867-4dbe-9285-80c327699864-rabbitmq-confd\") pod \"1b34fa5c-2867-4dbe-9285-80c327699864\" (UID: \"1b34fa5c-2867-4dbe-9285-80c327699864\") " Dec 03 14:34:41 crc kubenswrapper[4805]: I1203 14:34:41.719048 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1b34fa5c-2867-4dbe-9285-80c327699864-plugins-conf" (OuterVolumeSpecName: "plugins-conf") pod "1b34fa5c-2867-4dbe-9285-80c327699864" (UID: "1b34fa5c-2867-4dbe-9285-80c327699864"). InnerVolumeSpecName "plugins-conf". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 14:34:41 crc kubenswrapper[4805]: I1203 14:34:41.719797 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1b34fa5c-2867-4dbe-9285-80c327699864-rabbitmq-plugins" (OuterVolumeSpecName: "rabbitmq-plugins") pod "1b34fa5c-2867-4dbe-9285-80c327699864" (UID: "1b34fa5c-2867-4dbe-9285-80c327699864"). InnerVolumeSpecName "rabbitmq-plugins". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 14:34:41 crc kubenswrapper[4805]: I1203 14:34:41.720550 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1b34fa5c-2867-4dbe-9285-80c327699864-rabbitmq-erlang-cookie" (OuterVolumeSpecName: "rabbitmq-erlang-cookie") pod "1b34fa5c-2867-4dbe-9285-80c327699864" (UID: "1b34fa5c-2867-4dbe-9285-80c327699864"). InnerVolumeSpecName "rabbitmq-erlang-cookie". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 14:34:41 crc kubenswrapper[4805]: I1203 14:34:41.727978 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage03-crc" (OuterVolumeSpecName: "persistence") pod "1b34fa5c-2867-4dbe-9285-80c327699864" (UID: "1b34fa5c-2867-4dbe-9285-80c327699864"). InnerVolumeSpecName "local-storage03-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Dec 03 14:34:41 crc kubenswrapper[4805]: I1203 14:34:41.731125 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1b34fa5c-2867-4dbe-9285-80c327699864-erlang-cookie-secret" (OuterVolumeSpecName: "erlang-cookie-secret") pod "1b34fa5c-2867-4dbe-9285-80c327699864" (UID: "1b34fa5c-2867-4dbe-9285-80c327699864"). InnerVolumeSpecName "erlang-cookie-secret". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 14:34:41 crc kubenswrapper[4805]: I1203 14:34:41.731529 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1b34fa5c-2867-4dbe-9285-80c327699864-rabbitmq-tls" (OuterVolumeSpecName: "rabbitmq-tls") pod "1b34fa5c-2867-4dbe-9285-80c327699864" (UID: "1b34fa5c-2867-4dbe-9285-80c327699864"). InnerVolumeSpecName "rabbitmq-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 14:34:41 crc kubenswrapper[4805]: I1203 14:34:41.746416 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/downward-api/1b34fa5c-2867-4dbe-9285-80c327699864-pod-info" (OuterVolumeSpecName: "pod-info") pod "1b34fa5c-2867-4dbe-9285-80c327699864" (UID: "1b34fa5c-2867-4dbe-9285-80c327699864"). InnerVolumeSpecName "pod-info". PluginName "kubernetes.io/downward-api", VolumeGidValue "" Dec 03 14:34:41 crc kubenswrapper[4805]: I1203 14:34:41.746479 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1b34fa5c-2867-4dbe-9285-80c327699864-kube-api-access-tbx6c" (OuterVolumeSpecName: "kube-api-access-tbx6c") pod "1b34fa5c-2867-4dbe-9285-80c327699864" (UID: "1b34fa5c-2867-4dbe-9285-80c327699864"). InnerVolumeSpecName "kube-api-access-tbx6c". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 14:34:41 crc kubenswrapper[4805]: I1203 14:34:41.770288 4805 generic.go:334] "Generic (PLEG): container finished" podID="1b34fa5c-2867-4dbe-9285-80c327699864" containerID="b834218684c741c79335521d4745edfa20b11d70a0f812104cdab4f0a3846643" exitCode=0 Dec 03 14:34:41 crc kubenswrapper[4805]: I1203 14:34:41.770336 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"1b34fa5c-2867-4dbe-9285-80c327699864","Type":"ContainerDied","Data":"b834218684c741c79335521d4745edfa20b11d70a0f812104cdab4f0a3846643"} Dec 03 14:34:41 crc kubenswrapper[4805]: I1203 14:34:41.770363 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"1b34fa5c-2867-4dbe-9285-80c327699864","Type":"ContainerDied","Data":"ac82af3d82d8e8e05ac0c4a1dd4039bc0b98c7443be5ab903606c15c6c452a3b"} Dec 03 14:34:41 crc kubenswrapper[4805]: I1203 14:34:41.770378 4805 scope.go:117] "RemoveContainer" containerID="b834218684c741c79335521d4745edfa20b11d70a0f812104cdab4f0a3846643" Dec 03 14:34:41 crc kubenswrapper[4805]: I1203 14:34:41.770517 4805 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Dec 03 14:34:41 crc kubenswrapper[4805]: I1203 14:34:41.787315 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1b34fa5c-2867-4dbe-9285-80c327699864-config-data" (OuterVolumeSpecName: "config-data") pod "1b34fa5c-2867-4dbe-9285-80c327699864" (UID: "1b34fa5c-2867-4dbe-9285-80c327699864"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 14:34:41 crc kubenswrapper[4805]: I1203 14:34:41.814310 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1b34fa5c-2867-4dbe-9285-80c327699864-server-conf" (OuterVolumeSpecName: "server-conf") pod "1b34fa5c-2867-4dbe-9285-80c327699864" (UID: "1b34fa5c-2867-4dbe-9285-80c327699864"). InnerVolumeSpecName "server-conf". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 14:34:41 crc kubenswrapper[4805]: I1203 14:34:41.819150 4805 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/1b34fa5c-2867-4dbe-9285-80c327699864-rabbitmq-plugins\") on node \"crc\" DevicePath \"\"" Dec 03 14:34:41 crc kubenswrapper[4805]: I1203 14:34:41.819182 4805 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/1b34fa5c-2867-4dbe-9285-80c327699864-rabbitmq-erlang-cookie\") on node \"crc\" DevicePath \"\"" Dec 03 14:34:41 crc kubenswrapper[4805]: I1203 14:34:41.819193 4805 reconciler_common.go:293] "Volume detached for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/1b34fa5c-2867-4dbe-9285-80c327699864-server-conf\") on node \"crc\" DevicePath \"\"" Dec 03 14:34:41 crc kubenswrapper[4805]: I1203 14:34:41.819202 4805 reconciler_common.go:293] "Volume detached for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/1b34fa5c-2867-4dbe-9285-80c327699864-plugins-conf\") on node \"crc\" DevicePath \"\"" Dec 03 14:34:41 crc kubenswrapper[4805]: I1203 14:34:41.819212 4805 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/1b34fa5c-2867-4dbe-9285-80c327699864-rabbitmq-tls\") on node \"crc\" DevicePath \"\"" Dec 03 14:34:41 crc kubenswrapper[4805]: I1203 14:34:41.819220 4805 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tbx6c\" (UniqueName: \"kubernetes.io/projected/1b34fa5c-2867-4dbe-9285-80c327699864-kube-api-access-tbx6c\") on node \"crc\" DevicePath \"\"" Dec 03 14:34:41 crc kubenswrapper[4805]: I1203 14:34:41.819228 4805 reconciler_common.go:293] "Volume detached for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/1b34fa5c-2867-4dbe-9285-80c327699864-pod-info\") on node \"crc\" DevicePath \"\"" Dec 03 14:34:41 crc kubenswrapper[4805]: I1203 14:34:41.819252 4805 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") on node \"crc\" " Dec 03 14:34:41 crc kubenswrapper[4805]: I1203 14:34:41.819260 4805 reconciler_common.go:293] "Volume detached for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/1b34fa5c-2867-4dbe-9285-80c327699864-erlang-cookie-secret\") on node \"crc\" DevicePath \"\"" Dec 03 14:34:41 crc kubenswrapper[4805]: I1203 14:34:41.819268 4805 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/1b34fa5c-2867-4dbe-9285-80c327699864-config-data\") on node \"crc\" DevicePath \"\"" Dec 03 14:34:41 crc kubenswrapper[4805]: I1203 14:34:41.886721 4805 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage03-crc" (UniqueName: "kubernetes.io/local-volume/local-storage03-crc") on node "crc" Dec 03 14:34:41 crc kubenswrapper[4805]: I1203 14:34:41.893096 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1b34fa5c-2867-4dbe-9285-80c327699864-rabbitmq-confd" (OuterVolumeSpecName: "rabbitmq-confd") pod "1b34fa5c-2867-4dbe-9285-80c327699864" (UID: "1b34fa5c-2867-4dbe-9285-80c327699864"). InnerVolumeSpecName "rabbitmq-confd". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 14:34:41 crc kubenswrapper[4805]: I1203 14:34:41.905668 4805 scope.go:117] "RemoveContainer" containerID="2e0f381f35c1220d9cac0d2d4c6085458b076fdcc37439c164c7ebd5a322440b" Dec 03 14:34:41 crc kubenswrapper[4805]: I1203 14:34:41.924555 4805 reconciler_common.go:293] "Volume detached for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") on node \"crc\" DevicePath \"\"" Dec 03 14:34:41 crc kubenswrapper[4805]: I1203 14:34:41.924617 4805 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/1b34fa5c-2867-4dbe-9285-80c327699864-rabbitmq-confd\") on node \"crc\" DevicePath \"\"" Dec 03 14:34:42 crc kubenswrapper[4805]: I1203 14:34:42.059034 4805 scope.go:117] "RemoveContainer" containerID="b834218684c741c79335521d4745edfa20b11d70a0f812104cdab4f0a3846643" Dec 03 14:34:42 crc kubenswrapper[4805]: E1203 14:34:42.063090 4805 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b834218684c741c79335521d4745edfa20b11d70a0f812104cdab4f0a3846643\": container with ID starting with b834218684c741c79335521d4745edfa20b11d70a0f812104cdab4f0a3846643 not found: ID does not exist" containerID="b834218684c741c79335521d4745edfa20b11d70a0f812104cdab4f0a3846643" Dec 03 14:34:42 crc kubenswrapper[4805]: I1203 14:34:42.063136 4805 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b834218684c741c79335521d4745edfa20b11d70a0f812104cdab4f0a3846643"} err="failed to get container status \"b834218684c741c79335521d4745edfa20b11d70a0f812104cdab4f0a3846643\": rpc error: code = NotFound desc = could not find container \"b834218684c741c79335521d4745edfa20b11d70a0f812104cdab4f0a3846643\": container with ID starting with b834218684c741c79335521d4745edfa20b11d70a0f812104cdab4f0a3846643 not found: ID does not exist" Dec 03 14:34:42 crc kubenswrapper[4805]: I1203 14:34:42.063164 4805 scope.go:117] "RemoveContainer" containerID="2e0f381f35c1220d9cac0d2d4c6085458b076fdcc37439c164c7ebd5a322440b" Dec 03 14:34:42 crc kubenswrapper[4805]: E1203 14:34:42.066965 4805 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2e0f381f35c1220d9cac0d2d4c6085458b076fdcc37439c164c7ebd5a322440b\": container with ID starting with 2e0f381f35c1220d9cac0d2d4c6085458b076fdcc37439c164c7ebd5a322440b not found: ID does not exist" containerID="2e0f381f35c1220d9cac0d2d4c6085458b076fdcc37439c164c7ebd5a322440b" Dec 03 14:34:42 crc kubenswrapper[4805]: I1203 14:34:42.067006 4805 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2e0f381f35c1220d9cac0d2d4c6085458b076fdcc37439c164c7ebd5a322440b"} err="failed to get container status \"2e0f381f35c1220d9cac0d2d4c6085458b076fdcc37439c164c7ebd5a322440b\": rpc error: code = NotFound desc = could not find container \"2e0f381f35c1220d9cac0d2d4c6085458b076fdcc37439c164c7ebd5a322440b\": container with ID starting with 2e0f381f35c1220d9cac0d2d4c6085458b076fdcc37439c164c7ebd5a322440b not found: ID does not exist" Dec 03 14:34:42 crc kubenswrapper[4805]: I1203 14:34:42.128279 4805 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/rabbitmq-server-0"] Dec 03 14:34:42 crc kubenswrapper[4805]: I1203 14:34:42.140464 4805 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/rabbitmq-server-0"] Dec 03 14:34:42 crc kubenswrapper[4805]: I1203 14:34:42.196896 4805 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/rabbitmq-server-0"] Dec 03 14:34:42 crc kubenswrapper[4805]: E1203 14:34:42.197334 4805 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1b34fa5c-2867-4dbe-9285-80c327699864" containerName="setup-container" Dec 03 14:34:42 crc kubenswrapper[4805]: I1203 14:34:42.197354 4805 state_mem.go:107] "Deleted CPUSet assignment" podUID="1b34fa5c-2867-4dbe-9285-80c327699864" containerName="setup-container" Dec 03 14:34:42 crc kubenswrapper[4805]: E1203 14:34:42.197372 4805 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1b34fa5c-2867-4dbe-9285-80c327699864" containerName="rabbitmq" Dec 03 14:34:42 crc kubenswrapper[4805]: I1203 14:34:42.197378 4805 state_mem.go:107] "Deleted CPUSet assignment" podUID="1b34fa5c-2867-4dbe-9285-80c327699864" containerName="rabbitmq" Dec 03 14:34:42 crc kubenswrapper[4805]: I1203 14:34:42.197557 4805 memory_manager.go:354] "RemoveStaleState removing state" podUID="1b34fa5c-2867-4dbe-9285-80c327699864" containerName="rabbitmq" Dec 03 14:34:42 crc kubenswrapper[4805]: I1203 14:34:42.198506 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Dec 03 14:34:42 crc kubenswrapper[4805]: I1203 14:34:42.205872 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-erlang-cookie" Dec 03 14:34:42 crc kubenswrapper[4805]: I1203 14:34:42.206225 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-default-user" Dec 03 14:34:42 crc kubenswrapper[4805]: I1203 14:34:42.206439 4805 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-plugins-conf" Dec 03 14:34:42 crc kubenswrapper[4805]: I1203 14:34:42.206569 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-server-dockercfg-7db6l" Dec 03 14:34:42 crc kubenswrapper[4805]: I1203 14:34:42.206639 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-rabbitmq-svc" Dec 03 14:34:42 crc kubenswrapper[4805]: I1203 14:34:42.206682 4805 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-config-data" Dec 03 14:34:42 crc kubenswrapper[4805]: I1203 14:34:42.206812 4805 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-server-conf" Dec 03 14:34:42 crc kubenswrapper[4805]: I1203 14:34:42.218382 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-server-0"] Dec 03 14:34:42 crc kubenswrapper[4805]: I1203 14:34:42.335994 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/0664c16b-c578-4685-b1ba-98c5d5f3fe07-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"0664c16b-c578-4685-b1ba-98c5d5f3fe07\") " pod="openstack/rabbitmq-server-0" Dec 03 14:34:42 crc kubenswrapper[4805]: I1203 14:34:42.336304 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/0664c16b-c578-4685-b1ba-98c5d5f3fe07-pod-info\") pod \"rabbitmq-server-0\" (UID: \"0664c16b-c578-4685-b1ba-98c5d5f3fe07\") " pod="openstack/rabbitmq-server-0" Dec 03 14:34:42 crc kubenswrapper[4805]: I1203 14:34:42.336470 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/0664c16b-c578-4685-b1ba-98c5d5f3fe07-config-data\") pod \"rabbitmq-server-0\" (UID: \"0664c16b-c578-4685-b1ba-98c5d5f3fe07\") " pod="openstack/rabbitmq-server-0" Dec 03 14:34:42 crc kubenswrapper[4805]: I1203 14:34:42.336649 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/0664c16b-c578-4685-b1ba-98c5d5f3fe07-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"0664c16b-c578-4685-b1ba-98c5d5f3fe07\") " pod="openstack/rabbitmq-server-0" Dec 03 14:34:42 crc kubenswrapper[4805]: I1203 14:34:42.336775 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/0664c16b-c578-4685-b1ba-98c5d5f3fe07-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"0664c16b-c578-4685-b1ba-98c5d5f3fe07\") " pod="openstack/rabbitmq-server-0" Dec 03 14:34:42 crc kubenswrapper[4805]: I1203 14:34:42.336936 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"rabbitmq-server-0\" (UID: \"0664c16b-c578-4685-b1ba-98c5d5f3fe07\") " pod="openstack/rabbitmq-server-0" Dec 03 14:34:42 crc kubenswrapper[4805]: I1203 14:34:42.337299 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/0664c16b-c578-4685-b1ba-98c5d5f3fe07-server-conf\") pod \"rabbitmq-server-0\" (UID: \"0664c16b-c578-4685-b1ba-98c5d5f3fe07\") " pod="openstack/rabbitmq-server-0" Dec 03 14:34:42 crc kubenswrapper[4805]: I1203 14:34:42.337489 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/0664c16b-c578-4685-b1ba-98c5d5f3fe07-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"0664c16b-c578-4685-b1ba-98c5d5f3fe07\") " pod="openstack/rabbitmq-server-0" Dec 03 14:34:42 crc kubenswrapper[4805]: I1203 14:34:42.337642 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sqz6n\" (UniqueName: \"kubernetes.io/projected/0664c16b-c578-4685-b1ba-98c5d5f3fe07-kube-api-access-sqz6n\") pod \"rabbitmq-server-0\" (UID: \"0664c16b-c578-4685-b1ba-98c5d5f3fe07\") " pod="openstack/rabbitmq-server-0" Dec 03 14:34:42 crc kubenswrapper[4805]: I1203 14:34:42.337917 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/0664c16b-c578-4685-b1ba-98c5d5f3fe07-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"0664c16b-c578-4685-b1ba-98c5d5f3fe07\") " pod="openstack/rabbitmq-server-0" Dec 03 14:34:42 crc kubenswrapper[4805]: I1203 14:34:42.338058 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/0664c16b-c578-4685-b1ba-98c5d5f3fe07-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"0664c16b-c578-4685-b1ba-98c5d5f3fe07\") " pod="openstack/rabbitmq-server-0" Dec 03 14:34:42 crc kubenswrapper[4805]: I1203 14:34:42.364753 4805 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/rabbitmq-cell1-server-0" podUID="656f691c-1188-47b9-b8dd-b393a230602a" containerName="rabbitmq" probeResult="failure" output="dial tcp 10.217.0.99:5671: connect: connection refused" Dec 03 14:34:42 crc kubenswrapper[4805]: I1203 14:34:42.440070 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/0664c16b-c578-4685-b1ba-98c5d5f3fe07-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"0664c16b-c578-4685-b1ba-98c5d5f3fe07\") " pod="openstack/rabbitmq-server-0" Dec 03 14:34:42 crc kubenswrapper[4805]: I1203 14:34:42.440125 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/0664c16b-c578-4685-b1ba-98c5d5f3fe07-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"0664c16b-c578-4685-b1ba-98c5d5f3fe07\") " pod="openstack/rabbitmq-server-0" Dec 03 14:34:42 crc kubenswrapper[4805]: I1203 14:34:42.440156 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"rabbitmq-server-0\" (UID: \"0664c16b-c578-4685-b1ba-98c5d5f3fe07\") " pod="openstack/rabbitmq-server-0" Dec 03 14:34:42 crc kubenswrapper[4805]: I1203 14:34:42.440180 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/0664c16b-c578-4685-b1ba-98c5d5f3fe07-server-conf\") pod \"rabbitmq-server-0\" (UID: \"0664c16b-c578-4685-b1ba-98c5d5f3fe07\") " pod="openstack/rabbitmq-server-0" Dec 03 14:34:42 crc kubenswrapper[4805]: I1203 14:34:42.440224 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/0664c16b-c578-4685-b1ba-98c5d5f3fe07-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"0664c16b-c578-4685-b1ba-98c5d5f3fe07\") " pod="openstack/rabbitmq-server-0" Dec 03 14:34:42 crc kubenswrapper[4805]: I1203 14:34:42.440244 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sqz6n\" (UniqueName: \"kubernetes.io/projected/0664c16b-c578-4685-b1ba-98c5d5f3fe07-kube-api-access-sqz6n\") pod \"rabbitmq-server-0\" (UID: \"0664c16b-c578-4685-b1ba-98c5d5f3fe07\") " pod="openstack/rabbitmq-server-0" Dec 03 14:34:42 crc kubenswrapper[4805]: I1203 14:34:42.440288 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/0664c16b-c578-4685-b1ba-98c5d5f3fe07-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"0664c16b-c578-4685-b1ba-98c5d5f3fe07\") " pod="openstack/rabbitmq-server-0" Dec 03 14:34:42 crc kubenswrapper[4805]: I1203 14:34:42.440319 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/0664c16b-c578-4685-b1ba-98c5d5f3fe07-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"0664c16b-c578-4685-b1ba-98c5d5f3fe07\") " pod="openstack/rabbitmq-server-0" Dec 03 14:34:42 crc kubenswrapper[4805]: I1203 14:34:42.440339 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/0664c16b-c578-4685-b1ba-98c5d5f3fe07-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"0664c16b-c578-4685-b1ba-98c5d5f3fe07\") " pod="openstack/rabbitmq-server-0" Dec 03 14:34:42 crc kubenswrapper[4805]: I1203 14:34:42.440355 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/0664c16b-c578-4685-b1ba-98c5d5f3fe07-pod-info\") pod \"rabbitmq-server-0\" (UID: \"0664c16b-c578-4685-b1ba-98c5d5f3fe07\") " pod="openstack/rabbitmq-server-0" Dec 03 14:34:42 crc kubenswrapper[4805]: I1203 14:34:42.440374 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/0664c16b-c578-4685-b1ba-98c5d5f3fe07-config-data\") pod \"rabbitmq-server-0\" (UID: \"0664c16b-c578-4685-b1ba-98c5d5f3fe07\") " pod="openstack/rabbitmq-server-0" Dec 03 14:34:42 crc kubenswrapper[4805]: I1203 14:34:42.440418 4805 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"rabbitmq-server-0\" (UID: \"0664c16b-c578-4685-b1ba-98c5d5f3fe07\") device mount path \"/mnt/openstack/pv03\"" pod="openstack/rabbitmq-server-0" Dec 03 14:34:42 crc kubenswrapper[4805]: I1203 14:34:42.441003 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/0664c16b-c578-4685-b1ba-98c5d5f3fe07-config-data\") pod \"rabbitmq-server-0\" (UID: \"0664c16b-c578-4685-b1ba-98c5d5f3fe07\") " pod="openstack/rabbitmq-server-0" Dec 03 14:34:42 crc kubenswrapper[4805]: I1203 14:34:42.441010 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/0664c16b-c578-4685-b1ba-98c5d5f3fe07-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"0664c16b-c578-4685-b1ba-98c5d5f3fe07\") " pod="openstack/rabbitmq-server-0" Dec 03 14:34:42 crc kubenswrapper[4805]: I1203 14:34:42.441804 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/0664c16b-c578-4685-b1ba-98c5d5f3fe07-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"0664c16b-c578-4685-b1ba-98c5d5f3fe07\") " pod="openstack/rabbitmq-server-0" Dec 03 14:34:42 crc kubenswrapper[4805]: I1203 14:34:42.442011 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/0664c16b-c578-4685-b1ba-98c5d5f3fe07-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"0664c16b-c578-4685-b1ba-98c5d5f3fe07\") " pod="openstack/rabbitmq-server-0" Dec 03 14:34:42 crc kubenswrapper[4805]: I1203 14:34:42.442559 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/0664c16b-c578-4685-b1ba-98c5d5f3fe07-server-conf\") pod \"rabbitmq-server-0\" (UID: \"0664c16b-c578-4685-b1ba-98c5d5f3fe07\") " pod="openstack/rabbitmq-server-0" Dec 03 14:34:42 crc kubenswrapper[4805]: I1203 14:34:42.446680 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/0664c16b-c578-4685-b1ba-98c5d5f3fe07-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"0664c16b-c578-4685-b1ba-98c5d5f3fe07\") " pod="openstack/rabbitmq-server-0" Dec 03 14:34:42 crc kubenswrapper[4805]: I1203 14:34:42.454112 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/0664c16b-c578-4685-b1ba-98c5d5f3fe07-pod-info\") pod \"rabbitmq-server-0\" (UID: \"0664c16b-c578-4685-b1ba-98c5d5f3fe07\") " pod="openstack/rabbitmq-server-0" Dec 03 14:34:42 crc kubenswrapper[4805]: I1203 14:34:42.454390 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/0664c16b-c578-4685-b1ba-98c5d5f3fe07-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"0664c16b-c578-4685-b1ba-98c5d5f3fe07\") " pod="openstack/rabbitmq-server-0" Dec 03 14:34:42 crc kubenswrapper[4805]: I1203 14:34:42.458606 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/0664c16b-c578-4685-b1ba-98c5d5f3fe07-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"0664c16b-c578-4685-b1ba-98c5d5f3fe07\") " pod="openstack/rabbitmq-server-0" Dec 03 14:34:42 crc kubenswrapper[4805]: I1203 14:34:42.467915 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sqz6n\" (UniqueName: \"kubernetes.io/projected/0664c16b-c578-4685-b1ba-98c5d5f3fe07-kube-api-access-sqz6n\") pod \"rabbitmq-server-0\" (UID: \"0664c16b-c578-4685-b1ba-98c5d5f3fe07\") " pod="openstack/rabbitmq-server-0" Dec 03 14:34:42 crc kubenswrapper[4805]: I1203 14:34:42.492561 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"rabbitmq-server-0\" (UID: \"0664c16b-c578-4685-b1ba-98c5d5f3fe07\") " pod="openstack/rabbitmq-server-0" Dec 03 14:34:42 crc kubenswrapper[4805]: I1203 14:34:42.635384 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Dec 03 14:34:42 crc kubenswrapper[4805]: I1203 14:34:42.706588 4805 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1b34fa5c-2867-4dbe-9285-80c327699864" path="/var/lib/kubelet/pods/1b34fa5c-2867-4dbe-9285-80c327699864/volumes" Dec 03 14:34:42 crc kubenswrapper[4805]: I1203 14:34:42.811473 4805 generic.go:334] "Generic (PLEG): container finished" podID="656f691c-1188-47b9-b8dd-b393a230602a" containerID="770e52863be49086afe2bea1b59e31fac9f82d73b30ec6fe23a2c1d8d92fe910" exitCode=0 Dec 03 14:34:42 crc kubenswrapper[4805]: I1203 14:34:42.811596 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"656f691c-1188-47b9-b8dd-b393a230602a","Type":"ContainerDied","Data":"770e52863be49086afe2bea1b59e31fac9f82d73b30ec6fe23a2c1d8d92fe910"} Dec 03 14:34:43 crc kubenswrapper[4805]: I1203 14:34:43.144206 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-server-0"] Dec 03 14:34:43 crc kubenswrapper[4805]: I1203 14:34:43.151484 4805 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Dec 03 14:34:43 crc kubenswrapper[4805]: W1203 14:34:43.166325 4805 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod0664c16b_c578_4685_b1ba_98c5d5f3fe07.slice/crio-eb5c9bce585b7352c6b85303fbdf6d6778d2dccdea74d506e0523493054a2c9c WatchSource:0}: Error finding container eb5c9bce585b7352c6b85303fbdf6d6778d2dccdea74d506e0523493054a2c9c: Status 404 returned error can't find the container with id eb5c9bce585b7352c6b85303fbdf6d6778d2dccdea74d506e0523493054a2c9c Dec 03 14:34:43 crc kubenswrapper[4805]: I1203 14:34:43.257490 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/656f691c-1188-47b9-b8dd-b393a230602a-config-data\") pod \"656f691c-1188-47b9-b8dd-b393a230602a\" (UID: \"656f691c-1188-47b9-b8dd-b393a230602a\") " Dec 03 14:34:43 crc kubenswrapper[4805]: I1203 14:34:43.257550 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/656f691c-1188-47b9-b8dd-b393a230602a-server-conf\") pod \"656f691c-1188-47b9-b8dd-b393a230602a\" (UID: \"656f691c-1188-47b9-b8dd-b393a230602a\") " Dec 03 14:34:43 crc kubenswrapper[4805]: I1203 14:34:43.257662 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/656f691c-1188-47b9-b8dd-b393a230602a-rabbitmq-plugins\") pod \"656f691c-1188-47b9-b8dd-b393a230602a\" (UID: \"656f691c-1188-47b9-b8dd-b393a230602a\") " Dec 03 14:34:43 crc kubenswrapper[4805]: I1203 14:34:43.257798 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/656f691c-1188-47b9-b8dd-b393a230602a-erlang-cookie-secret\") pod \"656f691c-1188-47b9-b8dd-b393a230602a\" (UID: \"656f691c-1188-47b9-b8dd-b393a230602a\") " Dec 03 14:34:43 crc kubenswrapper[4805]: I1203 14:34:43.257865 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/656f691c-1188-47b9-b8dd-b393a230602a-plugins-conf\") pod \"656f691c-1188-47b9-b8dd-b393a230602a\" (UID: \"656f691c-1188-47b9-b8dd-b393a230602a\") " Dec 03 14:34:43 crc kubenswrapper[4805]: I1203 14:34:43.257904 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/656f691c-1188-47b9-b8dd-b393a230602a-pod-info\") pod \"656f691c-1188-47b9-b8dd-b393a230602a\" (UID: \"656f691c-1188-47b9-b8dd-b393a230602a\") " Dec 03 14:34:43 crc kubenswrapper[4805]: I1203 14:34:43.258020 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9qnqg\" (UniqueName: \"kubernetes.io/projected/656f691c-1188-47b9-b8dd-b393a230602a-kube-api-access-9qnqg\") pod \"656f691c-1188-47b9-b8dd-b393a230602a\" (UID: \"656f691c-1188-47b9-b8dd-b393a230602a\") " Dec 03 14:34:43 crc kubenswrapper[4805]: I1203 14:34:43.258056 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/656f691c-1188-47b9-b8dd-b393a230602a-rabbitmq-confd\") pod \"656f691c-1188-47b9-b8dd-b393a230602a\" (UID: \"656f691c-1188-47b9-b8dd-b393a230602a\") " Dec 03 14:34:43 crc kubenswrapper[4805]: I1203 14:34:43.258112 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/656f691c-1188-47b9-b8dd-b393a230602a-rabbitmq-erlang-cookie\") pod \"656f691c-1188-47b9-b8dd-b393a230602a\" (UID: \"656f691c-1188-47b9-b8dd-b393a230602a\") " Dec 03 14:34:43 crc kubenswrapper[4805]: I1203 14:34:43.258166 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/656f691c-1188-47b9-b8dd-b393a230602a-rabbitmq-tls\") pod \"656f691c-1188-47b9-b8dd-b393a230602a\" (UID: \"656f691c-1188-47b9-b8dd-b393a230602a\") " Dec 03 14:34:43 crc kubenswrapper[4805]: I1203 14:34:43.258227 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"persistence\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"656f691c-1188-47b9-b8dd-b393a230602a\" (UID: \"656f691c-1188-47b9-b8dd-b393a230602a\") " Dec 03 14:34:43 crc kubenswrapper[4805]: I1203 14:34:43.259178 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/656f691c-1188-47b9-b8dd-b393a230602a-plugins-conf" (OuterVolumeSpecName: "plugins-conf") pod "656f691c-1188-47b9-b8dd-b393a230602a" (UID: "656f691c-1188-47b9-b8dd-b393a230602a"). InnerVolumeSpecName "plugins-conf". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 14:34:43 crc kubenswrapper[4805]: I1203 14:34:43.259994 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/656f691c-1188-47b9-b8dd-b393a230602a-rabbitmq-plugins" (OuterVolumeSpecName: "rabbitmq-plugins") pod "656f691c-1188-47b9-b8dd-b393a230602a" (UID: "656f691c-1188-47b9-b8dd-b393a230602a"). InnerVolumeSpecName "rabbitmq-plugins". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 14:34:43 crc kubenswrapper[4805]: I1203 14:34:43.260520 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/656f691c-1188-47b9-b8dd-b393a230602a-rabbitmq-erlang-cookie" (OuterVolumeSpecName: "rabbitmq-erlang-cookie") pod "656f691c-1188-47b9-b8dd-b393a230602a" (UID: "656f691c-1188-47b9-b8dd-b393a230602a"). InnerVolumeSpecName "rabbitmq-erlang-cookie". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 14:34:43 crc kubenswrapper[4805]: I1203 14:34:43.263890 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/downward-api/656f691c-1188-47b9-b8dd-b393a230602a-pod-info" (OuterVolumeSpecName: "pod-info") pod "656f691c-1188-47b9-b8dd-b393a230602a" (UID: "656f691c-1188-47b9-b8dd-b393a230602a"). InnerVolumeSpecName "pod-info". PluginName "kubernetes.io/downward-api", VolumeGidValue "" Dec 03 14:34:43 crc kubenswrapper[4805]: I1203 14:34:43.266085 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/656f691c-1188-47b9-b8dd-b393a230602a-erlang-cookie-secret" (OuterVolumeSpecName: "erlang-cookie-secret") pod "656f691c-1188-47b9-b8dd-b393a230602a" (UID: "656f691c-1188-47b9-b8dd-b393a230602a"). InnerVolumeSpecName "erlang-cookie-secret". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 14:34:43 crc kubenswrapper[4805]: I1203 14:34:43.268966 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage01-crc" (OuterVolumeSpecName: "persistence") pod "656f691c-1188-47b9-b8dd-b393a230602a" (UID: "656f691c-1188-47b9-b8dd-b393a230602a"). InnerVolumeSpecName "local-storage01-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Dec 03 14:34:43 crc kubenswrapper[4805]: I1203 14:34:43.274616 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/656f691c-1188-47b9-b8dd-b393a230602a-kube-api-access-9qnqg" (OuterVolumeSpecName: "kube-api-access-9qnqg") pod "656f691c-1188-47b9-b8dd-b393a230602a" (UID: "656f691c-1188-47b9-b8dd-b393a230602a"). InnerVolumeSpecName "kube-api-access-9qnqg". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 14:34:43 crc kubenswrapper[4805]: I1203 14:34:43.274744 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/656f691c-1188-47b9-b8dd-b393a230602a-rabbitmq-tls" (OuterVolumeSpecName: "rabbitmq-tls") pod "656f691c-1188-47b9-b8dd-b393a230602a" (UID: "656f691c-1188-47b9-b8dd-b393a230602a"). InnerVolumeSpecName "rabbitmq-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 14:34:43 crc kubenswrapper[4805]: I1203 14:34:43.296566 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/656f691c-1188-47b9-b8dd-b393a230602a-config-data" (OuterVolumeSpecName: "config-data") pod "656f691c-1188-47b9-b8dd-b393a230602a" (UID: "656f691c-1188-47b9-b8dd-b393a230602a"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 14:34:43 crc kubenswrapper[4805]: I1203 14:34:43.342480 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/656f691c-1188-47b9-b8dd-b393a230602a-server-conf" (OuterVolumeSpecName: "server-conf") pod "656f691c-1188-47b9-b8dd-b393a230602a" (UID: "656f691c-1188-47b9-b8dd-b393a230602a"). InnerVolumeSpecName "server-conf". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 14:34:43 crc kubenswrapper[4805]: I1203 14:34:43.362080 4805 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/656f691c-1188-47b9-b8dd-b393a230602a-rabbitmq-plugins\") on node \"crc\" DevicePath \"\"" Dec 03 14:34:43 crc kubenswrapper[4805]: I1203 14:34:43.362136 4805 reconciler_common.go:293] "Volume detached for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/656f691c-1188-47b9-b8dd-b393a230602a-erlang-cookie-secret\") on node \"crc\" DevicePath \"\"" Dec 03 14:34:43 crc kubenswrapper[4805]: I1203 14:34:43.362150 4805 reconciler_common.go:293] "Volume detached for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/656f691c-1188-47b9-b8dd-b393a230602a-plugins-conf\") on node \"crc\" DevicePath \"\"" Dec 03 14:34:43 crc kubenswrapper[4805]: I1203 14:34:43.362159 4805 reconciler_common.go:293] "Volume detached for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/656f691c-1188-47b9-b8dd-b393a230602a-pod-info\") on node \"crc\" DevicePath \"\"" Dec 03 14:34:43 crc kubenswrapper[4805]: I1203 14:34:43.362170 4805 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9qnqg\" (UniqueName: \"kubernetes.io/projected/656f691c-1188-47b9-b8dd-b393a230602a-kube-api-access-9qnqg\") on node \"crc\" DevicePath \"\"" Dec 03 14:34:43 crc kubenswrapper[4805]: I1203 14:34:43.362182 4805 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/656f691c-1188-47b9-b8dd-b393a230602a-rabbitmq-erlang-cookie\") on node \"crc\" DevicePath \"\"" Dec 03 14:34:43 crc kubenswrapper[4805]: I1203 14:34:43.362192 4805 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/656f691c-1188-47b9-b8dd-b393a230602a-rabbitmq-tls\") on node \"crc\" DevicePath \"\"" Dec 03 14:34:43 crc kubenswrapper[4805]: I1203 14:34:43.362221 4805 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") on node \"crc\" " Dec 03 14:34:43 crc kubenswrapper[4805]: I1203 14:34:43.362232 4805 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/656f691c-1188-47b9-b8dd-b393a230602a-config-data\") on node \"crc\" DevicePath \"\"" Dec 03 14:34:43 crc kubenswrapper[4805]: I1203 14:34:43.362241 4805 reconciler_common.go:293] "Volume detached for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/656f691c-1188-47b9-b8dd-b393a230602a-server-conf\") on node \"crc\" DevicePath \"\"" Dec 03 14:34:43 crc kubenswrapper[4805]: I1203 14:34:43.394340 4805 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage01-crc" (UniqueName: "kubernetes.io/local-volume/local-storage01-crc") on node "crc" Dec 03 14:34:43 crc kubenswrapper[4805]: I1203 14:34:43.432946 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/656f691c-1188-47b9-b8dd-b393a230602a-rabbitmq-confd" (OuterVolumeSpecName: "rabbitmq-confd") pod "656f691c-1188-47b9-b8dd-b393a230602a" (UID: "656f691c-1188-47b9-b8dd-b393a230602a"). InnerVolumeSpecName "rabbitmq-confd". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 14:34:43 crc kubenswrapper[4805]: I1203 14:34:43.464759 4805 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/656f691c-1188-47b9-b8dd-b393a230602a-rabbitmq-confd\") on node \"crc\" DevicePath \"\"" Dec 03 14:34:43 crc kubenswrapper[4805]: I1203 14:34:43.464826 4805 reconciler_common.go:293] "Volume detached for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") on node \"crc\" DevicePath \"\"" Dec 03 14:34:43 crc kubenswrapper[4805]: I1203 14:34:43.823593 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"0664c16b-c578-4685-b1ba-98c5d5f3fe07","Type":"ContainerStarted","Data":"eb5c9bce585b7352c6b85303fbdf6d6778d2dccdea74d506e0523493054a2c9c"} Dec 03 14:34:43 crc kubenswrapper[4805]: I1203 14:34:43.828497 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"656f691c-1188-47b9-b8dd-b393a230602a","Type":"ContainerDied","Data":"2e965419aa5bcdde632d972d20b4f13ceff9f46a20482645583589d483f55028"} Dec 03 14:34:43 crc kubenswrapper[4805]: I1203 14:34:43.828579 4805 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Dec 03 14:34:43 crc kubenswrapper[4805]: I1203 14:34:43.828696 4805 scope.go:117] "RemoveContainer" containerID="770e52863be49086afe2bea1b59e31fac9f82d73b30ec6fe23a2c1d8d92fe910" Dec 03 14:34:43 crc kubenswrapper[4805]: I1203 14:34:43.851738 4805 scope.go:117] "RemoveContainer" containerID="0d085016bd6e7629b80b349c936428485e82d12ba8205b13e1f636404a6741cc" Dec 03 14:34:43 crc kubenswrapper[4805]: I1203 14:34:43.868623 4805 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Dec 03 14:34:43 crc kubenswrapper[4805]: I1203 14:34:43.879463 4805 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Dec 03 14:34:43 crc kubenswrapper[4805]: I1203 14:34:43.909819 4805 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Dec 03 14:34:43 crc kubenswrapper[4805]: E1203 14:34:43.910401 4805 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="656f691c-1188-47b9-b8dd-b393a230602a" containerName="rabbitmq" Dec 03 14:34:43 crc kubenswrapper[4805]: I1203 14:34:43.910421 4805 state_mem.go:107] "Deleted CPUSet assignment" podUID="656f691c-1188-47b9-b8dd-b393a230602a" containerName="rabbitmq" Dec 03 14:34:43 crc kubenswrapper[4805]: E1203 14:34:43.910447 4805 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="656f691c-1188-47b9-b8dd-b393a230602a" containerName="setup-container" Dec 03 14:34:43 crc kubenswrapper[4805]: I1203 14:34:43.910455 4805 state_mem.go:107] "Deleted CPUSet assignment" podUID="656f691c-1188-47b9-b8dd-b393a230602a" containerName="setup-container" Dec 03 14:34:43 crc kubenswrapper[4805]: I1203 14:34:43.910682 4805 memory_manager.go:354] "RemoveStaleState removing state" podUID="656f691c-1188-47b9-b8dd-b393a230602a" containerName="rabbitmq" Dec 03 14:34:43 crc kubenswrapper[4805]: I1203 14:34:43.911893 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Dec 03 14:34:43 crc kubenswrapper[4805]: I1203 14:34:43.919876 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Dec 03 14:34:43 crc kubenswrapper[4805]: I1203 14:34:43.921401 4805 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-config-data" Dec 03 14:34:43 crc kubenswrapper[4805]: I1203 14:34:43.921562 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-server-dockercfg-vpb8f" Dec 03 14:34:43 crc kubenswrapper[4805]: I1203 14:34:43.921444 4805 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-plugins-conf" Dec 03 14:34:43 crc kubenswrapper[4805]: I1203 14:34:43.921675 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-default-user" Dec 03 14:34:43 crc kubenswrapper[4805]: I1203 14:34:43.921476 4805 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-server-conf" Dec 03 14:34:43 crc kubenswrapper[4805]: I1203 14:34:43.921748 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-rabbitmq-cell1-svc" Dec 03 14:34:43 crc kubenswrapper[4805]: I1203 14:34:43.921802 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-erlang-cookie" Dec 03 14:34:44 crc kubenswrapper[4805]: I1203 14:34:44.077890 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/fff4a752-b61c-4c51-bc0b-428c9affbb00-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"fff4a752-b61c-4c51-bc0b-428c9affbb00\") " pod="openstack/rabbitmq-cell1-server-0" Dec 03 14:34:44 crc kubenswrapper[4805]: I1203 14:34:44.077972 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/fff4a752-b61c-4c51-bc0b-428c9affbb00-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"fff4a752-b61c-4c51-bc0b-428c9affbb00\") " pod="openstack/rabbitmq-cell1-server-0" Dec 03 14:34:44 crc kubenswrapper[4805]: I1203 14:34:44.078002 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"fff4a752-b61c-4c51-bc0b-428c9affbb00\") " pod="openstack/rabbitmq-cell1-server-0" Dec 03 14:34:44 crc kubenswrapper[4805]: I1203 14:34:44.078039 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/fff4a752-b61c-4c51-bc0b-428c9affbb00-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"fff4a752-b61c-4c51-bc0b-428c9affbb00\") " pod="openstack/rabbitmq-cell1-server-0" Dec 03 14:34:44 crc kubenswrapper[4805]: I1203 14:34:44.078322 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/fff4a752-b61c-4c51-bc0b-428c9affbb00-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"fff4a752-b61c-4c51-bc0b-428c9affbb00\") " pod="openstack/rabbitmq-cell1-server-0" Dec 03 14:34:44 crc kubenswrapper[4805]: I1203 14:34:44.078394 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/fff4a752-b61c-4c51-bc0b-428c9affbb00-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"fff4a752-b61c-4c51-bc0b-428c9affbb00\") " pod="openstack/rabbitmq-cell1-server-0" Dec 03 14:34:44 crc kubenswrapper[4805]: I1203 14:34:44.078485 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/fff4a752-b61c-4c51-bc0b-428c9affbb00-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"fff4a752-b61c-4c51-bc0b-428c9affbb00\") " pod="openstack/rabbitmq-cell1-server-0" Dec 03 14:34:44 crc kubenswrapper[4805]: I1203 14:34:44.078523 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dh492\" (UniqueName: \"kubernetes.io/projected/fff4a752-b61c-4c51-bc0b-428c9affbb00-kube-api-access-dh492\") pod \"rabbitmq-cell1-server-0\" (UID: \"fff4a752-b61c-4c51-bc0b-428c9affbb00\") " pod="openstack/rabbitmq-cell1-server-0" Dec 03 14:34:44 crc kubenswrapper[4805]: I1203 14:34:44.078574 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/fff4a752-b61c-4c51-bc0b-428c9affbb00-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"fff4a752-b61c-4c51-bc0b-428c9affbb00\") " pod="openstack/rabbitmq-cell1-server-0" Dec 03 14:34:44 crc kubenswrapper[4805]: I1203 14:34:44.078660 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/fff4a752-b61c-4c51-bc0b-428c9affbb00-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"fff4a752-b61c-4c51-bc0b-428c9affbb00\") " pod="openstack/rabbitmq-cell1-server-0" Dec 03 14:34:44 crc kubenswrapper[4805]: I1203 14:34:44.078707 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/fff4a752-b61c-4c51-bc0b-428c9affbb00-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"fff4a752-b61c-4c51-bc0b-428c9affbb00\") " pod="openstack/rabbitmq-cell1-server-0" Dec 03 14:34:44 crc kubenswrapper[4805]: I1203 14:34:44.180110 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/fff4a752-b61c-4c51-bc0b-428c9affbb00-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"fff4a752-b61c-4c51-bc0b-428c9affbb00\") " pod="openstack/rabbitmq-cell1-server-0" Dec 03 14:34:44 crc kubenswrapper[4805]: I1203 14:34:44.180973 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/fff4a752-b61c-4c51-bc0b-428c9affbb00-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"fff4a752-b61c-4c51-bc0b-428c9affbb00\") " pod="openstack/rabbitmq-cell1-server-0" Dec 03 14:34:44 crc kubenswrapper[4805]: I1203 14:34:44.181015 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/fff4a752-b61c-4c51-bc0b-428c9affbb00-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"fff4a752-b61c-4c51-bc0b-428c9affbb00\") " pod="openstack/rabbitmq-cell1-server-0" Dec 03 14:34:44 crc kubenswrapper[4805]: I1203 14:34:44.181062 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/fff4a752-b61c-4c51-bc0b-428c9affbb00-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"fff4a752-b61c-4c51-bc0b-428c9affbb00\") " pod="openstack/rabbitmq-cell1-server-0" Dec 03 14:34:44 crc kubenswrapper[4805]: I1203 14:34:44.181094 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dh492\" (UniqueName: \"kubernetes.io/projected/fff4a752-b61c-4c51-bc0b-428c9affbb00-kube-api-access-dh492\") pod \"rabbitmq-cell1-server-0\" (UID: \"fff4a752-b61c-4c51-bc0b-428c9affbb00\") " pod="openstack/rabbitmq-cell1-server-0" Dec 03 14:34:44 crc kubenswrapper[4805]: I1203 14:34:44.181131 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/fff4a752-b61c-4c51-bc0b-428c9affbb00-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"fff4a752-b61c-4c51-bc0b-428c9affbb00\") " pod="openstack/rabbitmq-cell1-server-0" Dec 03 14:34:44 crc kubenswrapper[4805]: I1203 14:34:44.181190 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/fff4a752-b61c-4c51-bc0b-428c9affbb00-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"fff4a752-b61c-4c51-bc0b-428c9affbb00\") " pod="openstack/rabbitmq-cell1-server-0" Dec 03 14:34:44 crc kubenswrapper[4805]: I1203 14:34:44.181212 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/fff4a752-b61c-4c51-bc0b-428c9affbb00-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"fff4a752-b61c-4c51-bc0b-428c9affbb00\") " pod="openstack/rabbitmq-cell1-server-0" Dec 03 14:34:44 crc kubenswrapper[4805]: I1203 14:34:44.181263 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/fff4a752-b61c-4c51-bc0b-428c9affbb00-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"fff4a752-b61c-4c51-bc0b-428c9affbb00\") " pod="openstack/rabbitmq-cell1-server-0" Dec 03 14:34:44 crc kubenswrapper[4805]: I1203 14:34:44.181287 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/fff4a752-b61c-4c51-bc0b-428c9affbb00-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"fff4a752-b61c-4c51-bc0b-428c9affbb00\") " pod="openstack/rabbitmq-cell1-server-0" Dec 03 14:34:44 crc kubenswrapper[4805]: I1203 14:34:44.181309 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"fff4a752-b61c-4c51-bc0b-428c9affbb00\") " pod="openstack/rabbitmq-cell1-server-0" Dec 03 14:34:44 crc kubenswrapper[4805]: I1203 14:34:44.181437 4805 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"fff4a752-b61c-4c51-bc0b-428c9affbb00\") device mount path \"/mnt/openstack/pv01\"" pod="openstack/rabbitmq-cell1-server-0" Dec 03 14:34:44 crc kubenswrapper[4805]: I1203 14:34:44.181562 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/fff4a752-b61c-4c51-bc0b-428c9affbb00-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"fff4a752-b61c-4c51-bc0b-428c9affbb00\") " pod="openstack/rabbitmq-cell1-server-0" Dec 03 14:34:44 crc kubenswrapper[4805]: I1203 14:34:44.182014 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/fff4a752-b61c-4c51-bc0b-428c9affbb00-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"fff4a752-b61c-4c51-bc0b-428c9affbb00\") " pod="openstack/rabbitmq-cell1-server-0" Dec 03 14:34:44 crc kubenswrapper[4805]: I1203 14:34:44.182123 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/fff4a752-b61c-4c51-bc0b-428c9affbb00-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"fff4a752-b61c-4c51-bc0b-428c9affbb00\") " pod="openstack/rabbitmq-cell1-server-0" Dec 03 14:34:44 crc kubenswrapper[4805]: I1203 14:34:44.182599 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/fff4a752-b61c-4c51-bc0b-428c9affbb00-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"fff4a752-b61c-4c51-bc0b-428c9affbb00\") " pod="openstack/rabbitmq-cell1-server-0" Dec 03 14:34:44 crc kubenswrapper[4805]: I1203 14:34:44.182717 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/fff4a752-b61c-4c51-bc0b-428c9affbb00-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"fff4a752-b61c-4c51-bc0b-428c9affbb00\") " pod="openstack/rabbitmq-cell1-server-0" Dec 03 14:34:44 crc kubenswrapper[4805]: I1203 14:34:44.186590 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/fff4a752-b61c-4c51-bc0b-428c9affbb00-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"fff4a752-b61c-4c51-bc0b-428c9affbb00\") " pod="openstack/rabbitmq-cell1-server-0" Dec 03 14:34:44 crc kubenswrapper[4805]: I1203 14:34:44.195445 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/fff4a752-b61c-4c51-bc0b-428c9affbb00-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"fff4a752-b61c-4c51-bc0b-428c9affbb00\") " pod="openstack/rabbitmq-cell1-server-0" Dec 03 14:34:44 crc kubenswrapper[4805]: I1203 14:34:44.196023 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/fff4a752-b61c-4c51-bc0b-428c9affbb00-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"fff4a752-b61c-4c51-bc0b-428c9affbb00\") " pod="openstack/rabbitmq-cell1-server-0" Dec 03 14:34:44 crc kubenswrapper[4805]: I1203 14:34:44.197385 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/fff4a752-b61c-4c51-bc0b-428c9affbb00-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"fff4a752-b61c-4c51-bc0b-428c9affbb00\") " pod="openstack/rabbitmq-cell1-server-0" Dec 03 14:34:44 crc kubenswrapper[4805]: I1203 14:34:44.199345 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dh492\" (UniqueName: \"kubernetes.io/projected/fff4a752-b61c-4c51-bc0b-428c9affbb00-kube-api-access-dh492\") pod \"rabbitmq-cell1-server-0\" (UID: \"fff4a752-b61c-4c51-bc0b-428c9affbb00\") " pod="openstack/rabbitmq-cell1-server-0" Dec 03 14:34:44 crc kubenswrapper[4805]: I1203 14:34:44.223926 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"fff4a752-b61c-4c51-bc0b-428c9affbb00\") " pod="openstack/rabbitmq-cell1-server-0" Dec 03 14:34:44 crc kubenswrapper[4805]: I1203 14:34:44.262353 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Dec 03 14:34:44 crc kubenswrapper[4805]: I1203 14:34:44.705504 4805 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="656f691c-1188-47b9-b8dd-b393a230602a" path="/var/lib/kubelet/pods/656f691c-1188-47b9-b8dd-b393a230602a/volumes" Dec 03 14:34:44 crc kubenswrapper[4805]: I1203 14:34:44.818501 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Dec 03 14:34:45 crc kubenswrapper[4805]: I1203 14:34:45.850275 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"0664c16b-c578-4685-b1ba-98c5d5f3fe07","Type":"ContainerStarted","Data":"e300f43553ffdf9bbe5023281eeb1b30e1667ecb203fdb8aba8b39577bb5944e"} Dec 03 14:34:45 crc kubenswrapper[4805]: I1203 14:34:45.852143 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"fff4a752-b61c-4c51-bc0b-428c9affbb00","Type":"ContainerStarted","Data":"1244ab844210f974ed28d7fa656cac32c72100fb12890052d5c9268681afade0"} Dec 03 14:34:46 crc kubenswrapper[4805]: I1203 14:34:46.444590 4805 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-68df85789f-29wwx"] Dec 03 14:34:46 crc kubenswrapper[4805]: I1203 14:34:46.446816 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-68df85789f-29wwx" Dec 03 14:34:46 crc kubenswrapper[4805]: I1203 14:34:46.455490 4805 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-edpm-ipam" Dec 03 14:34:46 crc kubenswrapper[4805]: I1203 14:34:46.456361 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-68df85789f-29wwx"] Dec 03 14:34:46 crc kubenswrapper[4805]: I1203 14:34:46.627728 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/73dd742e-b2cd-45ec-aa1e-b1150268512e-dns-svc\") pod \"dnsmasq-dns-68df85789f-29wwx\" (UID: \"73dd742e-b2cd-45ec-aa1e-b1150268512e\") " pod="openstack/dnsmasq-dns-68df85789f-29wwx" Dec 03 14:34:46 crc kubenswrapper[4805]: I1203 14:34:46.627863 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-k4tj5\" (UniqueName: \"kubernetes.io/projected/73dd742e-b2cd-45ec-aa1e-b1150268512e-kube-api-access-k4tj5\") pod \"dnsmasq-dns-68df85789f-29wwx\" (UID: \"73dd742e-b2cd-45ec-aa1e-b1150268512e\") " pod="openstack/dnsmasq-dns-68df85789f-29wwx" Dec 03 14:34:46 crc kubenswrapper[4805]: I1203 14:34:46.627910 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/73dd742e-b2cd-45ec-aa1e-b1150268512e-config\") pod \"dnsmasq-dns-68df85789f-29wwx\" (UID: \"73dd742e-b2cd-45ec-aa1e-b1150268512e\") " pod="openstack/dnsmasq-dns-68df85789f-29wwx" Dec 03 14:34:46 crc kubenswrapper[4805]: I1203 14:34:46.627938 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/73dd742e-b2cd-45ec-aa1e-b1150268512e-dns-swift-storage-0\") pod \"dnsmasq-dns-68df85789f-29wwx\" (UID: \"73dd742e-b2cd-45ec-aa1e-b1150268512e\") " pod="openstack/dnsmasq-dns-68df85789f-29wwx" Dec 03 14:34:46 crc kubenswrapper[4805]: I1203 14:34:46.627990 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/73dd742e-b2cd-45ec-aa1e-b1150268512e-ovsdbserver-sb\") pod \"dnsmasq-dns-68df85789f-29wwx\" (UID: \"73dd742e-b2cd-45ec-aa1e-b1150268512e\") " pod="openstack/dnsmasq-dns-68df85789f-29wwx" Dec 03 14:34:46 crc kubenswrapper[4805]: I1203 14:34:46.628033 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/73dd742e-b2cd-45ec-aa1e-b1150268512e-openstack-edpm-ipam\") pod \"dnsmasq-dns-68df85789f-29wwx\" (UID: \"73dd742e-b2cd-45ec-aa1e-b1150268512e\") " pod="openstack/dnsmasq-dns-68df85789f-29wwx" Dec 03 14:34:46 crc kubenswrapper[4805]: I1203 14:34:46.628082 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/73dd742e-b2cd-45ec-aa1e-b1150268512e-ovsdbserver-nb\") pod \"dnsmasq-dns-68df85789f-29wwx\" (UID: \"73dd742e-b2cd-45ec-aa1e-b1150268512e\") " pod="openstack/dnsmasq-dns-68df85789f-29wwx" Dec 03 14:34:46 crc kubenswrapper[4805]: I1203 14:34:46.729646 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/73dd742e-b2cd-45ec-aa1e-b1150268512e-dns-svc\") pod \"dnsmasq-dns-68df85789f-29wwx\" (UID: \"73dd742e-b2cd-45ec-aa1e-b1150268512e\") " pod="openstack/dnsmasq-dns-68df85789f-29wwx" Dec 03 14:34:46 crc kubenswrapper[4805]: I1203 14:34:46.730081 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-k4tj5\" (UniqueName: \"kubernetes.io/projected/73dd742e-b2cd-45ec-aa1e-b1150268512e-kube-api-access-k4tj5\") pod \"dnsmasq-dns-68df85789f-29wwx\" (UID: \"73dd742e-b2cd-45ec-aa1e-b1150268512e\") " pod="openstack/dnsmasq-dns-68df85789f-29wwx" Dec 03 14:34:46 crc kubenswrapper[4805]: I1203 14:34:46.730265 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/73dd742e-b2cd-45ec-aa1e-b1150268512e-config\") pod \"dnsmasq-dns-68df85789f-29wwx\" (UID: \"73dd742e-b2cd-45ec-aa1e-b1150268512e\") " pod="openstack/dnsmasq-dns-68df85789f-29wwx" Dec 03 14:34:46 crc kubenswrapper[4805]: I1203 14:34:46.730406 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/73dd742e-b2cd-45ec-aa1e-b1150268512e-dns-swift-storage-0\") pod \"dnsmasq-dns-68df85789f-29wwx\" (UID: \"73dd742e-b2cd-45ec-aa1e-b1150268512e\") " pod="openstack/dnsmasq-dns-68df85789f-29wwx" Dec 03 14:34:46 crc kubenswrapper[4805]: I1203 14:34:46.730621 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/73dd742e-b2cd-45ec-aa1e-b1150268512e-ovsdbserver-sb\") pod \"dnsmasq-dns-68df85789f-29wwx\" (UID: \"73dd742e-b2cd-45ec-aa1e-b1150268512e\") " pod="openstack/dnsmasq-dns-68df85789f-29wwx" Dec 03 14:34:46 crc kubenswrapper[4805]: I1203 14:34:46.730656 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/73dd742e-b2cd-45ec-aa1e-b1150268512e-dns-svc\") pod \"dnsmasq-dns-68df85789f-29wwx\" (UID: \"73dd742e-b2cd-45ec-aa1e-b1150268512e\") " pod="openstack/dnsmasq-dns-68df85789f-29wwx" Dec 03 14:34:46 crc kubenswrapper[4805]: I1203 14:34:46.730936 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/73dd742e-b2cd-45ec-aa1e-b1150268512e-openstack-edpm-ipam\") pod \"dnsmasq-dns-68df85789f-29wwx\" (UID: \"73dd742e-b2cd-45ec-aa1e-b1150268512e\") " pod="openstack/dnsmasq-dns-68df85789f-29wwx" Dec 03 14:34:46 crc kubenswrapper[4805]: I1203 14:34:46.731100 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/73dd742e-b2cd-45ec-aa1e-b1150268512e-ovsdbserver-nb\") pod \"dnsmasq-dns-68df85789f-29wwx\" (UID: \"73dd742e-b2cd-45ec-aa1e-b1150268512e\") " pod="openstack/dnsmasq-dns-68df85789f-29wwx" Dec 03 14:34:46 crc kubenswrapper[4805]: I1203 14:34:46.731356 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/73dd742e-b2cd-45ec-aa1e-b1150268512e-config\") pod \"dnsmasq-dns-68df85789f-29wwx\" (UID: \"73dd742e-b2cd-45ec-aa1e-b1150268512e\") " pod="openstack/dnsmasq-dns-68df85789f-29wwx" Dec 03 14:34:46 crc kubenswrapper[4805]: I1203 14:34:46.731363 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/73dd742e-b2cd-45ec-aa1e-b1150268512e-ovsdbserver-sb\") pod \"dnsmasq-dns-68df85789f-29wwx\" (UID: \"73dd742e-b2cd-45ec-aa1e-b1150268512e\") " pod="openstack/dnsmasq-dns-68df85789f-29wwx" Dec 03 14:34:46 crc kubenswrapper[4805]: I1203 14:34:46.731625 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/73dd742e-b2cd-45ec-aa1e-b1150268512e-openstack-edpm-ipam\") pod \"dnsmasq-dns-68df85789f-29wwx\" (UID: \"73dd742e-b2cd-45ec-aa1e-b1150268512e\") " pod="openstack/dnsmasq-dns-68df85789f-29wwx" Dec 03 14:34:46 crc kubenswrapper[4805]: I1203 14:34:46.731898 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/73dd742e-b2cd-45ec-aa1e-b1150268512e-ovsdbserver-nb\") pod \"dnsmasq-dns-68df85789f-29wwx\" (UID: \"73dd742e-b2cd-45ec-aa1e-b1150268512e\") " pod="openstack/dnsmasq-dns-68df85789f-29wwx" Dec 03 14:34:46 crc kubenswrapper[4805]: I1203 14:34:46.732209 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/73dd742e-b2cd-45ec-aa1e-b1150268512e-dns-swift-storage-0\") pod \"dnsmasq-dns-68df85789f-29wwx\" (UID: \"73dd742e-b2cd-45ec-aa1e-b1150268512e\") " pod="openstack/dnsmasq-dns-68df85789f-29wwx" Dec 03 14:34:46 crc kubenswrapper[4805]: I1203 14:34:46.752533 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-k4tj5\" (UniqueName: \"kubernetes.io/projected/73dd742e-b2cd-45ec-aa1e-b1150268512e-kube-api-access-k4tj5\") pod \"dnsmasq-dns-68df85789f-29wwx\" (UID: \"73dd742e-b2cd-45ec-aa1e-b1150268512e\") " pod="openstack/dnsmasq-dns-68df85789f-29wwx" Dec 03 14:34:46 crc kubenswrapper[4805]: I1203 14:34:46.778449 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-68df85789f-29wwx" Dec 03 14:34:46 crc kubenswrapper[4805]: I1203 14:34:46.866121 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"fff4a752-b61c-4c51-bc0b-428c9affbb00","Type":"ContainerStarted","Data":"418f4dce2966a5e8660971f3dbd6dec474e7dd2acbabb489ad7da5954db81bbd"} Dec 03 14:34:47 crc kubenswrapper[4805]: I1203 14:34:47.247912 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-68df85789f-29wwx"] Dec 03 14:34:47 crc kubenswrapper[4805]: I1203 14:34:47.880977 4805 generic.go:334] "Generic (PLEG): container finished" podID="73dd742e-b2cd-45ec-aa1e-b1150268512e" containerID="4c5b955c157e6526fb7001b9dcf47092a4be298cd678ca4e759ad5df83d05281" exitCode=0 Dec 03 14:34:47 crc kubenswrapper[4805]: I1203 14:34:47.881048 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-68df85789f-29wwx" event={"ID":"73dd742e-b2cd-45ec-aa1e-b1150268512e","Type":"ContainerDied","Data":"4c5b955c157e6526fb7001b9dcf47092a4be298cd678ca4e759ad5df83d05281"} Dec 03 14:34:47 crc kubenswrapper[4805]: I1203 14:34:47.882658 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-68df85789f-29wwx" event={"ID":"73dd742e-b2cd-45ec-aa1e-b1150268512e","Type":"ContainerStarted","Data":"a202f5026becbdc8c0fdbf3c0708e256750c9cbfae1d7fc60777e94f265f9fc7"} Dec 03 14:34:48 crc kubenswrapper[4805]: I1203 14:34:48.893112 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-68df85789f-29wwx" event={"ID":"73dd742e-b2cd-45ec-aa1e-b1150268512e","Type":"ContainerStarted","Data":"28e445ee9f92705b854ee480c3fca15d4410f94d6161eed351c3926a544ed1c4"} Dec 03 14:34:48 crc kubenswrapper[4805]: I1203 14:34:48.893506 4805 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-68df85789f-29wwx" Dec 03 14:34:48 crc kubenswrapper[4805]: I1203 14:34:48.922608 4805 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-68df85789f-29wwx" podStartSLOduration=2.9225890679999997 podStartE2EDuration="2.922589068s" podCreationTimestamp="2025-12-03 14:34:46 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 14:34:48.911602989 +0000 UTC m=+1518.574519912" watchObservedRunningTime="2025-12-03 14:34:48.922589068 +0000 UTC m=+1518.585505991" Dec 03 14:34:56 crc kubenswrapper[4805]: I1203 14:34:56.779818 4805 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-68df85789f-29wwx" Dec 03 14:34:56 crc kubenswrapper[4805]: I1203 14:34:56.840491 4805 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-79b5d74c8c-qmdpk"] Dec 03 14:34:56 crc kubenswrapper[4805]: I1203 14:34:56.840728 4805 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-79b5d74c8c-qmdpk" podUID="73f56cb0-f79a-4e60-ae70-f86f1595a8c9" containerName="dnsmasq-dns" containerID="cri-o://0b641002c33e06eb4a00fef3b57217534b7ea1fc136c66dcc4bf637adfa49dee" gracePeriod=10 Dec 03 14:34:56 crc kubenswrapper[4805]: I1203 14:34:56.981559 4805 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-bb85b8995-trmpv"] Dec 03 14:34:56 crc kubenswrapper[4805]: I1203 14:34:56.984223 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-bb85b8995-trmpv" Dec 03 14:34:56 crc kubenswrapper[4805]: I1203 14:34:56.984265 4805 generic.go:334] "Generic (PLEG): container finished" podID="73f56cb0-f79a-4e60-ae70-f86f1595a8c9" containerID="0b641002c33e06eb4a00fef3b57217534b7ea1fc136c66dcc4bf637adfa49dee" exitCode=0 Dec 03 14:34:56 crc kubenswrapper[4805]: I1203 14:34:56.984306 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-79b5d74c8c-qmdpk" event={"ID":"73f56cb0-f79a-4e60-ae70-f86f1595a8c9","Type":"ContainerDied","Data":"0b641002c33e06eb4a00fef3b57217534b7ea1fc136c66dcc4bf637adfa49dee"} Dec 03 14:34:57 crc kubenswrapper[4805]: I1203 14:34:57.021609 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-bb85b8995-trmpv"] Dec 03 14:34:57 crc kubenswrapper[4805]: I1203 14:34:57.121354 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5lm2z\" (UniqueName: \"kubernetes.io/projected/7d390cdb-18d5-49b3-8b24-1c2203442042-kube-api-access-5lm2z\") pod \"dnsmasq-dns-bb85b8995-trmpv\" (UID: \"7d390cdb-18d5-49b3-8b24-1c2203442042\") " pod="openstack/dnsmasq-dns-bb85b8995-trmpv" Dec 03 14:34:57 crc kubenswrapper[4805]: I1203 14:34:57.121401 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/7d390cdb-18d5-49b3-8b24-1c2203442042-ovsdbserver-nb\") pod \"dnsmasq-dns-bb85b8995-trmpv\" (UID: \"7d390cdb-18d5-49b3-8b24-1c2203442042\") " pod="openstack/dnsmasq-dns-bb85b8995-trmpv" Dec 03 14:34:57 crc kubenswrapper[4805]: I1203 14:34:57.121467 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/7d390cdb-18d5-49b3-8b24-1c2203442042-dns-swift-storage-0\") pod \"dnsmasq-dns-bb85b8995-trmpv\" (UID: \"7d390cdb-18d5-49b3-8b24-1c2203442042\") " pod="openstack/dnsmasq-dns-bb85b8995-trmpv" Dec 03 14:34:57 crc kubenswrapper[4805]: I1203 14:34:57.121530 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/7d390cdb-18d5-49b3-8b24-1c2203442042-ovsdbserver-sb\") pod \"dnsmasq-dns-bb85b8995-trmpv\" (UID: \"7d390cdb-18d5-49b3-8b24-1c2203442042\") " pod="openstack/dnsmasq-dns-bb85b8995-trmpv" Dec 03 14:34:57 crc kubenswrapper[4805]: I1203 14:34:57.121557 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/7d390cdb-18d5-49b3-8b24-1c2203442042-dns-svc\") pod \"dnsmasq-dns-bb85b8995-trmpv\" (UID: \"7d390cdb-18d5-49b3-8b24-1c2203442042\") " pod="openstack/dnsmasq-dns-bb85b8995-trmpv" Dec 03 14:34:57 crc kubenswrapper[4805]: I1203 14:34:57.121580 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/7d390cdb-18d5-49b3-8b24-1c2203442042-openstack-edpm-ipam\") pod \"dnsmasq-dns-bb85b8995-trmpv\" (UID: \"7d390cdb-18d5-49b3-8b24-1c2203442042\") " pod="openstack/dnsmasq-dns-bb85b8995-trmpv" Dec 03 14:34:57 crc kubenswrapper[4805]: I1203 14:34:57.121610 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7d390cdb-18d5-49b3-8b24-1c2203442042-config\") pod \"dnsmasq-dns-bb85b8995-trmpv\" (UID: \"7d390cdb-18d5-49b3-8b24-1c2203442042\") " pod="openstack/dnsmasq-dns-bb85b8995-trmpv" Dec 03 14:34:57 crc kubenswrapper[4805]: I1203 14:34:57.223418 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/7d390cdb-18d5-49b3-8b24-1c2203442042-openstack-edpm-ipam\") pod \"dnsmasq-dns-bb85b8995-trmpv\" (UID: \"7d390cdb-18d5-49b3-8b24-1c2203442042\") " pod="openstack/dnsmasq-dns-bb85b8995-trmpv" Dec 03 14:34:57 crc kubenswrapper[4805]: I1203 14:34:57.223497 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7d390cdb-18d5-49b3-8b24-1c2203442042-config\") pod \"dnsmasq-dns-bb85b8995-trmpv\" (UID: \"7d390cdb-18d5-49b3-8b24-1c2203442042\") " pod="openstack/dnsmasq-dns-bb85b8995-trmpv" Dec 03 14:34:57 crc kubenswrapper[4805]: I1203 14:34:57.223585 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5lm2z\" (UniqueName: \"kubernetes.io/projected/7d390cdb-18d5-49b3-8b24-1c2203442042-kube-api-access-5lm2z\") pod \"dnsmasq-dns-bb85b8995-trmpv\" (UID: \"7d390cdb-18d5-49b3-8b24-1c2203442042\") " pod="openstack/dnsmasq-dns-bb85b8995-trmpv" Dec 03 14:34:57 crc kubenswrapper[4805]: I1203 14:34:57.223612 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/7d390cdb-18d5-49b3-8b24-1c2203442042-ovsdbserver-nb\") pod \"dnsmasq-dns-bb85b8995-trmpv\" (UID: \"7d390cdb-18d5-49b3-8b24-1c2203442042\") " pod="openstack/dnsmasq-dns-bb85b8995-trmpv" Dec 03 14:34:57 crc kubenswrapper[4805]: I1203 14:34:57.223659 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/7d390cdb-18d5-49b3-8b24-1c2203442042-dns-swift-storage-0\") pod \"dnsmasq-dns-bb85b8995-trmpv\" (UID: \"7d390cdb-18d5-49b3-8b24-1c2203442042\") " pod="openstack/dnsmasq-dns-bb85b8995-trmpv" Dec 03 14:34:57 crc kubenswrapper[4805]: I1203 14:34:57.223739 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/7d390cdb-18d5-49b3-8b24-1c2203442042-ovsdbserver-sb\") pod \"dnsmasq-dns-bb85b8995-trmpv\" (UID: \"7d390cdb-18d5-49b3-8b24-1c2203442042\") " pod="openstack/dnsmasq-dns-bb85b8995-trmpv" Dec 03 14:34:57 crc kubenswrapper[4805]: I1203 14:34:57.223773 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/7d390cdb-18d5-49b3-8b24-1c2203442042-dns-svc\") pod \"dnsmasq-dns-bb85b8995-trmpv\" (UID: \"7d390cdb-18d5-49b3-8b24-1c2203442042\") " pod="openstack/dnsmasq-dns-bb85b8995-trmpv" Dec 03 14:34:57 crc kubenswrapper[4805]: I1203 14:34:57.224335 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/7d390cdb-18d5-49b3-8b24-1c2203442042-openstack-edpm-ipam\") pod \"dnsmasq-dns-bb85b8995-trmpv\" (UID: \"7d390cdb-18d5-49b3-8b24-1c2203442042\") " pod="openstack/dnsmasq-dns-bb85b8995-trmpv" Dec 03 14:34:57 crc kubenswrapper[4805]: I1203 14:34:57.224607 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/7d390cdb-18d5-49b3-8b24-1c2203442042-dns-svc\") pod \"dnsmasq-dns-bb85b8995-trmpv\" (UID: \"7d390cdb-18d5-49b3-8b24-1c2203442042\") " pod="openstack/dnsmasq-dns-bb85b8995-trmpv" Dec 03 14:34:57 crc kubenswrapper[4805]: I1203 14:34:57.224977 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/7d390cdb-18d5-49b3-8b24-1c2203442042-ovsdbserver-nb\") pod \"dnsmasq-dns-bb85b8995-trmpv\" (UID: \"7d390cdb-18d5-49b3-8b24-1c2203442042\") " pod="openstack/dnsmasq-dns-bb85b8995-trmpv" Dec 03 14:34:57 crc kubenswrapper[4805]: I1203 14:34:57.230013 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/7d390cdb-18d5-49b3-8b24-1c2203442042-ovsdbserver-sb\") pod \"dnsmasq-dns-bb85b8995-trmpv\" (UID: \"7d390cdb-18d5-49b3-8b24-1c2203442042\") " pod="openstack/dnsmasq-dns-bb85b8995-trmpv" Dec 03 14:34:57 crc kubenswrapper[4805]: I1203 14:34:57.230652 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7d390cdb-18d5-49b3-8b24-1c2203442042-config\") pod \"dnsmasq-dns-bb85b8995-trmpv\" (UID: \"7d390cdb-18d5-49b3-8b24-1c2203442042\") " pod="openstack/dnsmasq-dns-bb85b8995-trmpv" Dec 03 14:34:57 crc kubenswrapper[4805]: I1203 14:34:57.231014 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/7d390cdb-18d5-49b3-8b24-1c2203442042-dns-swift-storage-0\") pod \"dnsmasq-dns-bb85b8995-trmpv\" (UID: \"7d390cdb-18d5-49b3-8b24-1c2203442042\") " pod="openstack/dnsmasq-dns-bb85b8995-trmpv" Dec 03 14:34:57 crc kubenswrapper[4805]: I1203 14:34:57.251001 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5lm2z\" (UniqueName: \"kubernetes.io/projected/7d390cdb-18d5-49b3-8b24-1c2203442042-kube-api-access-5lm2z\") pod \"dnsmasq-dns-bb85b8995-trmpv\" (UID: \"7d390cdb-18d5-49b3-8b24-1c2203442042\") " pod="openstack/dnsmasq-dns-bb85b8995-trmpv" Dec 03 14:34:57 crc kubenswrapper[4805]: I1203 14:34:57.335339 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-bb85b8995-trmpv" Dec 03 14:34:57 crc kubenswrapper[4805]: I1203 14:34:57.482194 4805 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-79b5d74c8c-qmdpk" Dec 03 14:34:57 crc kubenswrapper[4805]: I1203 14:34:57.641921 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cbqwb\" (UniqueName: \"kubernetes.io/projected/73f56cb0-f79a-4e60-ae70-f86f1595a8c9-kube-api-access-cbqwb\") pod \"73f56cb0-f79a-4e60-ae70-f86f1595a8c9\" (UID: \"73f56cb0-f79a-4e60-ae70-f86f1595a8c9\") " Dec 03 14:34:57 crc kubenswrapper[4805]: I1203 14:34:57.642204 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/73f56cb0-f79a-4e60-ae70-f86f1595a8c9-config\") pod \"73f56cb0-f79a-4e60-ae70-f86f1595a8c9\" (UID: \"73f56cb0-f79a-4e60-ae70-f86f1595a8c9\") " Dec 03 14:34:57 crc kubenswrapper[4805]: I1203 14:34:57.642311 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/73f56cb0-f79a-4e60-ae70-f86f1595a8c9-ovsdbserver-sb\") pod \"73f56cb0-f79a-4e60-ae70-f86f1595a8c9\" (UID: \"73f56cb0-f79a-4e60-ae70-f86f1595a8c9\") " Dec 03 14:34:57 crc kubenswrapper[4805]: I1203 14:34:57.642371 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/73f56cb0-f79a-4e60-ae70-f86f1595a8c9-dns-svc\") pod \"73f56cb0-f79a-4e60-ae70-f86f1595a8c9\" (UID: \"73f56cb0-f79a-4e60-ae70-f86f1595a8c9\") " Dec 03 14:34:57 crc kubenswrapper[4805]: I1203 14:34:57.642402 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/73f56cb0-f79a-4e60-ae70-f86f1595a8c9-dns-swift-storage-0\") pod \"73f56cb0-f79a-4e60-ae70-f86f1595a8c9\" (UID: \"73f56cb0-f79a-4e60-ae70-f86f1595a8c9\") " Dec 03 14:34:57 crc kubenswrapper[4805]: I1203 14:34:57.642422 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/73f56cb0-f79a-4e60-ae70-f86f1595a8c9-ovsdbserver-nb\") pod \"73f56cb0-f79a-4e60-ae70-f86f1595a8c9\" (UID: \"73f56cb0-f79a-4e60-ae70-f86f1595a8c9\") " Dec 03 14:34:57 crc kubenswrapper[4805]: I1203 14:34:57.648120 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/73f56cb0-f79a-4e60-ae70-f86f1595a8c9-kube-api-access-cbqwb" (OuterVolumeSpecName: "kube-api-access-cbqwb") pod "73f56cb0-f79a-4e60-ae70-f86f1595a8c9" (UID: "73f56cb0-f79a-4e60-ae70-f86f1595a8c9"). InnerVolumeSpecName "kube-api-access-cbqwb". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 14:34:57 crc kubenswrapper[4805]: I1203 14:34:57.695656 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/73f56cb0-f79a-4e60-ae70-f86f1595a8c9-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "73f56cb0-f79a-4e60-ae70-f86f1595a8c9" (UID: "73f56cb0-f79a-4e60-ae70-f86f1595a8c9"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 14:34:57 crc kubenswrapper[4805]: I1203 14:34:57.698813 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/73f56cb0-f79a-4e60-ae70-f86f1595a8c9-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "73f56cb0-f79a-4e60-ae70-f86f1595a8c9" (UID: "73f56cb0-f79a-4e60-ae70-f86f1595a8c9"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 14:34:57 crc kubenswrapper[4805]: I1203 14:34:57.714069 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/73f56cb0-f79a-4e60-ae70-f86f1595a8c9-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "73f56cb0-f79a-4e60-ae70-f86f1595a8c9" (UID: "73f56cb0-f79a-4e60-ae70-f86f1595a8c9"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 14:34:57 crc kubenswrapper[4805]: I1203 14:34:57.715941 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/73f56cb0-f79a-4e60-ae70-f86f1595a8c9-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "73f56cb0-f79a-4e60-ae70-f86f1595a8c9" (UID: "73f56cb0-f79a-4e60-ae70-f86f1595a8c9"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 14:34:57 crc kubenswrapper[4805]: I1203 14:34:57.727467 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/73f56cb0-f79a-4e60-ae70-f86f1595a8c9-config" (OuterVolumeSpecName: "config") pod "73f56cb0-f79a-4e60-ae70-f86f1595a8c9" (UID: "73f56cb0-f79a-4e60-ae70-f86f1595a8c9"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 14:34:57 crc kubenswrapper[4805]: I1203 14:34:57.744577 4805 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cbqwb\" (UniqueName: \"kubernetes.io/projected/73f56cb0-f79a-4e60-ae70-f86f1595a8c9-kube-api-access-cbqwb\") on node \"crc\" DevicePath \"\"" Dec 03 14:34:57 crc kubenswrapper[4805]: I1203 14:34:57.744613 4805 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/73f56cb0-f79a-4e60-ae70-f86f1595a8c9-config\") on node \"crc\" DevicePath \"\"" Dec 03 14:34:57 crc kubenswrapper[4805]: I1203 14:34:57.744624 4805 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/73f56cb0-f79a-4e60-ae70-f86f1595a8c9-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Dec 03 14:34:57 crc kubenswrapper[4805]: I1203 14:34:57.744634 4805 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/73f56cb0-f79a-4e60-ae70-f86f1595a8c9-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 03 14:34:57 crc kubenswrapper[4805]: I1203 14:34:57.744641 4805 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/73f56cb0-f79a-4e60-ae70-f86f1595a8c9-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Dec 03 14:34:57 crc kubenswrapper[4805]: I1203 14:34:57.744649 4805 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/73f56cb0-f79a-4e60-ae70-f86f1595a8c9-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Dec 03 14:34:57 crc kubenswrapper[4805]: I1203 14:34:57.943749 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-bb85b8995-trmpv"] Dec 03 14:34:57 crc kubenswrapper[4805]: I1203 14:34:57.997908 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-bb85b8995-trmpv" event={"ID":"7d390cdb-18d5-49b3-8b24-1c2203442042","Type":"ContainerStarted","Data":"d6393821eed3f0d78fa8c16058f6764d7c196287546d173c0d482ec1ef277918"} Dec 03 14:34:58 crc kubenswrapper[4805]: I1203 14:34:58.000258 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-79b5d74c8c-qmdpk" event={"ID":"73f56cb0-f79a-4e60-ae70-f86f1595a8c9","Type":"ContainerDied","Data":"d1c6c41b5239625860972078b347eb1f56398f99e7eb6345fb420006a803c4c0"} Dec 03 14:34:58 crc kubenswrapper[4805]: I1203 14:34:58.000292 4805 scope.go:117] "RemoveContainer" containerID="0b641002c33e06eb4a00fef3b57217534b7ea1fc136c66dcc4bf637adfa49dee" Dec 03 14:34:58 crc kubenswrapper[4805]: I1203 14:34:58.000404 4805 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-79b5d74c8c-qmdpk" Dec 03 14:34:58 crc kubenswrapper[4805]: I1203 14:34:58.120997 4805 scope.go:117] "RemoveContainer" containerID="8d5a86cad745b1c5d43012a05e252adf40a5ecc7c85e407457caa86d1b0e9dc4" Dec 03 14:34:58 crc kubenswrapper[4805]: I1203 14:34:58.156133 4805 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-79b5d74c8c-qmdpk"] Dec 03 14:34:58 crc kubenswrapper[4805]: I1203 14:34:58.166668 4805 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-79b5d74c8c-qmdpk"] Dec 03 14:34:58 crc kubenswrapper[4805]: I1203 14:34:58.717306 4805 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="73f56cb0-f79a-4e60-ae70-f86f1595a8c9" path="/var/lib/kubelet/pods/73f56cb0-f79a-4e60-ae70-f86f1595a8c9/volumes" Dec 03 14:34:59 crc kubenswrapper[4805]: I1203 14:34:59.010804 4805 generic.go:334] "Generic (PLEG): container finished" podID="7d390cdb-18d5-49b3-8b24-1c2203442042" containerID="615c41b0688a95e613d3bf662043b06e6dd36567af6f11d12b3579f4ce11d8c8" exitCode=0 Dec 03 14:34:59 crc kubenswrapper[4805]: I1203 14:34:59.010987 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-bb85b8995-trmpv" event={"ID":"7d390cdb-18d5-49b3-8b24-1c2203442042","Type":"ContainerDied","Data":"615c41b0688a95e613d3bf662043b06e6dd36567af6f11d12b3579f4ce11d8c8"} Dec 03 14:35:00 crc kubenswrapper[4805]: I1203 14:35:00.027886 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-bb85b8995-trmpv" event={"ID":"7d390cdb-18d5-49b3-8b24-1c2203442042","Type":"ContainerStarted","Data":"3d535d0f2b23d66d7b81f880e2bba6c12c18a37983e93a15655d446692760ed3"} Dec 03 14:35:00 crc kubenswrapper[4805]: I1203 14:35:00.028717 4805 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-bb85b8995-trmpv" Dec 03 14:35:00 crc kubenswrapper[4805]: I1203 14:35:00.053610 4805 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-bb85b8995-trmpv" podStartSLOduration=4.053590642 podStartE2EDuration="4.053590642s" podCreationTimestamp="2025-12-03 14:34:56 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 14:35:00.051088114 +0000 UTC m=+1529.714005047" watchObservedRunningTime="2025-12-03 14:35:00.053590642 +0000 UTC m=+1529.716507565" Dec 03 14:35:07 crc kubenswrapper[4805]: I1203 14:35:07.338085 4805 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-bb85b8995-trmpv" Dec 03 14:35:07 crc kubenswrapper[4805]: I1203 14:35:07.437235 4805 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-68df85789f-29wwx"] Dec 03 14:35:07 crc kubenswrapper[4805]: I1203 14:35:07.439780 4805 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-68df85789f-29wwx" podUID="73dd742e-b2cd-45ec-aa1e-b1150268512e" containerName="dnsmasq-dns" containerID="cri-o://28e445ee9f92705b854ee480c3fca15d4410f94d6161eed351c3926a544ed1c4" gracePeriod=10 Dec 03 14:35:08 crc kubenswrapper[4805]: I1203 14:35:08.103639 4805 generic.go:334] "Generic (PLEG): container finished" podID="73dd742e-b2cd-45ec-aa1e-b1150268512e" containerID="28e445ee9f92705b854ee480c3fca15d4410f94d6161eed351c3926a544ed1c4" exitCode=0 Dec 03 14:35:08 crc kubenswrapper[4805]: I1203 14:35:08.103783 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-68df85789f-29wwx" event={"ID":"73dd742e-b2cd-45ec-aa1e-b1150268512e","Type":"ContainerDied","Data":"28e445ee9f92705b854ee480c3fca15d4410f94d6161eed351c3926a544ed1c4"} Dec 03 14:35:08 crc kubenswrapper[4805]: I1203 14:35:08.488991 4805 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-68df85789f-29wwx" Dec 03 14:35:08 crc kubenswrapper[4805]: I1203 14:35:08.555497 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/73dd742e-b2cd-45ec-aa1e-b1150268512e-openstack-edpm-ipam\") pod \"73dd742e-b2cd-45ec-aa1e-b1150268512e\" (UID: \"73dd742e-b2cd-45ec-aa1e-b1150268512e\") " Dec 03 14:35:08 crc kubenswrapper[4805]: I1203 14:35:08.555635 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/73dd742e-b2cd-45ec-aa1e-b1150268512e-ovsdbserver-sb\") pod \"73dd742e-b2cd-45ec-aa1e-b1150268512e\" (UID: \"73dd742e-b2cd-45ec-aa1e-b1150268512e\") " Dec 03 14:35:08 crc kubenswrapper[4805]: I1203 14:35:08.555670 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/73dd742e-b2cd-45ec-aa1e-b1150268512e-dns-svc\") pod \"73dd742e-b2cd-45ec-aa1e-b1150268512e\" (UID: \"73dd742e-b2cd-45ec-aa1e-b1150268512e\") " Dec 03 14:35:08 crc kubenswrapper[4805]: I1203 14:35:08.555732 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/73dd742e-b2cd-45ec-aa1e-b1150268512e-config\") pod \"73dd742e-b2cd-45ec-aa1e-b1150268512e\" (UID: \"73dd742e-b2cd-45ec-aa1e-b1150268512e\") " Dec 03 14:35:08 crc kubenswrapper[4805]: I1203 14:35:08.555768 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/73dd742e-b2cd-45ec-aa1e-b1150268512e-dns-swift-storage-0\") pod \"73dd742e-b2cd-45ec-aa1e-b1150268512e\" (UID: \"73dd742e-b2cd-45ec-aa1e-b1150268512e\") " Dec 03 14:35:08 crc kubenswrapper[4805]: I1203 14:35:08.555795 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-k4tj5\" (UniqueName: \"kubernetes.io/projected/73dd742e-b2cd-45ec-aa1e-b1150268512e-kube-api-access-k4tj5\") pod \"73dd742e-b2cd-45ec-aa1e-b1150268512e\" (UID: \"73dd742e-b2cd-45ec-aa1e-b1150268512e\") " Dec 03 14:35:08 crc kubenswrapper[4805]: I1203 14:35:08.555812 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/73dd742e-b2cd-45ec-aa1e-b1150268512e-ovsdbserver-nb\") pod \"73dd742e-b2cd-45ec-aa1e-b1150268512e\" (UID: \"73dd742e-b2cd-45ec-aa1e-b1150268512e\") " Dec 03 14:35:08 crc kubenswrapper[4805]: I1203 14:35:08.560997 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/73dd742e-b2cd-45ec-aa1e-b1150268512e-kube-api-access-k4tj5" (OuterVolumeSpecName: "kube-api-access-k4tj5") pod "73dd742e-b2cd-45ec-aa1e-b1150268512e" (UID: "73dd742e-b2cd-45ec-aa1e-b1150268512e"). InnerVolumeSpecName "kube-api-access-k4tj5". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 14:35:08 crc kubenswrapper[4805]: I1203 14:35:08.608730 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/73dd742e-b2cd-45ec-aa1e-b1150268512e-config" (OuterVolumeSpecName: "config") pod "73dd742e-b2cd-45ec-aa1e-b1150268512e" (UID: "73dd742e-b2cd-45ec-aa1e-b1150268512e"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 14:35:08 crc kubenswrapper[4805]: I1203 14:35:08.615479 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/73dd742e-b2cd-45ec-aa1e-b1150268512e-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "73dd742e-b2cd-45ec-aa1e-b1150268512e" (UID: "73dd742e-b2cd-45ec-aa1e-b1150268512e"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 14:35:08 crc kubenswrapper[4805]: I1203 14:35:08.616226 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/73dd742e-b2cd-45ec-aa1e-b1150268512e-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "73dd742e-b2cd-45ec-aa1e-b1150268512e" (UID: "73dd742e-b2cd-45ec-aa1e-b1150268512e"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 14:35:08 crc kubenswrapper[4805]: I1203 14:35:08.616793 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/73dd742e-b2cd-45ec-aa1e-b1150268512e-openstack-edpm-ipam" (OuterVolumeSpecName: "openstack-edpm-ipam") pod "73dd742e-b2cd-45ec-aa1e-b1150268512e" (UID: "73dd742e-b2cd-45ec-aa1e-b1150268512e"). InnerVolumeSpecName "openstack-edpm-ipam". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 14:35:08 crc kubenswrapper[4805]: I1203 14:35:08.617790 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/73dd742e-b2cd-45ec-aa1e-b1150268512e-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "73dd742e-b2cd-45ec-aa1e-b1150268512e" (UID: "73dd742e-b2cd-45ec-aa1e-b1150268512e"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 14:35:08 crc kubenswrapper[4805]: I1203 14:35:08.635783 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/73dd742e-b2cd-45ec-aa1e-b1150268512e-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "73dd742e-b2cd-45ec-aa1e-b1150268512e" (UID: "73dd742e-b2cd-45ec-aa1e-b1150268512e"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 14:35:08 crc kubenswrapper[4805]: I1203 14:35:08.657764 4805 reconciler_common.go:293] "Volume detached for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/73dd742e-b2cd-45ec-aa1e-b1150268512e-openstack-edpm-ipam\") on node \"crc\" DevicePath \"\"" Dec 03 14:35:08 crc kubenswrapper[4805]: I1203 14:35:08.657798 4805 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/73dd742e-b2cd-45ec-aa1e-b1150268512e-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Dec 03 14:35:08 crc kubenswrapper[4805]: I1203 14:35:08.657811 4805 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/73dd742e-b2cd-45ec-aa1e-b1150268512e-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 03 14:35:08 crc kubenswrapper[4805]: I1203 14:35:08.657822 4805 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/73dd742e-b2cd-45ec-aa1e-b1150268512e-config\") on node \"crc\" DevicePath \"\"" Dec 03 14:35:08 crc kubenswrapper[4805]: I1203 14:35:08.657830 4805 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/73dd742e-b2cd-45ec-aa1e-b1150268512e-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Dec 03 14:35:08 crc kubenswrapper[4805]: I1203 14:35:08.657852 4805 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/73dd742e-b2cd-45ec-aa1e-b1150268512e-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Dec 03 14:35:08 crc kubenswrapper[4805]: I1203 14:35:08.657863 4805 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-k4tj5\" (UniqueName: \"kubernetes.io/projected/73dd742e-b2cd-45ec-aa1e-b1150268512e-kube-api-access-k4tj5\") on node \"crc\" DevicePath \"\"" Dec 03 14:35:09 crc kubenswrapper[4805]: I1203 14:35:09.113178 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-68df85789f-29wwx" event={"ID":"73dd742e-b2cd-45ec-aa1e-b1150268512e","Type":"ContainerDied","Data":"a202f5026becbdc8c0fdbf3c0708e256750c9cbfae1d7fc60777e94f265f9fc7"} Dec 03 14:35:09 crc kubenswrapper[4805]: I1203 14:35:09.113257 4805 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-68df85789f-29wwx" Dec 03 14:35:09 crc kubenswrapper[4805]: I1203 14:35:09.113510 4805 scope.go:117] "RemoveContainer" containerID="28e445ee9f92705b854ee480c3fca15d4410f94d6161eed351c3926a544ed1c4" Dec 03 14:35:09 crc kubenswrapper[4805]: I1203 14:35:09.142046 4805 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-68df85789f-29wwx"] Dec 03 14:35:09 crc kubenswrapper[4805]: I1203 14:35:09.145178 4805 scope.go:117] "RemoveContainer" containerID="4c5b955c157e6526fb7001b9dcf47092a4be298cd678ca4e759ad5df83d05281" Dec 03 14:35:09 crc kubenswrapper[4805]: I1203 14:35:09.151568 4805 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-68df85789f-29wwx"] Dec 03 14:35:10 crc kubenswrapper[4805]: I1203 14:35:10.713894 4805 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="73dd742e-b2cd-45ec-aa1e-b1150268512e" path="/var/lib/kubelet/pods/73dd742e-b2cd-45ec-aa1e-b1150268512e/volumes" Dec 03 14:35:18 crc kubenswrapper[4805]: I1203 14:35:18.202294 4805 generic.go:334] "Generic (PLEG): container finished" podID="0664c16b-c578-4685-b1ba-98c5d5f3fe07" containerID="e300f43553ffdf9bbe5023281eeb1b30e1667ecb203fdb8aba8b39577bb5944e" exitCode=0 Dec 03 14:35:18 crc kubenswrapper[4805]: I1203 14:35:18.203309 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"0664c16b-c578-4685-b1ba-98c5d5f3fe07","Type":"ContainerDied","Data":"e300f43553ffdf9bbe5023281eeb1b30e1667ecb203fdb8aba8b39577bb5944e"} Dec 03 14:35:19 crc kubenswrapper[4805]: I1203 14:35:19.240729 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"0664c16b-c578-4685-b1ba-98c5d5f3fe07","Type":"ContainerStarted","Data":"51650324009aa1a9e1f92e3ea29fb1ba9fd023394b5194587c3062d900d7b2d4"} Dec 03 14:35:19 crc kubenswrapper[4805]: I1203 14:35:19.241507 4805 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/rabbitmq-server-0" Dec 03 14:35:19 crc kubenswrapper[4805]: I1203 14:35:19.245765 4805 generic.go:334] "Generic (PLEG): container finished" podID="fff4a752-b61c-4c51-bc0b-428c9affbb00" containerID="418f4dce2966a5e8660971f3dbd6dec474e7dd2acbabb489ad7da5954db81bbd" exitCode=0 Dec 03 14:35:19 crc kubenswrapper[4805]: I1203 14:35:19.245809 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"fff4a752-b61c-4c51-bc0b-428c9affbb00","Type":"ContainerDied","Data":"418f4dce2966a5e8660971f3dbd6dec474e7dd2acbabb489ad7da5954db81bbd"} Dec 03 14:35:19 crc kubenswrapper[4805]: I1203 14:35:19.282501 4805 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/rabbitmq-server-0" podStartSLOduration=37.282482832 podStartE2EDuration="37.282482832s" podCreationTimestamp="2025-12-03 14:34:42 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 14:35:19.263156556 +0000 UTC m=+1548.926073479" watchObservedRunningTime="2025-12-03 14:35:19.282482832 +0000 UTC m=+1548.945399755" Dec 03 14:35:20 crc kubenswrapper[4805]: I1203 14:35:20.257237 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"fff4a752-b61c-4c51-bc0b-428c9affbb00","Type":"ContainerStarted","Data":"c7e4ffea82b4369422295a7230c1085e1f171067c1532a5653914319fd8d5153"} Dec 03 14:35:20 crc kubenswrapper[4805]: I1203 14:35:20.258113 4805 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/rabbitmq-cell1-server-0" Dec 03 14:35:20 crc kubenswrapper[4805]: I1203 14:35:20.290916 4805 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/rabbitmq-cell1-server-0" podStartSLOduration=37.290874304 podStartE2EDuration="37.290874304s" podCreationTimestamp="2025-12-03 14:34:43 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 14:35:20.279185715 +0000 UTC m=+1549.942102638" watchObservedRunningTime="2025-12-03 14:35:20.290874304 +0000 UTC m=+1549.953791227" Dec 03 14:35:20 crc kubenswrapper[4805]: I1203 14:35:20.466223 4805 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-lghw6"] Dec 03 14:35:20 crc kubenswrapper[4805]: E1203 14:35:20.466633 4805 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="73f56cb0-f79a-4e60-ae70-f86f1595a8c9" containerName="init" Dec 03 14:35:20 crc kubenswrapper[4805]: I1203 14:35:20.466646 4805 state_mem.go:107] "Deleted CPUSet assignment" podUID="73f56cb0-f79a-4e60-ae70-f86f1595a8c9" containerName="init" Dec 03 14:35:20 crc kubenswrapper[4805]: E1203 14:35:20.466673 4805 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="73dd742e-b2cd-45ec-aa1e-b1150268512e" containerName="dnsmasq-dns" Dec 03 14:35:20 crc kubenswrapper[4805]: I1203 14:35:20.466679 4805 state_mem.go:107] "Deleted CPUSet assignment" podUID="73dd742e-b2cd-45ec-aa1e-b1150268512e" containerName="dnsmasq-dns" Dec 03 14:35:20 crc kubenswrapper[4805]: E1203 14:35:20.466692 4805 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="73dd742e-b2cd-45ec-aa1e-b1150268512e" containerName="init" Dec 03 14:35:20 crc kubenswrapper[4805]: I1203 14:35:20.466700 4805 state_mem.go:107] "Deleted CPUSet assignment" podUID="73dd742e-b2cd-45ec-aa1e-b1150268512e" containerName="init" Dec 03 14:35:20 crc kubenswrapper[4805]: E1203 14:35:20.466721 4805 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="73f56cb0-f79a-4e60-ae70-f86f1595a8c9" containerName="dnsmasq-dns" Dec 03 14:35:20 crc kubenswrapper[4805]: I1203 14:35:20.466727 4805 state_mem.go:107] "Deleted CPUSet assignment" podUID="73f56cb0-f79a-4e60-ae70-f86f1595a8c9" containerName="dnsmasq-dns" Dec 03 14:35:20 crc kubenswrapper[4805]: I1203 14:35:20.466951 4805 memory_manager.go:354] "RemoveStaleState removing state" podUID="73f56cb0-f79a-4e60-ae70-f86f1595a8c9" containerName="dnsmasq-dns" Dec 03 14:35:20 crc kubenswrapper[4805]: I1203 14:35:20.466976 4805 memory_manager.go:354] "RemoveStaleState removing state" podUID="73dd742e-b2cd-45ec-aa1e-b1150268512e" containerName="dnsmasq-dns" Dec 03 14:35:20 crc kubenswrapper[4805]: I1203 14:35:20.467642 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-lghw6" Dec 03 14:35:20 crc kubenswrapper[4805]: I1203 14:35:20.472301 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Dec 03 14:35:20 crc kubenswrapper[4805]: I1203 14:35:20.472515 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-n7jt4" Dec 03 14:35:20 crc kubenswrapper[4805]: I1203 14:35:20.472671 4805 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 03 14:35:20 crc kubenswrapper[4805]: I1203 14:35:20.472787 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Dec 03 14:35:20 crc kubenswrapper[4805]: I1203 14:35:20.477999 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-lghw6"] Dec 03 14:35:20 crc kubenswrapper[4805]: I1203 14:35:20.594232 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/a63daecb-effd-412b-9aeb-9400cd81d763-inventory\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-lghw6\" (UID: \"a63daecb-effd-412b-9aeb-9400cd81d763\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-lghw6" Dec 03 14:35:20 crc kubenswrapper[4805]: I1203 14:35:20.594310 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a63daecb-effd-412b-9aeb-9400cd81d763-repo-setup-combined-ca-bundle\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-lghw6\" (UID: \"a63daecb-effd-412b-9aeb-9400cd81d763\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-lghw6" Dec 03 14:35:20 crc kubenswrapper[4805]: I1203 14:35:20.594373 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lbtdr\" (UniqueName: \"kubernetes.io/projected/a63daecb-effd-412b-9aeb-9400cd81d763-kube-api-access-lbtdr\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-lghw6\" (UID: \"a63daecb-effd-412b-9aeb-9400cd81d763\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-lghw6" Dec 03 14:35:20 crc kubenswrapper[4805]: I1203 14:35:20.594584 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/a63daecb-effd-412b-9aeb-9400cd81d763-ssh-key\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-lghw6\" (UID: \"a63daecb-effd-412b-9aeb-9400cd81d763\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-lghw6" Dec 03 14:35:20 crc kubenswrapper[4805]: I1203 14:35:20.696901 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lbtdr\" (UniqueName: \"kubernetes.io/projected/a63daecb-effd-412b-9aeb-9400cd81d763-kube-api-access-lbtdr\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-lghw6\" (UID: \"a63daecb-effd-412b-9aeb-9400cd81d763\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-lghw6" Dec 03 14:35:20 crc kubenswrapper[4805]: I1203 14:35:20.697017 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/a63daecb-effd-412b-9aeb-9400cd81d763-ssh-key\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-lghw6\" (UID: \"a63daecb-effd-412b-9aeb-9400cd81d763\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-lghw6" Dec 03 14:35:20 crc kubenswrapper[4805]: I1203 14:35:20.697147 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/a63daecb-effd-412b-9aeb-9400cd81d763-inventory\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-lghw6\" (UID: \"a63daecb-effd-412b-9aeb-9400cd81d763\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-lghw6" Dec 03 14:35:20 crc kubenswrapper[4805]: I1203 14:35:20.697187 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a63daecb-effd-412b-9aeb-9400cd81d763-repo-setup-combined-ca-bundle\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-lghw6\" (UID: \"a63daecb-effd-412b-9aeb-9400cd81d763\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-lghw6" Dec 03 14:35:20 crc kubenswrapper[4805]: I1203 14:35:20.704976 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/a63daecb-effd-412b-9aeb-9400cd81d763-ssh-key\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-lghw6\" (UID: \"a63daecb-effd-412b-9aeb-9400cd81d763\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-lghw6" Dec 03 14:35:20 crc kubenswrapper[4805]: I1203 14:35:20.711468 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/a63daecb-effd-412b-9aeb-9400cd81d763-inventory\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-lghw6\" (UID: \"a63daecb-effd-412b-9aeb-9400cd81d763\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-lghw6" Dec 03 14:35:20 crc kubenswrapper[4805]: I1203 14:35:20.713020 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a63daecb-effd-412b-9aeb-9400cd81d763-repo-setup-combined-ca-bundle\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-lghw6\" (UID: \"a63daecb-effd-412b-9aeb-9400cd81d763\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-lghw6" Dec 03 14:35:20 crc kubenswrapper[4805]: I1203 14:35:20.721671 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lbtdr\" (UniqueName: \"kubernetes.io/projected/a63daecb-effd-412b-9aeb-9400cd81d763-kube-api-access-lbtdr\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-lghw6\" (UID: \"a63daecb-effd-412b-9aeb-9400cd81d763\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-lghw6" Dec 03 14:35:20 crc kubenswrapper[4805]: I1203 14:35:20.783690 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-lghw6" Dec 03 14:35:21 crc kubenswrapper[4805]: I1203 14:35:21.336308 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-lghw6"] Dec 03 14:35:21 crc kubenswrapper[4805]: W1203 14:35:21.341499 4805 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poda63daecb_effd_412b_9aeb_9400cd81d763.slice/crio-d56e76a2fb77453eac983ca3e8a1ee8955235508dcf90e54707a7434f75bd019 WatchSource:0}: Error finding container d56e76a2fb77453eac983ca3e8a1ee8955235508dcf90e54707a7434f75bd019: Status 404 returned error can't find the container with id d56e76a2fb77453eac983ca3e8a1ee8955235508dcf90e54707a7434f75bd019 Dec 03 14:35:22 crc kubenswrapper[4805]: I1203 14:35:22.278613 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-lghw6" event={"ID":"a63daecb-effd-412b-9aeb-9400cd81d763","Type":"ContainerStarted","Data":"d56e76a2fb77453eac983ca3e8a1ee8955235508dcf90e54707a7434f75bd019"} Dec 03 14:35:32 crc kubenswrapper[4805]: I1203 14:35:32.639866 4805 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/rabbitmq-server-0" podUID="0664c16b-c578-4685-b1ba-98c5d5f3fe07" containerName="rabbitmq" probeResult="failure" output="dial tcp 10.217.0.214:5671: connect: connection refused" Dec 03 14:35:32 crc kubenswrapper[4805]: I1203 14:35:32.877200 4805 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 03 14:35:33 crc kubenswrapper[4805]: I1203 14:35:33.386584 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-lghw6" event={"ID":"a63daecb-effd-412b-9aeb-9400cd81d763","Type":"ContainerStarted","Data":"d2e03afef0dcbb733fd2ef1df5b8c88dff92fa198c8dcbb43ef5c23e75606087"} Dec 03 14:35:33 crc kubenswrapper[4805]: I1203 14:35:33.406319 4805 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-lghw6" podStartSLOduration=1.875545127 podStartE2EDuration="13.406300548s" podCreationTimestamp="2025-12-03 14:35:20 +0000 UTC" firstStartedPulling="2025-12-03 14:35:21.34313707 +0000 UTC m=+1551.006053983" lastFinishedPulling="2025-12-03 14:35:32.873892471 +0000 UTC m=+1562.536809404" observedRunningTime="2025-12-03 14:35:33.401217891 +0000 UTC m=+1563.064134824" watchObservedRunningTime="2025-12-03 14:35:33.406300548 +0000 UTC m=+1563.069217471" Dec 03 14:35:34 crc kubenswrapper[4805]: I1203 14:35:34.268074 4805 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/rabbitmq-cell1-server-0" Dec 03 14:35:42 crc kubenswrapper[4805]: I1203 14:35:42.192629 4805 scope.go:117] "RemoveContainer" containerID="4aea414289608c29dbb3bf5bd2ab7f74e16d7e4dcfadde7ffd3d078e77fba603" Dec 03 14:35:42 crc kubenswrapper[4805]: I1203 14:35:42.639078 4805 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/rabbitmq-server-0" Dec 03 14:35:45 crc kubenswrapper[4805]: I1203 14:35:45.507095 4805 generic.go:334] "Generic (PLEG): container finished" podID="a63daecb-effd-412b-9aeb-9400cd81d763" containerID="d2e03afef0dcbb733fd2ef1df5b8c88dff92fa198c8dcbb43ef5c23e75606087" exitCode=0 Dec 03 14:35:45 crc kubenswrapper[4805]: I1203 14:35:45.507263 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-lghw6" event={"ID":"a63daecb-effd-412b-9aeb-9400cd81d763","Type":"ContainerDied","Data":"d2e03afef0dcbb733fd2ef1df5b8c88dff92fa198c8dcbb43ef5c23e75606087"} Dec 03 14:35:46 crc kubenswrapper[4805]: I1203 14:35:46.971061 4805 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-lghw6" Dec 03 14:35:47 crc kubenswrapper[4805]: I1203 14:35:47.134943 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/a63daecb-effd-412b-9aeb-9400cd81d763-ssh-key\") pod \"a63daecb-effd-412b-9aeb-9400cd81d763\" (UID: \"a63daecb-effd-412b-9aeb-9400cd81d763\") " Dec 03 14:35:47 crc kubenswrapper[4805]: I1203 14:35:47.135069 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a63daecb-effd-412b-9aeb-9400cd81d763-repo-setup-combined-ca-bundle\") pod \"a63daecb-effd-412b-9aeb-9400cd81d763\" (UID: \"a63daecb-effd-412b-9aeb-9400cd81d763\") " Dec 03 14:35:47 crc kubenswrapper[4805]: I1203 14:35:47.135219 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/a63daecb-effd-412b-9aeb-9400cd81d763-inventory\") pod \"a63daecb-effd-412b-9aeb-9400cd81d763\" (UID: \"a63daecb-effd-412b-9aeb-9400cd81d763\") " Dec 03 14:35:47 crc kubenswrapper[4805]: I1203 14:35:47.135291 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lbtdr\" (UniqueName: \"kubernetes.io/projected/a63daecb-effd-412b-9aeb-9400cd81d763-kube-api-access-lbtdr\") pod \"a63daecb-effd-412b-9aeb-9400cd81d763\" (UID: \"a63daecb-effd-412b-9aeb-9400cd81d763\") " Dec 03 14:35:47 crc kubenswrapper[4805]: I1203 14:35:47.141198 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a63daecb-effd-412b-9aeb-9400cd81d763-kube-api-access-lbtdr" (OuterVolumeSpecName: "kube-api-access-lbtdr") pod "a63daecb-effd-412b-9aeb-9400cd81d763" (UID: "a63daecb-effd-412b-9aeb-9400cd81d763"). InnerVolumeSpecName "kube-api-access-lbtdr". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 14:35:47 crc kubenswrapper[4805]: I1203 14:35:47.141371 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a63daecb-effd-412b-9aeb-9400cd81d763-repo-setup-combined-ca-bundle" (OuterVolumeSpecName: "repo-setup-combined-ca-bundle") pod "a63daecb-effd-412b-9aeb-9400cd81d763" (UID: "a63daecb-effd-412b-9aeb-9400cd81d763"). InnerVolumeSpecName "repo-setup-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 14:35:47 crc kubenswrapper[4805]: I1203 14:35:47.195532 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a63daecb-effd-412b-9aeb-9400cd81d763-inventory" (OuterVolumeSpecName: "inventory") pod "a63daecb-effd-412b-9aeb-9400cd81d763" (UID: "a63daecb-effd-412b-9aeb-9400cd81d763"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 14:35:47 crc kubenswrapper[4805]: I1203 14:35:47.196185 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a63daecb-effd-412b-9aeb-9400cd81d763-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "a63daecb-effd-412b-9aeb-9400cd81d763" (UID: "a63daecb-effd-412b-9aeb-9400cd81d763"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 14:35:47 crc kubenswrapper[4805]: I1203 14:35:47.237463 4805 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lbtdr\" (UniqueName: \"kubernetes.io/projected/a63daecb-effd-412b-9aeb-9400cd81d763-kube-api-access-lbtdr\") on node \"crc\" DevicePath \"\"" Dec 03 14:35:47 crc kubenswrapper[4805]: I1203 14:35:47.237533 4805 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/a63daecb-effd-412b-9aeb-9400cd81d763-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 03 14:35:47 crc kubenswrapper[4805]: I1203 14:35:47.237552 4805 reconciler_common.go:293] "Volume detached for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a63daecb-effd-412b-9aeb-9400cd81d763-repo-setup-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 14:35:47 crc kubenswrapper[4805]: I1203 14:35:47.237570 4805 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/a63daecb-effd-412b-9aeb-9400cd81d763-inventory\") on node \"crc\" DevicePath \"\"" Dec 03 14:35:47 crc kubenswrapper[4805]: I1203 14:35:47.536522 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-lghw6" event={"ID":"a63daecb-effd-412b-9aeb-9400cd81d763","Type":"ContainerDied","Data":"d56e76a2fb77453eac983ca3e8a1ee8955235508dcf90e54707a7434f75bd019"} Dec 03 14:35:47 crc kubenswrapper[4805]: I1203 14:35:47.536916 4805 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="d56e76a2fb77453eac983ca3e8a1ee8955235508dcf90e54707a7434f75bd019" Dec 03 14:35:47 crc kubenswrapper[4805]: I1203 14:35:47.536654 4805 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-lghw6" Dec 03 14:35:47 crc kubenswrapper[4805]: I1203 14:35:47.622812 4805 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/redhat-edpm-deployment-openstack-edpm-ipam-ltzhs"] Dec 03 14:35:47 crc kubenswrapper[4805]: E1203 14:35:47.623435 4805 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a63daecb-effd-412b-9aeb-9400cd81d763" containerName="repo-setup-edpm-deployment-openstack-edpm-ipam" Dec 03 14:35:47 crc kubenswrapper[4805]: I1203 14:35:47.623473 4805 state_mem.go:107] "Deleted CPUSet assignment" podUID="a63daecb-effd-412b-9aeb-9400cd81d763" containerName="repo-setup-edpm-deployment-openstack-edpm-ipam" Dec 03 14:35:47 crc kubenswrapper[4805]: I1203 14:35:47.624190 4805 memory_manager.go:354] "RemoveStaleState removing state" podUID="a63daecb-effd-412b-9aeb-9400cd81d763" containerName="repo-setup-edpm-deployment-openstack-edpm-ipam" Dec 03 14:35:47 crc kubenswrapper[4805]: I1203 14:35:47.625238 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-ltzhs" Dec 03 14:35:47 crc kubenswrapper[4805]: I1203 14:35:47.627341 4805 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 03 14:35:47 crc kubenswrapper[4805]: I1203 14:35:47.628294 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Dec 03 14:35:47 crc kubenswrapper[4805]: I1203 14:35:47.628426 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Dec 03 14:35:47 crc kubenswrapper[4805]: I1203 14:35:47.629092 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-n7jt4" Dec 03 14:35:47 crc kubenswrapper[4805]: I1203 14:35:47.639053 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/redhat-edpm-deployment-openstack-edpm-ipam-ltzhs"] Dec 03 14:35:47 crc kubenswrapper[4805]: I1203 14:35:47.646507 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/30551b1d-dce7-4c0f-85c9-300f32bf0fc5-ssh-key\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-ltzhs\" (UID: \"30551b1d-dce7-4c0f-85c9-300f32bf0fc5\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-ltzhs" Dec 03 14:35:47 crc kubenswrapper[4805]: I1203 14:35:47.646596 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/30551b1d-dce7-4c0f-85c9-300f32bf0fc5-inventory\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-ltzhs\" (UID: \"30551b1d-dce7-4c0f-85c9-300f32bf0fc5\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-ltzhs" Dec 03 14:35:47 crc kubenswrapper[4805]: I1203 14:35:47.646950 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ltcnf\" (UniqueName: \"kubernetes.io/projected/30551b1d-dce7-4c0f-85c9-300f32bf0fc5-kube-api-access-ltcnf\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-ltzhs\" (UID: \"30551b1d-dce7-4c0f-85c9-300f32bf0fc5\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-ltzhs" Dec 03 14:35:47 crc kubenswrapper[4805]: I1203 14:35:47.749803 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/30551b1d-dce7-4c0f-85c9-300f32bf0fc5-ssh-key\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-ltzhs\" (UID: \"30551b1d-dce7-4c0f-85c9-300f32bf0fc5\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-ltzhs" Dec 03 14:35:47 crc kubenswrapper[4805]: I1203 14:35:47.749918 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/30551b1d-dce7-4c0f-85c9-300f32bf0fc5-inventory\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-ltzhs\" (UID: \"30551b1d-dce7-4c0f-85c9-300f32bf0fc5\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-ltzhs" Dec 03 14:35:47 crc kubenswrapper[4805]: I1203 14:35:47.749981 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ltcnf\" (UniqueName: \"kubernetes.io/projected/30551b1d-dce7-4c0f-85c9-300f32bf0fc5-kube-api-access-ltcnf\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-ltzhs\" (UID: \"30551b1d-dce7-4c0f-85c9-300f32bf0fc5\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-ltzhs" Dec 03 14:35:47 crc kubenswrapper[4805]: I1203 14:35:47.757082 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/30551b1d-dce7-4c0f-85c9-300f32bf0fc5-inventory\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-ltzhs\" (UID: \"30551b1d-dce7-4c0f-85c9-300f32bf0fc5\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-ltzhs" Dec 03 14:35:47 crc kubenswrapper[4805]: I1203 14:35:47.757075 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/30551b1d-dce7-4c0f-85c9-300f32bf0fc5-ssh-key\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-ltzhs\" (UID: \"30551b1d-dce7-4c0f-85c9-300f32bf0fc5\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-ltzhs" Dec 03 14:35:47 crc kubenswrapper[4805]: I1203 14:35:47.771827 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ltcnf\" (UniqueName: \"kubernetes.io/projected/30551b1d-dce7-4c0f-85c9-300f32bf0fc5-kube-api-access-ltcnf\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-ltzhs\" (UID: \"30551b1d-dce7-4c0f-85c9-300f32bf0fc5\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-ltzhs" Dec 03 14:35:47 crc kubenswrapper[4805]: I1203 14:35:47.951120 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-ltzhs" Dec 03 14:35:48 crc kubenswrapper[4805]: I1203 14:35:48.587303 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/redhat-edpm-deployment-openstack-edpm-ipam-ltzhs"] Dec 03 14:35:48 crc kubenswrapper[4805]: W1203 14:35:48.590414 4805 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod30551b1d_dce7_4c0f_85c9_300f32bf0fc5.slice/crio-834d0e9031fe452e4d534f423ce3d28da75deb3caefcf36e3f35ef293aa93a8f WatchSource:0}: Error finding container 834d0e9031fe452e4d534f423ce3d28da75deb3caefcf36e3f35ef293aa93a8f: Status 404 returned error can't find the container with id 834d0e9031fe452e4d534f423ce3d28da75deb3caefcf36e3f35ef293aa93a8f Dec 03 14:35:49 crc kubenswrapper[4805]: I1203 14:35:49.555373 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-ltzhs" event={"ID":"30551b1d-dce7-4c0f-85c9-300f32bf0fc5","Type":"ContainerStarted","Data":"765b768434ffca4d93900550f8f42176fb35ca8bffe1e4e08dc6f8d18828e5bf"} Dec 03 14:35:49 crc kubenswrapper[4805]: I1203 14:35:49.555826 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-ltzhs" event={"ID":"30551b1d-dce7-4c0f-85c9-300f32bf0fc5","Type":"ContainerStarted","Data":"834d0e9031fe452e4d534f423ce3d28da75deb3caefcf36e3f35ef293aa93a8f"} Dec 03 14:35:49 crc kubenswrapper[4805]: I1203 14:35:49.582926 4805 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-ltzhs" podStartSLOduration=2.14332507 podStartE2EDuration="2.582890206s" podCreationTimestamp="2025-12-03 14:35:47 +0000 UTC" firstStartedPulling="2025-12-03 14:35:48.592947317 +0000 UTC m=+1578.255864240" lastFinishedPulling="2025-12-03 14:35:49.032512453 +0000 UTC m=+1578.695429376" observedRunningTime="2025-12-03 14:35:49.575541537 +0000 UTC m=+1579.238458460" watchObservedRunningTime="2025-12-03 14:35:49.582890206 +0000 UTC m=+1579.245807129" Dec 03 14:35:52 crc kubenswrapper[4805]: I1203 14:35:52.583299 4805 generic.go:334] "Generic (PLEG): container finished" podID="30551b1d-dce7-4c0f-85c9-300f32bf0fc5" containerID="765b768434ffca4d93900550f8f42176fb35ca8bffe1e4e08dc6f8d18828e5bf" exitCode=0 Dec 03 14:35:52 crc kubenswrapper[4805]: I1203 14:35:52.583380 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-ltzhs" event={"ID":"30551b1d-dce7-4c0f-85c9-300f32bf0fc5","Type":"ContainerDied","Data":"765b768434ffca4d93900550f8f42176fb35ca8bffe1e4e08dc6f8d18828e5bf"} Dec 03 14:35:54 crc kubenswrapper[4805]: I1203 14:35:54.003293 4805 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-ltzhs" Dec 03 14:35:54 crc kubenswrapper[4805]: I1203 14:35:54.179008 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ltcnf\" (UniqueName: \"kubernetes.io/projected/30551b1d-dce7-4c0f-85c9-300f32bf0fc5-kube-api-access-ltcnf\") pod \"30551b1d-dce7-4c0f-85c9-300f32bf0fc5\" (UID: \"30551b1d-dce7-4c0f-85c9-300f32bf0fc5\") " Dec 03 14:35:54 crc kubenswrapper[4805]: I1203 14:35:54.179162 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/30551b1d-dce7-4c0f-85c9-300f32bf0fc5-inventory\") pod \"30551b1d-dce7-4c0f-85c9-300f32bf0fc5\" (UID: \"30551b1d-dce7-4c0f-85c9-300f32bf0fc5\") " Dec 03 14:35:54 crc kubenswrapper[4805]: I1203 14:35:54.179259 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/30551b1d-dce7-4c0f-85c9-300f32bf0fc5-ssh-key\") pod \"30551b1d-dce7-4c0f-85c9-300f32bf0fc5\" (UID: \"30551b1d-dce7-4c0f-85c9-300f32bf0fc5\") " Dec 03 14:35:54 crc kubenswrapper[4805]: I1203 14:35:54.185022 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/30551b1d-dce7-4c0f-85c9-300f32bf0fc5-kube-api-access-ltcnf" (OuterVolumeSpecName: "kube-api-access-ltcnf") pod "30551b1d-dce7-4c0f-85c9-300f32bf0fc5" (UID: "30551b1d-dce7-4c0f-85c9-300f32bf0fc5"). InnerVolumeSpecName "kube-api-access-ltcnf". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 14:35:54 crc kubenswrapper[4805]: I1203 14:35:54.207533 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/30551b1d-dce7-4c0f-85c9-300f32bf0fc5-inventory" (OuterVolumeSpecName: "inventory") pod "30551b1d-dce7-4c0f-85c9-300f32bf0fc5" (UID: "30551b1d-dce7-4c0f-85c9-300f32bf0fc5"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 14:35:54 crc kubenswrapper[4805]: I1203 14:35:54.229818 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/30551b1d-dce7-4c0f-85c9-300f32bf0fc5-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "30551b1d-dce7-4c0f-85c9-300f32bf0fc5" (UID: "30551b1d-dce7-4c0f-85c9-300f32bf0fc5"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 14:35:54 crc kubenswrapper[4805]: I1203 14:35:54.281405 4805 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ltcnf\" (UniqueName: \"kubernetes.io/projected/30551b1d-dce7-4c0f-85c9-300f32bf0fc5-kube-api-access-ltcnf\") on node \"crc\" DevicePath \"\"" Dec 03 14:35:54 crc kubenswrapper[4805]: I1203 14:35:54.281450 4805 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/30551b1d-dce7-4c0f-85c9-300f32bf0fc5-inventory\") on node \"crc\" DevicePath \"\"" Dec 03 14:35:54 crc kubenswrapper[4805]: I1203 14:35:54.281465 4805 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/30551b1d-dce7-4c0f-85c9-300f32bf0fc5-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 03 14:35:54 crc kubenswrapper[4805]: I1203 14:35:54.603547 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-ltzhs" event={"ID":"30551b1d-dce7-4c0f-85c9-300f32bf0fc5","Type":"ContainerDied","Data":"834d0e9031fe452e4d534f423ce3d28da75deb3caefcf36e3f35ef293aa93a8f"} Dec 03 14:35:54 crc kubenswrapper[4805]: I1203 14:35:54.603635 4805 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-ltzhs" Dec 03 14:35:54 crc kubenswrapper[4805]: I1203 14:35:54.603646 4805 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="834d0e9031fe452e4d534f423ce3d28da75deb3caefcf36e3f35ef293aa93a8f" Dec 03 14:35:54 crc kubenswrapper[4805]: I1203 14:35:54.718990 4805 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-brwqg"] Dec 03 14:35:54 crc kubenswrapper[4805]: E1203 14:35:54.719320 4805 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="30551b1d-dce7-4c0f-85c9-300f32bf0fc5" containerName="redhat-edpm-deployment-openstack-edpm-ipam" Dec 03 14:35:54 crc kubenswrapper[4805]: I1203 14:35:54.719339 4805 state_mem.go:107] "Deleted CPUSet assignment" podUID="30551b1d-dce7-4c0f-85c9-300f32bf0fc5" containerName="redhat-edpm-deployment-openstack-edpm-ipam" Dec 03 14:35:54 crc kubenswrapper[4805]: I1203 14:35:54.719534 4805 memory_manager.go:354] "RemoveStaleState removing state" podUID="30551b1d-dce7-4c0f-85c9-300f32bf0fc5" containerName="redhat-edpm-deployment-openstack-edpm-ipam" Dec 03 14:35:54 crc kubenswrapper[4805]: I1203 14:35:54.720135 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-brwqg"] Dec 03 14:35:54 crc kubenswrapper[4805]: I1203 14:35:54.720264 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-brwqg" Dec 03 14:35:54 crc kubenswrapper[4805]: I1203 14:35:54.723573 4805 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 03 14:35:54 crc kubenswrapper[4805]: I1203 14:35:54.723811 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Dec 03 14:35:54 crc kubenswrapper[4805]: I1203 14:35:54.724073 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Dec 03 14:35:54 crc kubenswrapper[4805]: I1203 14:35:54.726475 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-n7jt4" Dec 03 14:35:54 crc kubenswrapper[4805]: I1203 14:35:54.797806 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/7dde2b60-d63b-4989-b76d-ec802af8f951-inventory\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-brwqg\" (UID: \"7dde2b60-d63b-4989-b76d-ec802af8f951\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-brwqg" Dec 03 14:35:54 crc kubenswrapper[4805]: I1203 14:35:54.797953 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/7dde2b60-d63b-4989-b76d-ec802af8f951-ssh-key\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-brwqg\" (UID: \"7dde2b60-d63b-4989-b76d-ec802af8f951\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-brwqg" Dec 03 14:35:54 crc kubenswrapper[4805]: I1203 14:35:54.797981 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lpvgf\" (UniqueName: \"kubernetes.io/projected/7dde2b60-d63b-4989-b76d-ec802af8f951-kube-api-access-lpvgf\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-brwqg\" (UID: \"7dde2b60-d63b-4989-b76d-ec802af8f951\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-brwqg" Dec 03 14:35:54 crc kubenswrapper[4805]: I1203 14:35:54.798022 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7dde2b60-d63b-4989-b76d-ec802af8f951-bootstrap-combined-ca-bundle\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-brwqg\" (UID: \"7dde2b60-d63b-4989-b76d-ec802af8f951\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-brwqg" Dec 03 14:35:54 crc kubenswrapper[4805]: I1203 14:35:54.899885 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/7dde2b60-d63b-4989-b76d-ec802af8f951-inventory\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-brwqg\" (UID: \"7dde2b60-d63b-4989-b76d-ec802af8f951\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-brwqg" Dec 03 14:35:54 crc kubenswrapper[4805]: I1203 14:35:54.900351 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/7dde2b60-d63b-4989-b76d-ec802af8f951-ssh-key\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-brwqg\" (UID: \"7dde2b60-d63b-4989-b76d-ec802af8f951\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-brwqg" Dec 03 14:35:54 crc kubenswrapper[4805]: I1203 14:35:54.900404 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lpvgf\" (UniqueName: \"kubernetes.io/projected/7dde2b60-d63b-4989-b76d-ec802af8f951-kube-api-access-lpvgf\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-brwqg\" (UID: \"7dde2b60-d63b-4989-b76d-ec802af8f951\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-brwqg" Dec 03 14:35:54 crc kubenswrapper[4805]: I1203 14:35:54.900460 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7dde2b60-d63b-4989-b76d-ec802af8f951-bootstrap-combined-ca-bundle\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-brwqg\" (UID: \"7dde2b60-d63b-4989-b76d-ec802af8f951\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-brwqg" Dec 03 14:35:54 crc kubenswrapper[4805]: I1203 14:35:54.905379 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/7dde2b60-d63b-4989-b76d-ec802af8f951-inventory\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-brwqg\" (UID: \"7dde2b60-d63b-4989-b76d-ec802af8f951\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-brwqg" Dec 03 14:35:54 crc kubenswrapper[4805]: I1203 14:35:54.906862 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/7dde2b60-d63b-4989-b76d-ec802af8f951-ssh-key\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-brwqg\" (UID: \"7dde2b60-d63b-4989-b76d-ec802af8f951\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-brwqg" Dec 03 14:35:54 crc kubenswrapper[4805]: I1203 14:35:54.907834 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7dde2b60-d63b-4989-b76d-ec802af8f951-bootstrap-combined-ca-bundle\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-brwqg\" (UID: \"7dde2b60-d63b-4989-b76d-ec802af8f951\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-brwqg" Dec 03 14:35:54 crc kubenswrapper[4805]: I1203 14:35:54.919821 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lpvgf\" (UniqueName: \"kubernetes.io/projected/7dde2b60-d63b-4989-b76d-ec802af8f951-kube-api-access-lpvgf\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-brwqg\" (UID: \"7dde2b60-d63b-4989-b76d-ec802af8f951\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-brwqg" Dec 03 14:35:55 crc kubenswrapper[4805]: I1203 14:35:55.036018 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-brwqg" Dec 03 14:35:56 crc kubenswrapper[4805]: W1203 14:35:55.564794 4805 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod7dde2b60_d63b_4989_b76d_ec802af8f951.slice/crio-3a7a60c546ec99e6ecd6b9cf68e1d5e56df25c09fbcc022dc0009e39abccbf5a WatchSource:0}: Error finding container 3a7a60c546ec99e6ecd6b9cf68e1d5e56df25c09fbcc022dc0009e39abccbf5a: Status 404 returned error can't find the container with id 3a7a60c546ec99e6ecd6b9cf68e1d5e56df25c09fbcc022dc0009e39abccbf5a Dec 03 14:35:56 crc kubenswrapper[4805]: I1203 14:35:55.568424 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-brwqg"] Dec 03 14:35:56 crc kubenswrapper[4805]: I1203 14:35:55.612819 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-brwqg" event={"ID":"7dde2b60-d63b-4989-b76d-ec802af8f951","Type":"ContainerStarted","Data":"3a7a60c546ec99e6ecd6b9cf68e1d5e56df25c09fbcc022dc0009e39abccbf5a"} Dec 03 14:35:56 crc kubenswrapper[4805]: I1203 14:35:56.627421 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-brwqg" event={"ID":"7dde2b60-d63b-4989-b76d-ec802af8f951","Type":"ContainerStarted","Data":"5d66200447b24991900225ca5ac22b1639afa2a7d2ce2cb67a537fd0865becbe"} Dec 03 14:35:56 crc kubenswrapper[4805]: I1203 14:35:56.648651 4805 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-brwqg" podStartSLOduration=2.222378783 podStartE2EDuration="2.648624329s" podCreationTimestamp="2025-12-03 14:35:54 +0000 UTC" firstStartedPulling="2025-12-03 14:35:55.571536574 +0000 UTC m=+1585.234453497" lastFinishedPulling="2025-12-03 14:35:55.99778212 +0000 UTC m=+1585.660699043" observedRunningTime="2025-12-03 14:35:56.64533662 +0000 UTC m=+1586.308253543" watchObservedRunningTime="2025-12-03 14:35:56.648624329 +0000 UTC m=+1586.311541252" Dec 03 14:36:42 crc kubenswrapper[4805]: I1203 14:36:42.463750 4805 scope.go:117] "RemoveContainer" containerID="182cee8becfec97e798fcb0944c192824cf5b87960b0a17801440f1a7101b6df" Dec 03 14:36:42 crc kubenswrapper[4805]: I1203 14:36:42.489383 4805 scope.go:117] "RemoveContainer" containerID="794a23f04151f6d0d061c884ab2a4ed2e4ba2f4abbcb5f8271c84b63773854dc" Dec 03 14:36:42 crc kubenswrapper[4805]: I1203 14:36:42.580011 4805 scope.go:117] "RemoveContainer" containerID="58d07340f3e9637400fe3855db1f26dfcf2ec01f9006fa34bbfe5a88846c9ff2" Dec 03 14:36:42 crc kubenswrapper[4805]: I1203 14:36:42.609821 4805 scope.go:117] "RemoveContainer" containerID="30eba44f20563e0ae1a1e3b18ccb56d1b633b85b21a348ad72c59383ee93daef" Dec 03 14:36:43 crc kubenswrapper[4805]: I1203 14:36:43.917475 4805 patch_prober.go:28] interesting pod/machine-config-daemon-gskh4 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 03 14:36:43 crc kubenswrapper[4805]: I1203 14:36:43.917928 4805 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-gskh4" podUID="6098937f-e3f6-45e8-a647-4994a79cd711" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 03 14:37:13 crc kubenswrapper[4805]: I1203 14:37:13.917596 4805 patch_prober.go:28] interesting pod/machine-config-daemon-gskh4 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 03 14:37:13 crc kubenswrapper[4805]: I1203 14:37:13.918184 4805 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-gskh4" podUID="6098937f-e3f6-45e8-a647-4994a79cd711" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 03 14:37:43 crc kubenswrapper[4805]: I1203 14:37:43.917672 4805 patch_prober.go:28] interesting pod/machine-config-daemon-gskh4 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 03 14:37:43 crc kubenswrapper[4805]: I1203 14:37:43.918187 4805 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-gskh4" podUID="6098937f-e3f6-45e8-a647-4994a79cd711" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 03 14:37:43 crc kubenswrapper[4805]: I1203 14:37:43.918231 4805 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-gskh4" Dec 03 14:37:43 crc kubenswrapper[4805]: I1203 14:37:43.919007 4805 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"748e11b7efe02ff9a381aeee0aa4ddad32272e1564bc028d779c442a7b7f2149"} pod="openshift-machine-config-operator/machine-config-daemon-gskh4" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 03 14:37:43 crc kubenswrapper[4805]: I1203 14:37:43.919072 4805 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-gskh4" podUID="6098937f-e3f6-45e8-a647-4994a79cd711" containerName="machine-config-daemon" containerID="cri-o://748e11b7efe02ff9a381aeee0aa4ddad32272e1564bc028d779c442a7b7f2149" gracePeriod=600 Dec 03 14:37:44 crc kubenswrapper[4805]: E1203 14:37:44.323259 4805 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-gskh4_openshift-machine-config-operator(6098937f-e3f6-45e8-a647-4994a79cd711)\"" pod="openshift-machine-config-operator/machine-config-daemon-gskh4" podUID="6098937f-e3f6-45e8-a647-4994a79cd711" Dec 03 14:37:44 crc kubenswrapper[4805]: I1203 14:37:44.611802 4805 generic.go:334] "Generic (PLEG): container finished" podID="6098937f-e3f6-45e8-a647-4994a79cd711" containerID="748e11b7efe02ff9a381aeee0aa4ddad32272e1564bc028d779c442a7b7f2149" exitCode=0 Dec 03 14:37:44 crc kubenswrapper[4805]: I1203 14:37:44.611875 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-gskh4" event={"ID":"6098937f-e3f6-45e8-a647-4994a79cd711","Type":"ContainerDied","Data":"748e11b7efe02ff9a381aeee0aa4ddad32272e1564bc028d779c442a7b7f2149"} Dec 03 14:37:44 crc kubenswrapper[4805]: I1203 14:37:44.611948 4805 scope.go:117] "RemoveContainer" containerID="a8c52725eecea0a7c49a5c7b85e82a2cd4d97ad1c78cf047a931df09ab5530e8" Dec 03 14:37:44 crc kubenswrapper[4805]: I1203 14:37:44.613335 4805 scope.go:117] "RemoveContainer" containerID="748e11b7efe02ff9a381aeee0aa4ddad32272e1564bc028d779c442a7b7f2149" Dec 03 14:37:44 crc kubenswrapper[4805]: E1203 14:37:44.613609 4805 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-gskh4_openshift-machine-config-operator(6098937f-e3f6-45e8-a647-4994a79cd711)\"" pod="openshift-machine-config-operator/machine-config-daemon-gskh4" podUID="6098937f-e3f6-45e8-a647-4994a79cd711" Dec 03 14:37:56 crc kubenswrapper[4805]: I1203 14:37:56.694756 4805 scope.go:117] "RemoveContainer" containerID="748e11b7efe02ff9a381aeee0aa4ddad32272e1564bc028d779c442a7b7f2149" Dec 03 14:37:56 crc kubenswrapper[4805]: E1203 14:37:56.695666 4805 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-gskh4_openshift-machine-config-operator(6098937f-e3f6-45e8-a647-4994a79cd711)\"" pod="openshift-machine-config-operator/machine-config-daemon-gskh4" podUID="6098937f-e3f6-45e8-a647-4994a79cd711" Dec 03 14:38:11 crc kubenswrapper[4805]: I1203 14:38:11.696187 4805 scope.go:117] "RemoveContainer" containerID="748e11b7efe02ff9a381aeee0aa4ddad32272e1564bc028d779c442a7b7f2149" Dec 03 14:38:11 crc kubenswrapper[4805]: E1203 14:38:11.697345 4805 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-gskh4_openshift-machine-config-operator(6098937f-e3f6-45e8-a647-4994a79cd711)\"" pod="openshift-machine-config-operator/machine-config-daemon-gskh4" podUID="6098937f-e3f6-45e8-a647-4994a79cd711" Dec 03 14:38:25 crc kubenswrapper[4805]: I1203 14:38:25.694944 4805 scope.go:117] "RemoveContainer" containerID="748e11b7efe02ff9a381aeee0aa4ddad32272e1564bc028d779c442a7b7f2149" Dec 03 14:38:25 crc kubenswrapper[4805]: E1203 14:38:25.695664 4805 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-gskh4_openshift-machine-config-operator(6098937f-e3f6-45e8-a647-4994a79cd711)\"" pod="openshift-machine-config-operator/machine-config-daemon-gskh4" podUID="6098937f-e3f6-45e8-a647-4994a79cd711" Dec 03 14:38:38 crc kubenswrapper[4805]: I1203 14:38:38.694809 4805 scope.go:117] "RemoveContainer" containerID="748e11b7efe02ff9a381aeee0aa4ddad32272e1564bc028d779c442a7b7f2149" Dec 03 14:38:38 crc kubenswrapper[4805]: E1203 14:38:38.696113 4805 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-gskh4_openshift-machine-config-operator(6098937f-e3f6-45e8-a647-4994a79cd711)\"" pod="openshift-machine-config-operator/machine-config-daemon-gskh4" podUID="6098937f-e3f6-45e8-a647-4994a79cd711" Dec 03 14:38:50 crc kubenswrapper[4805]: I1203 14:38:50.703167 4805 scope.go:117] "RemoveContainer" containerID="748e11b7efe02ff9a381aeee0aa4ddad32272e1564bc028d779c442a7b7f2149" Dec 03 14:38:50 crc kubenswrapper[4805]: E1203 14:38:50.703949 4805 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-gskh4_openshift-machine-config-operator(6098937f-e3f6-45e8-a647-4994a79cd711)\"" pod="openshift-machine-config-operator/machine-config-daemon-gskh4" podUID="6098937f-e3f6-45e8-a647-4994a79cd711" Dec 03 14:39:04 crc kubenswrapper[4805]: I1203 14:39:04.696290 4805 scope.go:117] "RemoveContainer" containerID="748e11b7efe02ff9a381aeee0aa4ddad32272e1564bc028d779c442a7b7f2149" Dec 03 14:39:04 crc kubenswrapper[4805]: E1203 14:39:04.697161 4805 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-gskh4_openshift-machine-config-operator(6098937f-e3f6-45e8-a647-4994a79cd711)\"" pod="openshift-machine-config-operator/machine-config-daemon-gskh4" podUID="6098937f-e3f6-45e8-a647-4994a79cd711" Dec 03 14:39:15 crc kubenswrapper[4805]: I1203 14:39:15.518264 4805 generic.go:334] "Generic (PLEG): container finished" podID="7dde2b60-d63b-4989-b76d-ec802af8f951" containerID="5d66200447b24991900225ca5ac22b1639afa2a7d2ce2cb67a537fd0865becbe" exitCode=0 Dec 03 14:39:15 crc kubenswrapper[4805]: I1203 14:39:15.518609 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-brwqg" event={"ID":"7dde2b60-d63b-4989-b76d-ec802af8f951","Type":"ContainerDied","Data":"5d66200447b24991900225ca5ac22b1639afa2a7d2ce2cb67a537fd0865becbe"} Dec 03 14:39:16 crc kubenswrapper[4805]: I1203 14:39:16.947151 4805 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-brwqg" Dec 03 14:39:17 crc kubenswrapper[4805]: I1203 14:39:17.043999 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lpvgf\" (UniqueName: \"kubernetes.io/projected/7dde2b60-d63b-4989-b76d-ec802af8f951-kube-api-access-lpvgf\") pod \"7dde2b60-d63b-4989-b76d-ec802af8f951\" (UID: \"7dde2b60-d63b-4989-b76d-ec802af8f951\") " Dec 03 14:39:17 crc kubenswrapper[4805]: I1203 14:39:17.044052 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7dde2b60-d63b-4989-b76d-ec802af8f951-bootstrap-combined-ca-bundle\") pod \"7dde2b60-d63b-4989-b76d-ec802af8f951\" (UID: \"7dde2b60-d63b-4989-b76d-ec802af8f951\") " Dec 03 14:39:17 crc kubenswrapper[4805]: I1203 14:39:17.044240 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/7dde2b60-d63b-4989-b76d-ec802af8f951-ssh-key\") pod \"7dde2b60-d63b-4989-b76d-ec802af8f951\" (UID: \"7dde2b60-d63b-4989-b76d-ec802af8f951\") " Dec 03 14:39:17 crc kubenswrapper[4805]: I1203 14:39:17.044297 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/7dde2b60-d63b-4989-b76d-ec802af8f951-inventory\") pod \"7dde2b60-d63b-4989-b76d-ec802af8f951\" (UID: \"7dde2b60-d63b-4989-b76d-ec802af8f951\") " Dec 03 14:39:17 crc kubenswrapper[4805]: I1203 14:39:17.050692 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7dde2b60-d63b-4989-b76d-ec802af8f951-bootstrap-combined-ca-bundle" (OuterVolumeSpecName: "bootstrap-combined-ca-bundle") pod "7dde2b60-d63b-4989-b76d-ec802af8f951" (UID: "7dde2b60-d63b-4989-b76d-ec802af8f951"). InnerVolumeSpecName "bootstrap-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 14:39:17 crc kubenswrapper[4805]: I1203 14:39:17.051152 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7dde2b60-d63b-4989-b76d-ec802af8f951-kube-api-access-lpvgf" (OuterVolumeSpecName: "kube-api-access-lpvgf") pod "7dde2b60-d63b-4989-b76d-ec802af8f951" (UID: "7dde2b60-d63b-4989-b76d-ec802af8f951"). InnerVolumeSpecName "kube-api-access-lpvgf". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 14:39:17 crc kubenswrapper[4805]: I1203 14:39:17.074230 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7dde2b60-d63b-4989-b76d-ec802af8f951-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "7dde2b60-d63b-4989-b76d-ec802af8f951" (UID: "7dde2b60-d63b-4989-b76d-ec802af8f951"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 14:39:17 crc kubenswrapper[4805]: I1203 14:39:17.075886 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7dde2b60-d63b-4989-b76d-ec802af8f951-inventory" (OuterVolumeSpecName: "inventory") pod "7dde2b60-d63b-4989-b76d-ec802af8f951" (UID: "7dde2b60-d63b-4989-b76d-ec802af8f951"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 14:39:17 crc kubenswrapper[4805]: I1203 14:39:17.146567 4805 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/7dde2b60-d63b-4989-b76d-ec802af8f951-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 03 14:39:17 crc kubenswrapper[4805]: I1203 14:39:17.146607 4805 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/7dde2b60-d63b-4989-b76d-ec802af8f951-inventory\") on node \"crc\" DevicePath \"\"" Dec 03 14:39:17 crc kubenswrapper[4805]: I1203 14:39:17.146659 4805 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lpvgf\" (UniqueName: \"kubernetes.io/projected/7dde2b60-d63b-4989-b76d-ec802af8f951-kube-api-access-lpvgf\") on node \"crc\" DevicePath \"\"" Dec 03 14:39:17 crc kubenswrapper[4805]: I1203 14:39:17.146671 4805 reconciler_common.go:293] "Volume detached for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7dde2b60-d63b-4989-b76d-ec802af8f951-bootstrap-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 14:39:17 crc kubenswrapper[4805]: I1203 14:39:17.536814 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-brwqg" event={"ID":"7dde2b60-d63b-4989-b76d-ec802af8f951","Type":"ContainerDied","Data":"3a7a60c546ec99e6ecd6b9cf68e1d5e56df25c09fbcc022dc0009e39abccbf5a"} Dec 03 14:39:17 crc kubenswrapper[4805]: I1203 14:39:17.536941 4805 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="3a7a60c546ec99e6ecd6b9cf68e1d5e56df25c09fbcc022dc0009e39abccbf5a" Dec 03 14:39:17 crc kubenswrapper[4805]: I1203 14:39:17.536862 4805 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-brwqg" Dec 03 14:39:17 crc kubenswrapper[4805]: I1203 14:39:17.619882 4805 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/download-cache-edpm-deployment-openstack-edpm-ipam-nvbww"] Dec 03 14:39:17 crc kubenswrapper[4805]: E1203 14:39:17.620383 4805 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7dde2b60-d63b-4989-b76d-ec802af8f951" containerName="bootstrap-edpm-deployment-openstack-edpm-ipam" Dec 03 14:39:17 crc kubenswrapper[4805]: I1203 14:39:17.620407 4805 state_mem.go:107] "Deleted CPUSet assignment" podUID="7dde2b60-d63b-4989-b76d-ec802af8f951" containerName="bootstrap-edpm-deployment-openstack-edpm-ipam" Dec 03 14:39:17 crc kubenswrapper[4805]: I1203 14:39:17.620624 4805 memory_manager.go:354] "RemoveStaleState removing state" podUID="7dde2b60-d63b-4989-b76d-ec802af8f951" containerName="bootstrap-edpm-deployment-openstack-edpm-ipam" Dec 03 14:39:17 crc kubenswrapper[4805]: I1203 14:39:17.621420 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-nvbww" Dec 03 14:39:17 crc kubenswrapper[4805]: I1203 14:39:17.623334 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-n7jt4" Dec 03 14:39:17 crc kubenswrapper[4805]: I1203 14:39:17.623787 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Dec 03 14:39:17 crc kubenswrapper[4805]: I1203 14:39:17.624110 4805 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 03 14:39:17 crc kubenswrapper[4805]: I1203 14:39:17.628974 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/download-cache-edpm-deployment-openstack-edpm-ipam-nvbww"] Dec 03 14:39:17 crc kubenswrapper[4805]: I1203 14:39:17.630257 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Dec 03 14:39:17 crc kubenswrapper[4805]: I1203 14:39:17.656180 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/a6f87a69-c7bb-44d8-9f65-82b1b4f8885a-inventory\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-nvbww\" (UID: \"a6f87a69-c7bb-44d8-9f65-82b1b4f8885a\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-nvbww" Dec 03 14:39:17 crc kubenswrapper[4805]: I1203 14:39:17.656230 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/a6f87a69-c7bb-44d8-9f65-82b1b4f8885a-ssh-key\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-nvbww\" (UID: \"a6f87a69-c7bb-44d8-9f65-82b1b4f8885a\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-nvbww" Dec 03 14:39:17 crc kubenswrapper[4805]: I1203 14:39:17.656326 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6slmn\" (UniqueName: \"kubernetes.io/projected/a6f87a69-c7bb-44d8-9f65-82b1b4f8885a-kube-api-access-6slmn\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-nvbww\" (UID: \"a6f87a69-c7bb-44d8-9f65-82b1b4f8885a\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-nvbww" Dec 03 14:39:17 crc kubenswrapper[4805]: I1203 14:39:17.694638 4805 scope.go:117] "RemoveContainer" containerID="748e11b7efe02ff9a381aeee0aa4ddad32272e1564bc028d779c442a7b7f2149" Dec 03 14:39:17 crc kubenswrapper[4805]: E1203 14:39:17.694895 4805 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-gskh4_openshift-machine-config-operator(6098937f-e3f6-45e8-a647-4994a79cd711)\"" pod="openshift-machine-config-operator/machine-config-daemon-gskh4" podUID="6098937f-e3f6-45e8-a647-4994a79cd711" Dec 03 14:39:17 crc kubenswrapper[4805]: I1203 14:39:17.757244 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6slmn\" (UniqueName: \"kubernetes.io/projected/a6f87a69-c7bb-44d8-9f65-82b1b4f8885a-kube-api-access-6slmn\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-nvbww\" (UID: \"a6f87a69-c7bb-44d8-9f65-82b1b4f8885a\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-nvbww" Dec 03 14:39:17 crc kubenswrapper[4805]: I1203 14:39:17.757360 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/a6f87a69-c7bb-44d8-9f65-82b1b4f8885a-inventory\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-nvbww\" (UID: \"a6f87a69-c7bb-44d8-9f65-82b1b4f8885a\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-nvbww" Dec 03 14:39:17 crc kubenswrapper[4805]: I1203 14:39:17.757400 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/a6f87a69-c7bb-44d8-9f65-82b1b4f8885a-ssh-key\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-nvbww\" (UID: \"a6f87a69-c7bb-44d8-9f65-82b1b4f8885a\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-nvbww" Dec 03 14:39:17 crc kubenswrapper[4805]: I1203 14:39:17.764048 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/a6f87a69-c7bb-44d8-9f65-82b1b4f8885a-ssh-key\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-nvbww\" (UID: \"a6f87a69-c7bb-44d8-9f65-82b1b4f8885a\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-nvbww" Dec 03 14:39:17 crc kubenswrapper[4805]: I1203 14:39:17.766445 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/a6f87a69-c7bb-44d8-9f65-82b1b4f8885a-inventory\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-nvbww\" (UID: \"a6f87a69-c7bb-44d8-9f65-82b1b4f8885a\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-nvbww" Dec 03 14:39:17 crc kubenswrapper[4805]: I1203 14:39:17.775981 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6slmn\" (UniqueName: \"kubernetes.io/projected/a6f87a69-c7bb-44d8-9f65-82b1b4f8885a-kube-api-access-6slmn\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-nvbww\" (UID: \"a6f87a69-c7bb-44d8-9f65-82b1b4f8885a\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-nvbww" Dec 03 14:39:17 crc kubenswrapper[4805]: I1203 14:39:17.946877 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-nvbww" Dec 03 14:39:18 crc kubenswrapper[4805]: I1203 14:39:18.469729 4805 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 03 14:39:18 crc kubenswrapper[4805]: I1203 14:39:18.474159 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/download-cache-edpm-deployment-openstack-edpm-ipam-nvbww"] Dec 03 14:39:18 crc kubenswrapper[4805]: I1203 14:39:18.545151 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-nvbww" event={"ID":"a6f87a69-c7bb-44d8-9f65-82b1b4f8885a","Type":"ContainerStarted","Data":"637c3576b583cfbdac29322815e9fba8a83f7bf9e8780480c5579adb46293846"} Dec 03 14:39:19 crc kubenswrapper[4805]: I1203 14:39:19.044422 4805 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-32eb-account-create-update-jmj7x"] Dec 03 14:39:19 crc kubenswrapper[4805]: I1203 14:39:19.053498 4805 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/placement-db-create-rrmbh"] Dec 03 14:39:19 crc kubenswrapper[4805]: I1203 14:39:19.063160 4805 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-db-create-bbs5s"] Dec 03 14:39:19 crc kubenswrapper[4805]: I1203 14:39:19.072211 4805 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-32eb-account-create-update-jmj7x"] Dec 03 14:39:19 crc kubenswrapper[4805]: I1203 14:39:19.081038 4805 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/placement-db-create-rrmbh"] Dec 03 14:39:19 crc kubenswrapper[4805]: I1203 14:39:19.088579 4805 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-db-create-bbs5s"] Dec 03 14:39:20 crc kubenswrapper[4805]: I1203 14:39:20.043222 4805 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/placement-7094-account-create-update-c7dm6"] Dec 03 14:39:20 crc kubenswrapper[4805]: I1203 14:39:20.052634 4805 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/placement-7094-account-create-update-c7dm6"] Dec 03 14:39:20 crc kubenswrapper[4805]: I1203 14:39:20.598261 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-nvbww" event={"ID":"a6f87a69-c7bb-44d8-9f65-82b1b4f8885a","Type":"ContainerStarted","Data":"1bcab277331e150c29b177bede8829495722413ffb111196309be06365d4d29a"} Dec 03 14:39:20 crc kubenswrapper[4805]: I1203 14:39:20.626341 4805 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-nvbww" podStartSLOduration=2.872382513 podStartE2EDuration="3.62631157s" podCreationTimestamp="2025-12-03 14:39:17 +0000 UTC" firstStartedPulling="2025-12-03 14:39:18.469498707 +0000 UTC m=+1788.132415630" lastFinishedPulling="2025-12-03 14:39:19.223427764 +0000 UTC m=+1788.886344687" observedRunningTime="2025-12-03 14:39:20.613874973 +0000 UTC m=+1790.276791896" watchObservedRunningTime="2025-12-03 14:39:20.62631157 +0000 UTC m=+1790.289228493" Dec 03 14:39:20 crc kubenswrapper[4805]: I1203 14:39:20.710115 4805 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="26a760ac-b299-4492-ab01-95db5e14ac97" path="/var/lib/kubelet/pods/26a760ac-b299-4492-ab01-95db5e14ac97/volumes" Dec 03 14:39:20 crc kubenswrapper[4805]: I1203 14:39:20.711166 4805 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4c56e040-47cb-4e7e-86d4-613e09cc3243" path="/var/lib/kubelet/pods/4c56e040-47cb-4e7e-86d4-613e09cc3243/volumes" Dec 03 14:39:20 crc kubenswrapper[4805]: I1203 14:39:20.711765 4805 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7895c06e-d5d7-4e5b-9f56-442cabf914bc" path="/var/lib/kubelet/pods/7895c06e-d5d7-4e5b-9f56-442cabf914bc/volumes" Dec 03 14:39:20 crc kubenswrapper[4805]: I1203 14:39:20.712349 4805 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a43c2a2a-291b-4422-9c06-cb448b722758" path="/var/lib/kubelet/pods/a43c2a2a-291b-4422-9c06-cb448b722758/volumes" Dec 03 14:39:21 crc kubenswrapper[4805]: I1203 14:39:21.045607 4805 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-de85-account-create-update-8gf95"] Dec 03 14:39:21 crc kubenswrapper[4805]: I1203 14:39:21.054074 4805 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-db-create-tbhgg"] Dec 03 14:39:21 crc kubenswrapper[4805]: I1203 14:39:21.061671 4805 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-db-create-tbhgg"] Dec 03 14:39:21 crc kubenswrapper[4805]: I1203 14:39:21.069047 4805 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-de85-account-create-update-8gf95"] Dec 03 14:39:22 crc kubenswrapper[4805]: I1203 14:39:22.704101 4805 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3e5d8271-48f9-4bc1-b35c-a37f919eefdd" path="/var/lib/kubelet/pods/3e5d8271-48f9-4bc1-b35c-a37f919eefdd/volumes" Dec 03 14:39:22 crc kubenswrapper[4805]: I1203 14:39:22.704689 4805 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8427fb09-203a-4093-807c-9f03061b6e07" path="/var/lib/kubelet/pods/8427fb09-203a-4093-807c-9f03061b6e07/volumes" Dec 03 14:39:28 crc kubenswrapper[4805]: I1203 14:39:28.696783 4805 scope.go:117] "RemoveContainer" containerID="748e11b7efe02ff9a381aeee0aa4ddad32272e1564bc028d779c442a7b7f2149" Dec 03 14:39:28 crc kubenswrapper[4805]: E1203 14:39:28.702232 4805 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-gskh4_openshift-machine-config-operator(6098937f-e3f6-45e8-a647-4994a79cd711)\"" pod="openshift-machine-config-operator/machine-config-daemon-gskh4" podUID="6098937f-e3f6-45e8-a647-4994a79cd711" Dec 03 14:39:39 crc kubenswrapper[4805]: I1203 14:39:39.271771 4805 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-t2xtl"] Dec 03 14:39:39 crc kubenswrapper[4805]: I1203 14:39:39.274704 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-t2xtl" Dec 03 14:39:39 crc kubenswrapper[4805]: I1203 14:39:39.284161 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-t2xtl"] Dec 03 14:39:39 crc kubenswrapper[4805]: I1203 14:39:39.379331 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6fdb2f98-919a-41db-8e05-53f7ea6f7d47-utilities\") pod \"certified-operators-t2xtl\" (UID: \"6fdb2f98-919a-41db-8e05-53f7ea6f7d47\") " pod="openshift-marketplace/certified-operators-t2xtl" Dec 03 14:39:39 crc kubenswrapper[4805]: I1203 14:39:39.379389 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6fdb2f98-919a-41db-8e05-53f7ea6f7d47-catalog-content\") pod \"certified-operators-t2xtl\" (UID: \"6fdb2f98-919a-41db-8e05-53f7ea6f7d47\") " pod="openshift-marketplace/certified-operators-t2xtl" Dec 03 14:39:39 crc kubenswrapper[4805]: I1203 14:39:39.379476 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bh9zz\" (UniqueName: \"kubernetes.io/projected/6fdb2f98-919a-41db-8e05-53f7ea6f7d47-kube-api-access-bh9zz\") pod \"certified-operators-t2xtl\" (UID: \"6fdb2f98-919a-41db-8e05-53f7ea6f7d47\") " pod="openshift-marketplace/certified-operators-t2xtl" Dec 03 14:39:39 crc kubenswrapper[4805]: I1203 14:39:39.481717 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bh9zz\" (UniqueName: \"kubernetes.io/projected/6fdb2f98-919a-41db-8e05-53f7ea6f7d47-kube-api-access-bh9zz\") pod \"certified-operators-t2xtl\" (UID: \"6fdb2f98-919a-41db-8e05-53f7ea6f7d47\") " pod="openshift-marketplace/certified-operators-t2xtl" Dec 03 14:39:39 crc kubenswrapper[4805]: I1203 14:39:39.481908 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6fdb2f98-919a-41db-8e05-53f7ea6f7d47-utilities\") pod \"certified-operators-t2xtl\" (UID: \"6fdb2f98-919a-41db-8e05-53f7ea6f7d47\") " pod="openshift-marketplace/certified-operators-t2xtl" Dec 03 14:39:39 crc kubenswrapper[4805]: I1203 14:39:39.481934 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6fdb2f98-919a-41db-8e05-53f7ea6f7d47-catalog-content\") pod \"certified-operators-t2xtl\" (UID: \"6fdb2f98-919a-41db-8e05-53f7ea6f7d47\") " pod="openshift-marketplace/certified-operators-t2xtl" Dec 03 14:39:39 crc kubenswrapper[4805]: I1203 14:39:39.482506 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6fdb2f98-919a-41db-8e05-53f7ea6f7d47-catalog-content\") pod \"certified-operators-t2xtl\" (UID: \"6fdb2f98-919a-41db-8e05-53f7ea6f7d47\") " pod="openshift-marketplace/certified-operators-t2xtl" Dec 03 14:39:39 crc kubenswrapper[4805]: I1203 14:39:39.482591 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6fdb2f98-919a-41db-8e05-53f7ea6f7d47-utilities\") pod \"certified-operators-t2xtl\" (UID: \"6fdb2f98-919a-41db-8e05-53f7ea6f7d47\") " pod="openshift-marketplace/certified-operators-t2xtl" Dec 03 14:39:39 crc kubenswrapper[4805]: I1203 14:39:39.509049 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bh9zz\" (UniqueName: \"kubernetes.io/projected/6fdb2f98-919a-41db-8e05-53f7ea6f7d47-kube-api-access-bh9zz\") pod \"certified-operators-t2xtl\" (UID: \"6fdb2f98-919a-41db-8e05-53f7ea6f7d47\") " pod="openshift-marketplace/certified-operators-t2xtl" Dec 03 14:39:39 crc kubenswrapper[4805]: I1203 14:39:39.603207 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-t2xtl" Dec 03 14:39:39 crc kubenswrapper[4805]: I1203 14:39:39.694698 4805 scope.go:117] "RemoveContainer" containerID="748e11b7efe02ff9a381aeee0aa4ddad32272e1564bc028d779c442a7b7f2149" Dec 03 14:39:39 crc kubenswrapper[4805]: E1203 14:39:39.695016 4805 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-gskh4_openshift-machine-config-operator(6098937f-e3f6-45e8-a647-4994a79cd711)\"" pod="openshift-machine-config-operator/machine-config-daemon-gskh4" podUID="6098937f-e3f6-45e8-a647-4994a79cd711" Dec 03 14:39:40 crc kubenswrapper[4805]: I1203 14:39:40.094405 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-t2xtl"] Dec 03 14:39:40 crc kubenswrapper[4805]: I1203 14:39:40.667093 4805 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-k26jl"] Dec 03 14:39:40 crc kubenswrapper[4805]: I1203 14:39:40.669723 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-k26jl" Dec 03 14:39:40 crc kubenswrapper[4805]: I1203 14:39:40.676229 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-k26jl"] Dec 03 14:39:40 crc kubenswrapper[4805]: I1203 14:39:40.810971 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/25ccfd1f-9135-4f16-974c-3ca2a748f114-utilities\") pod \"community-operators-k26jl\" (UID: \"25ccfd1f-9135-4f16-974c-3ca2a748f114\") " pod="openshift-marketplace/community-operators-k26jl" Dec 03 14:39:40 crc kubenswrapper[4805]: I1203 14:39:40.811018 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-twcsl\" (UniqueName: \"kubernetes.io/projected/25ccfd1f-9135-4f16-974c-3ca2a748f114-kube-api-access-twcsl\") pod \"community-operators-k26jl\" (UID: \"25ccfd1f-9135-4f16-974c-3ca2a748f114\") " pod="openshift-marketplace/community-operators-k26jl" Dec 03 14:39:40 crc kubenswrapper[4805]: I1203 14:39:40.811041 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/25ccfd1f-9135-4f16-974c-3ca2a748f114-catalog-content\") pod \"community-operators-k26jl\" (UID: \"25ccfd1f-9135-4f16-974c-3ca2a748f114\") " pod="openshift-marketplace/community-operators-k26jl" Dec 03 14:39:40 crc kubenswrapper[4805]: I1203 14:39:40.817681 4805 generic.go:334] "Generic (PLEG): container finished" podID="6fdb2f98-919a-41db-8e05-53f7ea6f7d47" containerID="4b5cd0864ee50b8c6e0eba9be23ee17d07f92a427c65fa45ca8f3888b44e42c3" exitCode=0 Dec 03 14:39:40 crc kubenswrapper[4805]: I1203 14:39:40.817727 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-t2xtl" event={"ID":"6fdb2f98-919a-41db-8e05-53f7ea6f7d47","Type":"ContainerDied","Data":"4b5cd0864ee50b8c6e0eba9be23ee17d07f92a427c65fa45ca8f3888b44e42c3"} Dec 03 14:39:40 crc kubenswrapper[4805]: I1203 14:39:40.817756 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-t2xtl" event={"ID":"6fdb2f98-919a-41db-8e05-53f7ea6f7d47","Type":"ContainerStarted","Data":"52210a5d1eec01b08559f4dbdd8e983174bc3d32a28e8c0b4877c1143d08de79"} Dec 03 14:39:40 crc kubenswrapper[4805]: I1203 14:39:40.913190 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/25ccfd1f-9135-4f16-974c-3ca2a748f114-utilities\") pod \"community-operators-k26jl\" (UID: \"25ccfd1f-9135-4f16-974c-3ca2a748f114\") " pod="openshift-marketplace/community-operators-k26jl" Dec 03 14:39:40 crc kubenswrapper[4805]: I1203 14:39:40.913268 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-twcsl\" (UniqueName: \"kubernetes.io/projected/25ccfd1f-9135-4f16-974c-3ca2a748f114-kube-api-access-twcsl\") pod \"community-operators-k26jl\" (UID: \"25ccfd1f-9135-4f16-974c-3ca2a748f114\") " pod="openshift-marketplace/community-operators-k26jl" Dec 03 14:39:40 crc kubenswrapper[4805]: I1203 14:39:40.913291 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/25ccfd1f-9135-4f16-974c-3ca2a748f114-catalog-content\") pod \"community-operators-k26jl\" (UID: \"25ccfd1f-9135-4f16-974c-3ca2a748f114\") " pod="openshift-marketplace/community-operators-k26jl" Dec 03 14:39:40 crc kubenswrapper[4805]: I1203 14:39:40.914462 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/25ccfd1f-9135-4f16-974c-3ca2a748f114-utilities\") pod \"community-operators-k26jl\" (UID: \"25ccfd1f-9135-4f16-974c-3ca2a748f114\") " pod="openshift-marketplace/community-operators-k26jl" Dec 03 14:39:40 crc kubenswrapper[4805]: I1203 14:39:40.914980 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/25ccfd1f-9135-4f16-974c-3ca2a748f114-catalog-content\") pod \"community-operators-k26jl\" (UID: \"25ccfd1f-9135-4f16-974c-3ca2a748f114\") " pod="openshift-marketplace/community-operators-k26jl" Dec 03 14:39:40 crc kubenswrapper[4805]: I1203 14:39:40.939535 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-twcsl\" (UniqueName: \"kubernetes.io/projected/25ccfd1f-9135-4f16-974c-3ca2a748f114-kube-api-access-twcsl\") pod \"community-operators-k26jl\" (UID: \"25ccfd1f-9135-4f16-974c-3ca2a748f114\") " pod="openshift-marketplace/community-operators-k26jl" Dec 03 14:39:40 crc kubenswrapper[4805]: I1203 14:39:40.989439 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-k26jl" Dec 03 14:39:41 crc kubenswrapper[4805]: W1203 14:39:41.526698 4805 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod25ccfd1f_9135_4f16_974c_3ca2a748f114.slice/crio-4431db9764d1778c69c65f8e5b6a8bdd16540161537c13907d6e0a73fa503016 WatchSource:0}: Error finding container 4431db9764d1778c69c65f8e5b6a8bdd16540161537c13907d6e0a73fa503016: Status 404 returned error can't find the container with id 4431db9764d1778c69c65f8e5b6a8bdd16540161537c13907d6e0a73fa503016 Dec 03 14:39:41 crc kubenswrapper[4805]: I1203 14:39:41.526939 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-k26jl"] Dec 03 14:39:41 crc kubenswrapper[4805]: I1203 14:39:41.679187 4805 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-6n94p"] Dec 03 14:39:41 crc kubenswrapper[4805]: I1203 14:39:41.681938 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-6n94p" Dec 03 14:39:41 crc kubenswrapper[4805]: I1203 14:39:41.692233 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-6n94p"] Dec 03 14:39:41 crc kubenswrapper[4805]: I1203 14:39:41.828371 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-k26jl" event={"ID":"25ccfd1f-9135-4f16-974c-3ca2a748f114","Type":"ContainerStarted","Data":"4431db9764d1778c69c65f8e5b6a8bdd16540161537c13907d6e0a73fa503016"} Dec 03 14:39:41 crc kubenswrapper[4805]: I1203 14:39:41.829881 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/52dcae87-e990-480d-886e-26cec1ee4da2-catalog-content\") pod \"redhat-operators-6n94p\" (UID: \"52dcae87-e990-480d-886e-26cec1ee4da2\") " pod="openshift-marketplace/redhat-operators-6n94p" Dec 03 14:39:41 crc kubenswrapper[4805]: I1203 14:39:41.829938 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/52dcae87-e990-480d-886e-26cec1ee4da2-utilities\") pod \"redhat-operators-6n94p\" (UID: \"52dcae87-e990-480d-886e-26cec1ee4da2\") " pod="openshift-marketplace/redhat-operators-6n94p" Dec 03 14:39:41 crc kubenswrapper[4805]: I1203 14:39:41.829958 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cb8jx\" (UniqueName: \"kubernetes.io/projected/52dcae87-e990-480d-886e-26cec1ee4da2-kube-api-access-cb8jx\") pod \"redhat-operators-6n94p\" (UID: \"52dcae87-e990-480d-886e-26cec1ee4da2\") " pod="openshift-marketplace/redhat-operators-6n94p" Dec 03 14:39:41 crc kubenswrapper[4805]: I1203 14:39:41.932058 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/52dcae87-e990-480d-886e-26cec1ee4da2-utilities\") pod \"redhat-operators-6n94p\" (UID: \"52dcae87-e990-480d-886e-26cec1ee4da2\") " pod="openshift-marketplace/redhat-operators-6n94p" Dec 03 14:39:41 crc kubenswrapper[4805]: I1203 14:39:41.932428 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cb8jx\" (UniqueName: \"kubernetes.io/projected/52dcae87-e990-480d-886e-26cec1ee4da2-kube-api-access-cb8jx\") pod \"redhat-operators-6n94p\" (UID: \"52dcae87-e990-480d-886e-26cec1ee4da2\") " pod="openshift-marketplace/redhat-operators-6n94p" Dec 03 14:39:41 crc kubenswrapper[4805]: I1203 14:39:41.932546 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/52dcae87-e990-480d-886e-26cec1ee4da2-utilities\") pod \"redhat-operators-6n94p\" (UID: \"52dcae87-e990-480d-886e-26cec1ee4da2\") " pod="openshift-marketplace/redhat-operators-6n94p" Dec 03 14:39:41 crc kubenswrapper[4805]: I1203 14:39:41.932647 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/52dcae87-e990-480d-886e-26cec1ee4da2-catalog-content\") pod \"redhat-operators-6n94p\" (UID: \"52dcae87-e990-480d-886e-26cec1ee4da2\") " pod="openshift-marketplace/redhat-operators-6n94p" Dec 03 14:39:41 crc kubenswrapper[4805]: I1203 14:39:41.932929 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/52dcae87-e990-480d-886e-26cec1ee4da2-catalog-content\") pod \"redhat-operators-6n94p\" (UID: \"52dcae87-e990-480d-886e-26cec1ee4da2\") " pod="openshift-marketplace/redhat-operators-6n94p" Dec 03 14:39:41 crc kubenswrapper[4805]: I1203 14:39:41.952970 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cb8jx\" (UniqueName: \"kubernetes.io/projected/52dcae87-e990-480d-886e-26cec1ee4da2-kube-api-access-cb8jx\") pod \"redhat-operators-6n94p\" (UID: \"52dcae87-e990-480d-886e-26cec1ee4da2\") " pod="openshift-marketplace/redhat-operators-6n94p" Dec 03 14:39:42 crc kubenswrapper[4805]: I1203 14:39:42.046287 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-6n94p" Dec 03 14:39:42 crc kubenswrapper[4805]: I1203 14:39:42.520874 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-6n94p"] Dec 03 14:39:42 crc kubenswrapper[4805]: W1203 14:39:42.538701 4805 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod52dcae87_e990_480d_886e_26cec1ee4da2.slice/crio-dfba0950e85a10783685ca65cbe4c1b02e6e2bb55ebf926c23008992e9296602 WatchSource:0}: Error finding container dfba0950e85a10783685ca65cbe4c1b02e6e2bb55ebf926c23008992e9296602: Status 404 returned error can't find the container with id dfba0950e85a10783685ca65cbe4c1b02e6e2bb55ebf926c23008992e9296602 Dec 03 14:39:42 crc kubenswrapper[4805]: I1203 14:39:42.744923 4805 scope.go:117] "RemoveContainer" containerID="3f40241eb2521d3baeb79026fd9c58f88640f341c2ed7f60275e5448221ebc87" Dec 03 14:39:42 crc kubenswrapper[4805]: I1203 14:39:42.777103 4805 scope.go:117] "RemoveContainer" containerID="802333011ddf70752382747945a287b572296e7e1e7e8075bb0350869fd7a9d8" Dec 03 14:39:42 crc kubenswrapper[4805]: I1203 14:39:42.804370 4805 scope.go:117] "RemoveContainer" containerID="45ec4e94a53ac4236a647cc33770987e9ea8a30d52b48b172a20c30c8795008e" Dec 03 14:39:42 crc kubenswrapper[4805]: I1203 14:39:42.827981 4805 scope.go:117] "RemoveContainer" containerID="cd504bae43e97694691ee6b43212fd15ddf1c14e98f67ec53f5407f32e37e20e" Dec 03 14:39:42 crc kubenswrapper[4805]: I1203 14:39:42.839699 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-6n94p" event={"ID":"52dcae87-e990-480d-886e-26cec1ee4da2","Type":"ContainerStarted","Data":"dfba0950e85a10783685ca65cbe4c1b02e6e2bb55ebf926c23008992e9296602"} Dec 03 14:39:42 crc kubenswrapper[4805]: I1203 14:39:42.841438 4805 generic.go:334] "Generic (PLEG): container finished" podID="25ccfd1f-9135-4f16-974c-3ca2a748f114" containerID="c749f07bbbebffba44171680b1ae68c8df99161fc4e73b8946a9e9b329d16a1a" exitCode=0 Dec 03 14:39:42 crc kubenswrapper[4805]: I1203 14:39:42.841546 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-k26jl" event={"ID":"25ccfd1f-9135-4f16-974c-3ca2a748f114","Type":"ContainerDied","Data":"c749f07bbbebffba44171680b1ae68c8df99161fc4e73b8946a9e9b329d16a1a"} Dec 03 14:39:42 crc kubenswrapper[4805]: I1203 14:39:42.851884 4805 scope.go:117] "RemoveContainer" containerID="7d87783d81055065efdd933bb95570fe72fbe1fbdce5e19e32bf1b28e8ae8f37" Dec 03 14:39:42 crc kubenswrapper[4805]: I1203 14:39:42.879279 4805 scope.go:117] "RemoveContainer" containerID="cac18d8710a21e00b23f7447c31eb2589e1fed3f4a88d556913bbe54a0d5d540" Dec 03 14:39:43 crc kubenswrapper[4805]: I1203 14:39:43.859982 4805 generic.go:334] "Generic (PLEG): container finished" podID="52dcae87-e990-480d-886e-26cec1ee4da2" containerID="76721e783cb35638599296cee3c853c939c1d9fc15ed4204f52e89dfe8595858" exitCode=0 Dec 03 14:39:43 crc kubenswrapper[4805]: I1203 14:39:43.860043 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-6n94p" event={"ID":"52dcae87-e990-480d-886e-26cec1ee4da2","Type":"ContainerDied","Data":"76721e783cb35638599296cee3c853c939c1d9fc15ed4204f52e89dfe8595858"} Dec 03 14:39:43 crc kubenswrapper[4805]: I1203 14:39:43.862365 4805 generic.go:334] "Generic (PLEG): container finished" podID="6fdb2f98-919a-41db-8e05-53f7ea6f7d47" containerID="7fa9951c60b2f655065c53fc9f9abfed4a3b4d560f6f4e7c50233d8d3e256879" exitCode=0 Dec 03 14:39:43 crc kubenswrapper[4805]: I1203 14:39:43.862421 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-t2xtl" event={"ID":"6fdb2f98-919a-41db-8e05-53f7ea6f7d47","Type":"ContainerDied","Data":"7fa9951c60b2f655065c53fc9f9abfed4a3b4d560f6f4e7c50233d8d3e256879"} Dec 03 14:39:47 crc kubenswrapper[4805]: I1203 14:39:47.915486 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-k26jl" event={"ID":"25ccfd1f-9135-4f16-974c-3ca2a748f114","Type":"ContainerStarted","Data":"73e87e96d612ee368c4cc4850d256ff34237c7d712bd3ea5ff03571c6dff648e"} Dec 03 14:39:48 crc kubenswrapper[4805]: I1203 14:39:48.034696 4805 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-db-sync-27jrz"] Dec 03 14:39:48 crc kubenswrapper[4805]: I1203 14:39:48.045005 4805 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-db-sync-27jrz"] Dec 03 14:39:48 crc kubenswrapper[4805]: I1203 14:39:48.731345 4805 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="285c21ec-cbb8-4cb7-b5c2-0b27656f7870" path="/var/lib/kubelet/pods/285c21ec-cbb8-4cb7-b5c2-0b27656f7870/volumes" Dec 03 14:39:48 crc kubenswrapper[4805]: I1203 14:39:48.930427 4805 generic.go:334] "Generic (PLEG): container finished" podID="25ccfd1f-9135-4f16-974c-3ca2a748f114" containerID="73e87e96d612ee368c4cc4850d256ff34237c7d712bd3ea5ff03571c6dff648e" exitCode=0 Dec 03 14:39:48 crc kubenswrapper[4805]: I1203 14:39:48.930527 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-k26jl" event={"ID":"25ccfd1f-9135-4f16-974c-3ca2a748f114","Type":"ContainerDied","Data":"73e87e96d612ee368c4cc4850d256ff34237c7d712bd3ea5ff03571c6dff648e"} Dec 03 14:39:51 crc kubenswrapper[4805]: I1203 14:39:51.694897 4805 scope.go:117] "RemoveContainer" containerID="748e11b7efe02ff9a381aeee0aa4ddad32272e1564bc028d779c442a7b7f2149" Dec 03 14:39:51 crc kubenswrapper[4805]: E1203 14:39:51.695370 4805 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-gskh4_openshift-machine-config-operator(6098937f-e3f6-45e8-a647-4994a79cd711)\"" pod="openshift-machine-config-operator/machine-config-daemon-gskh4" podUID="6098937f-e3f6-45e8-a647-4994a79cd711" Dec 03 14:39:57 crc kubenswrapper[4805]: I1203 14:39:57.003332 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-6n94p" event={"ID":"52dcae87-e990-480d-886e-26cec1ee4da2","Type":"ContainerStarted","Data":"d88dbbb436134ffc11776172134aba2f48fe52988989e95b3be0bfe335b2f185"} Dec 03 14:39:57 crc kubenswrapper[4805]: I1203 14:39:57.010552 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-t2xtl" event={"ID":"6fdb2f98-919a-41db-8e05-53f7ea6f7d47","Type":"ContainerStarted","Data":"2d42950e8cd75d00ff1e739eeefb562bc6e386075a577c1c7f42173695297ee8"} Dec 03 14:39:57 crc kubenswrapper[4805]: I1203 14:39:57.012804 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-k26jl" event={"ID":"25ccfd1f-9135-4f16-974c-3ca2a748f114","Type":"ContainerStarted","Data":"0b5b85279c008eb4d6dbc8f11cb2e28aa47725ab8517da2213b974ead341e7f6"} Dec 03 14:39:57 crc kubenswrapper[4805]: I1203 14:39:57.035784 4805 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-t2xtl" podStartSLOduration=3.5794732639999998 podStartE2EDuration="18.035767104s" podCreationTimestamp="2025-12-03 14:39:39 +0000 UTC" firstStartedPulling="2025-12-03 14:39:40.819252053 +0000 UTC m=+1810.482168976" lastFinishedPulling="2025-12-03 14:39:55.275545893 +0000 UTC m=+1824.938462816" observedRunningTime="2025-12-03 14:39:57.028752534 +0000 UTC m=+1826.691669467" watchObservedRunningTime="2025-12-03 14:39:57.035767104 +0000 UTC m=+1826.698684027" Dec 03 14:39:59 crc kubenswrapper[4805]: I1203 14:39:59.035612 4805 generic.go:334] "Generic (PLEG): container finished" podID="52dcae87-e990-480d-886e-26cec1ee4da2" containerID="d88dbbb436134ffc11776172134aba2f48fe52988989e95b3be0bfe335b2f185" exitCode=0 Dec 03 14:39:59 crc kubenswrapper[4805]: I1203 14:39:59.036079 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-6n94p" event={"ID":"52dcae87-e990-480d-886e-26cec1ee4da2","Type":"ContainerDied","Data":"d88dbbb436134ffc11776172134aba2f48fe52988989e95b3be0bfe335b2f185"} Dec 03 14:39:59 crc kubenswrapper[4805]: I1203 14:39:59.603631 4805 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-t2xtl" Dec 03 14:39:59 crc kubenswrapper[4805]: I1203 14:39:59.603685 4805 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-t2xtl" Dec 03 14:39:59 crc kubenswrapper[4805]: I1203 14:39:59.661962 4805 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-t2xtl" Dec 03 14:40:00 crc kubenswrapper[4805]: I1203 14:40:00.068266 4805 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-k26jl" podStartSLOduration=6.565865228 podStartE2EDuration="20.068245187s" podCreationTimestamp="2025-12-03 14:39:40 +0000 UTC" firstStartedPulling="2025-12-03 14:39:42.852100933 +0000 UTC m=+1812.515017856" lastFinishedPulling="2025-12-03 14:39:56.354480892 +0000 UTC m=+1826.017397815" observedRunningTime="2025-12-03 14:40:00.060738224 +0000 UTC m=+1829.723655157" watchObservedRunningTime="2025-12-03 14:40:00.068245187 +0000 UTC m=+1829.731162110" Dec 03 14:40:00 crc kubenswrapper[4805]: I1203 14:40:00.989699 4805 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-k26jl" Dec 03 14:40:00 crc kubenswrapper[4805]: I1203 14:40:00.989751 4805 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-k26jl" Dec 03 14:40:02 crc kubenswrapper[4805]: I1203 14:40:02.041186 4805 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/community-operators-k26jl" podUID="25ccfd1f-9135-4f16-974c-3ca2a748f114" containerName="registry-server" probeResult="failure" output=< Dec 03 14:40:02 crc kubenswrapper[4805]: timeout: failed to connect service ":50051" within 1s Dec 03 14:40:02 crc kubenswrapper[4805]: > Dec 03 14:40:03 crc kubenswrapper[4805]: I1203 14:40:03.044740 4805 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-db-create-q7qlk"] Dec 03 14:40:03 crc kubenswrapper[4805]: I1203 14:40:03.053224 4805 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-db-create-7bqcx"] Dec 03 14:40:03 crc kubenswrapper[4805]: I1203 14:40:03.061970 4805 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-db-create-mhn6v"] Dec 03 14:40:03 crc kubenswrapper[4805]: I1203 14:40:03.073301 4805 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-77e9-account-create-update-4mkc7"] Dec 03 14:40:03 crc kubenswrapper[4805]: I1203 14:40:03.087124 4805 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/heat-ee1c-account-create-update-g8lk4"] Dec 03 14:40:03 crc kubenswrapper[4805]: I1203 14:40:03.094906 4805 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-db-create-mhn6v"] Dec 03 14:40:03 crc kubenswrapper[4805]: I1203 14:40:03.103121 4805 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-db-create-7bqcx"] Dec 03 14:40:03 crc kubenswrapper[4805]: I1203 14:40:03.112792 4805 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-3c27-account-create-update-79pjf"] Dec 03 14:40:03 crc kubenswrapper[4805]: I1203 14:40:03.123961 4805 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-7ac1-account-create-update-6jk6w"] Dec 03 14:40:03 crc kubenswrapper[4805]: I1203 14:40:03.132963 4805 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/heat-db-create-ztcqj"] Dec 03 14:40:03 crc kubenswrapper[4805]: I1203 14:40:03.140514 4805 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-db-create-q7qlk"] Dec 03 14:40:03 crc kubenswrapper[4805]: I1203 14:40:03.147505 4805 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-3c27-account-create-update-79pjf"] Dec 03 14:40:03 crc kubenswrapper[4805]: I1203 14:40:03.154968 4805 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-77e9-account-create-update-4mkc7"] Dec 03 14:40:03 crc kubenswrapper[4805]: I1203 14:40:03.162420 4805 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-7ac1-account-create-update-6jk6w"] Dec 03 14:40:03 crc kubenswrapper[4805]: I1203 14:40:03.169481 4805 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/heat-db-create-ztcqj"] Dec 03 14:40:03 crc kubenswrapper[4805]: I1203 14:40:03.177826 4805 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/heat-ee1c-account-create-update-g8lk4"] Dec 03 14:40:04 crc kubenswrapper[4805]: I1203 14:40:04.707787 4805 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="10163a58-aa6e-4988-8b33-93f8cffa4cc0" path="/var/lib/kubelet/pods/10163a58-aa6e-4988-8b33-93f8cffa4cc0/volumes" Dec 03 14:40:04 crc kubenswrapper[4805]: I1203 14:40:04.710006 4805 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="18be72b0-25d0-41b4-8038-abd0fb38deba" path="/var/lib/kubelet/pods/18be72b0-25d0-41b4-8038-abd0fb38deba/volumes" Dec 03 14:40:04 crc kubenswrapper[4805]: I1203 14:40:04.711121 4805 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="696fec58-2abe-474b-956c-61228970cc74" path="/var/lib/kubelet/pods/696fec58-2abe-474b-956c-61228970cc74/volumes" Dec 03 14:40:04 crc kubenswrapper[4805]: I1203 14:40:04.711917 4805 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="99891296-b622-466d-95f7-bcf0e3ebe896" path="/var/lib/kubelet/pods/99891296-b622-466d-95f7-bcf0e3ebe896/volumes" Dec 03 14:40:04 crc kubenswrapper[4805]: I1203 14:40:04.713105 4805 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="aead85c7-af73-44d0-b4dc-5cd98d6439b3" path="/var/lib/kubelet/pods/aead85c7-af73-44d0-b4dc-5cd98d6439b3/volumes" Dec 03 14:40:04 crc kubenswrapper[4805]: I1203 14:40:04.713916 4805 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b27b4fde-29ff-421d-a6be-5d259bfe8016" path="/var/lib/kubelet/pods/b27b4fde-29ff-421d-a6be-5d259bfe8016/volumes" Dec 03 14:40:04 crc kubenswrapper[4805]: I1203 14:40:04.714626 4805 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d319f692-2f37-4a01-8e6c-bbf0e7b23c8e" path="/var/lib/kubelet/pods/d319f692-2f37-4a01-8e6c-bbf0e7b23c8e/volumes" Dec 03 14:40:04 crc kubenswrapper[4805]: I1203 14:40:04.715757 4805 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="eda36a78-7b3a-43f8-858e-fc6b8fdd442a" path="/var/lib/kubelet/pods/eda36a78-7b3a-43f8-858e-fc6b8fdd442a/volumes" Dec 03 14:40:05 crc kubenswrapper[4805]: I1203 14:40:05.092528 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-6n94p" event={"ID":"52dcae87-e990-480d-886e-26cec1ee4da2","Type":"ContainerStarted","Data":"61487cf96f4183128033c3a46cac125a221def87bd0acd6034da9d404a362660"} Dec 03 14:40:05 crc kubenswrapper[4805]: I1203 14:40:05.119608 4805 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-6n94p" podStartSLOduration=3.813890735 podStartE2EDuration="24.11958062s" podCreationTimestamp="2025-12-03 14:39:41 +0000 UTC" firstStartedPulling="2025-12-03 14:39:43.862030522 +0000 UTC m=+1813.524947445" lastFinishedPulling="2025-12-03 14:40:04.167720407 +0000 UTC m=+1833.830637330" observedRunningTime="2025-12-03 14:40:05.111635285 +0000 UTC m=+1834.774552208" watchObservedRunningTime="2025-12-03 14:40:05.11958062 +0000 UTC m=+1834.782497543" Dec 03 14:40:06 crc kubenswrapper[4805]: I1203 14:40:06.694626 4805 scope.go:117] "RemoveContainer" containerID="748e11b7efe02ff9a381aeee0aa4ddad32272e1564bc028d779c442a7b7f2149" Dec 03 14:40:06 crc kubenswrapper[4805]: E1203 14:40:06.695489 4805 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-gskh4_openshift-machine-config-operator(6098937f-e3f6-45e8-a647-4994a79cd711)\"" pod="openshift-machine-config-operator/machine-config-daemon-gskh4" podUID="6098937f-e3f6-45e8-a647-4994a79cd711" Dec 03 14:40:07 crc kubenswrapper[4805]: I1203 14:40:07.028991 4805 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-db-sync-75wdv"] Dec 03 14:40:07 crc kubenswrapper[4805]: I1203 14:40:07.036371 4805 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-db-sync-75wdv"] Dec 03 14:40:08 crc kubenswrapper[4805]: I1203 14:40:08.706358 4805 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c15336c7-d573-432f-8b4c-ba5a39c82cf8" path="/var/lib/kubelet/pods/c15336c7-d573-432f-8b4c-ba5a39c82cf8/volumes" Dec 03 14:40:09 crc kubenswrapper[4805]: I1203 14:40:09.654089 4805 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-t2xtl" Dec 03 14:40:09 crc kubenswrapper[4805]: I1203 14:40:09.703439 4805 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-t2xtl"] Dec 03 14:40:10 crc kubenswrapper[4805]: I1203 14:40:10.145439 4805 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-t2xtl" podUID="6fdb2f98-919a-41db-8e05-53f7ea6f7d47" containerName="registry-server" containerID="cri-o://2d42950e8cd75d00ff1e739eeefb562bc6e386075a577c1c7f42173695297ee8" gracePeriod=2 Dec 03 14:40:11 crc kubenswrapper[4805]: I1203 14:40:11.070908 4805 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-k26jl" Dec 03 14:40:11 crc kubenswrapper[4805]: I1203 14:40:11.127726 4805 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-k26jl" Dec 03 14:40:11 crc kubenswrapper[4805]: I1203 14:40:11.158490 4805 generic.go:334] "Generic (PLEG): container finished" podID="6fdb2f98-919a-41db-8e05-53f7ea6f7d47" containerID="2d42950e8cd75d00ff1e739eeefb562bc6e386075a577c1c7f42173695297ee8" exitCode=0 Dec 03 14:40:11 crc kubenswrapper[4805]: I1203 14:40:11.158541 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-t2xtl" event={"ID":"6fdb2f98-919a-41db-8e05-53f7ea6f7d47","Type":"ContainerDied","Data":"2d42950e8cd75d00ff1e739eeefb562bc6e386075a577c1c7f42173695297ee8"} Dec 03 14:40:11 crc kubenswrapper[4805]: I1203 14:40:11.264748 4805 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-t2xtl" Dec 03 14:40:11 crc kubenswrapper[4805]: I1203 14:40:11.316420 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6fdb2f98-919a-41db-8e05-53f7ea6f7d47-utilities\") pod \"6fdb2f98-919a-41db-8e05-53f7ea6f7d47\" (UID: \"6fdb2f98-919a-41db-8e05-53f7ea6f7d47\") " Dec 03 14:40:11 crc kubenswrapper[4805]: I1203 14:40:11.316561 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bh9zz\" (UniqueName: \"kubernetes.io/projected/6fdb2f98-919a-41db-8e05-53f7ea6f7d47-kube-api-access-bh9zz\") pod \"6fdb2f98-919a-41db-8e05-53f7ea6f7d47\" (UID: \"6fdb2f98-919a-41db-8e05-53f7ea6f7d47\") " Dec 03 14:40:11 crc kubenswrapper[4805]: I1203 14:40:11.316970 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6fdb2f98-919a-41db-8e05-53f7ea6f7d47-catalog-content\") pod \"6fdb2f98-919a-41db-8e05-53f7ea6f7d47\" (UID: \"6fdb2f98-919a-41db-8e05-53f7ea6f7d47\") " Dec 03 14:40:11 crc kubenswrapper[4805]: I1203 14:40:11.317033 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/6fdb2f98-919a-41db-8e05-53f7ea6f7d47-utilities" (OuterVolumeSpecName: "utilities") pod "6fdb2f98-919a-41db-8e05-53f7ea6f7d47" (UID: "6fdb2f98-919a-41db-8e05-53f7ea6f7d47"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 14:40:11 crc kubenswrapper[4805]: I1203 14:40:11.317939 4805 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6fdb2f98-919a-41db-8e05-53f7ea6f7d47-utilities\") on node \"crc\" DevicePath \"\"" Dec 03 14:40:11 crc kubenswrapper[4805]: I1203 14:40:11.328681 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6fdb2f98-919a-41db-8e05-53f7ea6f7d47-kube-api-access-bh9zz" (OuterVolumeSpecName: "kube-api-access-bh9zz") pod "6fdb2f98-919a-41db-8e05-53f7ea6f7d47" (UID: "6fdb2f98-919a-41db-8e05-53f7ea6f7d47"). InnerVolumeSpecName "kube-api-access-bh9zz". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 14:40:11 crc kubenswrapper[4805]: I1203 14:40:11.371316 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/6fdb2f98-919a-41db-8e05-53f7ea6f7d47-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "6fdb2f98-919a-41db-8e05-53f7ea6f7d47" (UID: "6fdb2f98-919a-41db-8e05-53f7ea6f7d47"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 14:40:11 crc kubenswrapper[4805]: I1203 14:40:11.418963 4805 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6fdb2f98-919a-41db-8e05-53f7ea6f7d47-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 03 14:40:11 crc kubenswrapper[4805]: I1203 14:40:11.418993 4805 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bh9zz\" (UniqueName: \"kubernetes.io/projected/6fdb2f98-919a-41db-8e05-53f7ea6f7d47-kube-api-access-bh9zz\") on node \"crc\" DevicePath \"\"" Dec 03 14:40:12 crc kubenswrapper[4805]: I1203 14:40:12.046999 4805 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-6n94p" Dec 03 14:40:12 crc kubenswrapper[4805]: I1203 14:40:12.047408 4805 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-6n94p" Dec 03 14:40:12 crc kubenswrapper[4805]: I1203 14:40:12.108326 4805 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-6n94p" Dec 03 14:40:12 crc kubenswrapper[4805]: I1203 14:40:12.170299 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-t2xtl" event={"ID":"6fdb2f98-919a-41db-8e05-53f7ea6f7d47","Type":"ContainerDied","Data":"52210a5d1eec01b08559f4dbdd8e983174bc3d32a28e8c0b4877c1143d08de79"} Dec 03 14:40:12 crc kubenswrapper[4805]: I1203 14:40:12.170334 4805 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-t2xtl" Dec 03 14:40:12 crc kubenswrapper[4805]: I1203 14:40:12.170401 4805 scope.go:117] "RemoveContainer" containerID="2d42950e8cd75d00ff1e739eeefb562bc6e386075a577c1c7f42173695297ee8" Dec 03 14:40:12 crc kubenswrapper[4805]: I1203 14:40:12.200245 4805 scope.go:117] "RemoveContainer" containerID="7fa9951c60b2f655065c53fc9f9abfed4a3b4d560f6f4e7c50233d8d3e256879" Dec 03 14:40:12 crc kubenswrapper[4805]: I1203 14:40:12.207558 4805 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-t2xtl"] Dec 03 14:40:12 crc kubenswrapper[4805]: I1203 14:40:12.216133 4805 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-t2xtl"] Dec 03 14:40:12 crc kubenswrapper[4805]: I1203 14:40:12.224416 4805 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-6n94p" Dec 03 14:40:12 crc kubenswrapper[4805]: I1203 14:40:12.239110 4805 scope.go:117] "RemoveContainer" containerID="4b5cd0864ee50b8c6e0eba9be23ee17d07f92a427c65fa45ca8f3888b44e42c3" Dec 03 14:40:12 crc kubenswrapper[4805]: I1203 14:40:12.277281 4805 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-k26jl"] Dec 03 14:40:12 crc kubenswrapper[4805]: I1203 14:40:12.277543 4805 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-k26jl" podUID="25ccfd1f-9135-4f16-974c-3ca2a748f114" containerName="registry-server" containerID="cri-o://0b5b85279c008eb4d6dbc8f11cb2e28aa47725ab8517da2213b974ead341e7f6" gracePeriod=2 Dec 03 14:40:12 crc kubenswrapper[4805]: I1203 14:40:12.708440 4805 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6fdb2f98-919a-41db-8e05-53f7ea6f7d47" path="/var/lib/kubelet/pods/6fdb2f98-919a-41db-8e05-53f7ea6f7d47/volumes" Dec 03 14:40:13 crc kubenswrapper[4805]: I1203 14:40:13.182298 4805 generic.go:334] "Generic (PLEG): container finished" podID="25ccfd1f-9135-4f16-974c-3ca2a748f114" containerID="0b5b85279c008eb4d6dbc8f11cb2e28aa47725ab8517da2213b974ead341e7f6" exitCode=0 Dec 03 14:40:13 crc kubenswrapper[4805]: I1203 14:40:13.182374 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-k26jl" event={"ID":"25ccfd1f-9135-4f16-974c-3ca2a748f114","Type":"ContainerDied","Data":"0b5b85279c008eb4d6dbc8f11cb2e28aa47725ab8517da2213b974ead341e7f6"} Dec 03 14:40:13 crc kubenswrapper[4805]: I1203 14:40:13.281999 4805 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-k26jl" Dec 03 14:40:13 crc kubenswrapper[4805]: I1203 14:40:13.363171 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/25ccfd1f-9135-4f16-974c-3ca2a748f114-utilities\") pod \"25ccfd1f-9135-4f16-974c-3ca2a748f114\" (UID: \"25ccfd1f-9135-4f16-974c-3ca2a748f114\") " Dec 03 14:40:13 crc kubenswrapper[4805]: I1203 14:40:13.363253 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/25ccfd1f-9135-4f16-974c-3ca2a748f114-catalog-content\") pod \"25ccfd1f-9135-4f16-974c-3ca2a748f114\" (UID: \"25ccfd1f-9135-4f16-974c-3ca2a748f114\") " Dec 03 14:40:13 crc kubenswrapper[4805]: I1203 14:40:13.363360 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-twcsl\" (UniqueName: \"kubernetes.io/projected/25ccfd1f-9135-4f16-974c-3ca2a748f114-kube-api-access-twcsl\") pod \"25ccfd1f-9135-4f16-974c-3ca2a748f114\" (UID: \"25ccfd1f-9135-4f16-974c-3ca2a748f114\") " Dec 03 14:40:13 crc kubenswrapper[4805]: I1203 14:40:13.363969 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/25ccfd1f-9135-4f16-974c-3ca2a748f114-utilities" (OuterVolumeSpecName: "utilities") pod "25ccfd1f-9135-4f16-974c-3ca2a748f114" (UID: "25ccfd1f-9135-4f16-974c-3ca2a748f114"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 14:40:13 crc kubenswrapper[4805]: I1203 14:40:13.370899 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/25ccfd1f-9135-4f16-974c-3ca2a748f114-kube-api-access-twcsl" (OuterVolumeSpecName: "kube-api-access-twcsl") pod "25ccfd1f-9135-4f16-974c-3ca2a748f114" (UID: "25ccfd1f-9135-4f16-974c-3ca2a748f114"). InnerVolumeSpecName "kube-api-access-twcsl". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 14:40:13 crc kubenswrapper[4805]: I1203 14:40:13.420466 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/25ccfd1f-9135-4f16-974c-3ca2a748f114-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "25ccfd1f-9135-4f16-974c-3ca2a748f114" (UID: "25ccfd1f-9135-4f16-974c-3ca2a748f114"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 14:40:13 crc kubenswrapper[4805]: I1203 14:40:13.467220 4805 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/25ccfd1f-9135-4f16-974c-3ca2a748f114-utilities\") on node \"crc\" DevicePath \"\"" Dec 03 14:40:13 crc kubenswrapper[4805]: I1203 14:40:13.467284 4805 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/25ccfd1f-9135-4f16-974c-3ca2a748f114-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 03 14:40:13 crc kubenswrapper[4805]: I1203 14:40:13.467303 4805 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-twcsl\" (UniqueName: \"kubernetes.io/projected/25ccfd1f-9135-4f16-974c-3ca2a748f114-kube-api-access-twcsl\") on node \"crc\" DevicePath \"\"" Dec 03 14:40:14 crc kubenswrapper[4805]: I1203 14:40:14.196149 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-k26jl" event={"ID":"25ccfd1f-9135-4f16-974c-3ca2a748f114","Type":"ContainerDied","Data":"4431db9764d1778c69c65f8e5b6a8bdd16540161537c13907d6e0a73fa503016"} Dec 03 14:40:14 crc kubenswrapper[4805]: I1203 14:40:14.196224 4805 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-k26jl" Dec 03 14:40:14 crc kubenswrapper[4805]: I1203 14:40:14.196236 4805 scope.go:117] "RemoveContainer" containerID="0b5b85279c008eb4d6dbc8f11cb2e28aa47725ab8517da2213b974ead341e7f6" Dec 03 14:40:14 crc kubenswrapper[4805]: I1203 14:40:14.220302 4805 scope.go:117] "RemoveContainer" containerID="73e87e96d612ee368c4cc4850d256ff34237c7d712bd3ea5ff03571c6dff648e" Dec 03 14:40:14 crc kubenswrapper[4805]: I1203 14:40:14.263453 4805 scope.go:117] "RemoveContainer" containerID="c749f07bbbebffba44171680b1ae68c8df99161fc4e73b8946a9e9b329d16a1a" Dec 03 14:40:14 crc kubenswrapper[4805]: I1203 14:40:14.263511 4805 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-k26jl"] Dec 03 14:40:14 crc kubenswrapper[4805]: I1203 14:40:14.274047 4805 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-k26jl"] Dec 03 14:40:14 crc kubenswrapper[4805]: I1203 14:40:14.675277 4805 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-6n94p"] Dec 03 14:40:14 crc kubenswrapper[4805]: I1203 14:40:14.675542 4805 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-6n94p" podUID="52dcae87-e990-480d-886e-26cec1ee4da2" containerName="registry-server" containerID="cri-o://61487cf96f4183128033c3a46cac125a221def87bd0acd6034da9d404a362660" gracePeriod=2 Dec 03 14:40:14 crc kubenswrapper[4805]: I1203 14:40:14.717880 4805 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="25ccfd1f-9135-4f16-974c-3ca2a748f114" path="/var/lib/kubelet/pods/25ccfd1f-9135-4f16-974c-3ca2a748f114/volumes" Dec 03 14:40:15 crc kubenswrapper[4805]: I1203 14:40:15.206425 4805 generic.go:334] "Generic (PLEG): container finished" podID="52dcae87-e990-480d-886e-26cec1ee4da2" containerID="61487cf96f4183128033c3a46cac125a221def87bd0acd6034da9d404a362660" exitCode=0 Dec 03 14:40:15 crc kubenswrapper[4805]: I1203 14:40:15.206480 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-6n94p" event={"ID":"52dcae87-e990-480d-886e-26cec1ee4da2","Type":"ContainerDied","Data":"61487cf96f4183128033c3a46cac125a221def87bd0acd6034da9d404a362660"} Dec 03 14:40:15 crc kubenswrapper[4805]: I1203 14:40:15.619673 4805 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-6n94p" Dec 03 14:40:15 crc kubenswrapper[4805]: I1203 14:40:15.709874 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cb8jx\" (UniqueName: \"kubernetes.io/projected/52dcae87-e990-480d-886e-26cec1ee4da2-kube-api-access-cb8jx\") pod \"52dcae87-e990-480d-886e-26cec1ee4da2\" (UID: \"52dcae87-e990-480d-886e-26cec1ee4da2\") " Dec 03 14:40:15 crc kubenswrapper[4805]: I1203 14:40:15.710089 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/52dcae87-e990-480d-886e-26cec1ee4da2-utilities\") pod \"52dcae87-e990-480d-886e-26cec1ee4da2\" (UID: \"52dcae87-e990-480d-886e-26cec1ee4da2\") " Dec 03 14:40:15 crc kubenswrapper[4805]: I1203 14:40:15.710131 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/52dcae87-e990-480d-886e-26cec1ee4da2-catalog-content\") pod \"52dcae87-e990-480d-886e-26cec1ee4da2\" (UID: \"52dcae87-e990-480d-886e-26cec1ee4da2\") " Dec 03 14:40:15 crc kubenswrapper[4805]: I1203 14:40:15.711344 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/52dcae87-e990-480d-886e-26cec1ee4da2-utilities" (OuterVolumeSpecName: "utilities") pod "52dcae87-e990-480d-886e-26cec1ee4da2" (UID: "52dcae87-e990-480d-886e-26cec1ee4da2"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 14:40:15 crc kubenswrapper[4805]: I1203 14:40:15.717257 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/52dcae87-e990-480d-886e-26cec1ee4da2-kube-api-access-cb8jx" (OuterVolumeSpecName: "kube-api-access-cb8jx") pod "52dcae87-e990-480d-886e-26cec1ee4da2" (UID: "52dcae87-e990-480d-886e-26cec1ee4da2"). InnerVolumeSpecName "kube-api-access-cb8jx". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 14:40:15 crc kubenswrapper[4805]: I1203 14:40:15.812777 4805 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cb8jx\" (UniqueName: \"kubernetes.io/projected/52dcae87-e990-480d-886e-26cec1ee4da2-kube-api-access-cb8jx\") on node \"crc\" DevicePath \"\"" Dec 03 14:40:15 crc kubenswrapper[4805]: I1203 14:40:15.813191 4805 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/52dcae87-e990-480d-886e-26cec1ee4da2-utilities\") on node \"crc\" DevicePath \"\"" Dec 03 14:40:15 crc kubenswrapper[4805]: I1203 14:40:15.840761 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/52dcae87-e990-480d-886e-26cec1ee4da2-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "52dcae87-e990-480d-886e-26cec1ee4da2" (UID: "52dcae87-e990-480d-886e-26cec1ee4da2"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 14:40:15 crc kubenswrapper[4805]: I1203 14:40:15.915110 4805 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/52dcae87-e990-480d-886e-26cec1ee4da2-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 03 14:40:16 crc kubenswrapper[4805]: I1203 14:40:16.222376 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-6n94p" event={"ID":"52dcae87-e990-480d-886e-26cec1ee4da2","Type":"ContainerDied","Data":"dfba0950e85a10783685ca65cbe4c1b02e6e2bb55ebf926c23008992e9296602"} Dec 03 14:40:16 crc kubenswrapper[4805]: I1203 14:40:16.222431 4805 scope.go:117] "RemoveContainer" containerID="61487cf96f4183128033c3a46cac125a221def87bd0acd6034da9d404a362660" Dec 03 14:40:16 crc kubenswrapper[4805]: I1203 14:40:16.222549 4805 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-6n94p" Dec 03 14:40:16 crc kubenswrapper[4805]: I1203 14:40:16.249064 4805 scope.go:117] "RemoveContainer" containerID="d88dbbb436134ffc11776172134aba2f48fe52988989e95b3be0bfe335b2f185" Dec 03 14:40:16 crc kubenswrapper[4805]: I1203 14:40:16.286019 4805 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-6n94p"] Dec 03 14:40:16 crc kubenswrapper[4805]: I1203 14:40:16.296237 4805 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-6n94p"] Dec 03 14:40:16 crc kubenswrapper[4805]: I1203 14:40:16.299829 4805 scope.go:117] "RemoveContainer" containerID="76721e783cb35638599296cee3c853c939c1d9fc15ed4204f52e89dfe8595858" Dec 03 14:40:16 crc kubenswrapper[4805]: I1203 14:40:16.710586 4805 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="52dcae87-e990-480d-886e-26cec1ee4da2" path="/var/lib/kubelet/pods/52dcae87-e990-480d-886e-26cec1ee4da2/volumes" Dec 03 14:40:18 crc kubenswrapper[4805]: I1203 14:40:18.694271 4805 scope.go:117] "RemoveContainer" containerID="748e11b7efe02ff9a381aeee0aa4ddad32272e1564bc028d779c442a7b7f2149" Dec 03 14:40:18 crc kubenswrapper[4805]: E1203 14:40:18.694777 4805 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-gskh4_openshift-machine-config-operator(6098937f-e3f6-45e8-a647-4994a79cd711)\"" pod="openshift-machine-config-operator/machine-config-daemon-gskh4" podUID="6098937f-e3f6-45e8-a647-4994a79cd711" Dec 03 14:40:30 crc kubenswrapper[4805]: I1203 14:40:30.701062 4805 scope.go:117] "RemoveContainer" containerID="748e11b7efe02ff9a381aeee0aa4ddad32272e1564bc028d779c442a7b7f2149" Dec 03 14:40:30 crc kubenswrapper[4805]: E1203 14:40:30.702100 4805 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-gskh4_openshift-machine-config-operator(6098937f-e3f6-45e8-a647-4994a79cd711)\"" pod="openshift-machine-config-operator/machine-config-daemon-gskh4" podUID="6098937f-e3f6-45e8-a647-4994a79cd711" Dec 03 14:40:42 crc kubenswrapper[4805]: I1203 14:40:42.695476 4805 scope.go:117] "RemoveContainer" containerID="748e11b7efe02ff9a381aeee0aa4ddad32272e1564bc028d779c442a7b7f2149" Dec 03 14:40:42 crc kubenswrapper[4805]: E1203 14:40:42.697456 4805 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-gskh4_openshift-machine-config-operator(6098937f-e3f6-45e8-a647-4994a79cd711)\"" pod="openshift-machine-config-operator/machine-config-daemon-gskh4" podUID="6098937f-e3f6-45e8-a647-4994a79cd711" Dec 03 14:40:43 crc kubenswrapper[4805]: I1203 14:40:43.205911 4805 scope.go:117] "RemoveContainer" containerID="9eb2ebbbba3ba07290543694db5cafa019298a402693fbd060c1c233ad4f8274" Dec 03 14:40:43 crc kubenswrapper[4805]: I1203 14:40:43.258790 4805 scope.go:117] "RemoveContainer" containerID="183d5ec2c5b2105cc148606bc4cb45cdcd4f8f29170a59cfd7ac84260c358bec" Dec 03 14:40:43 crc kubenswrapper[4805]: I1203 14:40:43.316704 4805 scope.go:117] "RemoveContainer" containerID="60d8a6dccfa76964665d7813e9a0bde148a8bb28ab2dce1beae33824fbdf6004" Dec 03 14:40:43 crc kubenswrapper[4805]: I1203 14:40:43.351804 4805 scope.go:117] "RemoveContainer" containerID="f85ba51232525929e730722508afdb428d6d624935f15d634a7f4ad1b9ec47b9" Dec 03 14:40:43 crc kubenswrapper[4805]: I1203 14:40:43.398041 4805 scope.go:117] "RemoveContainer" containerID="a71a86043818a7d764e21262d2b172946552b42e4e3f1bcaed5a22daee349963" Dec 03 14:40:43 crc kubenswrapper[4805]: I1203 14:40:43.447526 4805 scope.go:117] "RemoveContainer" containerID="f78221c81b7e774055adb1e6d9260b350822233e80e328702784f13d9d72a46d" Dec 03 14:40:43 crc kubenswrapper[4805]: I1203 14:40:43.518553 4805 scope.go:117] "RemoveContainer" containerID="1d8a116dbb3024b6dc8a70a7377f7cb462d152825caf107dd604cf92bafb4e4d" Dec 03 14:40:43 crc kubenswrapper[4805]: I1203 14:40:43.547874 4805 scope.go:117] "RemoveContainer" containerID="d881e747656aad5e3ca4bcf5fa4c04aa2b595cfd284ad8b751619771f81ad1f0" Dec 03 14:40:43 crc kubenswrapper[4805]: I1203 14:40:43.568041 4805 scope.go:117] "RemoveContainer" containerID="67b68344596f7862edfa448581c05599565990196ca30c2e7b72a99ec85c5a87" Dec 03 14:40:43 crc kubenswrapper[4805]: I1203 14:40:43.588476 4805 scope.go:117] "RemoveContainer" containerID="917876c33cc904807afef7d8edc87b2672e031fae006b99501746b08d7943964" Dec 03 14:40:47 crc kubenswrapper[4805]: I1203 14:40:47.062931 4805 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/placement-db-sync-sj4wt"] Dec 03 14:40:47 crc kubenswrapper[4805]: I1203 14:40:47.071813 4805 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-db-sync-b6k7w"] Dec 03 14:40:47 crc kubenswrapper[4805]: I1203 14:40:47.081054 4805 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/heat-db-sync-jvqmk"] Dec 03 14:40:47 crc kubenswrapper[4805]: I1203 14:40:47.091617 4805 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-bootstrap-74pdd"] Dec 03 14:40:47 crc kubenswrapper[4805]: I1203 14:40:47.102684 4805 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/placement-db-sync-sj4wt"] Dec 03 14:40:47 crc kubenswrapper[4805]: I1203 14:40:47.111650 4805 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-db-sync-b6k7w"] Dec 03 14:40:47 crc kubenswrapper[4805]: I1203 14:40:47.119786 4805 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/heat-db-sync-jvqmk"] Dec 03 14:40:47 crc kubenswrapper[4805]: I1203 14:40:47.127933 4805 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-bootstrap-74pdd"] Dec 03 14:40:48 crc kubenswrapper[4805]: I1203 14:40:48.705925 4805 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="14c327e3-e880-4b49-b366-364a0395cc55" path="/var/lib/kubelet/pods/14c327e3-e880-4b49-b366-364a0395cc55/volumes" Dec 03 14:40:48 crc kubenswrapper[4805]: I1203 14:40:48.706724 4805 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="32920f09-a728-4b35-a5d7-65dae2d70834" path="/var/lib/kubelet/pods/32920f09-a728-4b35-a5d7-65dae2d70834/volumes" Dec 03 14:40:48 crc kubenswrapper[4805]: I1203 14:40:48.707550 4805 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7c1b5525-de7b-4a8c-9062-bb9a68336989" path="/var/lib/kubelet/pods/7c1b5525-de7b-4a8c-9062-bb9a68336989/volumes" Dec 03 14:40:48 crc kubenswrapper[4805]: I1203 14:40:48.708927 4805 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b4e050ed-47f5-4d9f-a288-b6d2a7ab294a" path="/var/lib/kubelet/pods/b4e050ed-47f5-4d9f-a288-b6d2a7ab294a/volumes" Dec 03 14:40:56 crc kubenswrapper[4805]: I1203 14:40:56.694571 4805 scope.go:117] "RemoveContainer" containerID="748e11b7efe02ff9a381aeee0aa4ddad32272e1564bc028d779c442a7b7f2149" Dec 03 14:40:56 crc kubenswrapper[4805]: E1203 14:40:56.695370 4805 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-gskh4_openshift-machine-config-operator(6098937f-e3f6-45e8-a647-4994a79cd711)\"" pod="openshift-machine-config-operator/machine-config-daemon-gskh4" podUID="6098937f-e3f6-45e8-a647-4994a79cd711" Dec 03 14:41:01 crc kubenswrapper[4805]: I1203 14:41:01.658023 4805 generic.go:334] "Generic (PLEG): container finished" podID="a6f87a69-c7bb-44d8-9f65-82b1b4f8885a" containerID="1bcab277331e150c29b177bede8829495722413ffb111196309be06365d4d29a" exitCode=0 Dec 03 14:41:01 crc kubenswrapper[4805]: I1203 14:41:01.658127 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-nvbww" event={"ID":"a6f87a69-c7bb-44d8-9f65-82b1b4f8885a","Type":"ContainerDied","Data":"1bcab277331e150c29b177bede8829495722413ffb111196309be06365d4d29a"} Dec 03 14:41:03 crc kubenswrapper[4805]: I1203 14:41:03.116132 4805 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-nvbww" Dec 03 14:41:03 crc kubenswrapper[4805]: I1203 14:41:03.184849 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6slmn\" (UniqueName: \"kubernetes.io/projected/a6f87a69-c7bb-44d8-9f65-82b1b4f8885a-kube-api-access-6slmn\") pod \"a6f87a69-c7bb-44d8-9f65-82b1b4f8885a\" (UID: \"a6f87a69-c7bb-44d8-9f65-82b1b4f8885a\") " Dec 03 14:41:03 crc kubenswrapper[4805]: I1203 14:41:03.185010 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/a6f87a69-c7bb-44d8-9f65-82b1b4f8885a-ssh-key\") pod \"a6f87a69-c7bb-44d8-9f65-82b1b4f8885a\" (UID: \"a6f87a69-c7bb-44d8-9f65-82b1b4f8885a\") " Dec 03 14:41:03 crc kubenswrapper[4805]: I1203 14:41:03.185125 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/a6f87a69-c7bb-44d8-9f65-82b1b4f8885a-inventory\") pod \"a6f87a69-c7bb-44d8-9f65-82b1b4f8885a\" (UID: \"a6f87a69-c7bb-44d8-9f65-82b1b4f8885a\") " Dec 03 14:41:03 crc kubenswrapper[4805]: I1203 14:41:03.191157 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a6f87a69-c7bb-44d8-9f65-82b1b4f8885a-kube-api-access-6slmn" (OuterVolumeSpecName: "kube-api-access-6slmn") pod "a6f87a69-c7bb-44d8-9f65-82b1b4f8885a" (UID: "a6f87a69-c7bb-44d8-9f65-82b1b4f8885a"). InnerVolumeSpecName "kube-api-access-6slmn". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 14:41:03 crc kubenswrapper[4805]: I1203 14:41:03.211327 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a6f87a69-c7bb-44d8-9f65-82b1b4f8885a-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "a6f87a69-c7bb-44d8-9f65-82b1b4f8885a" (UID: "a6f87a69-c7bb-44d8-9f65-82b1b4f8885a"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 14:41:03 crc kubenswrapper[4805]: I1203 14:41:03.219083 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a6f87a69-c7bb-44d8-9f65-82b1b4f8885a-inventory" (OuterVolumeSpecName: "inventory") pod "a6f87a69-c7bb-44d8-9f65-82b1b4f8885a" (UID: "a6f87a69-c7bb-44d8-9f65-82b1b4f8885a"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 14:41:03 crc kubenswrapper[4805]: I1203 14:41:03.291479 4805 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6slmn\" (UniqueName: \"kubernetes.io/projected/a6f87a69-c7bb-44d8-9f65-82b1b4f8885a-kube-api-access-6slmn\") on node \"crc\" DevicePath \"\"" Dec 03 14:41:03 crc kubenswrapper[4805]: I1203 14:41:03.291526 4805 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/a6f87a69-c7bb-44d8-9f65-82b1b4f8885a-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 03 14:41:03 crc kubenswrapper[4805]: I1203 14:41:03.291573 4805 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/a6f87a69-c7bb-44d8-9f65-82b1b4f8885a-inventory\") on node \"crc\" DevicePath \"\"" Dec 03 14:41:03 crc kubenswrapper[4805]: I1203 14:41:03.676549 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-nvbww" event={"ID":"a6f87a69-c7bb-44d8-9f65-82b1b4f8885a","Type":"ContainerDied","Data":"637c3576b583cfbdac29322815e9fba8a83f7bf9e8780480c5579adb46293846"} Dec 03 14:41:03 crc kubenswrapper[4805]: I1203 14:41:03.676879 4805 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="637c3576b583cfbdac29322815e9fba8a83f7bf9e8780480c5579adb46293846" Dec 03 14:41:03 crc kubenswrapper[4805]: I1203 14:41:03.676680 4805 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-nvbww" Dec 03 14:41:03 crc kubenswrapper[4805]: I1203 14:41:03.769241 4805 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/configure-network-edpm-deployment-openstack-edpm-ipam-6wncz"] Dec 03 14:41:03 crc kubenswrapper[4805]: E1203 14:41:03.769611 4805 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="25ccfd1f-9135-4f16-974c-3ca2a748f114" containerName="registry-server" Dec 03 14:41:03 crc kubenswrapper[4805]: I1203 14:41:03.769631 4805 state_mem.go:107] "Deleted CPUSet assignment" podUID="25ccfd1f-9135-4f16-974c-3ca2a748f114" containerName="registry-server" Dec 03 14:41:03 crc kubenswrapper[4805]: E1203 14:41:03.769658 4805 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6fdb2f98-919a-41db-8e05-53f7ea6f7d47" containerName="extract-content" Dec 03 14:41:03 crc kubenswrapper[4805]: I1203 14:41:03.769667 4805 state_mem.go:107] "Deleted CPUSet assignment" podUID="6fdb2f98-919a-41db-8e05-53f7ea6f7d47" containerName="extract-content" Dec 03 14:41:03 crc kubenswrapper[4805]: E1203 14:41:03.769693 4805 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="25ccfd1f-9135-4f16-974c-3ca2a748f114" containerName="extract-content" Dec 03 14:41:03 crc kubenswrapper[4805]: I1203 14:41:03.769699 4805 state_mem.go:107] "Deleted CPUSet assignment" podUID="25ccfd1f-9135-4f16-974c-3ca2a748f114" containerName="extract-content" Dec 03 14:41:03 crc kubenswrapper[4805]: E1203 14:41:03.769712 4805 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="52dcae87-e990-480d-886e-26cec1ee4da2" containerName="registry-server" Dec 03 14:41:03 crc kubenswrapper[4805]: I1203 14:41:03.769718 4805 state_mem.go:107] "Deleted CPUSet assignment" podUID="52dcae87-e990-480d-886e-26cec1ee4da2" containerName="registry-server" Dec 03 14:41:03 crc kubenswrapper[4805]: E1203 14:41:03.769725 4805 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="25ccfd1f-9135-4f16-974c-3ca2a748f114" containerName="extract-utilities" Dec 03 14:41:03 crc kubenswrapper[4805]: I1203 14:41:03.769731 4805 state_mem.go:107] "Deleted CPUSet assignment" podUID="25ccfd1f-9135-4f16-974c-3ca2a748f114" containerName="extract-utilities" Dec 03 14:41:03 crc kubenswrapper[4805]: E1203 14:41:03.769744 4805 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="52dcae87-e990-480d-886e-26cec1ee4da2" containerName="extract-content" Dec 03 14:41:03 crc kubenswrapper[4805]: I1203 14:41:03.769749 4805 state_mem.go:107] "Deleted CPUSet assignment" podUID="52dcae87-e990-480d-886e-26cec1ee4da2" containerName="extract-content" Dec 03 14:41:03 crc kubenswrapper[4805]: E1203 14:41:03.769757 4805 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6fdb2f98-919a-41db-8e05-53f7ea6f7d47" containerName="extract-utilities" Dec 03 14:41:03 crc kubenswrapper[4805]: I1203 14:41:03.769762 4805 state_mem.go:107] "Deleted CPUSet assignment" podUID="6fdb2f98-919a-41db-8e05-53f7ea6f7d47" containerName="extract-utilities" Dec 03 14:41:03 crc kubenswrapper[4805]: E1203 14:41:03.769772 4805 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="52dcae87-e990-480d-886e-26cec1ee4da2" containerName="extract-utilities" Dec 03 14:41:03 crc kubenswrapper[4805]: I1203 14:41:03.769778 4805 state_mem.go:107] "Deleted CPUSet assignment" podUID="52dcae87-e990-480d-886e-26cec1ee4da2" containerName="extract-utilities" Dec 03 14:41:03 crc kubenswrapper[4805]: E1203 14:41:03.769786 4805 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6fdb2f98-919a-41db-8e05-53f7ea6f7d47" containerName="registry-server" Dec 03 14:41:03 crc kubenswrapper[4805]: I1203 14:41:03.769792 4805 state_mem.go:107] "Deleted CPUSet assignment" podUID="6fdb2f98-919a-41db-8e05-53f7ea6f7d47" containerName="registry-server" Dec 03 14:41:03 crc kubenswrapper[4805]: E1203 14:41:03.769803 4805 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a6f87a69-c7bb-44d8-9f65-82b1b4f8885a" containerName="download-cache-edpm-deployment-openstack-edpm-ipam" Dec 03 14:41:03 crc kubenswrapper[4805]: I1203 14:41:03.769810 4805 state_mem.go:107] "Deleted CPUSet assignment" podUID="a6f87a69-c7bb-44d8-9f65-82b1b4f8885a" containerName="download-cache-edpm-deployment-openstack-edpm-ipam" Dec 03 14:41:03 crc kubenswrapper[4805]: I1203 14:41:03.770018 4805 memory_manager.go:354] "RemoveStaleState removing state" podUID="25ccfd1f-9135-4f16-974c-3ca2a748f114" containerName="registry-server" Dec 03 14:41:03 crc kubenswrapper[4805]: I1203 14:41:03.770049 4805 memory_manager.go:354] "RemoveStaleState removing state" podUID="a6f87a69-c7bb-44d8-9f65-82b1b4f8885a" containerName="download-cache-edpm-deployment-openstack-edpm-ipam" Dec 03 14:41:03 crc kubenswrapper[4805]: I1203 14:41:03.770065 4805 memory_manager.go:354] "RemoveStaleState removing state" podUID="6fdb2f98-919a-41db-8e05-53f7ea6f7d47" containerName="registry-server" Dec 03 14:41:03 crc kubenswrapper[4805]: I1203 14:41:03.770077 4805 memory_manager.go:354] "RemoveStaleState removing state" podUID="52dcae87-e990-480d-886e-26cec1ee4da2" containerName="registry-server" Dec 03 14:41:03 crc kubenswrapper[4805]: I1203 14:41:03.770798 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-6wncz" Dec 03 14:41:03 crc kubenswrapper[4805]: I1203 14:41:03.773918 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Dec 03 14:41:03 crc kubenswrapper[4805]: I1203 14:41:03.773994 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Dec 03 14:41:03 crc kubenswrapper[4805]: I1203 14:41:03.774150 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-n7jt4" Dec 03 14:41:03 crc kubenswrapper[4805]: I1203 14:41:03.774148 4805 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 03 14:41:03 crc kubenswrapper[4805]: I1203 14:41:03.783806 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/configure-network-edpm-deployment-openstack-edpm-ipam-6wncz"] Dec 03 14:41:03 crc kubenswrapper[4805]: I1203 14:41:03.901732 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/546a9824-86ca-46ed-b6fa-3384de230115-inventory\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-6wncz\" (UID: \"546a9824-86ca-46ed-b6fa-3384de230115\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-6wncz" Dec 03 14:41:03 crc kubenswrapper[4805]: I1203 14:41:03.901784 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/546a9824-86ca-46ed-b6fa-3384de230115-ssh-key\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-6wncz\" (UID: \"546a9824-86ca-46ed-b6fa-3384de230115\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-6wncz" Dec 03 14:41:03 crc kubenswrapper[4805]: I1203 14:41:03.902389 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6tx42\" (UniqueName: \"kubernetes.io/projected/546a9824-86ca-46ed-b6fa-3384de230115-kube-api-access-6tx42\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-6wncz\" (UID: \"546a9824-86ca-46ed-b6fa-3384de230115\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-6wncz" Dec 03 14:41:04 crc kubenswrapper[4805]: I1203 14:41:04.004297 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6tx42\" (UniqueName: \"kubernetes.io/projected/546a9824-86ca-46ed-b6fa-3384de230115-kube-api-access-6tx42\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-6wncz\" (UID: \"546a9824-86ca-46ed-b6fa-3384de230115\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-6wncz" Dec 03 14:41:04 crc kubenswrapper[4805]: I1203 14:41:04.004485 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/546a9824-86ca-46ed-b6fa-3384de230115-inventory\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-6wncz\" (UID: \"546a9824-86ca-46ed-b6fa-3384de230115\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-6wncz" Dec 03 14:41:04 crc kubenswrapper[4805]: I1203 14:41:04.004512 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/546a9824-86ca-46ed-b6fa-3384de230115-ssh-key\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-6wncz\" (UID: \"546a9824-86ca-46ed-b6fa-3384de230115\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-6wncz" Dec 03 14:41:04 crc kubenswrapper[4805]: I1203 14:41:04.011250 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/546a9824-86ca-46ed-b6fa-3384de230115-ssh-key\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-6wncz\" (UID: \"546a9824-86ca-46ed-b6fa-3384de230115\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-6wncz" Dec 03 14:41:04 crc kubenswrapper[4805]: I1203 14:41:04.011598 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/546a9824-86ca-46ed-b6fa-3384de230115-inventory\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-6wncz\" (UID: \"546a9824-86ca-46ed-b6fa-3384de230115\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-6wncz" Dec 03 14:41:04 crc kubenswrapper[4805]: I1203 14:41:04.023628 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6tx42\" (UniqueName: \"kubernetes.io/projected/546a9824-86ca-46ed-b6fa-3384de230115-kube-api-access-6tx42\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-6wncz\" (UID: \"546a9824-86ca-46ed-b6fa-3384de230115\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-6wncz" Dec 03 14:41:04 crc kubenswrapper[4805]: I1203 14:41:04.089058 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-6wncz" Dec 03 14:41:04 crc kubenswrapper[4805]: I1203 14:41:04.657950 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/configure-network-edpm-deployment-openstack-edpm-ipam-6wncz"] Dec 03 14:41:04 crc kubenswrapper[4805]: I1203 14:41:04.692811 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-6wncz" event={"ID":"546a9824-86ca-46ed-b6fa-3384de230115","Type":"ContainerStarted","Data":"1ab2974779275161adfb99db2bc5c24687ece7988093c17d69b53f28ae55e4c8"} Dec 03 14:41:05 crc kubenswrapper[4805]: I1203 14:41:05.705702 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-6wncz" event={"ID":"546a9824-86ca-46ed-b6fa-3384de230115","Type":"ContainerStarted","Data":"bd084b27bbda1ca20870b373899dcaafa61a688a2f7b9b53271ea1f83ad1ef9c"} Dec 03 14:41:05 crc kubenswrapper[4805]: I1203 14:41:05.722530 4805 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-6wncz" podStartSLOduration=2.197449695 podStartE2EDuration="2.722506524s" podCreationTimestamp="2025-12-03 14:41:03 +0000 UTC" firstStartedPulling="2025-12-03 14:41:04.662494848 +0000 UTC m=+1894.325411771" lastFinishedPulling="2025-12-03 14:41:05.187551667 +0000 UTC m=+1894.850468600" observedRunningTime="2025-12-03 14:41:05.72198437 +0000 UTC m=+1895.384901293" watchObservedRunningTime="2025-12-03 14:41:05.722506524 +0000 UTC m=+1895.385423447" Dec 03 14:41:09 crc kubenswrapper[4805]: I1203 14:41:09.695362 4805 scope.go:117] "RemoveContainer" containerID="748e11b7efe02ff9a381aeee0aa4ddad32272e1564bc028d779c442a7b7f2149" Dec 03 14:41:09 crc kubenswrapper[4805]: E1203 14:41:09.696536 4805 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-gskh4_openshift-machine-config-operator(6098937f-e3f6-45e8-a647-4994a79cd711)\"" pod="openshift-machine-config-operator/machine-config-daemon-gskh4" podUID="6098937f-e3f6-45e8-a647-4994a79cd711" Dec 03 14:41:12 crc kubenswrapper[4805]: I1203 14:41:12.055409 4805 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-db-sync-74tz4"] Dec 03 14:41:12 crc kubenswrapper[4805]: I1203 14:41:12.065299 4805 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-db-sync-74tz4"] Dec 03 14:41:12 crc kubenswrapper[4805]: I1203 14:41:12.712135 4805 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="13ba03f0-acc2-4fb8-90a9-583b4f8694cb" path="/var/lib/kubelet/pods/13ba03f0-acc2-4fb8-90a9-583b4f8694cb/volumes" Dec 03 14:41:21 crc kubenswrapper[4805]: I1203 14:41:21.695107 4805 scope.go:117] "RemoveContainer" containerID="748e11b7efe02ff9a381aeee0aa4ddad32272e1564bc028d779c442a7b7f2149" Dec 03 14:41:21 crc kubenswrapper[4805]: E1203 14:41:21.697212 4805 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-gskh4_openshift-machine-config-operator(6098937f-e3f6-45e8-a647-4994a79cd711)\"" pod="openshift-machine-config-operator/machine-config-daemon-gskh4" podUID="6098937f-e3f6-45e8-a647-4994a79cd711" Dec 03 14:41:23 crc kubenswrapper[4805]: I1203 14:41:23.043593 4805 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-db-sync-fh9c2"] Dec 03 14:41:23 crc kubenswrapper[4805]: I1203 14:41:23.051817 4805 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-db-sync-fh9c2"] Dec 03 14:41:24 crc kubenswrapper[4805]: I1203 14:41:24.708484 4805 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="260717ec-f898-4ccd-8a4f-c7d6b14ade6e" path="/var/lib/kubelet/pods/260717ec-f898-4ccd-8a4f-c7d6b14ade6e/volumes" Dec 03 14:41:34 crc kubenswrapper[4805]: I1203 14:41:34.694708 4805 scope.go:117] "RemoveContainer" containerID="748e11b7efe02ff9a381aeee0aa4ddad32272e1564bc028d779c442a7b7f2149" Dec 03 14:41:34 crc kubenswrapper[4805]: E1203 14:41:34.696460 4805 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-gskh4_openshift-machine-config-operator(6098937f-e3f6-45e8-a647-4994a79cd711)\"" pod="openshift-machine-config-operator/machine-config-daemon-gskh4" podUID="6098937f-e3f6-45e8-a647-4994a79cd711" Dec 03 14:41:43 crc kubenswrapper[4805]: I1203 14:41:43.824142 4805 scope.go:117] "RemoveContainer" containerID="d44bffcdc75d0c2fe9722cdbe85e33a78e3ee76b7cbfda1d8e0f176918810006" Dec 03 14:41:43 crc kubenswrapper[4805]: I1203 14:41:43.863222 4805 scope.go:117] "RemoveContainer" containerID="c8040c823d6c387af893c10d68844e6a3f47022e19a18567b34538fe04cfa697" Dec 03 14:41:43 crc kubenswrapper[4805]: I1203 14:41:43.919368 4805 scope.go:117] "RemoveContainer" containerID="76a9c6e35bfc792baf01e7282b5b7319e4e528a8abaa7f19a90eb5f470bdeab4" Dec 03 14:41:43 crc kubenswrapper[4805]: I1203 14:41:43.955578 4805 scope.go:117] "RemoveContainer" containerID="3ae18d4a53cc1df81f75833ab27219aaab27ca80a4b5ccbb30481f900bcecd6f" Dec 03 14:41:44 crc kubenswrapper[4805]: I1203 14:41:44.013695 4805 scope.go:117] "RemoveContainer" containerID="baebfb4ff35a0b981da87af49115db6766414aa8463f0e39b209086ad47d1d10" Dec 03 14:41:44 crc kubenswrapper[4805]: I1203 14:41:44.069780 4805 scope.go:117] "RemoveContainer" containerID="c74bd6f2e58840be38059668ab74c768c407bb15b81d002cdfe1258aa66a081b" Dec 03 14:41:45 crc kubenswrapper[4805]: I1203 14:41:45.033417 4805 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-db-create-b8n7x"] Dec 03 14:41:45 crc kubenswrapper[4805]: I1203 14:41:45.044278 4805 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-db-create-b8n7x"] Dec 03 14:41:46 crc kubenswrapper[4805]: I1203 14:41:46.714452 4805 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="596937d0-4958-4d99-a677-698ca4b7d4eb" path="/var/lib/kubelet/pods/596937d0-4958-4d99-a677-698ca4b7d4eb/volumes" Dec 03 14:41:48 crc kubenswrapper[4805]: I1203 14:41:48.032270 4805 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-db-create-bc8gs"] Dec 03 14:41:48 crc kubenswrapper[4805]: I1203 14:41:48.042463 4805 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-db-create-z7rqx"] Dec 03 14:41:48 crc kubenswrapper[4805]: I1203 14:41:48.051293 4805 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-db-create-bc8gs"] Dec 03 14:41:48 crc kubenswrapper[4805]: I1203 14:41:48.059316 4805 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-db-create-z7rqx"] Dec 03 14:41:48 crc kubenswrapper[4805]: I1203 14:41:48.709291 4805 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1245f02d-543a-40d6-bc19-745a48249604" path="/var/lib/kubelet/pods/1245f02d-543a-40d6-bc19-745a48249604/volumes" Dec 03 14:41:48 crc kubenswrapper[4805]: I1203 14:41:48.710055 4805 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="649e0d3a-332c-43d7-90bd-8ad10015268f" path="/var/lib/kubelet/pods/649e0d3a-332c-43d7-90bd-8ad10015268f/volumes" Dec 03 14:41:49 crc kubenswrapper[4805]: I1203 14:41:49.044108 4805 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-f183-account-create-update-p7mnx"] Dec 03 14:41:49 crc kubenswrapper[4805]: I1203 14:41:49.057160 4805 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-1be1-account-create-update-58nc2"] Dec 03 14:41:49 crc kubenswrapper[4805]: I1203 14:41:49.064774 4805 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-f183-account-create-update-p7mnx"] Dec 03 14:41:49 crc kubenswrapper[4805]: I1203 14:41:49.072366 4805 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-1be1-account-create-update-58nc2"] Dec 03 14:41:49 crc kubenswrapper[4805]: I1203 14:41:49.079723 4805 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-7801-account-create-update-rjbpl"] Dec 03 14:41:49 crc kubenswrapper[4805]: I1203 14:41:49.088078 4805 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-7801-account-create-update-rjbpl"] Dec 03 14:41:49 crc kubenswrapper[4805]: I1203 14:41:49.694403 4805 scope.go:117] "RemoveContainer" containerID="748e11b7efe02ff9a381aeee0aa4ddad32272e1564bc028d779c442a7b7f2149" Dec 03 14:41:49 crc kubenswrapper[4805]: E1203 14:41:49.695278 4805 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-gskh4_openshift-machine-config-operator(6098937f-e3f6-45e8-a647-4994a79cd711)\"" pod="openshift-machine-config-operator/machine-config-daemon-gskh4" podUID="6098937f-e3f6-45e8-a647-4994a79cd711" Dec 03 14:41:50 crc kubenswrapper[4805]: I1203 14:41:50.710364 4805 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0c5c12d6-db52-4367-8898-e88fc8cdde41" path="/var/lib/kubelet/pods/0c5c12d6-db52-4367-8898-e88fc8cdde41/volumes" Dec 03 14:41:50 crc kubenswrapper[4805]: I1203 14:41:50.711154 4805 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1082c862-47c9-4382-b3e8-09e58fb5d85d" path="/var/lib/kubelet/pods/1082c862-47c9-4382-b3e8-09e58fb5d85d/volumes" Dec 03 14:41:50 crc kubenswrapper[4805]: I1203 14:41:50.711718 4805 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5c694d3e-75d4-40d4-924d-cfe9e9ef5ef7" path="/var/lib/kubelet/pods/5c694d3e-75d4-40d4-924d-cfe9e9ef5ef7/volumes" Dec 03 14:42:00 crc kubenswrapper[4805]: I1203 14:42:00.701517 4805 scope.go:117] "RemoveContainer" containerID="748e11b7efe02ff9a381aeee0aa4ddad32272e1564bc028d779c442a7b7f2149" Dec 03 14:42:00 crc kubenswrapper[4805]: E1203 14:42:00.702417 4805 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-gskh4_openshift-machine-config-operator(6098937f-e3f6-45e8-a647-4994a79cd711)\"" pod="openshift-machine-config-operator/machine-config-daemon-gskh4" podUID="6098937f-e3f6-45e8-a647-4994a79cd711" Dec 03 14:42:11 crc kubenswrapper[4805]: I1203 14:42:11.694564 4805 scope.go:117] "RemoveContainer" containerID="748e11b7efe02ff9a381aeee0aa4ddad32272e1564bc028d779c442a7b7f2149" Dec 03 14:42:11 crc kubenswrapper[4805]: E1203 14:42:11.696232 4805 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-gskh4_openshift-machine-config-operator(6098937f-e3f6-45e8-a647-4994a79cd711)\"" pod="openshift-machine-config-operator/machine-config-daemon-gskh4" podUID="6098937f-e3f6-45e8-a647-4994a79cd711" Dec 03 14:42:23 crc kubenswrapper[4805]: I1203 14:42:23.414047 4805 generic.go:334] "Generic (PLEG): container finished" podID="546a9824-86ca-46ed-b6fa-3384de230115" containerID="bd084b27bbda1ca20870b373899dcaafa61a688a2f7b9b53271ea1f83ad1ef9c" exitCode=0 Dec 03 14:42:23 crc kubenswrapper[4805]: I1203 14:42:23.414154 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-6wncz" event={"ID":"546a9824-86ca-46ed-b6fa-3384de230115","Type":"ContainerDied","Data":"bd084b27bbda1ca20870b373899dcaafa61a688a2f7b9b53271ea1f83ad1ef9c"} Dec 03 14:42:24 crc kubenswrapper[4805]: I1203 14:42:24.901104 4805 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-6wncz" Dec 03 14:42:24 crc kubenswrapper[4805]: I1203 14:42:24.976636 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/546a9824-86ca-46ed-b6fa-3384de230115-inventory\") pod \"546a9824-86ca-46ed-b6fa-3384de230115\" (UID: \"546a9824-86ca-46ed-b6fa-3384de230115\") " Dec 03 14:42:24 crc kubenswrapper[4805]: I1203 14:42:24.976688 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/546a9824-86ca-46ed-b6fa-3384de230115-ssh-key\") pod \"546a9824-86ca-46ed-b6fa-3384de230115\" (UID: \"546a9824-86ca-46ed-b6fa-3384de230115\") " Dec 03 14:42:24 crc kubenswrapper[4805]: I1203 14:42:24.976750 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6tx42\" (UniqueName: \"kubernetes.io/projected/546a9824-86ca-46ed-b6fa-3384de230115-kube-api-access-6tx42\") pod \"546a9824-86ca-46ed-b6fa-3384de230115\" (UID: \"546a9824-86ca-46ed-b6fa-3384de230115\") " Dec 03 14:42:24 crc kubenswrapper[4805]: I1203 14:42:24.983572 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/546a9824-86ca-46ed-b6fa-3384de230115-kube-api-access-6tx42" (OuterVolumeSpecName: "kube-api-access-6tx42") pod "546a9824-86ca-46ed-b6fa-3384de230115" (UID: "546a9824-86ca-46ed-b6fa-3384de230115"). InnerVolumeSpecName "kube-api-access-6tx42". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 14:42:25 crc kubenswrapper[4805]: I1203 14:42:25.007938 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/546a9824-86ca-46ed-b6fa-3384de230115-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "546a9824-86ca-46ed-b6fa-3384de230115" (UID: "546a9824-86ca-46ed-b6fa-3384de230115"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 14:42:25 crc kubenswrapper[4805]: I1203 14:42:25.011558 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/546a9824-86ca-46ed-b6fa-3384de230115-inventory" (OuterVolumeSpecName: "inventory") pod "546a9824-86ca-46ed-b6fa-3384de230115" (UID: "546a9824-86ca-46ed-b6fa-3384de230115"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 14:42:25 crc kubenswrapper[4805]: I1203 14:42:25.080657 4805 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/546a9824-86ca-46ed-b6fa-3384de230115-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 03 14:42:25 crc kubenswrapper[4805]: I1203 14:42:25.080696 4805 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6tx42\" (UniqueName: \"kubernetes.io/projected/546a9824-86ca-46ed-b6fa-3384de230115-kube-api-access-6tx42\") on node \"crc\" DevicePath \"\"" Dec 03 14:42:25 crc kubenswrapper[4805]: I1203 14:42:25.080712 4805 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/546a9824-86ca-46ed-b6fa-3384de230115-inventory\") on node \"crc\" DevicePath \"\"" Dec 03 14:42:25 crc kubenswrapper[4805]: I1203 14:42:25.438104 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-6wncz" event={"ID":"546a9824-86ca-46ed-b6fa-3384de230115","Type":"ContainerDied","Data":"1ab2974779275161adfb99db2bc5c24687ece7988093c17d69b53f28ae55e4c8"} Dec 03 14:42:25 crc kubenswrapper[4805]: I1203 14:42:25.438529 4805 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="1ab2974779275161adfb99db2bc5c24687ece7988093c17d69b53f28ae55e4c8" Dec 03 14:42:25 crc kubenswrapper[4805]: I1203 14:42:25.438170 4805 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-6wncz" Dec 03 14:42:25 crc kubenswrapper[4805]: I1203 14:42:25.524191 4805 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/validate-network-edpm-deployment-openstack-edpm-ipam-7px67"] Dec 03 14:42:25 crc kubenswrapper[4805]: E1203 14:42:25.528816 4805 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="546a9824-86ca-46ed-b6fa-3384de230115" containerName="configure-network-edpm-deployment-openstack-edpm-ipam" Dec 03 14:42:25 crc kubenswrapper[4805]: I1203 14:42:25.528872 4805 state_mem.go:107] "Deleted CPUSet assignment" podUID="546a9824-86ca-46ed-b6fa-3384de230115" containerName="configure-network-edpm-deployment-openstack-edpm-ipam" Dec 03 14:42:25 crc kubenswrapper[4805]: I1203 14:42:25.529261 4805 memory_manager.go:354] "RemoveStaleState removing state" podUID="546a9824-86ca-46ed-b6fa-3384de230115" containerName="configure-network-edpm-deployment-openstack-edpm-ipam" Dec 03 14:42:25 crc kubenswrapper[4805]: I1203 14:42:25.530185 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-7px67" Dec 03 14:42:25 crc kubenswrapper[4805]: I1203 14:42:25.534049 4805 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 03 14:42:25 crc kubenswrapper[4805]: I1203 14:42:25.534103 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Dec 03 14:42:25 crc kubenswrapper[4805]: I1203 14:42:25.535092 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-n7jt4" Dec 03 14:42:25 crc kubenswrapper[4805]: I1203 14:42:25.535357 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Dec 03 14:42:25 crc kubenswrapper[4805]: I1203 14:42:25.566195 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/validate-network-edpm-deployment-openstack-edpm-ipam-7px67"] Dec 03 14:42:25 crc kubenswrapper[4805]: I1203 14:42:25.591647 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/dd64758f-9fe5-4f00-8f60-4bba03a0f98a-inventory\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-7px67\" (UID: \"dd64758f-9fe5-4f00-8f60-4bba03a0f98a\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-7px67" Dec 03 14:42:25 crc kubenswrapper[4805]: I1203 14:42:25.591691 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/dd64758f-9fe5-4f00-8f60-4bba03a0f98a-ssh-key\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-7px67\" (UID: \"dd64758f-9fe5-4f00-8f60-4bba03a0f98a\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-7px67" Dec 03 14:42:25 crc kubenswrapper[4805]: I1203 14:42:25.591785 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-28xxm\" (UniqueName: \"kubernetes.io/projected/dd64758f-9fe5-4f00-8f60-4bba03a0f98a-kube-api-access-28xxm\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-7px67\" (UID: \"dd64758f-9fe5-4f00-8f60-4bba03a0f98a\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-7px67" Dec 03 14:42:25 crc kubenswrapper[4805]: I1203 14:42:25.693782 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/dd64758f-9fe5-4f00-8f60-4bba03a0f98a-inventory\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-7px67\" (UID: \"dd64758f-9fe5-4f00-8f60-4bba03a0f98a\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-7px67" Dec 03 14:42:25 crc kubenswrapper[4805]: I1203 14:42:25.693856 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/dd64758f-9fe5-4f00-8f60-4bba03a0f98a-ssh-key\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-7px67\" (UID: \"dd64758f-9fe5-4f00-8f60-4bba03a0f98a\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-7px67" Dec 03 14:42:25 crc kubenswrapper[4805]: I1203 14:42:25.693923 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-28xxm\" (UniqueName: \"kubernetes.io/projected/dd64758f-9fe5-4f00-8f60-4bba03a0f98a-kube-api-access-28xxm\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-7px67\" (UID: \"dd64758f-9fe5-4f00-8f60-4bba03a0f98a\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-7px67" Dec 03 14:42:25 crc kubenswrapper[4805]: I1203 14:42:25.695396 4805 scope.go:117] "RemoveContainer" containerID="748e11b7efe02ff9a381aeee0aa4ddad32272e1564bc028d779c442a7b7f2149" Dec 03 14:42:25 crc kubenswrapper[4805]: E1203 14:42:25.695780 4805 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-gskh4_openshift-machine-config-operator(6098937f-e3f6-45e8-a647-4994a79cd711)\"" pod="openshift-machine-config-operator/machine-config-daemon-gskh4" podUID="6098937f-e3f6-45e8-a647-4994a79cd711" Dec 03 14:42:25 crc kubenswrapper[4805]: I1203 14:42:25.699313 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/dd64758f-9fe5-4f00-8f60-4bba03a0f98a-inventory\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-7px67\" (UID: \"dd64758f-9fe5-4f00-8f60-4bba03a0f98a\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-7px67" Dec 03 14:42:25 crc kubenswrapper[4805]: I1203 14:42:25.700698 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/dd64758f-9fe5-4f00-8f60-4bba03a0f98a-ssh-key\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-7px67\" (UID: \"dd64758f-9fe5-4f00-8f60-4bba03a0f98a\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-7px67" Dec 03 14:42:25 crc kubenswrapper[4805]: I1203 14:42:25.714807 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-28xxm\" (UniqueName: \"kubernetes.io/projected/dd64758f-9fe5-4f00-8f60-4bba03a0f98a-kube-api-access-28xxm\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-7px67\" (UID: \"dd64758f-9fe5-4f00-8f60-4bba03a0f98a\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-7px67" Dec 03 14:42:25 crc kubenswrapper[4805]: I1203 14:42:25.849226 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-7px67" Dec 03 14:42:26 crc kubenswrapper[4805]: I1203 14:42:26.386425 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/validate-network-edpm-deployment-openstack-edpm-ipam-7px67"] Dec 03 14:42:26 crc kubenswrapper[4805]: W1203 14:42:26.390385 4805 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poddd64758f_9fe5_4f00_8f60_4bba03a0f98a.slice/crio-75e01d2d924e8b5bddee770d9a81d76cd682e249576964b9162b64eb59893e24 WatchSource:0}: Error finding container 75e01d2d924e8b5bddee770d9a81d76cd682e249576964b9162b64eb59893e24: Status 404 returned error can't find the container with id 75e01d2d924e8b5bddee770d9a81d76cd682e249576964b9162b64eb59893e24 Dec 03 14:42:26 crc kubenswrapper[4805]: I1203 14:42:26.447214 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-7px67" event={"ID":"dd64758f-9fe5-4f00-8f60-4bba03a0f98a","Type":"ContainerStarted","Data":"75e01d2d924e8b5bddee770d9a81d76cd682e249576964b9162b64eb59893e24"} Dec 03 14:42:28 crc kubenswrapper[4805]: I1203 14:42:28.466730 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-7px67" event={"ID":"dd64758f-9fe5-4f00-8f60-4bba03a0f98a","Type":"ContainerStarted","Data":"1550ff86130027e46c3614321fb5e00542e47fcdc0e81a35a399dd1e75e52ba8"} Dec 03 14:42:28 crc kubenswrapper[4805]: I1203 14:42:28.489781 4805 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-7px67" podStartSLOduration=2.348544946 podStartE2EDuration="3.489756747s" podCreationTimestamp="2025-12-03 14:42:25 +0000 UTC" firstStartedPulling="2025-12-03 14:42:26.39289923 +0000 UTC m=+1976.055816153" lastFinishedPulling="2025-12-03 14:42:27.534111031 +0000 UTC m=+1977.197027954" observedRunningTime="2025-12-03 14:42:28.484459434 +0000 UTC m=+1978.147376367" watchObservedRunningTime="2025-12-03 14:42:28.489756747 +0000 UTC m=+1978.152673670" Dec 03 14:42:32 crc kubenswrapper[4805]: I1203 14:42:32.508532 4805 generic.go:334] "Generic (PLEG): container finished" podID="dd64758f-9fe5-4f00-8f60-4bba03a0f98a" containerID="1550ff86130027e46c3614321fb5e00542e47fcdc0e81a35a399dd1e75e52ba8" exitCode=0 Dec 03 14:42:32 crc kubenswrapper[4805]: I1203 14:42:32.508654 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-7px67" event={"ID":"dd64758f-9fe5-4f00-8f60-4bba03a0f98a","Type":"ContainerDied","Data":"1550ff86130027e46c3614321fb5e00542e47fcdc0e81a35a399dd1e75e52ba8"} Dec 03 14:42:33 crc kubenswrapper[4805]: I1203 14:42:33.940554 4805 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-7px67" Dec 03 14:42:34 crc kubenswrapper[4805]: I1203 14:42:34.058870 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/dd64758f-9fe5-4f00-8f60-4bba03a0f98a-ssh-key\") pod \"dd64758f-9fe5-4f00-8f60-4bba03a0f98a\" (UID: \"dd64758f-9fe5-4f00-8f60-4bba03a0f98a\") " Dec 03 14:42:34 crc kubenswrapper[4805]: I1203 14:42:34.059138 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/dd64758f-9fe5-4f00-8f60-4bba03a0f98a-inventory\") pod \"dd64758f-9fe5-4f00-8f60-4bba03a0f98a\" (UID: \"dd64758f-9fe5-4f00-8f60-4bba03a0f98a\") " Dec 03 14:42:34 crc kubenswrapper[4805]: I1203 14:42:34.059218 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-28xxm\" (UniqueName: \"kubernetes.io/projected/dd64758f-9fe5-4f00-8f60-4bba03a0f98a-kube-api-access-28xxm\") pod \"dd64758f-9fe5-4f00-8f60-4bba03a0f98a\" (UID: \"dd64758f-9fe5-4f00-8f60-4bba03a0f98a\") " Dec 03 14:42:34 crc kubenswrapper[4805]: I1203 14:42:34.064610 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/dd64758f-9fe5-4f00-8f60-4bba03a0f98a-kube-api-access-28xxm" (OuterVolumeSpecName: "kube-api-access-28xxm") pod "dd64758f-9fe5-4f00-8f60-4bba03a0f98a" (UID: "dd64758f-9fe5-4f00-8f60-4bba03a0f98a"). InnerVolumeSpecName "kube-api-access-28xxm". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 14:42:34 crc kubenswrapper[4805]: I1203 14:42:34.087517 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/dd64758f-9fe5-4f00-8f60-4bba03a0f98a-inventory" (OuterVolumeSpecName: "inventory") pod "dd64758f-9fe5-4f00-8f60-4bba03a0f98a" (UID: "dd64758f-9fe5-4f00-8f60-4bba03a0f98a"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 14:42:34 crc kubenswrapper[4805]: I1203 14:42:34.088061 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/dd64758f-9fe5-4f00-8f60-4bba03a0f98a-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "dd64758f-9fe5-4f00-8f60-4bba03a0f98a" (UID: "dd64758f-9fe5-4f00-8f60-4bba03a0f98a"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 14:42:34 crc kubenswrapper[4805]: I1203 14:42:34.162121 4805 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/dd64758f-9fe5-4f00-8f60-4bba03a0f98a-inventory\") on node \"crc\" DevicePath \"\"" Dec 03 14:42:34 crc kubenswrapper[4805]: I1203 14:42:34.162183 4805 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-28xxm\" (UniqueName: \"kubernetes.io/projected/dd64758f-9fe5-4f00-8f60-4bba03a0f98a-kube-api-access-28xxm\") on node \"crc\" DevicePath \"\"" Dec 03 14:42:34 crc kubenswrapper[4805]: I1203 14:42:34.162201 4805 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/dd64758f-9fe5-4f00-8f60-4bba03a0f98a-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 03 14:42:34 crc kubenswrapper[4805]: I1203 14:42:34.532092 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-7px67" event={"ID":"dd64758f-9fe5-4f00-8f60-4bba03a0f98a","Type":"ContainerDied","Data":"75e01d2d924e8b5bddee770d9a81d76cd682e249576964b9162b64eb59893e24"} Dec 03 14:42:34 crc kubenswrapper[4805]: I1203 14:42:34.532133 4805 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="75e01d2d924e8b5bddee770d9a81d76cd682e249576964b9162b64eb59893e24" Dec 03 14:42:34 crc kubenswrapper[4805]: I1203 14:42:34.532148 4805 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-7px67" Dec 03 14:42:34 crc kubenswrapper[4805]: I1203 14:42:34.601656 4805 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/install-os-edpm-deployment-openstack-edpm-ipam-v79c2"] Dec 03 14:42:34 crc kubenswrapper[4805]: E1203 14:42:34.613820 4805 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="dd64758f-9fe5-4f00-8f60-4bba03a0f98a" containerName="validate-network-edpm-deployment-openstack-edpm-ipam" Dec 03 14:42:34 crc kubenswrapper[4805]: I1203 14:42:34.613882 4805 state_mem.go:107] "Deleted CPUSet assignment" podUID="dd64758f-9fe5-4f00-8f60-4bba03a0f98a" containerName="validate-network-edpm-deployment-openstack-edpm-ipam" Dec 03 14:42:34 crc kubenswrapper[4805]: I1203 14:42:34.614236 4805 memory_manager.go:354] "RemoveStaleState removing state" podUID="dd64758f-9fe5-4f00-8f60-4bba03a0f98a" containerName="validate-network-edpm-deployment-openstack-edpm-ipam" Dec 03 14:42:34 crc kubenswrapper[4805]: I1203 14:42:34.614908 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/install-os-edpm-deployment-openstack-edpm-ipam-v79c2"] Dec 03 14:42:34 crc kubenswrapper[4805]: I1203 14:42:34.615002 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-v79c2" Dec 03 14:42:34 crc kubenswrapper[4805]: I1203 14:42:34.618380 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Dec 03 14:42:34 crc kubenswrapper[4805]: I1203 14:42:34.618582 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-n7jt4" Dec 03 14:42:34 crc kubenswrapper[4805]: I1203 14:42:34.619380 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Dec 03 14:42:34 crc kubenswrapper[4805]: I1203 14:42:34.619612 4805 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 03 14:42:34 crc kubenswrapper[4805]: I1203 14:42:34.670821 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/b96a6fd3-b6a0-4580-9504-9ac4237ad37e-inventory\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-v79c2\" (UID: \"b96a6fd3-b6a0-4580-9504-9ac4237ad37e\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-v79c2" Dec 03 14:42:34 crc kubenswrapper[4805]: I1203 14:42:34.670967 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-s5bvr\" (UniqueName: \"kubernetes.io/projected/b96a6fd3-b6a0-4580-9504-9ac4237ad37e-kube-api-access-s5bvr\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-v79c2\" (UID: \"b96a6fd3-b6a0-4580-9504-9ac4237ad37e\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-v79c2" Dec 03 14:42:34 crc kubenswrapper[4805]: I1203 14:42:34.670996 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/b96a6fd3-b6a0-4580-9504-9ac4237ad37e-ssh-key\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-v79c2\" (UID: \"b96a6fd3-b6a0-4580-9504-9ac4237ad37e\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-v79c2" Dec 03 14:42:34 crc kubenswrapper[4805]: I1203 14:42:34.773045 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/b96a6fd3-b6a0-4580-9504-9ac4237ad37e-inventory\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-v79c2\" (UID: \"b96a6fd3-b6a0-4580-9504-9ac4237ad37e\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-v79c2" Dec 03 14:42:34 crc kubenswrapper[4805]: I1203 14:42:34.773107 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s5bvr\" (UniqueName: \"kubernetes.io/projected/b96a6fd3-b6a0-4580-9504-9ac4237ad37e-kube-api-access-s5bvr\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-v79c2\" (UID: \"b96a6fd3-b6a0-4580-9504-9ac4237ad37e\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-v79c2" Dec 03 14:42:34 crc kubenswrapper[4805]: I1203 14:42:34.773149 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/b96a6fd3-b6a0-4580-9504-9ac4237ad37e-ssh-key\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-v79c2\" (UID: \"b96a6fd3-b6a0-4580-9504-9ac4237ad37e\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-v79c2" Dec 03 14:42:34 crc kubenswrapper[4805]: I1203 14:42:34.779757 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/b96a6fd3-b6a0-4580-9504-9ac4237ad37e-ssh-key\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-v79c2\" (UID: \"b96a6fd3-b6a0-4580-9504-9ac4237ad37e\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-v79c2" Dec 03 14:42:34 crc kubenswrapper[4805]: I1203 14:42:34.780191 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/b96a6fd3-b6a0-4580-9504-9ac4237ad37e-inventory\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-v79c2\" (UID: \"b96a6fd3-b6a0-4580-9504-9ac4237ad37e\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-v79c2" Dec 03 14:42:34 crc kubenswrapper[4805]: I1203 14:42:34.795220 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s5bvr\" (UniqueName: \"kubernetes.io/projected/b96a6fd3-b6a0-4580-9504-9ac4237ad37e-kube-api-access-s5bvr\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-v79c2\" (UID: \"b96a6fd3-b6a0-4580-9504-9ac4237ad37e\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-v79c2" Dec 03 14:42:34 crc kubenswrapper[4805]: I1203 14:42:34.936686 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-v79c2" Dec 03 14:42:35 crc kubenswrapper[4805]: I1203 14:42:35.457952 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/install-os-edpm-deployment-openstack-edpm-ipam-v79c2"] Dec 03 14:42:35 crc kubenswrapper[4805]: I1203 14:42:35.558955 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-v79c2" event={"ID":"b96a6fd3-b6a0-4580-9504-9ac4237ad37e","Type":"ContainerStarted","Data":"b4d381cdffa9df560eeb03a7a07881ad3ad4ce9e89b36004788e1ece42488224"} Dec 03 14:42:36 crc kubenswrapper[4805]: I1203 14:42:36.568681 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-v79c2" event={"ID":"b96a6fd3-b6a0-4580-9504-9ac4237ad37e","Type":"ContainerStarted","Data":"2106282d69cb9c8ce4432a330e45180960aca07902585a41bf1df1c388d01654"} Dec 03 14:42:36 crc kubenswrapper[4805]: I1203 14:42:36.590107 4805 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-v79c2" podStartSLOduration=2.200057109 podStartE2EDuration="2.59008629s" podCreationTimestamp="2025-12-03 14:42:34 +0000 UTC" firstStartedPulling="2025-12-03 14:42:35.462557988 +0000 UTC m=+1985.125474911" lastFinishedPulling="2025-12-03 14:42:35.852587179 +0000 UTC m=+1985.515504092" observedRunningTime="2025-12-03 14:42:36.584788547 +0000 UTC m=+1986.247705470" watchObservedRunningTime="2025-12-03 14:42:36.59008629 +0000 UTC m=+1986.253003213" Dec 03 14:42:39 crc kubenswrapper[4805]: I1203 14:42:39.694530 4805 scope.go:117] "RemoveContainer" containerID="748e11b7efe02ff9a381aeee0aa4ddad32272e1564bc028d779c442a7b7f2149" Dec 03 14:42:39 crc kubenswrapper[4805]: E1203 14:42:39.695233 4805 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-gskh4_openshift-machine-config-operator(6098937f-e3f6-45e8-a647-4994a79cd711)\"" pod="openshift-machine-config-operator/machine-config-daemon-gskh4" podUID="6098937f-e3f6-45e8-a647-4994a79cd711" Dec 03 14:42:43 crc kubenswrapper[4805]: I1203 14:42:43.044116 4805 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-conductor-db-sync-dxrx8"] Dec 03 14:42:43 crc kubenswrapper[4805]: I1203 14:42:43.056234 4805 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-conductor-db-sync-dxrx8"] Dec 03 14:42:44 crc kubenswrapper[4805]: I1203 14:42:44.215252 4805 scope.go:117] "RemoveContainer" containerID="0ea9106d14a822eebf60173804f34994b649001c2b6d78c9d19a9f0dc68877ff" Dec 03 14:42:44 crc kubenswrapper[4805]: I1203 14:42:44.243972 4805 scope.go:117] "RemoveContainer" containerID="1ec90286975a522ff11e0855f75e40651d133383eee6046c19747bea819bb89d" Dec 03 14:42:44 crc kubenswrapper[4805]: I1203 14:42:44.296777 4805 scope.go:117] "RemoveContainer" containerID="40bf98d3f171d99179acd99a215011038ee2992502d009f80c701bf38be239b7" Dec 03 14:42:44 crc kubenswrapper[4805]: I1203 14:42:44.342143 4805 scope.go:117] "RemoveContainer" containerID="d095c57ea2be9ec1e49b311636249f5bb64be63c03bb0fa7f97830a9b479f1ab" Dec 03 14:42:44 crc kubenswrapper[4805]: I1203 14:42:44.391906 4805 scope.go:117] "RemoveContainer" containerID="832cf42c86bbb5add1b9f92b1884cf830d1330476cffa86cd81f8731103c155a" Dec 03 14:42:44 crc kubenswrapper[4805]: I1203 14:42:44.427260 4805 scope.go:117] "RemoveContainer" containerID="eda2ac15752bbe039e0576b6fba2c9ca5e27ae5381c8e8d3815692ad5a460f9b" Dec 03 14:42:44 crc kubenswrapper[4805]: I1203 14:42:44.714320 4805 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e56c5750-e0c6-4427-ae8c-ab4af0c9b7d0" path="/var/lib/kubelet/pods/e56c5750-e0c6-4427-ae8c-ab4af0c9b7d0/volumes" Dec 03 14:42:52 crc kubenswrapper[4805]: I1203 14:42:52.696084 4805 scope.go:117] "RemoveContainer" containerID="748e11b7efe02ff9a381aeee0aa4ddad32272e1564bc028d779c442a7b7f2149" Dec 03 14:42:57 crc kubenswrapper[4805]: I1203 14:42:57.774931 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-gskh4" event={"ID":"6098937f-e3f6-45e8-a647-4994a79cd711","Type":"ContainerStarted","Data":"339cd7bd4a35abfc7cd21edc654a8e6b31d75042fa72434198d7d952d589476a"} Dec 03 14:43:05 crc kubenswrapper[4805]: I1203 14:43:05.042503 4805 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-cell-mapping-bmkm9"] Dec 03 14:43:05 crc kubenswrapper[4805]: I1203 14:43:05.052785 4805 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-conductor-db-sync-42n9l"] Dec 03 14:43:05 crc kubenswrapper[4805]: I1203 14:43:05.062750 4805 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-conductor-db-sync-42n9l"] Dec 03 14:43:05 crc kubenswrapper[4805]: I1203 14:43:05.073057 4805 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-cell-mapping-bmkm9"] Dec 03 14:43:06 crc kubenswrapper[4805]: I1203 14:43:06.707150 4805 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="527667db-e5ab-4fe2-89a5-d9110602f1d2" path="/var/lib/kubelet/pods/527667db-e5ab-4fe2-89a5-d9110602f1d2/volumes" Dec 03 14:43:06 crc kubenswrapper[4805]: I1203 14:43:06.707970 4805 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="97cf9020-6417-4d08-9cdf-f38515b63d82" path="/var/lib/kubelet/pods/97cf9020-6417-4d08-9cdf-f38515b63d82/volumes" Dec 03 14:43:11 crc kubenswrapper[4805]: I1203 14:43:11.911164 4805 generic.go:334] "Generic (PLEG): container finished" podID="b96a6fd3-b6a0-4580-9504-9ac4237ad37e" containerID="2106282d69cb9c8ce4432a330e45180960aca07902585a41bf1df1c388d01654" exitCode=0 Dec 03 14:43:11 crc kubenswrapper[4805]: I1203 14:43:11.911263 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-v79c2" event={"ID":"b96a6fd3-b6a0-4580-9504-9ac4237ad37e","Type":"ContainerDied","Data":"2106282d69cb9c8ce4432a330e45180960aca07902585a41bf1df1c388d01654"} Dec 03 14:43:13 crc kubenswrapper[4805]: I1203 14:43:13.329925 4805 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-v79c2" Dec 03 14:43:13 crc kubenswrapper[4805]: I1203 14:43:13.422309 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/b96a6fd3-b6a0-4580-9504-9ac4237ad37e-ssh-key\") pod \"b96a6fd3-b6a0-4580-9504-9ac4237ad37e\" (UID: \"b96a6fd3-b6a0-4580-9504-9ac4237ad37e\") " Dec 03 14:43:13 crc kubenswrapper[4805]: I1203 14:43:13.422439 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-s5bvr\" (UniqueName: \"kubernetes.io/projected/b96a6fd3-b6a0-4580-9504-9ac4237ad37e-kube-api-access-s5bvr\") pod \"b96a6fd3-b6a0-4580-9504-9ac4237ad37e\" (UID: \"b96a6fd3-b6a0-4580-9504-9ac4237ad37e\") " Dec 03 14:43:13 crc kubenswrapper[4805]: I1203 14:43:13.422482 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/b96a6fd3-b6a0-4580-9504-9ac4237ad37e-inventory\") pod \"b96a6fd3-b6a0-4580-9504-9ac4237ad37e\" (UID: \"b96a6fd3-b6a0-4580-9504-9ac4237ad37e\") " Dec 03 14:43:13 crc kubenswrapper[4805]: I1203 14:43:13.429234 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b96a6fd3-b6a0-4580-9504-9ac4237ad37e-kube-api-access-s5bvr" (OuterVolumeSpecName: "kube-api-access-s5bvr") pod "b96a6fd3-b6a0-4580-9504-9ac4237ad37e" (UID: "b96a6fd3-b6a0-4580-9504-9ac4237ad37e"). InnerVolumeSpecName "kube-api-access-s5bvr". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 14:43:13 crc kubenswrapper[4805]: I1203 14:43:13.450631 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b96a6fd3-b6a0-4580-9504-9ac4237ad37e-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "b96a6fd3-b6a0-4580-9504-9ac4237ad37e" (UID: "b96a6fd3-b6a0-4580-9504-9ac4237ad37e"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 14:43:13 crc kubenswrapper[4805]: I1203 14:43:13.455454 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b96a6fd3-b6a0-4580-9504-9ac4237ad37e-inventory" (OuterVolumeSpecName: "inventory") pod "b96a6fd3-b6a0-4580-9504-9ac4237ad37e" (UID: "b96a6fd3-b6a0-4580-9504-9ac4237ad37e"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 14:43:13 crc kubenswrapper[4805]: I1203 14:43:13.524394 4805 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-s5bvr\" (UniqueName: \"kubernetes.io/projected/b96a6fd3-b6a0-4580-9504-9ac4237ad37e-kube-api-access-s5bvr\") on node \"crc\" DevicePath \"\"" Dec 03 14:43:13 crc kubenswrapper[4805]: I1203 14:43:13.524433 4805 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/b96a6fd3-b6a0-4580-9504-9ac4237ad37e-inventory\") on node \"crc\" DevicePath \"\"" Dec 03 14:43:13 crc kubenswrapper[4805]: I1203 14:43:13.524442 4805 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/b96a6fd3-b6a0-4580-9504-9ac4237ad37e-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 03 14:43:13 crc kubenswrapper[4805]: I1203 14:43:13.932974 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-v79c2" event={"ID":"b96a6fd3-b6a0-4580-9504-9ac4237ad37e","Type":"ContainerDied","Data":"b4d381cdffa9df560eeb03a7a07881ad3ad4ce9e89b36004788e1ece42488224"} Dec 03 14:43:13 crc kubenswrapper[4805]: I1203 14:43:13.933368 4805 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="b4d381cdffa9df560eeb03a7a07881ad3ad4ce9e89b36004788e1ece42488224" Dec 03 14:43:13 crc kubenswrapper[4805]: I1203 14:43:13.933448 4805 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-v79c2" Dec 03 14:43:14 crc kubenswrapper[4805]: I1203 14:43:14.011755 4805 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/configure-os-edpm-deployment-openstack-edpm-ipam-5bpz7"] Dec 03 14:43:14 crc kubenswrapper[4805]: E1203 14:43:14.012200 4805 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b96a6fd3-b6a0-4580-9504-9ac4237ad37e" containerName="install-os-edpm-deployment-openstack-edpm-ipam" Dec 03 14:43:14 crc kubenswrapper[4805]: I1203 14:43:14.012218 4805 state_mem.go:107] "Deleted CPUSet assignment" podUID="b96a6fd3-b6a0-4580-9504-9ac4237ad37e" containerName="install-os-edpm-deployment-openstack-edpm-ipam" Dec 03 14:43:14 crc kubenswrapper[4805]: I1203 14:43:14.012428 4805 memory_manager.go:354] "RemoveStaleState removing state" podUID="b96a6fd3-b6a0-4580-9504-9ac4237ad37e" containerName="install-os-edpm-deployment-openstack-edpm-ipam" Dec 03 14:43:14 crc kubenswrapper[4805]: I1203 14:43:14.013135 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-5bpz7" Dec 03 14:43:14 crc kubenswrapper[4805]: I1203 14:43:14.015713 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Dec 03 14:43:14 crc kubenswrapper[4805]: I1203 14:43:14.015890 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Dec 03 14:43:14 crc kubenswrapper[4805]: I1203 14:43:14.017326 4805 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 03 14:43:14 crc kubenswrapper[4805]: I1203 14:43:14.021728 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-n7jt4" Dec 03 14:43:14 crc kubenswrapper[4805]: I1203 14:43:14.030677 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/configure-os-edpm-deployment-openstack-edpm-ipam-5bpz7"] Dec 03 14:43:14 crc kubenswrapper[4805]: I1203 14:43:14.035912 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/2aa97cdd-3fbb-4e80-b82a-fb688c8c92d0-inventory\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-5bpz7\" (UID: \"2aa97cdd-3fbb-4e80-b82a-fb688c8c92d0\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-5bpz7" Dec 03 14:43:14 crc kubenswrapper[4805]: I1203 14:43:14.036014 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vzbmv\" (UniqueName: \"kubernetes.io/projected/2aa97cdd-3fbb-4e80-b82a-fb688c8c92d0-kube-api-access-vzbmv\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-5bpz7\" (UID: \"2aa97cdd-3fbb-4e80-b82a-fb688c8c92d0\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-5bpz7" Dec 03 14:43:14 crc kubenswrapper[4805]: I1203 14:43:14.036173 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/2aa97cdd-3fbb-4e80-b82a-fb688c8c92d0-ssh-key\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-5bpz7\" (UID: \"2aa97cdd-3fbb-4e80-b82a-fb688c8c92d0\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-5bpz7" Dec 03 14:43:14 crc kubenswrapper[4805]: I1203 14:43:14.137929 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vzbmv\" (UniqueName: \"kubernetes.io/projected/2aa97cdd-3fbb-4e80-b82a-fb688c8c92d0-kube-api-access-vzbmv\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-5bpz7\" (UID: \"2aa97cdd-3fbb-4e80-b82a-fb688c8c92d0\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-5bpz7" Dec 03 14:43:14 crc kubenswrapper[4805]: I1203 14:43:14.138154 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/2aa97cdd-3fbb-4e80-b82a-fb688c8c92d0-ssh-key\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-5bpz7\" (UID: \"2aa97cdd-3fbb-4e80-b82a-fb688c8c92d0\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-5bpz7" Dec 03 14:43:14 crc kubenswrapper[4805]: I1203 14:43:14.138216 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/2aa97cdd-3fbb-4e80-b82a-fb688c8c92d0-inventory\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-5bpz7\" (UID: \"2aa97cdd-3fbb-4e80-b82a-fb688c8c92d0\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-5bpz7" Dec 03 14:43:14 crc kubenswrapper[4805]: I1203 14:43:14.143165 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/2aa97cdd-3fbb-4e80-b82a-fb688c8c92d0-ssh-key\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-5bpz7\" (UID: \"2aa97cdd-3fbb-4e80-b82a-fb688c8c92d0\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-5bpz7" Dec 03 14:43:14 crc kubenswrapper[4805]: I1203 14:43:14.143253 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/2aa97cdd-3fbb-4e80-b82a-fb688c8c92d0-inventory\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-5bpz7\" (UID: \"2aa97cdd-3fbb-4e80-b82a-fb688c8c92d0\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-5bpz7" Dec 03 14:43:14 crc kubenswrapper[4805]: I1203 14:43:14.155146 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vzbmv\" (UniqueName: \"kubernetes.io/projected/2aa97cdd-3fbb-4e80-b82a-fb688c8c92d0-kube-api-access-vzbmv\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-5bpz7\" (UID: \"2aa97cdd-3fbb-4e80-b82a-fb688c8c92d0\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-5bpz7" Dec 03 14:43:14 crc kubenswrapper[4805]: I1203 14:43:14.333153 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-5bpz7" Dec 03 14:43:14 crc kubenswrapper[4805]: I1203 14:43:14.833816 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/configure-os-edpm-deployment-openstack-edpm-ipam-5bpz7"] Dec 03 14:43:14 crc kubenswrapper[4805]: I1203 14:43:14.943549 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-5bpz7" event={"ID":"2aa97cdd-3fbb-4e80-b82a-fb688c8c92d0","Type":"ContainerStarted","Data":"ee5cdad79b47792fdfc1cbe57126d8c262ba14df8a69bec3f074430f6f730a47"} Dec 03 14:43:15 crc kubenswrapper[4805]: I1203 14:43:15.952912 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-5bpz7" event={"ID":"2aa97cdd-3fbb-4e80-b82a-fb688c8c92d0","Type":"ContainerStarted","Data":"aefe1d37818c2bb62a20be03527b70fc1369e947210c594a12365d8531b28cb1"} Dec 03 14:43:15 crc kubenswrapper[4805]: I1203 14:43:15.974956 4805 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-5bpz7" podStartSLOduration=2.536582357 podStartE2EDuration="2.974940424s" podCreationTimestamp="2025-12-03 14:43:13 +0000 UTC" firstStartedPulling="2025-12-03 14:43:14.838766979 +0000 UTC m=+2024.501683902" lastFinishedPulling="2025-12-03 14:43:15.277125046 +0000 UTC m=+2024.940041969" observedRunningTime="2025-12-03 14:43:15.970633948 +0000 UTC m=+2025.633550871" watchObservedRunningTime="2025-12-03 14:43:15.974940424 +0000 UTC m=+2025.637857347" Dec 03 14:43:44 crc kubenswrapper[4805]: I1203 14:43:44.589706 4805 scope.go:117] "RemoveContainer" containerID="855ce4dd11138f5cb712a02c8377e47248650a1e2a6401e002ad634a4a558325" Dec 03 14:43:44 crc kubenswrapper[4805]: I1203 14:43:44.645311 4805 scope.go:117] "RemoveContainer" containerID="8adea027cee74f522316a8abd4ae6407b8f8c77128bc6cb459c352fdf09df968" Dec 03 14:43:44 crc kubenswrapper[4805]: I1203 14:43:44.695232 4805 scope.go:117] "RemoveContainer" containerID="3ebc3bcafd06fc724ece768dc68192d7ce17f5f76321fbde3f409727992f5ea4" Dec 03 14:43:50 crc kubenswrapper[4805]: I1203 14:43:50.058040 4805 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-cell-mapping-pkj2h"] Dec 03 14:43:50 crc kubenswrapper[4805]: I1203 14:43:50.068587 4805 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-cell-mapping-pkj2h"] Dec 03 14:43:50 crc kubenswrapper[4805]: I1203 14:43:50.707331 4805 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="cae0a70b-e47c-40ff-9b40-40a2aa570097" path="/var/lib/kubelet/pods/cae0a70b-e47c-40ff-9b40-40a2aa570097/volumes" Dec 03 14:44:04 crc kubenswrapper[4805]: I1203 14:44:04.379217 4805 generic.go:334] "Generic (PLEG): container finished" podID="2aa97cdd-3fbb-4e80-b82a-fb688c8c92d0" containerID="aefe1d37818c2bb62a20be03527b70fc1369e947210c594a12365d8531b28cb1" exitCode=0 Dec 03 14:44:04 crc kubenswrapper[4805]: I1203 14:44:04.379308 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-5bpz7" event={"ID":"2aa97cdd-3fbb-4e80-b82a-fb688c8c92d0","Type":"ContainerDied","Data":"aefe1d37818c2bb62a20be03527b70fc1369e947210c594a12365d8531b28cb1"} Dec 03 14:44:04 crc kubenswrapper[4805]: I1203 14:44:04.510370 4805 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-6m8lt"] Dec 03 14:44:04 crc kubenswrapper[4805]: I1203 14:44:04.512755 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-6m8lt" Dec 03 14:44:04 crc kubenswrapper[4805]: I1203 14:44:04.521872 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-6m8lt"] Dec 03 14:44:04 crc kubenswrapper[4805]: I1203 14:44:04.675274 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1737023c-c5dc-41b4-b11d-e031f4e67d7a-utilities\") pod \"redhat-marketplace-6m8lt\" (UID: \"1737023c-c5dc-41b4-b11d-e031f4e67d7a\") " pod="openshift-marketplace/redhat-marketplace-6m8lt" Dec 03 14:44:04 crc kubenswrapper[4805]: I1203 14:44:04.675342 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-j6lwd\" (UniqueName: \"kubernetes.io/projected/1737023c-c5dc-41b4-b11d-e031f4e67d7a-kube-api-access-j6lwd\") pod \"redhat-marketplace-6m8lt\" (UID: \"1737023c-c5dc-41b4-b11d-e031f4e67d7a\") " pod="openshift-marketplace/redhat-marketplace-6m8lt" Dec 03 14:44:04 crc kubenswrapper[4805]: I1203 14:44:04.675417 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1737023c-c5dc-41b4-b11d-e031f4e67d7a-catalog-content\") pod \"redhat-marketplace-6m8lt\" (UID: \"1737023c-c5dc-41b4-b11d-e031f4e67d7a\") " pod="openshift-marketplace/redhat-marketplace-6m8lt" Dec 03 14:44:04 crc kubenswrapper[4805]: I1203 14:44:04.777275 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1737023c-c5dc-41b4-b11d-e031f4e67d7a-utilities\") pod \"redhat-marketplace-6m8lt\" (UID: \"1737023c-c5dc-41b4-b11d-e031f4e67d7a\") " pod="openshift-marketplace/redhat-marketplace-6m8lt" Dec 03 14:44:04 crc kubenswrapper[4805]: I1203 14:44:04.777350 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-j6lwd\" (UniqueName: \"kubernetes.io/projected/1737023c-c5dc-41b4-b11d-e031f4e67d7a-kube-api-access-j6lwd\") pod \"redhat-marketplace-6m8lt\" (UID: \"1737023c-c5dc-41b4-b11d-e031f4e67d7a\") " pod="openshift-marketplace/redhat-marketplace-6m8lt" Dec 03 14:44:04 crc kubenswrapper[4805]: I1203 14:44:04.777429 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1737023c-c5dc-41b4-b11d-e031f4e67d7a-catalog-content\") pod \"redhat-marketplace-6m8lt\" (UID: \"1737023c-c5dc-41b4-b11d-e031f4e67d7a\") " pod="openshift-marketplace/redhat-marketplace-6m8lt" Dec 03 14:44:04 crc kubenswrapper[4805]: I1203 14:44:04.777917 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1737023c-c5dc-41b4-b11d-e031f4e67d7a-utilities\") pod \"redhat-marketplace-6m8lt\" (UID: \"1737023c-c5dc-41b4-b11d-e031f4e67d7a\") " pod="openshift-marketplace/redhat-marketplace-6m8lt" Dec 03 14:44:04 crc kubenswrapper[4805]: I1203 14:44:04.777935 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1737023c-c5dc-41b4-b11d-e031f4e67d7a-catalog-content\") pod \"redhat-marketplace-6m8lt\" (UID: \"1737023c-c5dc-41b4-b11d-e031f4e67d7a\") " pod="openshift-marketplace/redhat-marketplace-6m8lt" Dec 03 14:44:04 crc kubenswrapper[4805]: I1203 14:44:04.804793 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-j6lwd\" (UniqueName: \"kubernetes.io/projected/1737023c-c5dc-41b4-b11d-e031f4e67d7a-kube-api-access-j6lwd\") pod \"redhat-marketplace-6m8lt\" (UID: \"1737023c-c5dc-41b4-b11d-e031f4e67d7a\") " pod="openshift-marketplace/redhat-marketplace-6m8lt" Dec 03 14:44:04 crc kubenswrapper[4805]: I1203 14:44:04.851293 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-6m8lt" Dec 03 14:44:05 crc kubenswrapper[4805]: I1203 14:44:05.332959 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-6m8lt"] Dec 03 14:44:05 crc kubenswrapper[4805]: I1203 14:44:05.391741 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-6m8lt" event={"ID":"1737023c-c5dc-41b4-b11d-e031f4e67d7a","Type":"ContainerStarted","Data":"7f354d9cda99e0a0cd18edd18fd45b4f156b57ed4fa246de087903f206e7e106"} Dec 03 14:44:05 crc kubenswrapper[4805]: I1203 14:44:05.714099 4805 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-5bpz7" Dec 03 14:44:05 crc kubenswrapper[4805]: I1203 14:44:05.798631 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/2aa97cdd-3fbb-4e80-b82a-fb688c8c92d0-inventory\") pod \"2aa97cdd-3fbb-4e80-b82a-fb688c8c92d0\" (UID: \"2aa97cdd-3fbb-4e80-b82a-fb688c8c92d0\") " Dec 03 14:44:05 crc kubenswrapper[4805]: I1203 14:44:05.799132 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/2aa97cdd-3fbb-4e80-b82a-fb688c8c92d0-ssh-key\") pod \"2aa97cdd-3fbb-4e80-b82a-fb688c8c92d0\" (UID: \"2aa97cdd-3fbb-4e80-b82a-fb688c8c92d0\") " Dec 03 14:44:05 crc kubenswrapper[4805]: I1203 14:44:05.799169 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vzbmv\" (UniqueName: \"kubernetes.io/projected/2aa97cdd-3fbb-4e80-b82a-fb688c8c92d0-kube-api-access-vzbmv\") pod \"2aa97cdd-3fbb-4e80-b82a-fb688c8c92d0\" (UID: \"2aa97cdd-3fbb-4e80-b82a-fb688c8c92d0\") " Dec 03 14:44:05 crc kubenswrapper[4805]: I1203 14:44:05.804470 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2aa97cdd-3fbb-4e80-b82a-fb688c8c92d0-kube-api-access-vzbmv" (OuterVolumeSpecName: "kube-api-access-vzbmv") pod "2aa97cdd-3fbb-4e80-b82a-fb688c8c92d0" (UID: "2aa97cdd-3fbb-4e80-b82a-fb688c8c92d0"). InnerVolumeSpecName "kube-api-access-vzbmv". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 14:44:05 crc kubenswrapper[4805]: I1203 14:44:05.834954 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2aa97cdd-3fbb-4e80-b82a-fb688c8c92d0-inventory" (OuterVolumeSpecName: "inventory") pod "2aa97cdd-3fbb-4e80-b82a-fb688c8c92d0" (UID: "2aa97cdd-3fbb-4e80-b82a-fb688c8c92d0"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 14:44:05 crc kubenswrapper[4805]: I1203 14:44:05.836112 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2aa97cdd-3fbb-4e80-b82a-fb688c8c92d0-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "2aa97cdd-3fbb-4e80-b82a-fb688c8c92d0" (UID: "2aa97cdd-3fbb-4e80-b82a-fb688c8c92d0"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 14:44:05 crc kubenswrapper[4805]: I1203 14:44:05.900362 4805 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/2aa97cdd-3fbb-4e80-b82a-fb688c8c92d0-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 03 14:44:05 crc kubenswrapper[4805]: I1203 14:44:05.900396 4805 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vzbmv\" (UniqueName: \"kubernetes.io/projected/2aa97cdd-3fbb-4e80-b82a-fb688c8c92d0-kube-api-access-vzbmv\") on node \"crc\" DevicePath \"\"" Dec 03 14:44:05 crc kubenswrapper[4805]: I1203 14:44:05.900409 4805 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/2aa97cdd-3fbb-4e80-b82a-fb688c8c92d0-inventory\") on node \"crc\" DevicePath \"\"" Dec 03 14:44:06 crc kubenswrapper[4805]: I1203 14:44:06.400238 4805 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-5bpz7" Dec 03 14:44:06 crc kubenswrapper[4805]: I1203 14:44:06.400232 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-5bpz7" event={"ID":"2aa97cdd-3fbb-4e80-b82a-fb688c8c92d0","Type":"ContainerDied","Data":"ee5cdad79b47792fdfc1cbe57126d8c262ba14df8a69bec3f074430f6f730a47"} Dec 03 14:44:06 crc kubenswrapper[4805]: I1203 14:44:06.400338 4805 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="ee5cdad79b47792fdfc1cbe57126d8c262ba14df8a69bec3f074430f6f730a47" Dec 03 14:44:06 crc kubenswrapper[4805]: I1203 14:44:06.401891 4805 generic.go:334] "Generic (PLEG): container finished" podID="1737023c-c5dc-41b4-b11d-e031f4e67d7a" containerID="bc681223f9f69de6f7324bac7b7155e8568406036e2f6cbfc863fa9932ebda03" exitCode=0 Dec 03 14:44:06 crc kubenswrapper[4805]: I1203 14:44:06.401936 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-6m8lt" event={"ID":"1737023c-c5dc-41b4-b11d-e031f4e67d7a","Type":"ContainerDied","Data":"bc681223f9f69de6f7324bac7b7155e8568406036e2f6cbfc863fa9932ebda03"} Dec 03 14:44:06 crc kubenswrapper[4805]: I1203 14:44:06.489013 4805 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ssh-known-hosts-edpm-deployment-hzqxj"] Dec 03 14:44:06 crc kubenswrapper[4805]: E1203 14:44:06.489868 4805 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2aa97cdd-3fbb-4e80-b82a-fb688c8c92d0" containerName="configure-os-edpm-deployment-openstack-edpm-ipam" Dec 03 14:44:06 crc kubenswrapper[4805]: I1203 14:44:06.489898 4805 state_mem.go:107] "Deleted CPUSet assignment" podUID="2aa97cdd-3fbb-4e80-b82a-fb688c8c92d0" containerName="configure-os-edpm-deployment-openstack-edpm-ipam" Dec 03 14:44:06 crc kubenswrapper[4805]: I1203 14:44:06.490206 4805 memory_manager.go:354] "RemoveStaleState removing state" podUID="2aa97cdd-3fbb-4e80-b82a-fb688c8c92d0" containerName="configure-os-edpm-deployment-openstack-edpm-ipam" Dec 03 14:44:06 crc kubenswrapper[4805]: I1203 14:44:06.491047 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ssh-known-hosts-edpm-deployment-hzqxj" Dec 03 14:44:06 crc kubenswrapper[4805]: I1203 14:44:06.493341 4805 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 03 14:44:06 crc kubenswrapper[4805]: I1203 14:44:06.493667 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Dec 03 14:44:06 crc kubenswrapper[4805]: I1203 14:44:06.493955 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-n7jt4" Dec 03 14:44:06 crc kubenswrapper[4805]: I1203 14:44:06.496337 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Dec 03 14:44:06 crc kubenswrapper[4805]: I1203 14:44:06.509830 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ssh-known-hosts-edpm-deployment-hzqxj"] Dec 03 14:44:06 crc kubenswrapper[4805]: I1203 14:44:06.511945 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/5221f418-44bd-4db1-96c8-788e0df137c7-ssh-key-openstack-edpm-ipam\") pod \"ssh-known-hosts-edpm-deployment-hzqxj\" (UID: \"5221f418-44bd-4db1-96c8-788e0df137c7\") " pod="openstack/ssh-known-hosts-edpm-deployment-hzqxj" Dec 03 14:44:06 crc kubenswrapper[4805]: I1203 14:44:06.512056 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/5221f418-44bd-4db1-96c8-788e0df137c7-inventory-0\") pod \"ssh-known-hosts-edpm-deployment-hzqxj\" (UID: \"5221f418-44bd-4db1-96c8-788e0df137c7\") " pod="openstack/ssh-known-hosts-edpm-deployment-hzqxj" Dec 03 14:44:06 crc kubenswrapper[4805]: I1203 14:44:06.512361 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xsrkl\" (UniqueName: \"kubernetes.io/projected/5221f418-44bd-4db1-96c8-788e0df137c7-kube-api-access-xsrkl\") pod \"ssh-known-hosts-edpm-deployment-hzqxj\" (UID: \"5221f418-44bd-4db1-96c8-788e0df137c7\") " pod="openstack/ssh-known-hosts-edpm-deployment-hzqxj" Dec 03 14:44:06 crc kubenswrapper[4805]: I1203 14:44:06.613856 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/5221f418-44bd-4db1-96c8-788e0df137c7-ssh-key-openstack-edpm-ipam\") pod \"ssh-known-hosts-edpm-deployment-hzqxj\" (UID: \"5221f418-44bd-4db1-96c8-788e0df137c7\") " pod="openstack/ssh-known-hosts-edpm-deployment-hzqxj" Dec 03 14:44:06 crc kubenswrapper[4805]: I1203 14:44:06.614004 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/5221f418-44bd-4db1-96c8-788e0df137c7-inventory-0\") pod \"ssh-known-hosts-edpm-deployment-hzqxj\" (UID: \"5221f418-44bd-4db1-96c8-788e0df137c7\") " pod="openstack/ssh-known-hosts-edpm-deployment-hzqxj" Dec 03 14:44:06 crc kubenswrapper[4805]: I1203 14:44:06.614099 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xsrkl\" (UniqueName: \"kubernetes.io/projected/5221f418-44bd-4db1-96c8-788e0df137c7-kube-api-access-xsrkl\") pod \"ssh-known-hosts-edpm-deployment-hzqxj\" (UID: \"5221f418-44bd-4db1-96c8-788e0df137c7\") " pod="openstack/ssh-known-hosts-edpm-deployment-hzqxj" Dec 03 14:44:06 crc kubenswrapper[4805]: I1203 14:44:06.619745 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/5221f418-44bd-4db1-96c8-788e0df137c7-ssh-key-openstack-edpm-ipam\") pod \"ssh-known-hosts-edpm-deployment-hzqxj\" (UID: \"5221f418-44bd-4db1-96c8-788e0df137c7\") " pod="openstack/ssh-known-hosts-edpm-deployment-hzqxj" Dec 03 14:44:06 crc kubenswrapper[4805]: I1203 14:44:06.619754 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/5221f418-44bd-4db1-96c8-788e0df137c7-inventory-0\") pod \"ssh-known-hosts-edpm-deployment-hzqxj\" (UID: \"5221f418-44bd-4db1-96c8-788e0df137c7\") " pod="openstack/ssh-known-hosts-edpm-deployment-hzqxj" Dec 03 14:44:06 crc kubenswrapper[4805]: I1203 14:44:06.630642 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xsrkl\" (UniqueName: \"kubernetes.io/projected/5221f418-44bd-4db1-96c8-788e0df137c7-kube-api-access-xsrkl\") pod \"ssh-known-hosts-edpm-deployment-hzqxj\" (UID: \"5221f418-44bd-4db1-96c8-788e0df137c7\") " pod="openstack/ssh-known-hosts-edpm-deployment-hzqxj" Dec 03 14:44:06 crc kubenswrapper[4805]: I1203 14:44:06.806763 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ssh-known-hosts-edpm-deployment-hzqxj" Dec 03 14:44:07 crc kubenswrapper[4805]: I1203 14:44:07.314331 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ssh-known-hosts-edpm-deployment-hzqxj"] Dec 03 14:44:07 crc kubenswrapper[4805]: W1203 14:44:07.314613 4805 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod5221f418_44bd_4db1_96c8_788e0df137c7.slice/crio-ab5a5143371ccb9fdefccf000b66e6b7aaf8865ccc2f184003710cbc85d31479 WatchSource:0}: Error finding container ab5a5143371ccb9fdefccf000b66e6b7aaf8865ccc2f184003710cbc85d31479: Status 404 returned error can't find the container with id ab5a5143371ccb9fdefccf000b66e6b7aaf8865ccc2f184003710cbc85d31479 Dec 03 14:44:07 crc kubenswrapper[4805]: I1203 14:44:07.420065 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ssh-known-hosts-edpm-deployment-hzqxj" event={"ID":"5221f418-44bd-4db1-96c8-788e0df137c7","Type":"ContainerStarted","Data":"ab5a5143371ccb9fdefccf000b66e6b7aaf8865ccc2f184003710cbc85d31479"} Dec 03 14:44:07 crc kubenswrapper[4805]: I1203 14:44:07.426132 4805 generic.go:334] "Generic (PLEG): container finished" podID="1737023c-c5dc-41b4-b11d-e031f4e67d7a" containerID="f05f785bd4da68f8b30682d25646ba4a0460ebb6718765b82395b7f2189febc1" exitCode=0 Dec 03 14:44:07 crc kubenswrapper[4805]: I1203 14:44:07.426180 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-6m8lt" event={"ID":"1737023c-c5dc-41b4-b11d-e031f4e67d7a","Type":"ContainerDied","Data":"f05f785bd4da68f8b30682d25646ba4a0460ebb6718765b82395b7f2189febc1"} Dec 03 14:44:08 crc kubenswrapper[4805]: I1203 14:44:08.440180 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ssh-known-hosts-edpm-deployment-hzqxj" event={"ID":"5221f418-44bd-4db1-96c8-788e0df137c7","Type":"ContainerStarted","Data":"ea0def5bf69998c46286e1454d896ea32462a6ab9aa5aa4a250581097aecbb93"} Dec 03 14:44:08 crc kubenswrapper[4805]: I1203 14:44:08.443246 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-6m8lt" event={"ID":"1737023c-c5dc-41b4-b11d-e031f4e67d7a","Type":"ContainerStarted","Data":"db19dc7bd2fd1b0f62e18b0d3abce4a707aa1570793f588e24ac7cc7609562c9"} Dec 03 14:44:08 crc kubenswrapper[4805]: I1203 14:44:08.455659 4805 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ssh-known-hosts-edpm-deployment-hzqxj" podStartSLOduration=2.048519451 podStartE2EDuration="2.45564297s" podCreationTimestamp="2025-12-03 14:44:06 +0000 UTC" firstStartedPulling="2025-12-03 14:44:07.318148565 +0000 UTC m=+2076.981065498" lastFinishedPulling="2025-12-03 14:44:07.725272094 +0000 UTC m=+2077.388189017" observedRunningTime="2025-12-03 14:44:08.453537603 +0000 UTC m=+2078.116454546" watchObservedRunningTime="2025-12-03 14:44:08.45564297 +0000 UTC m=+2078.118559903" Dec 03 14:44:08 crc kubenswrapper[4805]: I1203 14:44:08.478604 4805 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-6m8lt" podStartSLOduration=3.04916874 podStartE2EDuration="4.478580321s" podCreationTimestamp="2025-12-03 14:44:04 +0000 UTC" firstStartedPulling="2025-12-03 14:44:06.404109111 +0000 UTC m=+2076.067026024" lastFinishedPulling="2025-12-03 14:44:07.833520682 +0000 UTC m=+2077.496437605" observedRunningTime="2025-12-03 14:44:08.469053503 +0000 UTC m=+2078.131970456" watchObservedRunningTime="2025-12-03 14:44:08.478580321 +0000 UTC m=+2078.141497254" Dec 03 14:44:14 crc kubenswrapper[4805]: I1203 14:44:14.852031 4805 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-6m8lt" Dec 03 14:44:14 crc kubenswrapper[4805]: I1203 14:44:14.852603 4805 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-6m8lt" Dec 03 14:44:14 crc kubenswrapper[4805]: I1203 14:44:14.901122 4805 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-6m8lt" Dec 03 14:44:15 crc kubenswrapper[4805]: I1203 14:44:15.502764 4805 generic.go:334] "Generic (PLEG): container finished" podID="5221f418-44bd-4db1-96c8-788e0df137c7" containerID="ea0def5bf69998c46286e1454d896ea32462a6ab9aa5aa4a250581097aecbb93" exitCode=0 Dec 03 14:44:15 crc kubenswrapper[4805]: I1203 14:44:15.504238 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ssh-known-hosts-edpm-deployment-hzqxj" event={"ID":"5221f418-44bd-4db1-96c8-788e0df137c7","Type":"ContainerDied","Data":"ea0def5bf69998c46286e1454d896ea32462a6ab9aa5aa4a250581097aecbb93"} Dec 03 14:44:15 crc kubenswrapper[4805]: I1203 14:44:15.573261 4805 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-6m8lt" Dec 03 14:44:15 crc kubenswrapper[4805]: I1203 14:44:15.616784 4805 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-6m8lt"] Dec 03 14:44:16 crc kubenswrapper[4805]: I1203 14:44:16.909077 4805 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ssh-known-hosts-edpm-deployment-hzqxj" Dec 03 14:44:16 crc kubenswrapper[4805]: I1203 14:44:16.928149 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/5221f418-44bd-4db1-96c8-788e0df137c7-ssh-key-openstack-edpm-ipam\") pod \"5221f418-44bd-4db1-96c8-788e0df137c7\" (UID: \"5221f418-44bd-4db1-96c8-788e0df137c7\") " Dec 03 14:44:16 crc kubenswrapper[4805]: I1203 14:44:16.928782 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/5221f418-44bd-4db1-96c8-788e0df137c7-inventory-0\") pod \"5221f418-44bd-4db1-96c8-788e0df137c7\" (UID: \"5221f418-44bd-4db1-96c8-788e0df137c7\") " Dec 03 14:44:16 crc kubenswrapper[4805]: I1203 14:44:16.928864 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xsrkl\" (UniqueName: \"kubernetes.io/projected/5221f418-44bd-4db1-96c8-788e0df137c7-kube-api-access-xsrkl\") pod \"5221f418-44bd-4db1-96c8-788e0df137c7\" (UID: \"5221f418-44bd-4db1-96c8-788e0df137c7\") " Dec 03 14:44:16 crc kubenswrapper[4805]: I1203 14:44:16.934343 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5221f418-44bd-4db1-96c8-788e0df137c7-kube-api-access-xsrkl" (OuterVolumeSpecName: "kube-api-access-xsrkl") pod "5221f418-44bd-4db1-96c8-788e0df137c7" (UID: "5221f418-44bd-4db1-96c8-788e0df137c7"). InnerVolumeSpecName "kube-api-access-xsrkl". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 14:44:16 crc kubenswrapper[4805]: I1203 14:44:16.957209 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5221f418-44bd-4db1-96c8-788e0df137c7-ssh-key-openstack-edpm-ipam" (OuterVolumeSpecName: "ssh-key-openstack-edpm-ipam") pod "5221f418-44bd-4db1-96c8-788e0df137c7" (UID: "5221f418-44bd-4db1-96c8-788e0df137c7"). InnerVolumeSpecName "ssh-key-openstack-edpm-ipam". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 14:44:16 crc kubenswrapper[4805]: I1203 14:44:16.958493 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5221f418-44bd-4db1-96c8-788e0df137c7-inventory-0" (OuterVolumeSpecName: "inventory-0") pod "5221f418-44bd-4db1-96c8-788e0df137c7" (UID: "5221f418-44bd-4db1-96c8-788e0df137c7"). InnerVolumeSpecName "inventory-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 14:44:17 crc kubenswrapper[4805]: I1203 14:44:17.032905 4805 reconciler_common.go:293] "Volume detached for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/5221f418-44bd-4db1-96c8-788e0df137c7-ssh-key-openstack-edpm-ipam\") on node \"crc\" DevicePath \"\"" Dec 03 14:44:17 crc kubenswrapper[4805]: I1203 14:44:17.033022 4805 reconciler_common.go:293] "Volume detached for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/5221f418-44bd-4db1-96c8-788e0df137c7-inventory-0\") on node \"crc\" DevicePath \"\"" Dec 03 14:44:17 crc kubenswrapper[4805]: I1203 14:44:17.033033 4805 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xsrkl\" (UniqueName: \"kubernetes.io/projected/5221f418-44bd-4db1-96c8-788e0df137c7-kube-api-access-xsrkl\") on node \"crc\" DevicePath \"\"" Dec 03 14:44:17 crc kubenswrapper[4805]: I1203 14:44:17.522563 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ssh-known-hosts-edpm-deployment-hzqxj" event={"ID":"5221f418-44bd-4db1-96c8-788e0df137c7","Type":"ContainerDied","Data":"ab5a5143371ccb9fdefccf000b66e6b7aaf8865ccc2f184003710cbc85d31479"} Dec 03 14:44:17 crc kubenswrapper[4805]: I1203 14:44:17.522940 4805 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="ab5a5143371ccb9fdefccf000b66e6b7aaf8865ccc2f184003710cbc85d31479" Dec 03 14:44:17 crc kubenswrapper[4805]: I1203 14:44:17.522729 4805 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-6m8lt" podUID="1737023c-c5dc-41b4-b11d-e031f4e67d7a" containerName="registry-server" containerID="cri-o://db19dc7bd2fd1b0f62e18b0d3abce4a707aa1570793f588e24ac7cc7609562c9" gracePeriod=2 Dec 03 14:44:17 crc kubenswrapper[4805]: I1203 14:44:17.522602 4805 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ssh-known-hosts-edpm-deployment-hzqxj" Dec 03 14:44:17 crc kubenswrapper[4805]: I1203 14:44:17.642449 4805 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/run-os-edpm-deployment-openstack-edpm-ipam-lp9h9"] Dec 03 14:44:17 crc kubenswrapper[4805]: E1203 14:44:17.642797 4805 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5221f418-44bd-4db1-96c8-788e0df137c7" containerName="ssh-known-hosts-edpm-deployment" Dec 03 14:44:17 crc kubenswrapper[4805]: I1203 14:44:17.642812 4805 state_mem.go:107] "Deleted CPUSet assignment" podUID="5221f418-44bd-4db1-96c8-788e0df137c7" containerName="ssh-known-hosts-edpm-deployment" Dec 03 14:44:17 crc kubenswrapper[4805]: I1203 14:44:17.643008 4805 memory_manager.go:354] "RemoveStaleState removing state" podUID="5221f418-44bd-4db1-96c8-788e0df137c7" containerName="ssh-known-hosts-edpm-deployment" Dec 03 14:44:17 crc kubenswrapper[4805]: I1203 14:44:17.643639 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-lp9h9" Dec 03 14:44:17 crc kubenswrapper[4805]: I1203 14:44:17.646499 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Dec 03 14:44:17 crc kubenswrapper[4805]: I1203 14:44:17.646935 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Dec 03 14:44:17 crc kubenswrapper[4805]: I1203 14:44:17.647974 4805 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 03 14:44:17 crc kubenswrapper[4805]: I1203 14:44:17.648422 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-n7jt4" Dec 03 14:44:17 crc kubenswrapper[4805]: I1203 14:44:17.660719 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/run-os-edpm-deployment-openstack-edpm-ipam-lp9h9"] Dec 03 14:44:17 crc kubenswrapper[4805]: I1203 14:44:17.744984 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/3183d9eb-f2b5-4ac6-8cb1-917cff8a8ded-inventory\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-lp9h9\" (UID: \"3183d9eb-f2b5-4ac6-8cb1-917cff8a8ded\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-lp9h9" Dec 03 14:44:17 crc kubenswrapper[4805]: I1203 14:44:17.745087 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/3183d9eb-f2b5-4ac6-8cb1-917cff8a8ded-ssh-key\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-lp9h9\" (UID: \"3183d9eb-f2b5-4ac6-8cb1-917cff8a8ded\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-lp9h9" Dec 03 14:44:17 crc kubenswrapper[4805]: I1203 14:44:17.745115 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dmhsm\" (UniqueName: \"kubernetes.io/projected/3183d9eb-f2b5-4ac6-8cb1-917cff8a8ded-kube-api-access-dmhsm\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-lp9h9\" (UID: \"3183d9eb-f2b5-4ac6-8cb1-917cff8a8ded\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-lp9h9" Dec 03 14:44:17 crc kubenswrapper[4805]: I1203 14:44:17.846529 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/3183d9eb-f2b5-4ac6-8cb1-917cff8a8ded-inventory\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-lp9h9\" (UID: \"3183d9eb-f2b5-4ac6-8cb1-917cff8a8ded\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-lp9h9" Dec 03 14:44:17 crc kubenswrapper[4805]: I1203 14:44:17.846598 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/3183d9eb-f2b5-4ac6-8cb1-917cff8a8ded-ssh-key\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-lp9h9\" (UID: \"3183d9eb-f2b5-4ac6-8cb1-917cff8a8ded\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-lp9h9" Dec 03 14:44:17 crc kubenswrapper[4805]: I1203 14:44:17.846625 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dmhsm\" (UniqueName: \"kubernetes.io/projected/3183d9eb-f2b5-4ac6-8cb1-917cff8a8ded-kube-api-access-dmhsm\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-lp9h9\" (UID: \"3183d9eb-f2b5-4ac6-8cb1-917cff8a8ded\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-lp9h9" Dec 03 14:44:17 crc kubenswrapper[4805]: I1203 14:44:17.851434 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/3183d9eb-f2b5-4ac6-8cb1-917cff8a8ded-ssh-key\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-lp9h9\" (UID: \"3183d9eb-f2b5-4ac6-8cb1-917cff8a8ded\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-lp9h9" Dec 03 14:44:17 crc kubenswrapper[4805]: I1203 14:44:17.855554 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/3183d9eb-f2b5-4ac6-8cb1-917cff8a8ded-inventory\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-lp9h9\" (UID: \"3183d9eb-f2b5-4ac6-8cb1-917cff8a8ded\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-lp9h9" Dec 03 14:44:17 crc kubenswrapper[4805]: I1203 14:44:17.864606 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dmhsm\" (UniqueName: \"kubernetes.io/projected/3183d9eb-f2b5-4ac6-8cb1-917cff8a8ded-kube-api-access-dmhsm\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-lp9h9\" (UID: \"3183d9eb-f2b5-4ac6-8cb1-917cff8a8ded\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-lp9h9" Dec 03 14:44:18 crc kubenswrapper[4805]: I1203 14:44:18.015452 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-lp9h9" Dec 03 14:44:18 crc kubenswrapper[4805]: I1203 14:44:18.180051 4805 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-6m8lt" Dec 03 14:44:18 crc kubenswrapper[4805]: I1203 14:44:18.255450 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-j6lwd\" (UniqueName: \"kubernetes.io/projected/1737023c-c5dc-41b4-b11d-e031f4e67d7a-kube-api-access-j6lwd\") pod \"1737023c-c5dc-41b4-b11d-e031f4e67d7a\" (UID: \"1737023c-c5dc-41b4-b11d-e031f4e67d7a\") " Dec 03 14:44:18 crc kubenswrapper[4805]: I1203 14:44:18.255622 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1737023c-c5dc-41b4-b11d-e031f4e67d7a-utilities\") pod \"1737023c-c5dc-41b4-b11d-e031f4e67d7a\" (UID: \"1737023c-c5dc-41b4-b11d-e031f4e67d7a\") " Dec 03 14:44:18 crc kubenswrapper[4805]: I1203 14:44:18.255670 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1737023c-c5dc-41b4-b11d-e031f4e67d7a-catalog-content\") pod \"1737023c-c5dc-41b4-b11d-e031f4e67d7a\" (UID: \"1737023c-c5dc-41b4-b11d-e031f4e67d7a\") " Dec 03 14:44:18 crc kubenswrapper[4805]: I1203 14:44:18.256691 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1737023c-c5dc-41b4-b11d-e031f4e67d7a-utilities" (OuterVolumeSpecName: "utilities") pod "1737023c-c5dc-41b4-b11d-e031f4e67d7a" (UID: "1737023c-c5dc-41b4-b11d-e031f4e67d7a"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 14:44:18 crc kubenswrapper[4805]: I1203 14:44:18.260316 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1737023c-c5dc-41b4-b11d-e031f4e67d7a-kube-api-access-j6lwd" (OuterVolumeSpecName: "kube-api-access-j6lwd") pod "1737023c-c5dc-41b4-b11d-e031f4e67d7a" (UID: "1737023c-c5dc-41b4-b11d-e031f4e67d7a"). InnerVolumeSpecName "kube-api-access-j6lwd". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 14:44:18 crc kubenswrapper[4805]: I1203 14:44:18.272822 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1737023c-c5dc-41b4-b11d-e031f4e67d7a-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "1737023c-c5dc-41b4-b11d-e031f4e67d7a" (UID: "1737023c-c5dc-41b4-b11d-e031f4e67d7a"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 14:44:18 crc kubenswrapper[4805]: I1203 14:44:18.357629 4805 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1737023c-c5dc-41b4-b11d-e031f4e67d7a-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 03 14:44:18 crc kubenswrapper[4805]: I1203 14:44:18.357670 4805 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-j6lwd\" (UniqueName: \"kubernetes.io/projected/1737023c-c5dc-41b4-b11d-e031f4e67d7a-kube-api-access-j6lwd\") on node \"crc\" DevicePath \"\"" Dec 03 14:44:18 crc kubenswrapper[4805]: I1203 14:44:18.357686 4805 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1737023c-c5dc-41b4-b11d-e031f4e67d7a-utilities\") on node \"crc\" DevicePath \"\"" Dec 03 14:44:18 crc kubenswrapper[4805]: W1203 14:44:18.526403 4805 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod3183d9eb_f2b5_4ac6_8cb1_917cff8a8ded.slice/crio-134fcbdf450d30048c7dded30101512d7e798f4fe7a1a52c6d93609b6688910a WatchSource:0}: Error finding container 134fcbdf450d30048c7dded30101512d7e798f4fe7a1a52c6d93609b6688910a: Status 404 returned error can't find the container with id 134fcbdf450d30048c7dded30101512d7e798f4fe7a1a52c6d93609b6688910a Dec 03 14:44:18 crc kubenswrapper[4805]: I1203 14:44:18.528035 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/run-os-edpm-deployment-openstack-edpm-ipam-lp9h9"] Dec 03 14:44:18 crc kubenswrapper[4805]: I1203 14:44:18.529368 4805 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 03 14:44:18 crc kubenswrapper[4805]: I1203 14:44:18.533583 4805 generic.go:334] "Generic (PLEG): container finished" podID="1737023c-c5dc-41b4-b11d-e031f4e67d7a" containerID="db19dc7bd2fd1b0f62e18b0d3abce4a707aa1570793f588e24ac7cc7609562c9" exitCode=0 Dec 03 14:44:18 crc kubenswrapper[4805]: I1203 14:44:18.533633 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-6m8lt" event={"ID":"1737023c-c5dc-41b4-b11d-e031f4e67d7a","Type":"ContainerDied","Data":"db19dc7bd2fd1b0f62e18b0d3abce4a707aa1570793f588e24ac7cc7609562c9"} Dec 03 14:44:18 crc kubenswrapper[4805]: I1203 14:44:18.533663 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-6m8lt" event={"ID":"1737023c-c5dc-41b4-b11d-e031f4e67d7a","Type":"ContainerDied","Data":"7f354d9cda99e0a0cd18edd18fd45b4f156b57ed4fa246de087903f206e7e106"} Dec 03 14:44:18 crc kubenswrapper[4805]: I1203 14:44:18.533685 4805 scope.go:117] "RemoveContainer" containerID="db19dc7bd2fd1b0f62e18b0d3abce4a707aa1570793f588e24ac7cc7609562c9" Dec 03 14:44:18 crc kubenswrapper[4805]: I1203 14:44:18.533818 4805 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-6m8lt" Dec 03 14:44:18 crc kubenswrapper[4805]: I1203 14:44:18.565304 4805 scope.go:117] "RemoveContainer" containerID="f05f785bd4da68f8b30682d25646ba4a0460ebb6718765b82395b7f2189febc1" Dec 03 14:44:18 crc kubenswrapper[4805]: I1203 14:44:18.586512 4805 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-6m8lt"] Dec 03 14:44:18 crc kubenswrapper[4805]: I1203 14:44:18.592076 4805 scope.go:117] "RemoveContainer" containerID="bc681223f9f69de6f7324bac7b7155e8568406036e2f6cbfc863fa9932ebda03" Dec 03 14:44:18 crc kubenswrapper[4805]: I1203 14:44:18.594390 4805 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-6m8lt"] Dec 03 14:44:18 crc kubenswrapper[4805]: I1203 14:44:18.613385 4805 scope.go:117] "RemoveContainer" containerID="db19dc7bd2fd1b0f62e18b0d3abce4a707aa1570793f588e24ac7cc7609562c9" Dec 03 14:44:18 crc kubenswrapper[4805]: E1203 14:44:18.613878 4805 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"db19dc7bd2fd1b0f62e18b0d3abce4a707aa1570793f588e24ac7cc7609562c9\": container with ID starting with db19dc7bd2fd1b0f62e18b0d3abce4a707aa1570793f588e24ac7cc7609562c9 not found: ID does not exist" containerID="db19dc7bd2fd1b0f62e18b0d3abce4a707aa1570793f588e24ac7cc7609562c9" Dec 03 14:44:18 crc kubenswrapper[4805]: I1203 14:44:18.614004 4805 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"db19dc7bd2fd1b0f62e18b0d3abce4a707aa1570793f588e24ac7cc7609562c9"} err="failed to get container status \"db19dc7bd2fd1b0f62e18b0d3abce4a707aa1570793f588e24ac7cc7609562c9\": rpc error: code = NotFound desc = could not find container \"db19dc7bd2fd1b0f62e18b0d3abce4a707aa1570793f588e24ac7cc7609562c9\": container with ID starting with db19dc7bd2fd1b0f62e18b0d3abce4a707aa1570793f588e24ac7cc7609562c9 not found: ID does not exist" Dec 03 14:44:18 crc kubenswrapper[4805]: I1203 14:44:18.614126 4805 scope.go:117] "RemoveContainer" containerID="f05f785bd4da68f8b30682d25646ba4a0460ebb6718765b82395b7f2189febc1" Dec 03 14:44:18 crc kubenswrapper[4805]: E1203 14:44:18.614700 4805 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f05f785bd4da68f8b30682d25646ba4a0460ebb6718765b82395b7f2189febc1\": container with ID starting with f05f785bd4da68f8b30682d25646ba4a0460ebb6718765b82395b7f2189febc1 not found: ID does not exist" containerID="f05f785bd4da68f8b30682d25646ba4a0460ebb6718765b82395b7f2189febc1" Dec 03 14:44:18 crc kubenswrapper[4805]: I1203 14:44:18.615367 4805 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f05f785bd4da68f8b30682d25646ba4a0460ebb6718765b82395b7f2189febc1"} err="failed to get container status \"f05f785bd4da68f8b30682d25646ba4a0460ebb6718765b82395b7f2189febc1\": rpc error: code = NotFound desc = could not find container \"f05f785bd4da68f8b30682d25646ba4a0460ebb6718765b82395b7f2189febc1\": container with ID starting with f05f785bd4da68f8b30682d25646ba4a0460ebb6718765b82395b7f2189febc1 not found: ID does not exist" Dec 03 14:44:18 crc kubenswrapper[4805]: I1203 14:44:18.615518 4805 scope.go:117] "RemoveContainer" containerID="bc681223f9f69de6f7324bac7b7155e8568406036e2f6cbfc863fa9932ebda03" Dec 03 14:44:18 crc kubenswrapper[4805]: E1203 14:44:18.616021 4805 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"bc681223f9f69de6f7324bac7b7155e8568406036e2f6cbfc863fa9932ebda03\": container with ID starting with bc681223f9f69de6f7324bac7b7155e8568406036e2f6cbfc863fa9932ebda03 not found: ID does not exist" containerID="bc681223f9f69de6f7324bac7b7155e8568406036e2f6cbfc863fa9932ebda03" Dec 03 14:44:18 crc kubenswrapper[4805]: I1203 14:44:18.616135 4805 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"bc681223f9f69de6f7324bac7b7155e8568406036e2f6cbfc863fa9932ebda03"} err="failed to get container status \"bc681223f9f69de6f7324bac7b7155e8568406036e2f6cbfc863fa9932ebda03\": rpc error: code = NotFound desc = could not find container \"bc681223f9f69de6f7324bac7b7155e8568406036e2f6cbfc863fa9932ebda03\": container with ID starting with bc681223f9f69de6f7324bac7b7155e8568406036e2f6cbfc863fa9932ebda03 not found: ID does not exist" Dec 03 14:44:18 crc kubenswrapper[4805]: I1203 14:44:18.705280 4805 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1737023c-c5dc-41b4-b11d-e031f4e67d7a" path="/var/lib/kubelet/pods/1737023c-c5dc-41b4-b11d-e031f4e67d7a/volumes" Dec 03 14:44:19 crc kubenswrapper[4805]: I1203 14:44:19.544734 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-lp9h9" event={"ID":"3183d9eb-f2b5-4ac6-8cb1-917cff8a8ded","Type":"ContainerStarted","Data":"134fcbdf450d30048c7dded30101512d7e798f4fe7a1a52c6d93609b6688910a"} Dec 03 14:44:20 crc kubenswrapper[4805]: I1203 14:44:20.554059 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-lp9h9" event={"ID":"3183d9eb-f2b5-4ac6-8cb1-917cff8a8ded","Type":"ContainerStarted","Data":"7eaeb93cd926d41ac3c8d3f556e605c942064d24528753b0e96df3b813ac413b"} Dec 03 14:44:20 crc kubenswrapper[4805]: I1203 14:44:20.572097 4805 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-lp9h9" podStartSLOduration=2.7752973340000002 podStartE2EDuration="3.572078348s" podCreationTimestamp="2025-12-03 14:44:17 +0000 UTC" firstStartedPulling="2025-12-03 14:44:18.529130278 +0000 UTC m=+2088.192047201" lastFinishedPulling="2025-12-03 14:44:19.325911292 +0000 UTC m=+2088.988828215" observedRunningTime="2025-12-03 14:44:20.567900275 +0000 UTC m=+2090.230817198" watchObservedRunningTime="2025-12-03 14:44:20.572078348 +0000 UTC m=+2090.234995271" Dec 03 14:44:27 crc kubenswrapper[4805]: I1203 14:44:27.621616 4805 generic.go:334] "Generic (PLEG): container finished" podID="3183d9eb-f2b5-4ac6-8cb1-917cff8a8ded" containerID="7eaeb93cd926d41ac3c8d3f556e605c942064d24528753b0e96df3b813ac413b" exitCode=0 Dec 03 14:44:27 crc kubenswrapper[4805]: I1203 14:44:27.621758 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-lp9h9" event={"ID":"3183d9eb-f2b5-4ac6-8cb1-917cff8a8ded","Type":"ContainerDied","Data":"7eaeb93cd926d41ac3c8d3f556e605c942064d24528753b0e96df3b813ac413b"} Dec 03 14:44:29 crc kubenswrapper[4805]: I1203 14:44:29.042282 4805 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-lp9h9" Dec 03 14:44:29 crc kubenswrapper[4805]: I1203 14:44:29.163832 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dmhsm\" (UniqueName: \"kubernetes.io/projected/3183d9eb-f2b5-4ac6-8cb1-917cff8a8ded-kube-api-access-dmhsm\") pod \"3183d9eb-f2b5-4ac6-8cb1-917cff8a8ded\" (UID: \"3183d9eb-f2b5-4ac6-8cb1-917cff8a8ded\") " Dec 03 14:44:29 crc kubenswrapper[4805]: I1203 14:44:29.164137 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/3183d9eb-f2b5-4ac6-8cb1-917cff8a8ded-ssh-key\") pod \"3183d9eb-f2b5-4ac6-8cb1-917cff8a8ded\" (UID: \"3183d9eb-f2b5-4ac6-8cb1-917cff8a8ded\") " Dec 03 14:44:29 crc kubenswrapper[4805]: I1203 14:44:29.164206 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/3183d9eb-f2b5-4ac6-8cb1-917cff8a8ded-inventory\") pod \"3183d9eb-f2b5-4ac6-8cb1-917cff8a8ded\" (UID: \"3183d9eb-f2b5-4ac6-8cb1-917cff8a8ded\") " Dec 03 14:44:29 crc kubenswrapper[4805]: I1203 14:44:29.172272 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3183d9eb-f2b5-4ac6-8cb1-917cff8a8ded-kube-api-access-dmhsm" (OuterVolumeSpecName: "kube-api-access-dmhsm") pod "3183d9eb-f2b5-4ac6-8cb1-917cff8a8ded" (UID: "3183d9eb-f2b5-4ac6-8cb1-917cff8a8ded"). InnerVolumeSpecName "kube-api-access-dmhsm". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 14:44:29 crc kubenswrapper[4805]: I1203 14:44:29.193156 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3183d9eb-f2b5-4ac6-8cb1-917cff8a8ded-inventory" (OuterVolumeSpecName: "inventory") pod "3183d9eb-f2b5-4ac6-8cb1-917cff8a8ded" (UID: "3183d9eb-f2b5-4ac6-8cb1-917cff8a8ded"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 14:44:29 crc kubenswrapper[4805]: I1203 14:44:29.218285 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3183d9eb-f2b5-4ac6-8cb1-917cff8a8ded-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "3183d9eb-f2b5-4ac6-8cb1-917cff8a8ded" (UID: "3183d9eb-f2b5-4ac6-8cb1-917cff8a8ded"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 14:44:29 crc kubenswrapper[4805]: I1203 14:44:29.267161 4805 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/3183d9eb-f2b5-4ac6-8cb1-917cff8a8ded-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 03 14:44:29 crc kubenswrapper[4805]: I1203 14:44:29.267200 4805 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/3183d9eb-f2b5-4ac6-8cb1-917cff8a8ded-inventory\") on node \"crc\" DevicePath \"\"" Dec 03 14:44:29 crc kubenswrapper[4805]: I1203 14:44:29.267213 4805 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dmhsm\" (UniqueName: \"kubernetes.io/projected/3183d9eb-f2b5-4ac6-8cb1-917cff8a8ded-kube-api-access-dmhsm\") on node \"crc\" DevicePath \"\"" Dec 03 14:44:29 crc kubenswrapper[4805]: I1203 14:44:29.642466 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-lp9h9" event={"ID":"3183d9eb-f2b5-4ac6-8cb1-917cff8a8ded","Type":"ContainerDied","Data":"134fcbdf450d30048c7dded30101512d7e798f4fe7a1a52c6d93609b6688910a"} Dec 03 14:44:29 crc kubenswrapper[4805]: I1203 14:44:29.642524 4805 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="134fcbdf450d30048c7dded30101512d7e798f4fe7a1a52c6d93609b6688910a" Dec 03 14:44:29 crc kubenswrapper[4805]: I1203 14:44:29.642608 4805 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-lp9h9" Dec 03 14:44:29 crc kubenswrapper[4805]: I1203 14:44:29.728726 4805 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-k5scz"] Dec 03 14:44:29 crc kubenswrapper[4805]: E1203 14:44:29.729159 4805 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1737023c-c5dc-41b4-b11d-e031f4e67d7a" containerName="extract-content" Dec 03 14:44:29 crc kubenswrapper[4805]: I1203 14:44:29.729175 4805 state_mem.go:107] "Deleted CPUSet assignment" podUID="1737023c-c5dc-41b4-b11d-e031f4e67d7a" containerName="extract-content" Dec 03 14:44:29 crc kubenswrapper[4805]: E1203 14:44:29.729189 4805 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3183d9eb-f2b5-4ac6-8cb1-917cff8a8ded" containerName="run-os-edpm-deployment-openstack-edpm-ipam" Dec 03 14:44:29 crc kubenswrapper[4805]: I1203 14:44:29.729197 4805 state_mem.go:107] "Deleted CPUSet assignment" podUID="3183d9eb-f2b5-4ac6-8cb1-917cff8a8ded" containerName="run-os-edpm-deployment-openstack-edpm-ipam" Dec 03 14:44:29 crc kubenswrapper[4805]: E1203 14:44:29.729206 4805 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1737023c-c5dc-41b4-b11d-e031f4e67d7a" containerName="extract-utilities" Dec 03 14:44:29 crc kubenswrapper[4805]: I1203 14:44:29.729212 4805 state_mem.go:107] "Deleted CPUSet assignment" podUID="1737023c-c5dc-41b4-b11d-e031f4e67d7a" containerName="extract-utilities" Dec 03 14:44:29 crc kubenswrapper[4805]: E1203 14:44:29.729238 4805 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1737023c-c5dc-41b4-b11d-e031f4e67d7a" containerName="registry-server" Dec 03 14:44:29 crc kubenswrapper[4805]: I1203 14:44:29.729245 4805 state_mem.go:107] "Deleted CPUSet assignment" podUID="1737023c-c5dc-41b4-b11d-e031f4e67d7a" containerName="registry-server" Dec 03 14:44:29 crc kubenswrapper[4805]: I1203 14:44:29.729434 4805 memory_manager.go:354] "RemoveStaleState removing state" podUID="3183d9eb-f2b5-4ac6-8cb1-917cff8a8ded" containerName="run-os-edpm-deployment-openstack-edpm-ipam" Dec 03 14:44:29 crc kubenswrapper[4805]: I1203 14:44:29.729448 4805 memory_manager.go:354] "RemoveStaleState removing state" podUID="1737023c-c5dc-41b4-b11d-e031f4e67d7a" containerName="registry-server" Dec 03 14:44:29 crc kubenswrapper[4805]: I1203 14:44:29.730135 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-k5scz" Dec 03 14:44:29 crc kubenswrapper[4805]: I1203 14:44:29.733011 4805 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 03 14:44:29 crc kubenswrapper[4805]: I1203 14:44:29.733440 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-n7jt4" Dec 03 14:44:29 crc kubenswrapper[4805]: I1203 14:44:29.733965 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Dec 03 14:44:29 crc kubenswrapper[4805]: I1203 14:44:29.735654 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Dec 03 14:44:29 crc kubenswrapper[4805]: I1203 14:44:29.737418 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-k5scz"] Dec 03 14:44:29 crc kubenswrapper[4805]: I1203 14:44:29.878623 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/ba862dc9-b0f4-4ba3-888a-5eae1601dfdb-inventory\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-k5scz\" (UID: \"ba862dc9-b0f4-4ba3-888a-5eae1601dfdb\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-k5scz" Dec 03 14:44:29 crc kubenswrapper[4805]: I1203 14:44:29.878706 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/ba862dc9-b0f4-4ba3-888a-5eae1601dfdb-ssh-key\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-k5scz\" (UID: \"ba862dc9-b0f4-4ba3-888a-5eae1601dfdb\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-k5scz" Dec 03 14:44:29 crc kubenswrapper[4805]: I1203 14:44:29.878790 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gxv2j\" (UniqueName: \"kubernetes.io/projected/ba862dc9-b0f4-4ba3-888a-5eae1601dfdb-kube-api-access-gxv2j\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-k5scz\" (UID: \"ba862dc9-b0f4-4ba3-888a-5eae1601dfdb\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-k5scz" Dec 03 14:44:29 crc kubenswrapper[4805]: I1203 14:44:29.980965 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/ba862dc9-b0f4-4ba3-888a-5eae1601dfdb-inventory\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-k5scz\" (UID: \"ba862dc9-b0f4-4ba3-888a-5eae1601dfdb\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-k5scz" Dec 03 14:44:29 crc kubenswrapper[4805]: I1203 14:44:29.981067 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/ba862dc9-b0f4-4ba3-888a-5eae1601dfdb-ssh-key\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-k5scz\" (UID: \"ba862dc9-b0f4-4ba3-888a-5eae1601dfdb\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-k5scz" Dec 03 14:44:29 crc kubenswrapper[4805]: I1203 14:44:29.981127 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gxv2j\" (UniqueName: \"kubernetes.io/projected/ba862dc9-b0f4-4ba3-888a-5eae1601dfdb-kube-api-access-gxv2j\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-k5scz\" (UID: \"ba862dc9-b0f4-4ba3-888a-5eae1601dfdb\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-k5scz" Dec 03 14:44:29 crc kubenswrapper[4805]: I1203 14:44:29.984667 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/ba862dc9-b0f4-4ba3-888a-5eae1601dfdb-inventory\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-k5scz\" (UID: \"ba862dc9-b0f4-4ba3-888a-5eae1601dfdb\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-k5scz" Dec 03 14:44:29 crc kubenswrapper[4805]: I1203 14:44:29.985707 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/ba862dc9-b0f4-4ba3-888a-5eae1601dfdb-ssh-key\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-k5scz\" (UID: \"ba862dc9-b0f4-4ba3-888a-5eae1601dfdb\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-k5scz" Dec 03 14:44:29 crc kubenswrapper[4805]: I1203 14:44:29.997873 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gxv2j\" (UniqueName: \"kubernetes.io/projected/ba862dc9-b0f4-4ba3-888a-5eae1601dfdb-kube-api-access-gxv2j\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-k5scz\" (UID: \"ba862dc9-b0f4-4ba3-888a-5eae1601dfdb\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-k5scz" Dec 03 14:44:30 crc kubenswrapper[4805]: I1203 14:44:30.054291 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-k5scz" Dec 03 14:44:30 crc kubenswrapper[4805]: I1203 14:44:30.596793 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-k5scz"] Dec 03 14:44:30 crc kubenswrapper[4805]: I1203 14:44:30.650610 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-k5scz" event={"ID":"ba862dc9-b0f4-4ba3-888a-5eae1601dfdb","Type":"ContainerStarted","Data":"2a98d64ef6187afd9d5ecc1314e0830668beb89a29ec0b594cac68663e65be4f"} Dec 03 14:44:31 crc kubenswrapper[4805]: I1203 14:44:31.130249 4805 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 03 14:44:31 crc kubenswrapper[4805]: I1203 14:44:31.659747 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-k5scz" event={"ID":"ba862dc9-b0f4-4ba3-888a-5eae1601dfdb","Type":"ContainerStarted","Data":"44d4875fd1ed8e773e0364e6e414b6d47f726d4c30d68e9e7fef6fa7b4f9686d"} Dec 03 14:44:31 crc kubenswrapper[4805]: I1203 14:44:31.680994 4805 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-k5scz" podStartSLOduration=2.158156158 podStartE2EDuration="2.680967758s" podCreationTimestamp="2025-12-03 14:44:29 +0000 UTC" firstStartedPulling="2025-12-03 14:44:30.604828093 +0000 UTC m=+2100.267745016" lastFinishedPulling="2025-12-03 14:44:31.127639693 +0000 UTC m=+2100.790556616" observedRunningTime="2025-12-03 14:44:31.677557766 +0000 UTC m=+2101.340474709" watchObservedRunningTime="2025-12-03 14:44:31.680967758 +0000 UTC m=+2101.343884691" Dec 03 14:44:40 crc kubenswrapper[4805]: I1203 14:44:40.746385 4805 generic.go:334] "Generic (PLEG): container finished" podID="ba862dc9-b0f4-4ba3-888a-5eae1601dfdb" containerID="44d4875fd1ed8e773e0364e6e414b6d47f726d4c30d68e9e7fef6fa7b4f9686d" exitCode=0 Dec 03 14:44:40 crc kubenswrapper[4805]: I1203 14:44:40.746508 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-k5scz" event={"ID":"ba862dc9-b0f4-4ba3-888a-5eae1601dfdb","Type":"ContainerDied","Data":"44d4875fd1ed8e773e0364e6e414b6d47f726d4c30d68e9e7fef6fa7b4f9686d"} Dec 03 14:44:42 crc kubenswrapper[4805]: I1203 14:44:42.156965 4805 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-k5scz" Dec 03 14:44:42 crc kubenswrapper[4805]: I1203 14:44:42.329217 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/ba862dc9-b0f4-4ba3-888a-5eae1601dfdb-ssh-key\") pod \"ba862dc9-b0f4-4ba3-888a-5eae1601dfdb\" (UID: \"ba862dc9-b0f4-4ba3-888a-5eae1601dfdb\") " Dec 03 14:44:42 crc kubenswrapper[4805]: I1203 14:44:42.329421 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gxv2j\" (UniqueName: \"kubernetes.io/projected/ba862dc9-b0f4-4ba3-888a-5eae1601dfdb-kube-api-access-gxv2j\") pod \"ba862dc9-b0f4-4ba3-888a-5eae1601dfdb\" (UID: \"ba862dc9-b0f4-4ba3-888a-5eae1601dfdb\") " Dec 03 14:44:42 crc kubenswrapper[4805]: I1203 14:44:42.329485 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/ba862dc9-b0f4-4ba3-888a-5eae1601dfdb-inventory\") pod \"ba862dc9-b0f4-4ba3-888a-5eae1601dfdb\" (UID: \"ba862dc9-b0f4-4ba3-888a-5eae1601dfdb\") " Dec 03 14:44:42 crc kubenswrapper[4805]: I1203 14:44:42.335402 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ba862dc9-b0f4-4ba3-888a-5eae1601dfdb-kube-api-access-gxv2j" (OuterVolumeSpecName: "kube-api-access-gxv2j") pod "ba862dc9-b0f4-4ba3-888a-5eae1601dfdb" (UID: "ba862dc9-b0f4-4ba3-888a-5eae1601dfdb"). InnerVolumeSpecName "kube-api-access-gxv2j". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 14:44:42 crc kubenswrapper[4805]: I1203 14:44:42.357114 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ba862dc9-b0f4-4ba3-888a-5eae1601dfdb-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "ba862dc9-b0f4-4ba3-888a-5eae1601dfdb" (UID: "ba862dc9-b0f4-4ba3-888a-5eae1601dfdb"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 14:44:42 crc kubenswrapper[4805]: I1203 14:44:42.377294 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ba862dc9-b0f4-4ba3-888a-5eae1601dfdb-inventory" (OuterVolumeSpecName: "inventory") pod "ba862dc9-b0f4-4ba3-888a-5eae1601dfdb" (UID: "ba862dc9-b0f4-4ba3-888a-5eae1601dfdb"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 14:44:42 crc kubenswrapper[4805]: I1203 14:44:42.431912 4805 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gxv2j\" (UniqueName: \"kubernetes.io/projected/ba862dc9-b0f4-4ba3-888a-5eae1601dfdb-kube-api-access-gxv2j\") on node \"crc\" DevicePath \"\"" Dec 03 14:44:42 crc kubenswrapper[4805]: I1203 14:44:42.431960 4805 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/ba862dc9-b0f4-4ba3-888a-5eae1601dfdb-inventory\") on node \"crc\" DevicePath \"\"" Dec 03 14:44:42 crc kubenswrapper[4805]: I1203 14:44:42.431972 4805 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/ba862dc9-b0f4-4ba3-888a-5eae1601dfdb-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 03 14:44:42 crc kubenswrapper[4805]: I1203 14:44:42.776259 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-k5scz" event={"ID":"ba862dc9-b0f4-4ba3-888a-5eae1601dfdb","Type":"ContainerDied","Data":"2a98d64ef6187afd9d5ecc1314e0830668beb89a29ec0b594cac68663e65be4f"} Dec 03 14:44:42 crc kubenswrapper[4805]: I1203 14:44:42.776306 4805 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="2a98d64ef6187afd9d5ecc1314e0830668beb89a29ec0b594cac68663e65be4f" Dec 03 14:44:42 crc kubenswrapper[4805]: I1203 14:44:42.776329 4805 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-k5scz" Dec 03 14:44:42 crc kubenswrapper[4805]: I1203 14:44:42.859337 4805 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/install-certs-edpm-deployment-openstack-edpm-ipam-tw4dn"] Dec 03 14:44:42 crc kubenswrapper[4805]: E1203 14:44:42.859754 4805 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ba862dc9-b0f4-4ba3-888a-5eae1601dfdb" containerName="reboot-os-edpm-deployment-openstack-edpm-ipam" Dec 03 14:44:42 crc kubenswrapper[4805]: I1203 14:44:42.859781 4805 state_mem.go:107] "Deleted CPUSet assignment" podUID="ba862dc9-b0f4-4ba3-888a-5eae1601dfdb" containerName="reboot-os-edpm-deployment-openstack-edpm-ipam" Dec 03 14:44:42 crc kubenswrapper[4805]: I1203 14:44:42.860034 4805 memory_manager.go:354] "RemoveStaleState removing state" podUID="ba862dc9-b0f4-4ba3-888a-5eae1601dfdb" containerName="reboot-os-edpm-deployment-openstack-edpm-ipam" Dec 03 14:44:42 crc kubenswrapper[4805]: I1203 14:44:42.860795 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-tw4dn" Dec 03 14:44:42 crc kubenswrapper[4805]: I1203 14:44:42.863389 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-libvirt-default-certs-0" Dec 03 14:44:42 crc kubenswrapper[4805]: I1203 14:44:42.863570 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-telemetry-default-certs-0" Dec 03 14:44:42 crc kubenswrapper[4805]: I1203 14:44:42.863679 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-ovn-default-certs-0" Dec 03 14:44:42 crc kubenswrapper[4805]: I1203 14:44:42.865576 4805 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 03 14:44:42 crc kubenswrapper[4805]: I1203 14:44:42.865608 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Dec 03 14:44:42 crc kubenswrapper[4805]: I1203 14:44:42.866860 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-n7jt4" Dec 03 14:44:42 crc kubenswrapper[4805]: I1203 14:44:42.866860 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-neutron-metadata-default-certs-0" Dec 03 14:44:42 crc kubenswrapper[4805]: I1203 14:44:42.867351 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Dec 03 14:44:42 crc kubenswrapper[4805]: I1203 14:44:42.872330 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/install-certs-edpm-deployment-openstack-edpm-ipam-tw4dn"] Dec 03 14:44:43 crc kubenswrapper[4805]: I1203 14:44:43.043638 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9cd62272-1077-4e13-8e99-c158fbb8ba1a-neutron-metadata-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-tw4dn\" (UID: \"9cd62272-1077-4e13-8e99-c158fbb8ba1a\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-tw4dn" Dec 03 14:44:43 crc kubenswrapper[4805]: I1203 14:44:43.043701 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9cd62272-1077-4e13-8e99-c158fbb8ba1a-libvirt-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-tw4dn\" (UID: \"9cd62272-1077-4e13-8e99-c158fbb8ba1a\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-tw4dn" Dec 03 14:44:43 crc kubenswrapper[4805]: I1203 14:44:43.043756 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-edpm-ipam-neutron-metadata-default-certs-0\" (UniqueName: \"kubernetes.io/projected/9cd62272-1077-4e13-8e99-c158fbb8ba1a-openstack-edpm-ipam-neutron-metadata-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-tw4dn\" (UID: \"9cd62272-1077-4e13-8e99-c158fbb8ba1a\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-tw4dn" Dec 03 14:44:43 crc kubenswrapper[4805]: I1203 14:44:43.043800 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/9cd62272-1077-4e13-8e99-c158fbb8ba1a-inventory\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-tw4dn\" (UID: \"9cd62272-1077-4e13-8e99-c158fbb8ba1a\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-tw4dn" Dec 03 14:44:43 crc kubenswrapper[4805]: I1203 14:44:43.043885 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9cd62272-1077-4e13-8e99-c158fbb8ba1a-bootstrap-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-tw4dn\" (UID: \"9cd62272-1077-4e13-8e99-c158fbb8ba1a\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-tw4dn" Dec 03 14:44:43 crc kubenswrapper[4805]: I1203 14:44:43.043939 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9cd62272-1077-4e13-8e99-c158fbb8ba1a-ovn-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-tw4dn\" (UID: \"9cd62272-1077-4e13-8e99-c158fbb8ba1a\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-tw4dn" Dec 03 14:44:43 crc kubenswrapper[4805]: I1203 14:44:43.043969 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-b6rfz\" (UniqueName: \"kubernetes.io/projected/9cd62272-1077-4e13-8e99-c158fbb8ba1a-kube-api-access-b6rfz\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-tw4dn\" (UID: \"9cd62272-1077-4e13-8e99-c158fbb8ba1a\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-tw4dn" Dec 03 14:44:43 crc kubenswrapper[4805]: I1203 14:44:43.043997 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-edpm-ipam-telemetry-default-certs-0\" (UniqueName: \"kubernetes.io/projected/9cd62272-1077-4e13-8e99-c158fbb8ba1a-openstack-edpm-ipam-telemetry-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-tw4dn\" (UID: \"9cd62272-1077-4e13-8e99-c158fbb8ba1a\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-tw4dn" Dec 03 14:44:43 crc kubenswrapper[4805]: I1203 14:44:43.044028 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/9cd62272-1077-4e13-8e99-c158fbb8ba1a-ssh-key\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-tw4dn\" (UID: \"9cd62272-1077-4e13-8e99-c158fbb8ba1a\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-tw4dn" Dec 03 14:44:43 crc kubenswrapper[4805]: I1203 14:44:43.044059 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9cd62272-1077-4e13-8e99-c158fbb8ba1a-telemetry-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-tw4dn\" (UID: \"9cd62272-1077-4e13-8e99-c158fbb8ba1a\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-tw4dn" Dec 03 14:44:43 crc kubenswrapper[4805]: I1203 14:44:43.044102 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-edpm-ipam-libvirt-default-certs-0\" (UniqueName: \"kubernetes.io/projected/9cd62272-1077-4e13-8e99-c158fbb8ba1a-openstack-edpm-ipam-libvirt-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-tw4dn\" (UID: \"9cd62272-1077-4e13-8e99-c158fbb8ba1a\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-tw4dn" Dec 03 14:44:43 crc kubenswrapper[4805]: I1203 14:44:43.044128 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9cd62272-1077-4e13-8e99-c158fbb8ba1a-repo-setup-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-tw4dn\" (UID: \"9cd62272-1077-4e13-8e99-c158fbb8ba1a\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-tw4dn" Dec 03 14:44:43 crc kubenswrapper[4805]: I1203 14:44:43.044186 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-edpm-ipam-ovn-default-certs-0\" (UniqueName: \"kubernetes.io/projected/9cd62272-1077-4e13-8e99-c158fbb8ba1a-openstack-edpm-ipam-ovn-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-tw4dn\" (UID: \"9cd62272-1077-4e13-8e99-c158fbb8ba1a\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-tw4dn" Dec 03 14:44:43 crc kubenswrapper[4805]: I1203 14:44:43.044209 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9cd62272-1077-4e13-8e99-c158fbb8ba1a-nova-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-tw4dn\" (UID: \"9cd62272-1077-4e13-8e99-c158fbb8ba1a\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-tw4dn" Dec 03 14:44:43 crc kubenswrapper[4805]: I1203 14:44:43.146278 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9cd62272-1077-4e13-8e99-c158fbb8ba1a-telemetry-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-tw4dn\" (UID: \"9cd62272-1077-4e13-8e99-c158fbb8ba1a\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-tw4dn" Dec 03 14:44:43 crc kubenswrapper[4805]: I1203 14:44:43.146342 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-edpm-ipam-libvirt-default-certs-0\" (UniqueName: \"kubernetes.io/projected/9cd62272-1077-4e13-8e99-c158fbb8ba1a-openstack-edpm-ipam-libvirt-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-tw4dn\" (UID: \"9cd62272-1077-4e13-8e99-c158fbb8ba1a\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-tw4dn" Dec 03 14:44:43 crc kubenswrapper[4805]: I1203 14:44:43.146367 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9cd62272-1077-4e13-8e99-c158fbb8ba1a-repo-setup-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-tw4dn\" (UID: \"9cd62272-1077-4e13-8e99-c158fbb8ba1a\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-tw4dn" Dec 03 14:44:43 crc kubenswrapper[4805]: I1203 14:44:43.146403 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9cd62272-1077-4e13-8e99-c158fbb8ba1a-nova-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-tw4dn\" (UID: \"9cd62272-1077-4e13-8e99-c158fbb8ba1a\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-tw4dn" Dec 03 14:44:43 crc kubenswrapper[4805]: I1203 14:44:43.146422 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-edpm-ipam-ovn-default-certs-0\" (UniqueName: \"kubernetes.io/projected/9cd62272-1077-4e13-8e99-c158fbb8ba1a-openstack-edpm-ipam-ovn-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-tw4dn\" (UID: \"9cd62272-1077-4e13-8e99-c158fbb8ba1a\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-tw4dn" Dec 03 14:44:43 crc kubenswrapper[4805]: I1203 14:44:43.146504 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9cd62272-1077-4e13-8e99-c158fbb8ba1a-neutron-metadata-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-tw4dn\" (UID: \"9cd62272-1077-4e13-8e99-c158fbb8ba1a\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-tw4dn" Dec 03 14:44:43 crc kubenswrapper[4805]: I1203 14:44:43.146528 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9cd62272-1077-4e13-8e99-c158fbb8ba1a-libvirt-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-tw4dn\" (UID: \"9cd62272-1077-4e13-8e99-c158fbb8ba1a\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-tw4dn" Dec 03 14:44:43 crc kubenswrapper[4805]: I1203 14:44:43.146555 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-edpm-ipam-neutron-metadata-default-certs-0\" (UniqueName: \"kubernetes.io/projected/9cd62272-1077-4e13-8e99-c158fbb8ba1a-openstack-edpm-ipam-neutron-metadata-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-tw4dn\" (UID: \"9cd62272-1077-4e13-8e99-c158fbb8ba1a\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-tw4dn" Dec 03 14:44:43 crc kubenswrapper[4805]: I1203 14:44:43.146576 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/9cd62272-1077-4e13-8e99-c158fbb8ba1a-inventory\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-tw4dn\" (UID: \"9cd62272-1077-4e13-8e99-c158fbb8ba1a\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-tw4dn" Dec 03 14:44:43 crc kubenswrapper[4805]: I1203 14:44:43.146599 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9cd62272-1077-4e13-8e99-c158fbb8ba1a-bootstrap-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-tw4dn\" (UID: \"9cd62272-1077-4e13-8e99-c158fbb8ba1a\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-tw4dn" Dec 03 14:44:43 crc kubenswrapper[4805]: I1203 14:44:43.146634 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9cd62272-1077-4e13-8e99-c158fbb8ba1a-ovn-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-tw4dn\" (UID: \"9cd62272-1077-4e13-8e99-c158fbb8ba1a\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-tw4dn" Dec 03 14:44:43 crc kubenswrapper[4805]: I1203 14:44:43.146655 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-b6rfz\" (UniqueName: \"kubernetes.io/projected/9cd62272-1077-4e13-8e99-c158fbb8ba1a-kube-api-access-b6rfz\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-tw4dn\" (UID: \"9cd62272-1077-4e13-8e99-c158fbb8ba1a\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-tw4dn" Dec 03 14:44:43 crc kubenswrapper[4805]: I1203 14:44:43.146677 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-edpm-ipam-telemetry-default-certs-0\" (UniqueName: \"kubernetes.io/projected/9cd62272-1077-4e13-8e99-c158fbb8ba1a-openstack-edpm-ipam-telemetry-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-tw4dn\" (UID: \"9cd62272-1077-4e13-8e99-c158fbb8ba1a\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-tw4dn" Dec 03 14:44:43 crc kubenswrapper[4805]: I1203 14:44:43.146696 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/9cd62272-1077-4e13-8e99-c158fbb8ba1a-ssh-key\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-tw4dn\" (UID: \"9cd62272-1077-4e13-8e99-c158fbb8ba1a\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-tw4dn" Dec 03 14:44:43 crc kubenswrapper[4805]: I1203 14:44:43.151007 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9cd62272-1077-4e13-8e99-c158fbb8ba1a-telemetry-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-tw4dn\" (UID: \"9cd62272-1077-4e13-8e99-c158fbb8ba1a\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-tw4dn" Dec 03 14:44:43 crc kubenswrapper[4805]: I1203 14:44:43.157686 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9cd62272-1077-4e13-8e99-c158fbb8ba1a-libvirt-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-tw4dn\" (UID: \"9cd62272-1077-4e13-8e99-c158fbb8ba1a\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-tw4dn" Dec 03 14:44:43 crc kubenswrapper[4805]: I1203 14:44:43.157949 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/9cd62272-1077-4e13-8e99-c158fbb8ba1a-ssh-key\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-tw4dn\" (UID: \"9cd62272-1077-4e13-8e99-c158fbb8ba1a\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-tw4dn" Dec 03 14:44:43 crc kubenswrapper[4805]: I1203 14:44:43.160354 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9cd62272-1077-4e13-8e99-c158fbb8ba1a-nova-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-tw4dn\" (UID: \"9cd62272-1077-4e13-8e99-c158fbb8ba1a\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-tw4dn" Dec 03 14:44:43 crc kubenswrapper[4805]: I1203 14:44:43.161159 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-edpm-ipam-ovn-default-certs-0\" (UniqueName: \"kubernetes.io/projected/9cd62272-1077-4e13-8e99-c158fbb8ba1a-openstack-edpm-ipam-ovn-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-tw4dn\" (UID: \"9cd62272-1077-4e13-8e99-c158fbb8ba1a\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-tw4dn" Dec 03 14:44:43 crc kubenswrapper[4805]: I1203 14:44:43.161893 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9cd62272-1077-4e13-8e99-c158fbb8ba1a-repo-setup-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-tw4dn\" (UID: \"9cd62272-1077-4e13-8e99-c158fbb8ba1a\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-tw4dn" Dec 03 14:44:43 crc kubenswrapper[4805]: I1203 14:44:43.163907 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/9cd62272-1077-4e13-8e99-c158fbb8ba1a-inventory\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-tw4dn\" (UID: \"9cd62272-1077-4e13-8e99-c158fbb8ba1a\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-tw4dn" Dec 03 14:44:43 crc kubenswrapper[4805]: I1203 14:44:43.167651 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-edpm-ipam-telemetry-default-certs-0\" (UniqueName: \"kubernetes.io/projected/9cd62272-1077-4e13-8e99-c158fbb8ba1a-openstack-edpm-ipam-telemetry-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-tw4dn\" (UID: \"9cd62272-1077-4e13-8e99-c158fbb8ba1a\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-tw4dn" Dec 03 14:44:43 crc kubenswrapper[4805]: I1203 14:44:43.168173 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9cd62272-1077-4e13-8e99-c158fbb8ba1a-bootstrap-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-tw4dn\" (UID: \"9cd62272-1077-4e13-8e99-c158fbb8ba1a\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-tw4dn" Dec 03 14:44:43 crc kubenswrapper[4805]: I1203 14:44:43.169855 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-edpm-ipam-neutron-metadata-default-certs-0\" (UniqueName: \"kubernetes.io/projected/9cd62272-1077-4e13-8e99-c158fbb8ba1a-openstack-edpm-ipam-neutron-metadata-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-tw4dn\" (UID: \"9cd62272-1077-4e13-8e99-c158fbb8ba1a\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-tw4dn" Dec 03 14:44:43 crc kubenswrapper[4805]: I1203 14:44:43.171345 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9cd62272-1077-4e13-8e99-c158fbb8ba1a-neutron-metadata-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-tw4dn\" (UID: \"9cd62272-1077-4e13-8e99-c158fbb8ba1a\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-tw4dn" Dec 03 14:44:43 crc kubenswrapper[4805]: I1203 14:44:43.171493 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-edpm-ipam-libvirt-default-certs-0\" (UniqueName: \"kubernetes.io/projected/9cd62272-1077-4e13-8e99-c158fbb8ba1a-openstack-edpm-ipam-libvirt-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-tw4dn\" (UID: \"9cd62272-1077-4e13-8e99-c158fbb8ba1a\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-tw4dn" Dec 03 14:44:43 crc kubenswrapper[4805]: I1203 14:44:43.174687 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9cd62272-1077-4e13-8e99-c158fbb8ba1a-ovn-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-tw4dn\" (UID: \"9cd62272-1077-4e13-8e99-c158fbb8ba1a\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-tw4dn" Dec 03 14:44:43 crc kubenswrapper[4805]: I1203 14:44:43.178115 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-b6rfz\" (UniqueName: \"kubernetes.io/projected/9cd62272-1077-4e13-8e99-c158fbb8ba1a-kube-api-access-b6rfz\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-tw4dn\" (UID: \"9cd62272-1077-4e13-8e99-c158fbb8ba1a\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-tw4dn" Dec 03 14:44:43 crc kubenswrapper[4805]: I1203 14:44:43.180930 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-tw4dn" Dec 03 14:44:43 crc kubenswrapper[4805]: I1203 14:44:43.712114 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/install-certs-edpm-deployment-openstack-edpm-ipam-tw4dn"] Dec 03 14:44:43 crc kubenswrapper[4805]: I1203 14:44:43.785646 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-tw4dn" event={"ID":"9cd62272-1077-4e13-8e99-c158fbb8ba1a","Type":"ContainerStarted","Data":"2aef129fd3f3732cb554fc9fa9b5e3f0bbb4a79d0922efc68d40c5bc0f7355fb"} Dec 03 14:44:44 crc kubenswrapper[4805]: I1203 14:44:44.797897 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-tw4dn" event={"ID":"9cd62272-1077-4e13-8e99-c158fbb8ba1a","Type":"ContainerStarted","Data":"29c9b68e053f254afe0751a7456bc810567c2ddf318fee18c7b6db38b0c4c254"} Dec 03 14:44:44 crc kubenswrapper[4805]: I1203 14:44:44.805771 4805 scope.go:117] "RemoveContainer" containerID="d384c8b48d1a97e7c0c06910a1049b12545dc37afe379f1e5d0a1b926448aa6b" Dec 03 14:44:44 crc kubenswrapper[4805]: I1203 14:44:44.827355 4805 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-tw4dn" podStartSLOduration=2.445601189 podStartE2EDuration="2.82733656s" podCreationTimestamp="2025-12-03 14:44:42 +0000 UTC" firstStartedPulling="2025-12-03 14:44:43.718828519 +0000 UTC m=+2113.381745452" lastFinishedPulling="2025-12-03 14:44:44.1005639 +0000 UTC m=+2113.763480823" observedRunningTime="2025-12-03 14:44:44.814990886 +0000 UTC m=+2114.477907829" watchObservedRunningTime="2025-12-03 14:44:44.82733656 +0000 UTC m=+2114.490253483" Dec 03 14:44:59 crc kubenswrapper[4805]: I1203 14:44:59.789996 4805 prober.go:107] "Probe failed" probeType="Liveness" pod="openstack/ceilometer-0" podUID="aee26d6d-7304-4001-90f8-c0ed886a3cda" containerName="ceilometer-central-agent" probeResult="failure" output="command timed out" Dec 03 14:45:00 crc kubenswrapper[4805]: I1203 14:45:00.163995 4805 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29412885-nz8rt"] Dec 03 14:45:00 crc kubenswrapper[4805]: I1203 14:45:00.166081 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29412885-nz8rt" Dec 03 14:45:00 crc kubenswrapper[4805]: I1203 14:45:00.169096 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Dec 03 14:45:00 crc kubenswrapper[4805]: I1203 14:45:00.169398 4805 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Dec 03 14:45:00 crc kubenswrapper[4805]: I1203 14:45:00.184263 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29412885-nz8rt"] Dec 03 14:45:00 crc kubenswrapper[4805]: I1203 14:45:00.327232 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/9b9690b7-2a97-4043-a398-071a05fc9aea-config-volume\") pod \"collect-profiles-29412885-nz8rt\" (UID: \"9b9690b7-2a97-4043-a398-071a05fc9aea\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29412885-nz8rt" Dec 03 14:45:00 crc kubenswrapper[4805]: I1203 14:45:00.327500 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/9b9690b7-2a97-4043-a398-071a05fc9aea-secret-volume\") pod \"collect-profiles-29412885-nz8rt\" (UID: \"9b9690b7-2a97-4043-a398-071a05fc9aea\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29412885-nz8rt" Dec 03 14:45:00 crc kubenswrapper[4805]: I1203 14:45:00.327561 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-577fw\" (UniqueName: \"kubernetes.io/projected/9b9690b7-2a97-4043-a398-071a05fc9aea-kube-api-access-577fw\") pod \"collect-profiles-29412885-nz8rt\" (UID: \"9b9690b7-2a97-4043-a398-071a05fc9aea\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29412885-nz8rt" Dec 03 14:45:00 crc kubenswrapper[4805]: I1203 14:45:00.430121 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/9b9690b7-2a97-4043-a398-071a05fc9aea-secret-volume\") pod \"collect-profiles-29412885-nz8rt\" (UID: \"9b9690b7-2a97-4043-a398-071a05fc9aea\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29412885-nz8rt" Dec 03 14:45:00 crc kubenswrapper[4805]: I1203 14:45:00.430174 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-577fw\" (UniqueName: \"kubernetes.io/projected/9b9690b7-2a97-4043-a398-071a05fc9aea-kube-api-access-577fw\") pod \"collect-profiles-29412885-nz8rt\" (UID: \"9b9690b7-2a97-4043-a398-071a05fc9aea\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29412885-nz8rt" Dec 03 14:45:00 crc kubenswrapper[4805]: I1203 14:45:00.430312 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/9b9690b7-2a97-4043-a398-071a05fc9aea-config-volume\") pod \"collect-profiles-29412885-nz8rt\" (UID: \"9b9690b7-2a97-4043-a398-071a05fc9aea\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29412885-nz8rt" Dec 03 14:45:00 crc kubenswrapper[4805]: I1203 14:45:00.431183 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/9b9690b7-2a97-4043-a398-071a05fc9aea-config-volume\") pod \"collect-profiles-29412885-nz8rt\" (UID: \"9b9690b7-2a97-4043-a398-071a05fc9aea\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29412885-nz8rt" Dec 03 14:45:00 crc kubenswrapper[4805]: I1203 14:45:00.441390 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/9b9690b7-2a97-4043-a398-071a05fc9aea-secret-volume\") pod \"collect-profiles-29412885-nz8rt\" (UID: \"9b9690b7-2a97-4043-a398-071a05fc9aea\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29412885-nz8rt" Dec 03 14:45:00 crc kubenswrapper[4805]: I1203 14:45:00.456660 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-577fw\" (UniqueName: \"kubernetes.io/projected/9b9690b7-2a97-4043-a398-071a05fc9aea-kube-api-access-577fw\") pod \"collect-profiles-29412885-nz8rt\" (UID: \"9b9690b7-2a97-4043-a398-071a05fc9aea\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29412885-nz8rt" Dec 03 14:45:00 crc kubenswrapper[4805]: I1203 14:45:00.531370 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29412885-nz8rt" Dec 03 14:45:01 crc kubenswrapper[4805]: I1203 14:45:01.000209 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29412885-nz8rt"] Dec 03 14:45:01 crc kubenswrapper[4805]: W1203 14:45:01.005929 4805 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod9b9690b7_2a97_4043_a398_071a05fc9aea.slice/crio-d1803793459ad32f76807da72fd71818c40429dee1b8db4fc9aa5cf1e13ae77e WatchSource:0}: Error finding container d1803793459ad32f76807da72fd71818c40429dee1b8db4fc9aa5cf1e13ae77e: Status 404 returned error can't find the container with id d1803793459ad32f76807da72fd71818c40429dee1b8db4fc9aa5cf1e13ae77e Dec 03 14:45:01 crc kubenswrapper[4805]: I1203 14:45:01.969370 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29412885-nz8rt" event={"ID":"9b9690b7-2a97-4043-a398-071a05fc9aea","Type":"ContainerStarted","Data":"d1803793459ad32f76807da72fd71818c40429dee1b8db4fc9aa5cf1e13ae77e"} Dec 03 14:45:04 crc kubenswrapper[4805]: I1203 14:45:04.789613 4805 prober.go:107] "Probe failed" probeType="Liveness" pod="openstack/ceilometer-0" podUID="aee26d6d-7304-4001-90f8-c0ed886a3cda" containerName="ceilometer-central-agent" probeResult="failure" output="command timed out" Dec 03 14:45:09 crc kubenswrapper[4805]: I1203 14:45:09.794141 4805 prober.go:107] "Probe failed" probeType="Liveness" pod="openstack/ceilometer-0" podUID="aee26d6d-7304-4001-90f8-c0ed886a3cda" containerName="ceilometer-central-agent" probeResult="failure" output="command timed out" Dec 03 14:45:09 crc kubenswrapper[4805]: I1203 14:45:09.794586 4805 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openstack/ceilometer-0" Dec 03 14:45:09 crc kubenswrapper[4805]: I1203 14:45:09.795681 4805 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="ceilometer-central-agent" containerStatusID={"Type":"cri-o","ID":"73941e0d6546b90dd2830c242d337c821bfefb112e9c263e904f7c35bfea49e4"} pod="openstack/ceilometer-0" containerMessage="Container ceilometer-central-agent failed liveness probe, will be restarted" Dec 03 14:45:09 crc kubenswrapper[4805]: I1203 14:45:09.795884 4805 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="aee26d6d-7304-4001-90f8-c0ed886a3cda" containerName="ceilometer-central-agent" containerID="cri-o://73941e0d6546b90dd2830c242d337c821bfefb112e9c263e904f7c35bfea49e4" gracePeriod=30 Dec 03 14:45:13 crc kubenswrapper[4805]: I1203 14:45:13.917168 4805 patch_prober.go:28] interesting pod/machine-config-daemon-gskh4 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 03 14:45:13 crc kubenswrapper[4805]: I1203 14:45:13.918053 4805 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-gskh4" podUID="6098937f-e3f6-45e8-a647-4994a79cd711" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 03 14:45:24 crc kubenswrapper[4805]: I1203 14:45:24.794623 4805 prober.go:107] "Probe failed" probeType="Liveness" pod="openstack/ceilometer-0" podUID="aee26d6d-7304-4001-90f8-c0ed886a3cda" containerName="ceilometer-notification-agent" probeResult="failure" output="command timed out" Dec 03 14:45:43 crc kubenswrapper[4805]: I1203 14:45:43.917913 4805 patch_prober.go:28] interesting pod/machine-config-daemon-gskh4 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 03 14:45:43 crc kubenswrapper[4805]: I1203 14:45:43.918539 4805 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-gskh4" podUID="6098937f-e3f6-45e8-a647-4994a79cd711" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 03 14:45:46 crc kubenswrapper[4805]: I1203 14:45:46.336547 4805 patch_prober.go:28] interesting pod/controller-manager-797bfc7f65-k7wn4 container/controller-manager namespace/openshift-controller-manager: Liveness probe status=failure output="Get \"https://10.217.0.63:8443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Dec 03 14:45:46 crc kubenswrapper[4805]: I1203 14:45:46.336957 4805 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-controller-manager/controller-manager-797bfc7f65-k7wn4" podUID="98b69e82-9cbe-4d01-9a43-e8d94f911a3f" containerName="controller-manager" probeResult="failure" output="Get \"https://10.217.0.63:8443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Dec 03 14:45:50 crc kubenswrapper[4805]: I1203 14:45:50.492853 4805 prober.go:107] "Probe failed" probeType="Liveness" pod="openstack/ceilometer-0" podUID="aee26d6d-7304-4001-90f8-c0ed886a3cda" containerName="ceilometer-notification-agent" probeResult="failure" output=< Dec 03 14:45:50 crc kubenswrapper[4805]: Unkown error: Expecting value: line 1 column 1 (char 0) Dec 03 14:45:50 crc kubenswrapper[4805]: > Dec 03 14:46:03 crc kubenswrapper[4805]: I1203 14:46:03.915312 4805 prober.go:107] "Probe failed" probeType="Liveness" pod="metallb-system/frr-k8s-jsrqx" podUID="2ba805a2-6c47-4649-b8dc-eb92fb614437" containerName="frr" probeResult="failure" output="Get \"http://127.0.0.1:7573/livez\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 03 14:46:05 crc kubenswrapper[4805]: I1203 14:46:05.334576 4805 generic.go:334] "Generic (PLEG): container finished" podID="aee26d6d-7304-4001-90f8-c0ed886a3cda" containerID="73941e0d6546b90dd2830c242d337c821bfefb112e9c263e904f7c35bfea49e4" exitCode=-1 Dec 03 14:46:05 crc kubenswrapper[4805]: I1203 14:46:05.334634 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"aee26d6d-7304-4001-90f8-c0ed886a3cda","Type":"ContainerDied","Data":"73941e0d6546b90dd2830c242d337c821bfefb112e9c263e904f7c35bfea49e4"} Dec 03 14:46:13 crc kubenswrapper[4805]: I1203 14:46:13.917340 4805 patch_prober.go:28] interesting pod/machine-config-daemon-gskh4 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 03 14:46:13 crc kubenswrapper[4805]: I1203 14:46:13.917959 4805 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-gskh4" podUID="6098937f-e3f6-45e8-a647-4994a79cd711" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 03 14:46:13 crc kubenswrapper[4805]: I1203 14:46:13.918012 4805 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-gskh4" Dec 03 14:46:13 crc kubenswrapper[4805]: I1203 14:46:13.918751 4805 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"339cd7bd4a35abfc7cd21edc654a8e6b31d75042fa72434198d7d952d589476a"} pod="openshift-machine-config-operator/machine-config-daemon-gskh4" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 03 14:46:13 crc kubenswrapper[4805]: I1203 14:46:13.918816 4805 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-gskh4" podUID="6098937f-e3f6-45e8-a647-4994a79cd711" containerName="machine-config-daemon" containerID="cri-o://339cd7bd4a35abfc7cd21edc654a8e6b31d75042fa72434198d7d952d589476a" gracePeriod=600 Dec 03 14:46:19 crc kubenswrapper[4805]: I1203 14:46:19.730227 4805 prober.go:107] "Probe failed" probeType="Liveness" pod="openstack/ceilometer-0" podUID="aee26d6d-7304-4001-90f8-c0ed886a3cda" containerName="ceilometer-notification-agent" probeResult="failure" output=< Dec 03 14:46:19 crc kubenswrapper[4805]: Unkown error: Expecting value: line 1 column 1 (char 0) Dec 03 14:46:19 crc kubenswrapper[4805]: > Dec 03 14:46:19 crc kubenswrapper[4805]: I1203 14:46:19.730789 4805 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openstack/ceilometer-0" Dec 03 14:46:22 crc kubenswrapper[4805]: I1203 14:46:22.215172 4805 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-config-operator_machine-config-daemon-gskh4_6098937f-e3f6-45e8-a647-4994a79cd711/machine-config-daemon/8.log" Dec 03 14:46:22 crc kubenswrapper[4805]: I1203 14:46:22.217575 4805 generic.go:334] "Generic (PLEG): container finished" podID="6098937f-e3f6-45e8-a647-4994a79cd711" containerID="339cd7bd4a35abfc7cd21edc654a8e6b31d75042fa72434198d7d952d589476a" exitCode=-1 Dec 03 14:46:22 crc kubenswrapper[4805]: I1203 14:46:22.217624 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-gskh4" event={"ID":"6098937f-e3f6-45e8-a647-4994a79cd711","Type":"ContainerDied","Data":"339cd7bd4a35abfc7cd21edc654a8e6b31d75042fa72434198d7d952d589476a"} Dec 03 14:46:22 crc kubenswrapper[4805]: I1203 14:46:22.217661 4805 scope.go:117] "RemoveContainer" containerID="748e11b7efe02ff9a381aeee0aa4ddad32272e1564bc028d779c442a7b7f2149" Dec 03 14:46:25 crc kubenswrapper[4805]: I1203 14:46:25.255798 4805 generic.go:334] "Generic (PLEG): container finished" podID="9b9690b7-2a97-4043-a398-071a05fc9aea" containerID="52bf529a439b949affea292e2c996d7d0ec64acd9b851e33f7bd9cfde8608c0c" exitCode=0 Dec 03 14:46:25 crc kubenswrapper[4805]: I1203 14:46:25.255918 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29412885-nz8rt" event={"ID":"9b9690b7-2a97-4043-a398-071a05fc9aea","Type":"ContainerDied","Data":"52bf529a439b949affea292e2c996d7d0ec64acd9b851e33f7bd9cfde8608c0c"} Dec 03 14:46:25 crc kubenswrapper[4805]: I1203 14:46:25.262056 4805 generic.go:334] "Generic (PLEG): container finished" podID="9cd62272-1077-4e13-8e99-c158fbb8ba1a" containerID="29c9b68e053f254afe0751a7456bc810567c2ddf318fee18c7b6db38b0c4c254" exitCode=0 Dec 03 14:46:25 crc kubenswrapper[4805]: I1203 14:46:25.262117 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-tw4dn" event={"ID":"9cd62272-1077-4e13-8e99-c158fbb8ba1a","Type":"ContainerDied","Data":"29c9b68e053f254afe0751a7456bc810567c2ddf318fee18c7b6db38b0c4c254"} Dec 03 14:46:26 crc kubenswrapper[4805]: I1203 14:46:26.273305 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-gskh4" event={"ID":"6098937f-e3f6-45e8-a647-4994a79cd711","Type":"ContainerStarted","Data":"fec9f6923538948ef3e55903bb8e108d1f94095a493c6b5117232e21f41ab624"} Dec 03 14:46:26 crc kubenswrapper[4805]: I1203 14:46:26.761048 4805 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29412885-nz8rt" Dec 03 14:46:26 crc kubenswrapper[4805]: I1203 14:46:26.772804 4805 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-tw4dn" Dec 03 14:46:26 crc kubenswrapper[4805]: I1203 14:46:26.883673 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-edpm-ipam-neutron-metadata-default-certs-0\" (UniqueName: \"kubernetes.io/projected/9cd62272-1077-4e13-8e99-c158fbb8ba1a-openstack-edpm-ipam-neutron-metadata-default-certs-0\") pod \"9cd62272-1077-4e13-8e99-c158fbb8ba1a\" (UID: \"9cd62272-1077-4e13-8e99-c158fbb8ba1a\") " Dec 03 14:46:26 crc kubenswrapper[4805]: I1203 14:46:26.883746 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9cd62272-1077-4e13-8e99-c158fbb8ba1a-ovn-combined-ca-bundle\") pod \"9cd62272-1077-4e13-8e99-c158fbb8ba1a\" (UID: \"9cd62272-1077-4e13-8e99-c158fbb8ba1a\") " Dec 03 14:46:26 crc kubenswrapper[4805]: I1203 14:46:26.883880 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9cd62272-1077-4e13-8e99-c158fbb8ba1a-nova-combined-ca-bundle\") pod \"9cd62272-1077-4e13-8e99-c158fbb8ba1a\" (UID: \"9cd62272-1077-4e13-8e99-c158fbb8ba1a\") " Dec 03 14:46:26 crc kubenswrapper[4805]: I1203 14:46:26.883906 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-edpm-ipam-ovn-default-certs-0\" (UniqueName: \"kubernetes.io/projected/9cd62272-1077-4e13-8e99-c158fbb8ba1a-openstack-edpm-ipam-ovn-default-certs-0\") pod \"9cd62272-1077-4e13-8e99-c158fbb8ba1a\" (UID: \"9cd62272-1077-4e13-8e99-c158fbb8ba1a\") " Dec 03 14:46:26 crc kubenswrapper[4805]: I1203 14:46:26.883946 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-edpm-ipam-telemetry-default-certs-0\" (UniqueName: \"kubernetes.io/projected/9cd62272-1077-4e13-8e99-c158fbb8ba1a-openstack-edpm-ipam-telemetry-default-certs-0\") pod \"9cd62272-1077-4e13-8e99-c158fbb8ba1a\" (UID: \"9cd62272-1077-4e13-8e99-c158fbb8ba1a\") " Dec 03 14:46:26 crc kubenswrapper[4805]: I1203 14:46:26.884012 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9cd62272-1077-4e13-8e99-c158fbb8ba1a-telemetry-combined-ca-bundle\") pod \"9cd62272-1077-4e13-8e99-c158fbb8ba1a\" (UID: \"9cd62272-1077-4e13-8e99-c158fbb8ba1a\") " Dec 03 14:46:26 crc kubenswrapper[4805]: I1203 14:46:26.884074 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/9b9690b7-2a97-4043-a398-071a05fc9aea-secret-volume\") pod \"9b9690b7-2a97-4043-a398-071a05fc9aea\" (UID: \"9b9690b7-2a97-4043-a398-071a05fc9aea\") " Dec 03 14:46:26 crc kubenswrapper[4805]: I1203 14:46:26.884099 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9cd62272-1077-4e13-8e99-c158fbb8ba1a-neutron-metadata-combined-ca-bundle\") pod \"9cd62272-1077-4e13-8e99-c158fbb8ba1a\" (UID: \"9cd62272-1077-4e13-8e99-c158fbb8ba1a\") " Dec 03 14:46:26 crc kubenswrapper[4805]: I1203 14:46:26.884143 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9cd62272-1077-4e13-8e99-c158fbb8ba1a-libvirt-combined-ca-bundle\") pod \"9cd62272-1077-4e13-8e99-c158fbb8ba1a\" (UID: \"9cd62272-1077-4e13-8e99-c158fbb8ba1a\") " Dec 03 14:46:26 crc kubenswrapper[4805]: I1203 14:46:26.884178 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-edpm-ipam-libvirt-default-certs-0\" (UniqueName: \"kubernetes.io/projected/9cd62272-1077-4e13-8e99-c158fbb8ba1a-openstack-edpm-ipam-libvirt-default-certs-0\") pod \"9cd62272-1077-4e13-8e99-c158fbb8ba1a\" (UID: \"9cd62272-1077-4e13-8e99-c158fbb8ba1a\") " Dec 03 14:46:26 crc kubenswrapper[4805]: I1203 14:46:26.884230 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/9cd62272-1077-4e13-8e99-c158fbb8ba1a-inventory\") pod \"9cd62272-1077-4e13-8e99-c158fbb8ba1a\" (UID: \"9cd62272-1077-4e13-8e99-c158fbb8ba1a\") " Dec 03 14:46:26 crc kubenswrapper[4805]: I1203 14:46:26.884253 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9cd62272-1077-4e13-8e99-c158fbb8ba1a-repo-setup-combined-ca-bundle\") pod \"9cd62272-1077-4e13-8e99-c158fbb8ba1a\" (UID: \"9cd62272-1077-4e13-8e99-c158fbb8ba1a\") " Dec 03 14:46:26 crc kubenswrapper[4805]: I1203 14:46:26.884279 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/9cd62272-1077-4e13-8e99-c158fbb8ba1a-ssh-key\") pod \"9cd62272-1077-4e13-8e99-c158fbb8ba1a\" (UID: \"9cd62272-1077-4e13-8e99-c158fbb8ba1a\") " Dec 03 14:46:26 crc kubenswrapper[4805]: I1203 14:46:26.884315 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9cd62272-1077-4e13-8e99-c158fbb8ba1a-bootstrap-combined-ca-bundle\") pod \"9cd62272-1077-4e13-8e99-c158fbb8ba1a\" (UID: \"9cd62272-1077-4e13-8e99-c158fbb8ba1a\") " Dec 03 14:46:26 crc kubenswrapper[4805]: I1203 14:46:26.884338 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-577fw\" (UniqueName: \"kubernetes.io/projected/9b9690b7-2a97-4043-a398-071a05fc9aea-kube-api-access-577fw\") pod \"9b9690b7-2a97-4043-a398-071a05fc9aea\" (UID: \"9b9690b7-2a97-4043-a398-071a05fc9aea\") " Dec 03 14:46:26 crc kubenswrapper[4805]: I1203 14:46:26.884360 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/9b9690b7-2a97-4043-a398-071a05fc9aea-config-volume\") pod \"9b9690b7-2a97-4043-a398-071a05fc9aea\" (UID: \"9b9690b7-2a97-4043-a398-071a05fc9aea\") " Dec 03 14:46:26 crc kubenswrapper[4805]: I1203 14:46:26.884404 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-b6rfz\" (UniqueName: \"kubernetes.io/projected/9cd62272-1077-4e13-8e99-c158fbb8ba1a-kube-api-access-b6rfz\") pod \"9cd62272-1077-4e13-8e99-c158fbb8ba1a\" (UID: \"9cd62272-1077-4e13-8e99-c158fbb8ba1a\") " Dec 03 14:46:26 crc kubenswrapper[4805]: I1203 14:46:26.893559 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9cd62272-1077-4e13-8e99-c158fbb8ba1a-telemetry-combined-ca-bundle" (OuterVolumeSpecName: "telemetry-combined-ca-bundle") pod "9cd62272-1077-4e13-8e99-c158fbb8ba1a" (UID: "9cd62272-1077-4e13-8e99-c158fbb8ba1a"). InnerVolumeSpecName "telemetry-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 14:46:26 crc kubenswrapper[4805]: I1203 14:46:26.893637 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9cd62272-1077-4e13-8e99-c158fbb8ba1a-kube-api-access-b6rfz" (OuterVolumeSpecName: "kube-api-access-b6rfz") pod "9cd62272-1077-4e13-8e99-c158fbb8ba1a" (UID: "9cd62272-1077-4e13-8e99-c158fbb8ba1a"). InnerVolumeSpecName "kube-api-access-b6rfz". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 14:46:26 crc kubenswrapper[4805]: I1203 14:46:26.893969 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9cd62272-1077-4e13-8e99-c158fbb8ba1a-libvirt-combined-ca-bundle" (OuterVolumeSpecName: "libvirt-combined-ca-bundle") pod "9cd62272-1077-4e13-8e99-c158fbb8ba1a" (UID: "9cd62272-1077-4e13-8e99-c158fbb8ba1a"). InnerVolumeSpecName "libvirt-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 14:46:26 crc kubenswrapper[4805]: I1203 14:46:26.894898 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9cd62272-1077-4e13-8e99-c158fbb8ba1a-openstack-edpm-ipam-neutron-metadata-default-certs-0" (OuterVolumeSpecName: "openstack-edpm-ipam-neutron-metadata-default-certs-0") pod "9cd62272-1077-4e13-8e99-c158fbb8ba1a" (UID: "9cd62272-1077-4e13-8e99-c158fbb8ba1a"). InnerVolumeSpecName "openstack-edpm-ipam-neutron-metadata-default-certs-0". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 14:46:26 crc kubenswrapper[4805]: I1203 14:46:26.895045 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9cd62272-1077-4e13-8e99-c158fbb8ba1a-ovn-combined-ca-bundle" (OuterVolumeSpecName: "ovn-combined-ca-bundle") pod "9cd62272-1077-4e13-8e99-c158fbb8ba1a" (UID: "9cd62272-1077-4e13-8e99-c158fbb8ba1a"). InnerVolumeSpecName "ovn-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 14:46:26 crc kubenswrapper[4805]: I1203 14:46:26.895587 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9b9690b7-2a97-4043-a398-071a05fc9aea-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "9b9690b7-2a97-4043-a398-071a05fc9aea" (UID: "9b9690b7-2a97-4043-a398-071a05fc9aea"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 14:46:26 crc kubenswrapper[4805]: I1203 14:46:26.897118 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9b9690b7-2a97-4043-a398-071a05fc9aea-config-volume" (OuterVolumeSpecName: "config-volume") pod "9b9690b7-2a97-4043-a398-071a05fc9aea" (UID: "9b9690b7-2a97-4043-a398-071a05fc9aea"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 14:46:26 crc kubenswrapper[4805]: I1203 14:46:26.898203 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9cd62272-1077-4e13-8e99-c158fbb8ba1a-neutron-metadata-combined-ca-bundle" (OuterVolumeSpecName: "neutron-metadata-combined-ca-bundle") pod "9cd62272-1077-4e13-8e99-c158fbb8ba1a" (UID: "9cd62272-1077-4e13-8e99-c158fbb8ba1a"). InnerVolumeSpecName "neutron-metadata-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 14:46:26 crc kubenswrapper[4805]: I1203 14:46:26.898822 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9cd62272-1077-4e13-8e99-c158fbb8ba1a-openstack-edpm-ipam-libvirt-default-certs-0" (OuterVolumeSpecName: "openstack-edpm-ipam-libvirt-default-certs-0") pod "9cd62272-1077-4e13-8e99-c158fbb8ba1a" (UID: "9cd62272-1077-4e13-8e99-c158fbb8ba1a"). InnerVolumeSpecName "openstack-edpm-ipam-libvirt-default-certs-0". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 14:46:26 crc kubenswrapper[4805]: I1203 14:46:26.900133 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9cd62272-1077-4e13-8e99-c158fbb8ba1a-openstack-edpm-ipam-telemetry-default-certs-0" (OuterVolumeSpecName: "openstack-edpm-ipam-telemetry-default-certs-0") pod "9cd62272-1077-4e13-8e99-c158fbb8ba1a" (UID: "9cd62272-1077-4e13-8e99-c158fbb8ba1a"). InnerVolumeSpecName "openstack-edpm-ipam-telemetry-default-certs-0". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 14:46:26 crc kubenswrapper[4805]: I1203 14:46:26.900204 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9b9690b7-2a97-4043-a398-071a05fc9aea-kube-api-access-577fw" (OuterVolumeSpecName: "kube-api-access-577fw") pod "9b9690b7-2a97-4043-a398-071a05fc9aea" (UID: "9b9690b7-2a97-4043-a398-071a05fc9aea"). InnerVolumeSpecName "kube-api-access-577fw". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 14:46:26 crc kubenswrapper[4805]: I1203 14:46:26.900227 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9cd62272-1077-4e13-8e99-c158fbb8ba1a-openstack-edpm-ipam-ovn-default-certs-0" (OuterVolumeSpecName: "openstack-edpm-ipam-ovn-default-certs-0") pod "9cd62272-1077-4e13-8e99-c158fbb8ba1a" (UID: "9cd62272-1077-4e13-8e99-c158fbb8ba1a"). InnerVolumeSpecName "openstack-edpm-ipam-ovn-default-certs-0". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 14:46:26 crc kubenswrapper[4805]: I1203 14:46:26.900499 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9cd62272-1077-4e13-8e99-c158fbb8ba1a-bootstrap-combined-ca-bundle" (OuterVolumeSpecName: "bootstrap-combined-ca-bundle") pod "9cd62272-1077-4e13-8e99-c158fbb8ba1a" (UID: "9cd62272-1077-4e13-8e99-c158fbb8ba1a"). InnerVolumeSpecName "bootstrap-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 14:46:26 crc kubenswrapper[4805]: I1203 14:46:26.900686 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9cd62272-1077-4e13-8e99-c158fbb8ba1a-nova-combined-ca-bundle" (OuterVolumeSpecName: "nova-combined-ca-bundle") pod "9cd62272-1077-4e13-8e99-c158fbb8ba1a" (UID: "9cd62272-1077-4e13-8e99-c158fbb8ba1a"). InnerVolumeSpecName "nova-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 14:46:26 crc kubenswrapper[4805]: I1203 14:46:26.901400 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9cd62272-1077-4e13-8e99-c158fbb8ba1a-repo-setup-combined-ca-bundle" (OuterVolumeSpecName: "repo-setup-combined-ca-bundle") pod "9cd62272-1077-4e13-8e99-c158fbb8ba1a" (UID: "9cd62272-1077-4e13-8e99-c158fbb8ba1a"). InnerVolumeSpecName "repo-setup-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 14:46:26 crc kubenswrapper[4805]: I1203 14:46:26.923380 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9cd62272-1077-4e13-8e99-c158fbb8ba1a-inventory" (OuterVolumeSpecName: "inventory") pod "9cd62272-1077-4e13-8e99-c158fbb8ba1a" (UID: "9cd62272-1077-4e13-8e99-c158fbb8ba1a"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 14:46:26 crc kubenswrapper[4805]: I1203 14:46:26.925095 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9cd62272-1077-4e13-8e99-c158fbb8ba1a-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "9cd62272-1077-4e13-8e99-c158fbb8ba1a" (UID: "9cd62272-1077-4e13-8e99-c158fbb8ba1a"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 14:46:26 crc kubenswrapper[4805]: I1203 14:46:26.986831 4805 reconciler_common.go:293] "Volume detached for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9cd62272-1077-4e13-8e99-c158fbb8ba1a-libvirt-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 14:46:26 crc kubenswrapper[4805]: I1203 14:46:26.986932 4805 reconciler_common.go:293] "Volume detached for volume \"openstack-edpm-ipam-libvirt-default-certs-0\" (UniqueName: \"kubernetes.io/projected/9cd62272-1077-4e13-8e99-c158fbb8ba1a-openstack-edpm-ipam-libvirt-default-certs-0\") on node \"crc\" DevicePath \"\"" Dec 03 14:46:26 crc kubenswrapper[4805]: I1203 14:46:26.986946 4805 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/9cd62272-1077-4e13-8e99-c158fbb8ba1a-inventory\") on node \"crc\" DevicePath \"\"" Dec 03 14:46:26 crc kubenswrapper[4805]: I1203 14:46:26.986968 4805 reconciler_common.go:293] "Volume detached for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9cd62272-1077-4e13-8e99-c158fbb8ba1a-repo-setup-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 14:46:26 crc kubenswrapper[4805]: I1203 14:46:26.986977 4805 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/9cd62272-1077-4e13-8e99-c158fbb8ba1a-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 03 14:46:26 crc kubenswrapper[4805]: I1203 14:46:26.986986 4805 reconciler_common.go:293] "Volume detached for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9cd62272-1077-4e13-8e99-c158fbb8ba1a-bootstrap-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 14:46:26 crc kubenswrapper[4805]: I1203 14:46:26.986994 4805 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-577fw\" (UniqueName: \"kubernetes.io/projected/9b9690b7-2a97-4043-a398-071a05fc9aea-kube-api-access-577fw\") on node \"crc\" DevicePath \"\"" Dec 03 14:46:26 crc kubenswrapper[4805]: I1203 14:46:26.987005 4805 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/9b9690b7-2a97-4043-a398-071a05fc9aea-config-volume\") on node \"crc\" DevicePath \"\"" Dec 03 14:46:26 crc kubenswrapper[4805]: I1203 14:46:26.987012 4805 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-b6rfz\" (UniqueName: \"kubernetes.io/projected/9cd62272-1077-4e13-8e99-c158fbb8ba1a-kube-api-access-b6rfz\") on node \"crc\" DevicePath \"\"" Dec 03 14:46:26 crc kubenswrapper[4805]: I1203 14:46:26.987021 4805 reconciler_common.go:293] "Volume detached for volume \"openstack-edpm-ipam-neutron-metadata-default-certs-0\" (UniqueName: \"kubernetes.io/projected/9cd62272-1077-4e13-8e99-c158fbb8ba1a-openstack-edpm-ipam-neutron-metadata-default-certs-0\") on node \"crc\" DevicePath \"\"" Dec 03 14:46:26 crc kubenswrapper[4805]: I1203 14:46:26.987031 4805 reconciler_common.go:293] "Volume detached for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9cd62272-1077-4e13-8e99-c158fbb8ba1a-ovn-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 14:46:26 crc kubenswrapper[4805]: I1203 14:46:26.987039 4805 reconciler_common.go:293] "Volume detached for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9cd62272-1077-4e13-8e99-c158fbb8ba1a-nova-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 14:46:26 crc kubenswrapper[4805]: I1203 14:46:26.987048 4805 reconciler_common.go:293] "Volume detached for volume \"openstack-edpm-ipam-ovn-default-certs-0\" (UniqueName: \"kubernetes.io/projected/9cd62272-1077-4e13-8e99-c158fbb8ba1a-openstack-edpm-ipam-ovn-default-certs-0\") on node \"crc\" DevicePath \"\"" Dec 03 14:46:26 crc kubenswrapper[4805]: I1203 14:46:26.987060 4805 reconciler_common.go:293] "Volume detached for volume \"openstack-edpm-ipam-telemetry-default-certs-0\" (UniqueName: \"kubernetes.io/projected/9cd62272-1077-4e13-8e99-c158fbb8ba1a-openstack-edpm-ipam-telemetry-default-certs-0\") on node \"crc\" DevicePath \"\"" Dec 03 14:46:26 crc kubenswrapper[4805]: I1203 14:46:26.987069 4805 reconciler_common.go:293] "Volume detached for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9cd62272-1077-4e13-8e99-c158fbb8ba1a-telemetry-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 14:46:26 crc kubenswrapper[4805]: I1203 14:46:26.987078 4805 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/9b9690b7-2a97-4043-a398-071a05fc9aea-secret-volume\") on node \"crc\" DevicePath \"\"" Dec 03 14:46:26 crc kubenswrapper[4805]: I1203 14:46:26.987096 4805 reconciler_common.go:293] "Volume detached for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9cd62272-1077-4e13-8e99-c158fbb8ba1a-neutron-metadata-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 14:46:27 crc kubenswrapper[4805]: I1203 14:46:27.282867 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-tw4dn" event={"ID":"9cd62272-1077-4e13-8e99-c158fbb8ba1a","Type":"ContainerDied","Data":"2aef129fd3f3732cb554fc9fa9b5e3f0bbb4a79d0922efc68d40c5bc0f7355fb"} Dec 03 14:46:27 crc kubenswrapper[4805]: I1203 14:46:27.283237 4805 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="2aef129fd3f3732cb554fc9fa9b5e3f0bbb4a79d0922efc68d40c5bc0f7355fb" Dec 03 14:46:27 crc kubenswrapper[4805]: I1203 14:46:27.283356 4805 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-tw4dn" Dec 03 14:46:27 crc kubenswrapper[4805]: I1203 14:46:27.285992 4805 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29412885-nz8rt" Dec 03 14:46:27 crc kubenswrapper[4805]: I1203 14:46:27.285980 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29412885-nz8rt" event={"ID":"9b9690b7-2a97-4043-a398-071a05fc9aea","Type":"ContainerDied","Data":"d1803793459ad32f76807da72fd71818c40429dee1b8db4fc9aa5cf1e13ae77e"} Dec 03 14:46:27 crc kubenswrapper[4805]: I1203 14:46:27.286197 4805 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="d1803793459ad32f76807da72fd71818c40429dee1b8db4fc9aa5cf1e13ae77e" Dec 03 14:46:27 crc kubenswrapper[4805]: I1203 14:46:27.290257 4805 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="ceilometer-notification-agent" containerStatusID={"Type":"cri-o","ID":"c6e36473c9af1c0d702a0744c34ad51f28c604114568f3a35c2d7eaf0de6e461"} pod="openstack/ceilometer-0" containerMessage="Container ceilometer-notification-agent failed liveness probe, will be restarted" Dec 03 14:46:27 crc kubenswrapper[4805]: I1203 14:46:27.290585 4805 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="aee26d6d-7304-4001-90f8-c0ed886a3cda" containerName="ceilometer-notification-agent" containerID="cri-o://c6e36473c9af1c0d702a0744c34ad51f28c604114568f3a35c2d7eaf0de6e461" gracePeriod=30 Dec 03 14:46:27 crc kubenswrapper[4805]: I1203 14:46:27.290671 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"aee26d6d-7304-4001-90f8-c0ed886a3cda","Type":"ContainerStarted","Data":"c5da23817a0fadd49f2754c6b0126b250c6c2cee661c805c0a922ca37d9d5a30"} Dec 03 14:46:27 crc kubenswrapper[4805]: I1203 14:46:27.461146 4805 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-edpm-deployment-openstack-edpm-ipam-hbx5q"] Dec 03 14:46:27 crc kubenswrapper[4805]: E1203 14:46:27.461769 4805 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9b9690b7-2a97-4043-a398-071a05fc9aea" containerName="collect-profiles" Dec 03 14:46:27 crc kubenswrapper[4805]: I1203 14:46:27.461857 4805 state_mem.go:107] "Deleted CPUSet assignment" podUID="9b9690b7-2a97-4043-a398-071a05fc9aea" containerName="collect-profiles" Dec 03 14:46:27 crc kubenswrapper[4805]: E1203 14:46:27.461955 4805 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9cd62272-1077-4e13-8e99-c158fbb8ba1a" containerName="install-certs-edpm-deployment-openstack-edpm-ipam" Dec 03 14:46:27 crc kubenswrapper[4805]: I1203 14:46:27.462010 4805 state_mem.go:107] "Deleted CPUSet assignment" podUID="9cd62272-1077-4e13-8e99-c158fbb8ba1a" containerName="install-certs-edpm-deployment-openstack-edpm-ipam" Dec 03 14:46:27 crc kubenswrapper[4805]: I1203 14:46:27.462242 4805 memory_manager.go:354] "RemoveStaleState removing state" podUID="9b9690b7-2a97-4043-a398-071a05fc9aea" containerName="collect-profiles" Dec 03 14:46:27 crc kubenswrapper[4805]: I1203 14:46:27.462307 4805 memory_manager.go:354] "RemoveStaleState removing state" podUID="9cd62272-1077-4e13-8e99-c158fbb8ba1a" containerName="install-certs-edpm-deployment-openstack-edpm-ipam" Dec 03 14:46:27 crc kubenswrapper[4805]: I1203 14:46:27.462945 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-hbx5q" Dec 03 14:46:27 crc kubenswrapper[4805]: I1203 14:46:27.467956 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-n7jt4" Dec 03 14:46:27 crc kubenswrapper[4805]: I1203 14:46:27.468250 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Dec 03 14:46:27 crc kubenswrapper[4805]: I1203 14:46:27.468075 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Dec 03 14:46:27 crc kubenswrapper[4805]: I1203 14:46:27.468708 4805 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 03 14:46:27 crc kubenswrapper[4805]: I1203 14:46:27.468895 4805 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovncontroller-config" Dec 03 14:46:27 crc kubenswrapper[4805]: I1203 14:46:27.482986 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-edpm-deployment-openstack-edpm-ipam-hbx5q"] Dec 03 14:46:27 crc kubenswrapper[4805]: I1203 14:46:27.598221 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/af960cbb-259d-409e-9140-2d66af349200-ssh-key\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-hbx5q\" (UID: \"af960cbb-259d-409e-9140-2d66af349200\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-hbx5q" Dec 03 14:46:27 crc kubenswrapper[4805]: I1203 14:46:27.598371 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/af960cbb-259d-409e-9140-2d66af349200-ovn-combined-ca-bundle\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-hbx5q\" (UID: \"af960cbb-259d-409e-9140-2d66af349200\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-hbx5q" Dec 03 14:46:27 crc kubenswrapper[4805]: I1203 14:46:27.598583 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovncontroller-config-0\" (UniqueName: \"kubernetes.io/configmap/af960cbb-259d-409e-9140-2d66af349200-ovncontroller-config-0\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-hbx5q\" (UID: \"af960cbb-259d-409e-9140-2d66af349200\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-hbx5q" Dec 03 14:46:27 crc kubenswrapper[4805]: I1203 14:46:27.598768 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/af960cbb-259d-409e-9140-2d66af349200-inventory\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-hbx5q\" (UID: \"af960cbb-259d-409e-9140-2d66af349200\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-hbx5q" Dec 03 14:46:27 crc kubenswrapper[4805]: I1203 14:46:27.598995 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8rnqb\" (UniqueName: \"kubernetes.io/projected/af960cbb-259d-409e-9140-2d66af349200-kube-api-access-8rnqb\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-hbx5q\" (UID: \"af960cbb-259d-409e-9140-2d66af349200\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-hbx5q" Dec 03 14:46:27 crc kubenswrapper[4805]: I1203 14:46:27.700571 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/af960cbb-259d-409e-9140-2d66af349200-ssh-key\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-hbx5q\" (UID: \"af960cbb-259d-409e-9140-2d66af349200\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-hbx5q" Dec 03 14:46:27 crc kubenswrapper[4805]: I1203 14:46:27.700642 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/af960cbb-259d-409e-9140-2d66af349200-ovn-combined-ca-bundle\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-hbx5q\" (UID: \"af960cbb-259d-409e-9140-2d66af349200\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-hbx5q" Dec 03 14:46:27 crc kubenswrapper[4805]: I1203 14:46:27.700679 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovncontroller-config-0\" (UniqueName: \"kubernetes.io/configmap/af960cbb-259d-409e-9140-2d66af349200-ovncontroller-config-0\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-hbx5q\" (UID: \"af960cbb-259d-409e-9140-2d66af349200\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-hbx5q" Dec 03 14:46:27 crc kubenswrapper[4805]: I1203 14:46:27.700724 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/af960cbb-259d-409e-9140-2d66af349200-inventory\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-hbx5q\" (UID: \"af960cbb-259d-409e-9140-2d66af349200\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-hbx5q" Dec 03 14:46:27 crc kubenswrapper[4805]: I1203 14:46:27.700778 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8rnqb\" (UniqueName: \"kubernetes.io/projected/af960cbb-259d-409e-9140-2d66af349200-kube-api-access-8rnqb\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-hbx5q\" (UID: \"af960cbb-259d-409e-9140-2d66af349200\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-hbx5q" Dec 03 14:46:27 crc kubenswrapper[4805]: I1203 14:46:27.702418 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovncontroller-config-0\" (UniqueName: \"kubernetes.io/configmap/af960cbb-259d-409e-9140-2d66af349200-ovncontroller-config-0\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-hbx5q\" (UID: \"af960cbb-259d-409e-9140-2d66af349200\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-hbx5q" Dec 03 14:46:27 crc kubenswrapper[4805]: I1203 14:46:27.707561 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/af960cbb-259d-409e-9140-2d66af349200-ovn-combined-ca-bundle\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-hbx5q\" (UID: \"af960cbb-259d-409e-9140-2d66af349200\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-hbx5q" Dec 03 14:46:27 crc kubenswrapper[4805]: I1203 14:46:27.717107 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/af960cbb-259d-409e-9140-2d66af349200-inventory\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-hbx5q\" (UID: \"af960cbb-259d-409e-9140-2d66af349200\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-hbx5q" Dec 03 14:46:27 crc kubenswrapper[4805]: I1203 14:46:27.717569 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/af960cbb-259d-409e-9140-2d66af349200-ssh-key\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-hbx5q\" (UID: \"af960cbb-259d-409e-9140-2d66af349200\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-hbx5q" Dec 03 14:46:27 crc kubenswrapper[4805]: I1203 14:46:27.721205 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8rnqb\" (UniqueName: \"kubernetes.io/projected/af960cbb-259d-409e-9140-2d66af349200-kube-api-access-8rnqb\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-hbx5q\" (UID: \"af960cbb-259d-409e-9140-2d66af349200\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-hbx5q" Dec 03 14:46:27 crc kubenswrapper[4805]: I1203 14:46:27.782068 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-hbx5q" Dec 03 14:46:27 crc kubenswrapper[4805]: I1203 14:46:27.853592 4805 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29412840-gj7pr"] Dec 03 14:46:27 crc kubenswrapper[4805]: I1203 14:46:27.876696 4805 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29412840-gj7pr"] Dec 03 14:46:28 crc kubenswrapper[4805]: W1203 14:46:28.375869 4805 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podaf960cbb_259d_409e_9140_2d66af349200.slice/crio-248e4a387b77e040f38b6ff56719bd999d4da5e34fc3f04eab2bb2c7efb26faa WatchSource:0}: Error finding container 248e4a387b77e040f38b6ff56719bd999d4da5e34fc3f04eab2bb2c7efb26faa: Status 404 returned error can't find the container with id 248e4a387b77e040f38b6ff56719bd999d4da5e34fc3f04eab2bb2c7efb26faa Dec 03 14:46:28 crc kubenswrapper[4805]: I1203 14:46:28.375880 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-edpm-deployment-openstack-edpm-ipam-hbx5q"] Dec 03 14:46:28 crc kubenswrapper[4805]: I1203 14:46:28.709055 4805 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8b954d96-a62c-434e-a340-8446d778c545" path="/var/lib/kubelet/pods/8b954d96-a62c-434e-a340-8446d778c545/volumes" Dec 03 14:46:29 crc kubenswrapper[4805]: I1203 14:46:29.309927 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-hbx5q" event={"ID":"af960cbb-259d-409e-9140-2d66af349200","Type":"ContainerStarted","Data":"e7911b592437f4191d5e8481d23241b41d1b50b1df9adf6903e96d279f2c689f"} Dec 03 14:46:29 crc kubenswrapper[4805]: I1203 14:46:29.310278 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-hbx5q" event={"ID":"af960cbb-259d-409e-9140-2d66af349200","Type":"ContainerStarted","Data":"248e4a387b77e040f38b6ff56719bd999d4da5e34fc3f04eab2bb2c7efb26faa"} Dec 03 14:46:29 crc kubenswrapper[4805]: I1203 14:46:29.336643 4805 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-hbx5q" podStartSLOduration=1.8502366829999999 podStartE2EDuration="2.336620896s" podCreationTimestamp="2025-12-03 14:46:27 +0000 UTC" firstStartedPulling="2025-12-03 14:46:28.378749962 +0000 UTC m=+2218.041666885" lastFinishedPulling="2025-12-03 14:46:28.865134175 +0000 UTC m=+2218.528051098" observedRunningTime="2025-12-03 14:46:29.326748119 +0000 UTC m=+2218.989665042" watchObservedRunningTime="2025-12-03 14:46:29.336620896 +0000 UTC m=+2218.999537819" Dec 03 14:46:30 crc kubenswrapper[4805]: I1203 14:46:30.324889 4805 generic.go:334] "Generic (PLEG): container finished" podID="aee26d6d-7304-4001-90f8-c0ed886a3cda" containerID="c6e36473c9af1c0d702a0744c34ad51f28c604114568f3a35c2d7eaf0de6e461" exitCode=0 Dec 03 14:46:30 crc kubenswrapper[4805]: I1203 14:46:30.325647 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"aee26d6d-7304-4001-90f8-c0ed886a3cda","Type":"ContainerDied","Data":"c6e36473c9af1c0d702a0744c34ad51f28c604114568f3a35c2d7eaf0de6e461"} Dec 03 14:46:32 crc kubenswrapper[4805]: I1203 14:46:32.351236 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"aee26d6d-7304-4001-90f8-c0ed886a3cda","Type":"ContainerStarted","Data":"5166a2fd0b33128b5411b9ce35f22d267a1f26fddd2c64c87f9709551024044c"} Dec 03 14:46:44 crc kubenswrapper[4805]: I1203 14:46:44.918761 4805 scope.go:117] "RemoveContainer" containerID="8f4f233b53a4e215b29e4cb9baef3ccf262fa81a4f6bddc1a28d3a846ab83332" Dec 03 14:47:47 crc kubenswrapper[4805]: I1203 14:47:47.100898 4805 generic.go:334] "Generic (PLEG): container finished" podID="af960cbb-259d-409e-9140-2d66af349200" containerID="e7911b592437f4191d5e8481d23241b41d1b50b1df9adf6903e96d279f2c689f" exitCode=0 Dec 03 14:47:47 crc kubenswrapper[4805]: I1203 14:47:47.100952 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-hbx5q" event={"ID":"af960cbb-259d-409e-9140-2d66af349200","Type":"ContainerDied","Data":"e7911b592437f4191d5e8481d23241b41d1b50b1df9adf6903e96d279f2c689f"} Dec 03 14:47:48 crc kubenswrapper[4805]: I1203 14:47:48.600236 4805 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-hbx5q" Dec 03 14:47:48 crc kubenswrapper[4805]: I1203 14:47:48.750199 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/af960cbb-259d-409e-9140-2d66af349200-ovn-combined-ca-bundle\") pod \"af960cbb-259d-409e-9140-2d66af349200\" (UID: \"af960cbb-259d-409e-9140-2d66af349200\") " Dec 03 14:47:48 crc kubenswrapper[4805]: I1203 14:47:48.750825 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/af960cbb-259d-409e-9140-2d66af349200-inventory\") pod \"af960cbb-259d-409e-9140-2d66af349200\" (UID: \"af960cbb-259d-409e-9140-2d66af349200\") " Dec 03 14:47:48 crc kubenswrapper[4805]: I1203 14:47:48.751498 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovncontroller-config-0\" (UniqueName: \"kubernetes.io/configmap/af960cbb-259d-409e-9140-2d66af349200-ovncontroller-config-0\") pod \"af960cbb-259d-409e-9140-2d66af349200\" (UID: \"af960cbb-259d-409e-9140-2d66af349200\") " Dec 03 14:47:48 crc kubenswrapper[4805]: I1203 14:47:48.751749 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/af960cbb-259d-409e-9140-2d66af349200-ssh-key\") pod \"af960cbb-259d-409e-9140-2d66af349200\" (UID: \"af960cbb-259d-409e-9140-2d66af349200\") " Dec 03 14:47:48 crc kubenswrapper[4805]: I1203 14:47:48.752136 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8rnqb\" (UniqueName: \"kubernetes.io/projected/af960cbb-259d-409e-9140-2d66af349200-kube-api-access-8rnqb\") pod \"af960cbb-259d-409e-9140-2d66af349200\" (UID: \"af960cbb-259d-409e-9140-2d66af349200\") " Dec 03 14:47:48 crc kubenswrapper[4805]: I1203 14:47:48.759077 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/af960cbb-259d-409e-9140-2d66af349200-kube-api-access-8rnqb" (OuterVolumeSpecName: "kube-api-access-8rnqb") pod "af960cbb-259d-409e-9140-2d66af349200" (UID: "af960cbb-259d-409e-9140-2d66af349200"). InnerVolumeSpecName "kube-api-access-8rnqb". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 14:47:48 crc kubenswrapper[4805]: I1203 14:47:48.762059 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/af960cbb-259d-409e-9140-2d66af349200-ovn-combined-ca-bundle" (OuterVolumeSpecName: "ovn-combined-ca-bundle") pod "af960cbb-259d-409e-9140-2d66af349200" (UID: "af960cbb-259d-409e-9140-2d66af349200"). InnerVolumeSpecName "ovn-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 14:47:48 crc kubenswrapper[4805]: I1203 14:47:48.783368 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/af960cbb-259d-409e-9140-2d66af349200-ovncontroller-config-0" (OuterVolumeSpecName: "ovncontroller-config-0") pod "af960cbb-259d-409e-9140-2d66af349200" (UID: "af960cbb-259d-409e-9140-2d66af349200"). InnerVolumeSpecName "ovncontroller-config-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 14:47:48 crc kubenswrapper[4805]: I1203 14:47:48.790109 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/af960cbb-259d-409e-9140-2d66af349200-inventory" (OuterVolumeSpecName: "inventory") pod "af960cbb-259d-409e-9140-2d66af349200" (UID: "af960cbb-259d-409e-9140-2d66af349200"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 14:47:48 crc kubenswrapper[4805]: I1203 14:47:48.798601 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/af960cbb-259d-409e-9140-2d66af349200-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "af960cbb-259d-409e-9140-2d66af349200" (UID: "af960cbb-259d-409e-9140-2d66af349200"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 14:47:48 crc kubenswrapper[4805]: I1203 14:47:48.855793 4805 reconciler_common.go:293] "Volume detached for volume \"ovncontroller-config-0\" (UniqueName: \"kubernetes.io/configmap/af960cbb-259d-409e-9140-2d66af349200-ovncontroller-config-0\") on node \"crc\" DevicePath \"\"" Dec 03 14:47:48 crc kubenswrapper[4805]: I1203 14:47:48.856198 4805 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/af960cbb-259d-409e-9140-2d66af349200-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 03 14:47:48 crc kubenswrapper[4805]: I1203 14:47:48.856217 4805 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8rnqb\" (UniqueName: \"kubernetes.io/projected/af960cbb-259d-409e-9140-2d66af349200-kube-api-access-8rnqb\") on node \"crc\" DevicePath \"\"" Dec 03 14:47:48 crc kubenswrapper[4805]: I1203 14:47:48.856232 4805 reconciler_common.go:293] "Volume detached for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/af960cbb-259d-409e-9140-2d66af349200-ovn-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 14:47:48 crc kubenswrapper[4805]: I1203 14:47:48.856248 4805 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/af960cbb-259d-409e-9140-2d66af349200-inventory\") on node \"crc\" DevicePath \"\"" Dec 03 14:47:49 crc kubenswrapper[4805]: I1203 14:47:49.123808 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-hbx5q" event={"ID":"af960cbb-259d-409e-9140-2d66af349200","Type":"ContainerDied","Data":"248e4a387b77e040f38b6ff56719bd999d4da5e34fc3f04eab2bb2c7efb26faa"} Dec 03 14:47:49 crc kubenswrapper[4805]: I1203 14:47:49.123877 4805 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="248e4a387b77e040f38b6ff56719bd999d4da5e34fc3f04eab2bb2c7efb26faa" Dec 03 14:47:49 crc kubenswrapper[4805]: I1203 14:47:49.123939 4805 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-hbx5q" Dec 03 14:47:49 crc kubenswrapper[4805]: I1203 14:47:49.259805 4805 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-s4fzz"] Dec 03 14:47:49 crc kubenswrapper[4805]: E1203 14:47:49.260746 4805 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="af960cbb-259d-409e-9140-2d66af349200" containerName="ovn-edpm-deployment-openstack-edpm-ipam" Dec 03 14:47:49 crc kubenswrapper[4805]: I1203 14:47:49.260782 4805 state_mem.go:107] "Deleted CPUSet assignment" podUID="af960cbb-259d-409e-9140-2d66af349200" containerName="ovn-edpm-deployment-openstack-edpm-ipam" Dec 03 14:47:49 crc kubenswrapper[4805]: I1203 14:47:49.261126 4805 memory_manager.go:354] "RemoveStaleState removing state" podUID="af960cbb-259d-409e-9140-2d66af349200" containerName="ovn-edpm-deployment-openstack-edpm-ipam" Dec 03 14:47:49 crc kubenswrapper[4805]: I1203 14:47:49.262384 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-s4fzz" Dec 03 14:47:49 crc kubenswrapper[4805]: I1203 14:47:49.265191 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Dec 03 14:47:49 crc kubenswrapper[4805]: I1203 14:47:49.265455 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-neutron-config" Dec 03 14:47:49 crc kubenswrapper[4805]: I1203 14:47:49.265752 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Dec 03 14:47:49 crc kubenswrapper[4805]: I1203 14:47:49.265986 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-ovn-metadata-agent-neutron-config" Dec 03 14:47:49 crc kubenswrapper[4805]: I1203 14:47:49.266327 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-n7jt4" Dec 03 14:47:49 crc kubenswrapper[4805]: I1203 14:47:49.266539 4805 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 03 14:47:49 crc kubenswrapper[4805]: I1203 14:47:49.273303 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-s4fzz"] Dec 03 14:47:49 crc kubenswrapper[4805]: I1203 14:47:49.365612 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/779c86b3-bd17-4893-b976-e668607675f2-neutron-metadata-combined-ca-bundle\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-s4fzz\" (UID: \"779c86b3-bd17-4893-b976-e668607675f2\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-s4fzz" Dec 03 14:47:49 crc kubenswrapper[4805]: I1203 14:47:49.365715 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/779c86b3-bd17-4893-b976-e668607675f2-inventory\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-s4fzz\" (UID: \"779c86b3-bd17-4893-b976-e668607675f2\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-s4fzz" Dec 03 14:47:49 crc kubenswrapper[4805]: I1203 14:47:49.365755 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-48q87\" (UniqueName: \"kubernetes.io/projected/779c86b3-bd17-4893-b976-e668607675f2-kube-api-access-48q87\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-s4fzz\" (UID: \"779c86b3-bd17-4893-b976-e668607675f2\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-s4fzz" Dec 03 14:47:49 crc kubenswrapper[4805]: I1203 14:47:49.365906 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/779c86b3-bd17-4893-b976-e668607675f2-ssh-key\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-s4fzz\" (UID: \"779c86b3-bd17-4893-b976-e668607675f2\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-s4fzz" Dec 03 14:47:49 crc kubenswrapper[4805]: I1203 14:47:49.365956 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-metadata-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/779c86b3-bd17-4893-b976-e668607675f2-nova-metadata-neutron-config-0\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-s4fzz\" (UID: \"779c86b3-bd17-4893-b976-e668607675f2\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-s4fzz" Dec 03 14:47:49 crc kubenswrapper[4805]: I1203 14:47:49.366016 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"neutron-ovn-metadata-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/779c86b3-bd17-4893-b976-e668607675f2-neutron-ovn-metadata-agent-neutron-config-0\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-s4fzz\" (UID: \"779c86b3-bd17-4893-b976-e668607675f2\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-s4fzz" Dec 03 14:47:49 crc kubenswrapper[4805]: I1203 14:47:49.468924 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/779c86b3-bd17-4893-b976-e668607675f2-neutron-metadata-combined-ca-bundle\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-s4fzz\" (UID: \"779c86b3-bd17-4893-b976-e668607675f2\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-s4fzz" Dec 03 14:47:49 crc kubenswrapper[4805]: I1203 14:47:49.469031 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/779c86b3-bd17-4893-b976-e668607675f2-inventory\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-s4fzz\" (UID: \"779c86b3-bd17-4893-b976-e668607675f2\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-s4fzz" Dec 03 14:47:49 crc kubenswrapper[4805]: I1203 14:47:49.469061 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-48q87\" (UniqueName: \"kubernetes.io/projected/779c86b3-bd17-4893-b976-e668607675f2-kube-api-access-48q87\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-s4fzz\" (UID: \"779c86b3-bd17-4893-b976-e668607675f2\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-s4fzz" Dec 03 14:47:49 crc kubenswrapper[4805]: I1203 14:47:49.469162 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/779c86b3-bd17-4893-b976-e668607675f2-ssh-key\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-s4fzz\" (UID: \"779c86b3-bd17-4893-b976-e668607675f2\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-s4fzz" Dec 03 14:47:49 crc kubenswrapper[4805]: I1203 14:47:49.469214 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-metadata-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/779c86b3-bd17-4893-b976-e668607675f2-nova-metadata-neutron-config-0\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-s4fzz\" (UID: \"779c86b3-bd17-4893-b976-e668607675f2\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-s4fzz" Dec 03 14:47:49 crc kubenswrapper[4805]: I1203 14:47:49.469288 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"neutron-ovn-metadata-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/779c86b3-bd17-4893-b976-e668607675f2-neutron-ovn-metadata-agent-neutron-config-0\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-s4fzz\" (UID: \"779c86b3-bd17-4893-b976-e668607675f2\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-s4fzz" Dec 03 14:47:49 crc kubenswrapper[4805]: I1203 14:47:49.476172 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-metadata-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/779c86b3-bd17-4893-b976-e668607675f2-nova-metadata-neutron-config-0\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-s4fzz\" (UID: \"779c86b3-bd17-4893-b976-e668607675f2\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-s4fzz" Dec 03 14:47:49 crc kubenswrapper[4805]: I1203 14:47:49.476746 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/779c86b3-bd17-4893-b976-e668607675f2-ssh-key\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-s4fzz\" (UID: \"779c86b3-bd17-4893-b976-e668607675f2\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-s4fzz" Dec 03 14:47:49 crc kubenswrapper[4805]: I1203 14:47:49.480610 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/779c86b3-bd17-4893-b976-e668607675f2-inventory\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-s4fzz\" (UID: \"779c86b3-bd17-4893-b976-e668607675f2\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-s4fzz" Dec 03 14:47:49 crc kubenswrapper[4805]: I1203 14:47:49.481025 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"neutron-ovn-metadata-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/779c86b3-bd17-4893-b976-e668607675f2-neutron-ovn-metadata-agent-neutron-config-0\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-s4fzz\" (UID: \"779c86b3-bd17-4893-b976-e668607675f2\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-s4fzz" Dec 03 14:47:49 crc kubenswrapper[4805]: I1203 14:47:49.496775 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/779c86b3-bd17-4893-b976-e668607675f2-neutron-metadata-combined-ca-bundle\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-s4fzz\" (UID: \"779c86b3-bd17-4893-b976-e668607675f2\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-s4fzz" Dec 03 14:47:49 crc kubenswrapper[4805]: I1203 14:47:49.497738 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-48q87\" (UniqueName: \"kubernetes.io/projected/779c86b3-bd17-4893-b976-e668607675f2-kube-api-access-48q87\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-s4fzz\" (UID: \"779c86b3-bd17-4893-b976-e668607675f2\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-s4fzz" Dec 03 14:47:49 crc kubenswrapper[4805]: I1203 14:47:49.594780 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-s4fzz" Dec 03 14:47:50 crc kubenswrapper[4805]: I1203 14:47:50.121935 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-s4fzz"] Dec 03 14:47:51 crc kubenswrapper[4805]: I1203 14:47:51.144435 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-s4fzz" event={"ID":"779c86b3-bd17-4893-b976-e668607675f2","Type":"ContainerStarted","Data":"79e4a00e1b9791bc7523892e41f41356bf74c7fef432856f6276c4cfec1324e8"} Dec 03 14:47:51 crc kubenswrapper[4805]: I1203 14:47:51.144936 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-s4fzz" event={"ID":"779c86b3-bd17-4893-b976-e668607675f2","Type":"ContainerStarted","Data":"0b9f460b11b65c1b0aba919caf423129d014457c50086c02f8adaa5da0813dec"} Dec 03 14:47:51 crc kubenswrapper[4805]: I1203 14:47:51.169991 4805 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-s4fzz" podStartSLOduration=1.633471434 podStartE2EDuration="2.169968054s" podCreationTimestamp="2025-12-03 14:47:49 +0000 UTC" firstStartedPulling="2025-12-03 14:47:50.133300237 +0000 UTC m=+2299.796217150" lastFinishedPulling="2025-12-03 14:47:50.669796827 +0000 UTC m=+2300.332713770" observedRunningTime="2025-12-03 14:47:51.161283059 +0000 UTC m=+2300.824200002" watchObservedRunningTime="2025-12-03 14:47:51.169968054 +0000 UTC m=+2300.832884987" Dec 03 14:48:37 crc kubenswrapper[4805]: I1203 14:48:37.610303 4805 generic.go:334] "Generic (PLEG): container finished" podID="779c86b3-bd17-4893-b976-e668607675f2" containerID="79e4a00e1b9791bc7523892e41f41356bf74c7fef432856f6276c4cfec1324e8" exitCode=0 Dec 03 14:48:37 crc kubenswrapper[4805]: I1203 14:48:37.610406 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-s4fzz" event={"ID":"779c86b3-bd17-4893-b976-e668607675f2","Type":"ContainerDied","Data":"79e4a00e1b9791bc7523892e41f41356bf74c7fef432856f6276c4cfec1324e8"} Dec 03 14:48:39 crc kubenswrapper[4805]: I1203 14:48:39.037334 4805 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-s4fzz" Dec 03 14:48:39 crc kubenswrapper[4805]: I1203 14:48:39.065390 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-metadata-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/779c86b3-bd17-4893-b976-e668607675f2-nova-metadata-neutron-config-0\") pod \"779c86b3-bd17-4893-b976-e668607675f2\" (UID: \"779c86b3-bd17-4893-b976-e668607675f2\") " Dec 03 14:48:39 crc kubenswrapper[4805]: I1203 14:48:39.065441 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-48q87\" (UniqueName: \"kubernetes.io/projected/779c86b3-bd17-4893-b976-e668607675f2-kube-api-access-48q87\") pod \"779c86b3-bd17-4893-b976-e668607675f2\" (UID: \"779c86b3-bd17-4893-b976-e668607675f2\") " Dec 03 14:48:39 crc kubenswrapper[4805]: I1203 14:48:39.071118 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/779c86b3-bd17-4893-b976-e668607675f2-kube-api-access-48q87" (OuterVolumeSpecName: "kube-api-access-48q87") pod "779c86b3-bd17-4893-b976-e668607675f2" (UID: "779c86b3-bd17-4893-b976-e668607675f2"). InnerVolumeSpecName "kube-api-access-48q87". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 14:48:39 crc kubenswrapper[4805]: I1203 14:48:39.094631 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/779c86b3-bd17-4893-b976-e668607675f2-nova-metadata-neutron-config-0" (OuterVolumeSpecName: "nova-metadata-neutron-config-0") pod "779c86b3-bd17-4893-b976-e668607675f2" (UID: "779c86b3-bd17-4893-b976-e668607675f2"). InnerVolumeSpecName "nova-metadata-neutron-config-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 14:48:39 crc kubenswrapper[4805]: I1203 14:48:39.166798 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/779c86b3-bd17-4893-b976-e668607675f2-inventory\") pod \"779c86b3-bd17-4893-b976-e668607675f2\" (UID: \"779c86b3-bd17-4893-b976-e668607675f2\") " Dec 03 14:48:39 crc kubenswrapper[4805]: I1203 14:48:39.167184 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/779c86b3-bd17-4893-b976-e668607675f2-ssh-key\") pod \"779c86b3-bd17-4893-b976-e668607675f2\" (UID: \"779c86b3-bd17-4893-b976-e668607675f2\") " Dec 03 14:48:39 crc kubenswrapper[4805]: I1203 14:48:39.167215 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/779c86b3-bd17-4893-b976-e668607675f2-neutron-metadata-combined-ca-bundle\") pod \"779c86b3-bd17-4893-b976-e668607675f2\" (UID: \"779c86b3-bd17-4893-b976-e668607675f2\") " Dec 03 14:48:39 crc kubenswrapper[4805]: I1203 14:48:39.167286 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"neutron-ovn-metadata-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/779c86b3-bd17-4893-b976-e668607675f2-neutron-ovn-metadata-agent-neutron-config-0\") pod \"779c86b3-bd17-4893-b976-e668607675f2\" (UID: \"779c86b3-bd17-4893-b976-e668607675f2\") " Dec 03 14:48:39 crc kubenswrapper[4805]: I1203 14:48:39.167675 4805 reconciler_common.go:293] "Volume detached for volume \"nova-metadata-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/779c86b3-bd17-4893-b976-e668607675f2-nova-metadata-neutron-config-0\") on node \"crc\" DevicePath \"\"" Dec 03 14:48:39 crc kubenswrapper[4805]: I1203 14:48:39.167696 4805 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-48q87\" (UniqueName: \"kubernetes.io/projected/779c86b3-bd17-4893-b976-e668607675f2-kube-api-access-48q87\") on node \"crc\" DevicePath \"\"" Dec 03 14:48:39 crc kubenswrapper[4805]: I1203 14:48:39.171381 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/779c86b3-bd17-4893-b976-e668607675f2-neutron-metadata-combined-ca-bundle" (OuterVolumeSpecName: "neutron-metadata-combined-ca-bundle") pod "779c86b3-bd17-4893-b976-e668607675f2" (UID: "779c86b3-bd17-4893-b976-e668607675f2"). InnerVolumeSpecName "neutron-metadata-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 14:48:39 crc kubenswrapper[4805]: I1203 14:48:39.193252 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/779c86b3-bd17-4893-b976-e668607675f2-inventory" (OuterVolumeSpecName: "inventory") pod "779c86b3-bd17-4893-b976-e668607675f2" (UID: "779c86b3-bd17-4893-b976-e668607675f2"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 14:48:39 crc kubenswrapper[4805]: I1203 14:48:39.212069 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/779c86b3-bd17-4893-b976-e668607675f2-neutron-ovn-metadata-agent-neutron-config-0" (OuterVolumeSpecName: "neutron-ovn-metadata-agent-neutron-config-0") pod "779c86b3-bd17-4893-b976-e668607675f2" (UID: "779c86b3-bd17-4893-b976-e668607675f2"). InnerVolumeSpecName "neutron-ovn-metadata-agent-neutron-config-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 14:48:39 crc kubenswrapper[4805]: I1203 14:48:39.218343 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/779c86b3-bd17-4893-b976-e668607675f2-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "779c86b3-bd17-4893-b976-e668607675f2" (UID: "779c86b3-bd17-4893-b976-e668607675f2"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 14:48:39 crc kubenswrapper[4805]: I1203 14:48:39.269971 4805 reconciler_common.go:293] "Volume detached for volume \"neutron-ovn-metadata-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/779c86b3-bd17-4893-b976-e668607675f2-neutron-ovn-metadata-agent-neutron-config-0\") on node \"crc\" DevicePath \"\"" Dec 03 14:48:39 crc kubenswrapper[4805]: I1203 14:48:39.270004 4805 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/779c86b3-bd17-4893-b976-e668607675f2-inventory\") on node \"crc\" DevicePath \"\"" Dec 03 14:48:39 crc kubenswrapper[4805]: I1203 14:48:39.270014 4805 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/779c86b3-bd17-4893-b976-e668607675f2-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 03 14:48:39 crc kubenswrapper[4805]: I1203 14:48:39.270024 4805 reconciler_common.go:293] "Volume detached for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/779c86b3-bd17-4893-b976-e668607675f2-neutron-metadata-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 14:48:39 crc kubenswrapper[4805]: I1203 14:48:39.641532 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-s4fzz" event={"ID":"779c86b3-bd17-4893-b976-e668607675f2","Type":"ContainerDied","Data":"0b9f460b11b65c1b0aba919caf423129d014457c50086c02f8adaa5da0813dec"} Dec 03 14:48:39 crc kubenswrapper[4805]: I1203 14:48:39.641597 4805 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="0b9f460b11b65c1b0aba919caf423129d014457c50086c02f8adaa5da0813dec" Dec 03 14:48:39 crc kubenswrapper[4805]: I1203 14:48:39.641631 4805 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-s4fzz" Dec 03 14:48:39 crc kubenswrapper[4805]: I1203 14:48:39.762673 4805 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/libvirt-edpm-deployment-openstack-edpm-ipam-lmlms"] Dec 03 14:48:39 crc kubenswrapper[4805]: E1203 14:48:39.763053 4805 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="779c86b3-bd17-4893-b976-e668607675f2" containerName="neutron-metadata-edpm-deployment-openstack-edpm-ipam" Dec 03 14:48:39 crc kubenswrapper[4805]: I1203 14:48:39.763069 4805 state_mem.go:107] "Deleted CPUSet assignment" podUID="779c86b3-bd17-4893-b976-e668607675f2" containerName="neutron-metadata-edpm-deployment-openstack-edpm-ipam" Dec 03 14:48:39 crc kubenswrapper[4805]: I1203 14:48:39.763280 4805 memory_manager.go:354] "RemoveStaleState removing state" podUID="779c86b3-bd17-4893-b976-e668607675f2" containerName="neutron-metadata-edpm-deployment-openstack-edpm-ipam" Dec 03 14:48:39 crc kubenswrapper[4805]: I1203 14:48:39.763924 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-lmlms" Dec 03 14:48:39 crc kubenswrapper[4805]: I1203 14:48:39.774150 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Dec 03 14:48:39 crc kubenswrapper[4805]: I1203 14:48:39.774483 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"libvirt-secret" Dec 03 14:48:39 crc kubenswrapper[4805]: I1203 14:48:39.778253 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-n7jt4" Dec 03 14:48:39 crc kubenswrapper[4805]: I1203 14:48:39.778270 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Dec 03 14:48:39 crc kubenswrapper[4805]: I1203 14:48:39.778268 4805 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 03 14:48:39 crc kubenswrapper[4805]: I1203 14:48:39.793976 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/1776968b-6613-4b93-bb53-a91c3b6c3d2c-ssh-key\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-lmlms\" (UID: \"1776968b-6613-4b93-bb53-a91c3b6c3d2c\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-lmlms" Dec 03 14:48:39 crc kubenswrapper[4805]: I1203 14:48:39.794066 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1776968b-6613-4b93-bb53-a91c3b6c3d2c-libvirt-combined-ca-bundle\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-lmlms\" (UID: \"1776968b-6613-4b93-bb53-a91c3b6c3d2c\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-lmlms" Dec 03 14:48:39 crc kubenswrapper[4805]: I1203 14:48:39.794207 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/1776968b-6613-4b93-bb53-a91c3b6c3d2c-inventory\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-lmlms\" (UID: \"1776968b-6613-4b93-bb53-a91c3b6c3d2c\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-lmlms" Dec 03 14:48:39 crc kubenswrapper[4805]: I1203 14:48:39.794245 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"libvirt-secret-0\" (UniqueName: \"kubernetes.io/secret/1776968b-6613-4b93-bb53-a91c3b6c3d2c-libvirt-secret-0\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-lmlms\" (UID: \"1776968b-6613-4b93-bb53-a91c3b6c3d2c\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-lmlms" Dec 03 14:48:39 crc kubenswrapper[4805]: I1203 14:48:39.794384 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-szhk8\" (UniqueName: \"kubernetes.io/projected/1776968b-6613-4b93-bb53-a91c3b6c3d2c-kube-api-access-szhk8\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-lmlms\" (UID: \"1776968b-6613-4b93-bb53-a91c3b6c3d2c\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-lmlms" Dec 03 14:48:39 crc kubenswrapper[4805]: I1203 14:48:39.811205 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/libvirt-edpm-deployment-openstack-edpm-ipam-lmlms"] Dec 03 14:48:39 crc kubenswrapper[4805]: I1203 14:48:39.896109 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/1776968b-6613-4b93-bb53-a91c3b6c3d2c-inventory\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-lmlms\" (UID: \"1776968b-6613-4b93-bb53-a91c3b6c3d2c\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-lmlms" Dec 03 14:48:39 crc kubenswrapper[4805]: I1203 14:48:39.896172 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"libvirt-secret-0\" (UniqueName: \"kubernetes.io/secret/1776968b-6613-4b93-bb53-a91c3b6c3d2c-libvirt-secret-0\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-lmlms\" (UID: \"1776968b-6613-4b93-bb53-a91c3b6c3d2c\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-lmlms" Dec 03 14:48:39 crc kubenswrapper[4805]: I1203 14:48:39.897076 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-szhk8\" (UniqueName: \"kubernetes.io/projected/1776968b-6613-4b93-bb53-a91c3b6c3d2c-kube-api-access-szhk8\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-lmlms\" (UID: \"1776968b-6613-4b93-bb53-a91c3b6c3d2c\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-lmlms" Dec 03 14:48:39 crc kubenswrapper[4805]: I1203 14:48:39.897221 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/1776968b-6613-4b93-bb53-a91c3b6c3d2c-ssh-key\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-lmlms\" (UID: \"1776968b-6613-4b93-bb53-a91c3b6c3d2c\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-lmlms" Dec 03 14:48:39 crc kubenswrapper[4805]: I1203 14:48:39.897247 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1776968b-6613-4b93-bb53-a91c3b6c3d2c-libvirt-combined-ca-bundle\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-lmlms\" (UID: \"1776968b-6613-4b93-bb53-a91c3b6c3d2c\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-lmlms" Dec 03 14:48:39 crc kubenswrapper[4805]: I1203 14:48:39.903566 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/1776968b-6613-4b93-bb53-a91c3b6c3d2c-inventory\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-lmlms\" (UID: \"1776968b-6613-4b93-bb53-a91c3b6c3d2c\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-lmlms" Dec 03 14:48:39 crc kubenswrapper[4805]: I1203 14:48:39.903729 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/1776968b-6613-4b93-bb53-a91c3b6c3d2c-ssh-key\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-lmlms\" (UID: \"1776968b-6613-4b93-bb53-a91c3b6c3d2c\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-lmlms" Dec 03 14:48:39 crc kubenswrapper[4805]: I1203 14:48:39.903874 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"libvirt-secret-0\" (UniqueName: \"kubernetes.io/secret/1776968b-6613-4b93-bb53-a91c3b6c3d2c-libvirt-secret-0\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-lmlms\" (UID: \"1776968b-6613-4b93-bb53-a91c3b6c3d2c\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-lmlms" Dec 03 14:48:39 crc kubenswrapper[4805]: I1203 14:48:39.905314 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1776968b-6613-4b93-bb53-a91c3b6c3d2c-libvirt-combined-ca-bundle\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-lmlms\" (UID: \"1776968b-6613-4b93-bb53-a91c3b6c3d2c\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-lmlms" Dec 03 14:48:39 crc kubenswrapper[4805]: I1203 14:48:39.912742 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-szhk8\" (UniqueName: \"kubernetes.io/projected/1776968b-6613-4b93-bb53-a91c3b6c3d2c-kube-api-access-szhk8\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-lmlms\" (UID: \"1776968b-6613-4b93-bb53-a91c3b6c3d2c\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-lmlms" Dec 03 14:48:40 crc kubenswrapper[4805]: I1203 14:48:40.104249 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-lmlms" Dec 03 14:48:40 crc kubenswrapper[4805]: I1203 14:48:40.630360 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/libvirt-edpm-deployment-openstack-edpm-ipam-lmlms"] Dec 03 14:48:40 crc kubenswrapper[4805]: W1203 14:48:40.637951 4805 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod1776968b_6613_4b93_bb53_a91c3b6c3d2c.slice/crio-84a9a8ddff972f2cded1372841d93a4d2004ba1dd9d9d7ae1e81848f9043f078 WatchSource:0}: Error finding container 84a9a8ddff972f2cded1372841d93a4d2004ba1dd9d9d7ae1e81848f9043f078: Status 404 returned error can't find the container with id 84a9a8ddff972f2cded1372841d93a4d2004ba1dd9d9d7ae1e81848f9043f078 Dec 03 14:48:40 crc kubenswrapper[4805]: I1203 14:48:40.656375 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-lmlms" event={"ID":"1776968b-6613-4b93-bb53-a91c3b6c3d2c","Type":"ContainerStarted","Data":"84a9a8ddff972f2cded1372841d93a4d2004ba1dd9d9d7ae1e81848f9043f078"} Dec 03 14:48:42 crc kubenswrapper[4805]: I1203 14:48:42.677440 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-lmlms" event={"ID":"1776968b-6613-4b93-bb53-a91c3b6c3d2c","Type":"ContainerStarted","Data":"b388e6b00c345fb123ed444bc2e736f841bc1e5e7980d411065a9f38a5783437"} Dec 03 14:48:42 crc kubenswrapper[4805]: I1203 14:48:42.699546 4805 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-lmlms" podStartSLOduration=2.849289026 podStartE2EDuration="3.699522116s" podCreationTimestamp="2025-12-03 14:48:39 +0000 UTC" firstStartedPulling="2025-12-03 14:48:40.641288401 +0000 UTC m=+2350.304205324" lastFinishedPulling="2025-12-03 14:48:41.491521491 +0000 UTC m=+2351.154438414" observedRunningTime="2025-12-03 14:48:42.694901891 +0000 UTC m=+2352.357818824" watchObservedRunningTime="2025-12-03 14:48:42.699522116 +0000 UTC m=+2352.362439039" Dec 03 14:48:43 crc kubenswrapper[4805]: I1203 14:48:43.917368 4805 patch_prober.go:28] interesting pod/machine-config-daemon-gskh4 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 03 14:48:43 crc kubenswrapper[4805]: I1203 14:48:43.917761 4805 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-gskh4" podUID="6098937f-e3f6-45e8-a647-4994a79cd711" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 03 14:49:13 crc kubenswrapper[4805]: I1203 14:49:13.917739 4805 patch_prober.go:28] interesting pod/machine-config-daemon-gskh4 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 03 14:49:13 crc kubenswrapper[4805]: I1203 14:49:13.921624 4805 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-gskh4" podUID="6098937f-e3f6-45e8-a647-4994a79cd711" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 03 14:49:43 crc kubenswrapper[4805]: I1203 14:49:43.918083 4805 patch_prober.go:28] interesting pod/machine-config-daemon-gskh4 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 03 14:49:43 crc kubenswrapper[4805]: I1203 14:49:43.918948 4805 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-gskh4" podUID="6098937f-e3f6-45e8-a647-4994a79cd711" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 03 14:49:43 crc kubenswrapper[4805]: I1203 14:49:43.919035 4805 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-gskh4" Dec 03 14:49:43 crc kubenswrapper[4805]: I1203 14:49:43.920292 4805 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"fec9f6923538948ef3e55903bb8e108d1f94095a493c6b5117232e21f41ab624"} pod="openshift-machine-config-operator/machine-config-daemon-gskh4" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 03 14:49:43 crc kubenswrapper[4805]: I1203 14:49:43.920397 4805 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-gskh4" podUID="6098937f-e3f6-45e8-a647-4994a79cd711" containerName="machine-config-daemon" containerID="cri-o://fec9f6923538948ef3e55903bb8e108d1f94095a493c6b5117232e21f41ab624" gracePeriod=600 Dec 03 14:49:44 crc kubenswrapper[4805]: E1203 14:49:44.048286 4805 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-gskh4_openshift-machine-config-operator(6098937f-e3f6-45e8-a647-4994a79cd711)\"" pod="openshift-machine-config-operator/machine-config-daemon-gskh4" podUID="6098937f-e3f6-45e8-a647-4994a79cd711" Dec 03 14:49:44 crc kubenswrapper[4805]: I1203 14:49:44.279831 4805 generic.go:334] "Generic (PLEG): container finished" podID="6098937f-e3f6-45e8-a647-4994a79cd711" containerID="fec9f6923538948ef3e55903bb8e108d1f94095a493c6b5117232e21f41ab624" exitCode=0 Dec 03 14:49:44 crc kubenswrapper[4805]: I1203 14:49:44.280379 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-gskh4" event={"ID":"6098937f-e3f6-45e8-a647-4994a79cd711","Type":"ContainerDied","Data":"fec9f6923538948ef3e55903bb8e108d1f94095a493c6b5117232e21f41ab624"} Dec 03 14:49:44 crc kubenswrapper[4805]: I1203 14:49:44.280585 4805 scope.go:117] "RemoveContainer" containerID="339cd7bd4a35abfc7cd21edc654a8e6b31d75042fa72434198d7d952d589476a" Dec 03 14:49:44 crc kubenswrapper[4805]: I1203 14:49:44.282218 4805 scope.go:117] "RemoveContainer" containerID="fec9f6923538948ef3e55903bb8e108d1f94095a493c6b5117232e21f41ab624" Dec 03 14:49:44 crc kubenswrapper[4805]: E1203 14:49:44.283352 4805 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-gskh4_openshift-machine-config-operator(6098937f-e3f6-45e8-a647-4994a79cd711)\"" pod="openshift-machine-config-operator/machine-config-daemon-gskh4" podUID="6098937f-e3f6-45e8-a647-4994a79cd711" Dec 03 14:49:51 crc kubenswrapper[4805]: I1203 14:49:51.997448 4805 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-64css"] Dec 03 14:49:52 crc kubenswrapper[4805]: I1203 14:49:52.000444 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-64css" Dec 03 14:49:52 crc kubenswrapper[4805]: I1203 14:49:52.008542 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-64css"] Dec 03 14:49:52 crc kubenswrapper[4805]: I1203 14:49:52.073478 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6h5fl\" (UniqueName: \"kubernetes.io/projected/e1211b72-d818-49f8-88a1-4f371a660930-kube-api-access-6h5fl\") pod \"redhat-operators-64css\" (UID: \"e1211b72-d818-49f8-88a1-4f371a660930\") " pod="openshift-marketplace/redhat-operators-64css" Dec 03 14:49:52 crc kubenswrapper[4805]: I1203 14:49:52.073607 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e1211b72-d818-49f8-88a1-4f371a660930-utilities\") pod \"redhat-operators-64css\" (UID: \"e1211b72-d818-49f8-88a1-4f371a660930\") " pod="openshift-marketplace/redhat-operators-64css" Dec 03 14:49:52 crc kubenswrapper[4805]: I1203 14:49:52.073685 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e1211b72-d818-49f8-88a1-4f371a660930-catalog-content\") pod \"redhat-operators-64css\" (UID: \"e1211b72-d818-49f8-88a1-4f371a660930\") " pod="openshift-marketplace/redhat-operators-64css" Dec 03 14:49:52 crc kubenswrapper[4805]: I1203 14:49:52.175486 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6h5fl\" (UniqueName: \"kubernetes.io/projected/e1211b72-d818-49f8-88a1-4f371a660930-kube-api-access-6h5fl\") pod \"redhat-operators-64css\" (UID: \"e1211b72-d818-49f8-88a1-4f371a660930\") " pod="openshift-marketplace/redhat-operators-64css" Dec 03 14:49:52 crc kubenswrapper[4805]: I1203 14:49:52.175547 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e1211b72-d818-49f8-88a1-4f371a660930-utilities\") pod \"redhat-operators-64css\" (UID: \"e1211b72-d818-49f8-88a1-4f371a660930\") " pod="openshift-marketplace/redhat-operators-64css" Dec 03 14:49:52 crc kubenswrapper[4805]: I1203 14:49:52.175585 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e1211b72-d818-49f8-88a1-4f371a660930-catalog-content\") pod \"redhat-operators-64css\" (UID: \"e1211b72-d818-49f8-88a1-4f371a660930\") " pod="openshift-marketplace/redhat-operators-64css" Dec 03 14:49:52 crc kubenswrapper[4805]: I1203 14:49:52.176150 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e1211b72-d818-49f8-88a1-4f371a660930-catalog-content\") pod \"redhat-operators-64css\" (UID: \"e1211b72-d818-49f8-88a1-4f371a660930\") " pod="openshift-marketplace/redhat-operators-64css" Dec 03 14:49:52 crc kubenswrapper[4805]: I1203 14:49:52.176191 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e1211b72-d818-49f8-88a1-4f371a660930-utilities\") pod \"redhat-operators-64css\" (UID: \"e1211b72-d818-49f8-88a1-4f371a660930\") " pod="openshift-marketplace/redhat-operators-64css" Dec 03 14:49:52 crc kubenswrapper[4805]: I1203 14:49:52.201248 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6h5fl\" (UniqueName: \"kubernetes.io/projected/e1211b72-d818-49f8-88a1-4f371a660930-kube-api-access-6h5fl\") pod \"redhat-operators-64css\" (UID: \"e1211b72-d818-49f8-88a1-4f371a660930\") " pod="openshift-marketplace/redhat-operators-64css" Dec 03 14:49:52 crc kubenswrapper[4805]: I1203 14:49:52.343692 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-64css" Dec 03 14:49:52 crc kubenswrapper[4805]: I1203 14:49:52.824515 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-64css"] Dec 03 14:49:53 crc kubenswrapper[4805]: I1203 14:49:53.383140 4805 generic.go:334] "Generic (PLEG): container finished" podID="e1211b72-d818-49f8-88a1-4f371a660930" containerID="119b5a60716eb65e3c2c1d10ce2cda5e341c71c0a9442c88e4a2076747fd9e6e" exitCode=0 Dec 03 14:49:53 crc kubenswrapper[4805]: I1203 14:49:53.383215 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-64css" event={"ID":"e1211b72-d818-49f8-88a1-4f371a660930","Type":"ContainerDied","Data":"119b5a60716eb65e3c2c1d10ce2cda5e341c71c0a9442c88e4a2076747fd9e6e"} Dec 03 14:49:53 crc kubenswrapper[4805]: I1203 14:49:53.383538 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-64css" event={"ID":"e1211b72-d818-49f8-88a1-4f371a660930","Type":"ContainerStarted","Data":"cc660ef338ffebbf20a37491dffbb058ac8aac823e34699acdb49d60d2a19d04"} Dec 03 14:49:53 crc kubenswrapper[4805]: I1203 14:49:53.385190 4805 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 03 14:49:53 crc kubenswrapper[4805]: I1203 14:49:53.392170 4805 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-fkwj5"] Dec 03 14:49:53 crc kubenswrapper[4805]: I1203 14:49:53.394096 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-fkwj5" Dec 03 14:49:53 crc kubenswrapper[4805]: I1203 14:49:53.409232 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-fkwj5"] Dec 03 14:49:53 crc kubenswrapper[4805]: I1203 14:49:53.499374 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7b48ed82-1c33-4361-9c24-7ac4a3ff8715-catalog-content\") pod \"certified-operators-fkwj5\" (UID: \"7b48ed82-1c33-4361-9c24-7ac4a3ff8715\") " pod="openshift-marketplace/certified-operators-fkwj5" Dec 03 14:49:53 crc kubenswrapper[4805]: I1203 14:49:53.499793 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zlnmw\" (UniqueName: \"kubernetes.io/projected/7b48ed82-1c33-4361-9c24-7ac4a3ff8715-kube-api-access-zlnmw\") pod \"certified-operators-fkwj5\" (UID: \"7b48ed82-1c33-4361-9c24-7ac4a3ff8715\") " pod="openshift-marketplace/certified-operators-fkwj5" Dec 03 14:49:53 crc kubenswrapper[4805]: I1203 14:49:53.499857 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7b48ed82-1c33-4361-9c24-7ac4a3ff8715-utilities\") pod \"certified-operators-fkwj5\" (UID: \"7b48ed82-1c33-4361-9c24-7ac4a3ff8715\") " pod="openshift-marketplace/certified-operators-fkwj5" Dec 03 14:49:53 crc kubenswrapper[4805]: I1203 14:49:53.601746 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7b48ed82-1c33-4361-9c24-7ac4a3ff8715-utilities\") pod \"certified-operators-fkwj5\" (UID: \"7b48ed82-1c33-4361-9c24-7ac4a3ff8715\") " pod="openshift-marketplace/certified-operators-fkwj5" Dec 03 14:49:53 crc kubenswrapper[4805]: I1203 14:49:53.601973 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7b48ed82-1c33-4361-9c24-7ac4a3ff8715-catalog-content\") pod \"certified-operators-fkwj5\" (UID: \"7b48ed82-1c33-4361-9c24-7ac4a3ff8715\") " pod="openshift-marketplace/certified-operators-fkwj5" Dec 03 14:49:53 crc kubenswrapper[4805]: I1203 14:49:53.602122 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zlnmw\" (UniqueName: \"kubernetes.io/projected/7b48ed82-1c33-4361-9c24-7ac4a3ff8715-kube-api-access-zlnmw\") pod \"certified-operators-fkwj5\" (UID: \"7b48ed82-1c33-4361-9c24-7ac4a3ff8715\") " pod="openshift-marketplace/certified-operators-fkwj5" Dec 03 14:49:53 crc kubenswrapper[4805]: I1203 14:49:53.602414 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7b48ed82-1c33-4361-9c24-7ac4a3ff8715-utilities\") pod \"certified-operators-fkwj5\" (UID: \"7b48ed82-1c33-4361-9c24-7ac4a3ff8715\") " pod="openshift-marketplace/certified-operators-fkwj5" Dec 03 14:49:53 crc kubenswrapper[4805]: I1203 14:49:53.602459 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7b48ed82-1c33-4361-9c24-7ac4a3ff8715-catalog-content\") pod \"certified-operators-fkwj5\" (UID: \"7b48ed82-1c33-4361-9c24-7ac4a3ff8715\") " pod="openshift-marketplace/certified-operators-fkwj5" Dec 03 14:49:53 crc kubenswrapper[4805]: I1203 14:49:53.624743 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zlnmw\" (UniqueName: \"kubernetes.io/projected/7b48ed82-1c33-4361-9c24-7ac4a3ff8715-kube-api-access-zlnmw\") pod \"certified-operators-fkwj5\" (UID: \"7b48ed82-1c33-4361-9c24-7ac4a3ff8715\") " pod="openshift-marketplace/certified-operators-fkwj5" Dec 03 14:49:53 crc kubenswrapper[4805]: I1203 14:49:53.721290 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-fkwj5" Dec 03 14:49:54 crc kubenswrapper[4805]: I1203 14:49:54.053472 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-fkwj5"] Dec 03 14:49:54 crc kubenswrapper[4805]: W1203 14:49:54.061469 4805 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod7b48ed82_1c33_4361_9c24_7ac4a3ff8715.slice/crio-0d0c3997e74fa7c67e49c7a19d58ad0e6eb68e070761cf6f27b78af146685c62 WatchSource:0}: Error finding container 0d0c3997e74fa7c67e49c7a19d58ad0e6eb68e070761cf6f27b78af146685c62: Status 404 returned error can't find the container with id 0d0c3997e74fa7c67e49c7a19d58ad0e6eb68e070761cf6f27b78af146685c62 Dec 03 14:49:54 crc kubenswrapper[4805]: I1203 14:49:54.387964 4805 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-n2pkd"] Dec 03 14:49:54 crc kubenswrapper[4805]: I1203 14:49:54.390211 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-n2pkd" Dec 03 14:49:54 crc kubenswrapper[4805]: I1203 14:49:54.420624 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-n2pkd"] Dec 03 14:49:54 crc kubenswrapper[4805]: I1203 14:49:54.437395 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-64css" event={"ID":"e1211b72-d818-49f8-88a1-4f371a660930","Type":"ContainerStarted","Data":"249fc24c61ca87c693a097dcc1f1f4ff1e2316bd4cab15254b2f2988a6330181"} Dec 03 14:49:54 crc kubenswrapper[4805]: I1203 14:49:54.442964 4805 generic.go:334] "Generic (PLEG): container finished" podID="7b48ed82-1c33-4361-9c24-7ac4a3ff8715" containerID="16fb611feed7a9e6917e4609f89a801fef9f9c130c6ce0fa6adca48860d97915" exitCode=0 Dec 03 14:49:54 crc kubenswrapper[4805]: I1203 14:49:54.443003 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-fkwj5" event={"ID":"7b48ed82-1c33-4361-9c24-7ac4a3ff8715","Type":"ContainerDied","Data":"16fb611feed7a9e6917e4609f89a801fef9f9c130c6ce0fa6adca48860d97915"} Dec 03 14:49:54 crc kubenswrapper[4805]: I1203 14:49:54.443029 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-fkwj5" event={"ID":"7b48ed82-1c33-4361-9c24-7ac4a3ff8715","Type":"ContainerStarted","Data":"0d0c3997e74fa7c67e49c7a19d58ad0e6eb68e070761cf6f27b78af146685c62"} Dec 03 14:49:54 crc kubenswrapper[4805]: I1203 14:49:54.523421 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/dcceed7d-71aa-432f-94b3-94ce65599157-utilities\") pod \"community-operators-n2pkd\" (UID: \"dcceed7d-71aa-432f-94b3-94ce65599157\") " pod="openshift-marketplace/community-operators-n2pkd" Dec 03 14:49:54 crc kubenswrapper[4805]: I1203 14:49:54.523811 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-m6sqm\" (UniqueName: \"kubernetes.io/projected/dcceed7d-71aa-432f-94b3-94ce65599157-kube-api-access-m6sqm\") pod \"community-operators-n2pkd\" (UID: \"dcceed7d-71aa-432f-94b3-94ce65599157\") " pod="openshift-marketplace/community-operators-n2pkd" Dec 03 14:49:54 crc kubenswrapper[4805]: I1203 14:49:54.523866 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/dcceed7d-71aa-432f-94b3-94ce65599157-catalog-content\") pod \"community-operators-n2pkd\" (UID: \"dcceed7d-71aa-432f-94b3-94ce65599157\") " pod="openshift-marketplace/community-operators-n2pkd" Dec 03 14:49:54 crc kubenswrapper[4805]: I1203 14:49:54.625384 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/dcceed7d-71aa-432f-94b3-94ce65599157-utilities\") pod \"community-operators-n2pkd\" (UID: \"dcceed7d-71aa-432f-94b3-94ce65599157\") " pod="openshift-marketplace/community-operators-n2pkd" Dec 03 14:49:54 crc kubenswrapper[4805]: I1203 14:49:54.625507 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-m6sqm\" (UniqueName: \"kubernetes.io/projected/dcceed7d-71aa-432f-94b3-94ce65599157-kube-api-access-m6sqm\") pod \"community-operators-n2pkd\" (UID: \"dcceed7d-71aa-432f-94b3-94ce65599157\") " pod="openshift-marketplace/community-operators-n2pkd" Dec 03 14:49:54 crc kubenswrapper[4805]: I1203 14:49:54.625541 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/dcceed7d-71aa-432f-94b3-94ce65599157-catalog-content\") pod \"community-operators-n2pkd\" (UID: \"dcceed7d-71aa-432f-94b3-94ce65599157\") " pod="openshift-marketplace/community-operators-n2pkd" Dec 03 14:49:54 crc kubenswrapper[4805]: I1203 14:49:54.625967 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/dcceed7d-71aa-432f-94b3-94ce65599157-catalog-content\") pod \"community-operators-n2pkd\" (UID: \"dcceed7d-71aa-432f-94b3-94ce65599157\") " pod="openshift-marketplace/community-operators-n2pkd" Dec 03 14:49:54 crc kubenswrapper[4805]: I1203 14:49:54.626711 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/dcceed7d-71aa-432f-94b3-94ce65599157-utilities\") pod \"community-operators-n2pkd\" (UID: \"dcceed7d-71aa-432f-94b3-94ce65599157\") " pod="openshift-marketplace/community-operators-n2pkd" Dec 03 14:49:54 crc kubenswrapper[4805]: I1203 14:49:54.656056 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-m6sqm\" (UniqueName: \"kubernetes.io/projected/dcceed7d-71aa-432f-94b3-94ce65599157-kube-api-access-m6sqm\") pod \"community-operators-n2pkd\" (UID: \"dcceed7d-71aa-432f-94b3-94ce65599157\") " pod="openshift-marketplace/community-operators-n2pkd" Dec 03 14:49:54 crc kubenswrapper[4805]: I1203 14:49:54.754980 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-n2pkd" Dec 03 14:49:55 crc kubenswrapper[4805]: I1203 14:49:55.290138 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-n2pkd"] Dec 03 14:49:55 crc kubenswrapper[4805]: W1203 14:49:55.299437 4805 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-poddcceed7d_71aa_432f_94b3_94ce65599157.slice/crio-4f5ad3a7634b6148b208698721aaa1a1bc201d630d888a8394e0d89956b3365d WatchSource:0}: Error finding container 4f5ad3a7634b6148b208698721aaa1a1bc201d630d888a8394e0d89956b3365d: Status 404 returned error can't find the container with id 4f5ad3a7634b6148b208698721aaa1a1bc201d630d888a8394e0d89956b3365d Dec 03 14:49:55 crc kubenswrapper[4805]: I1203 14:49:55.460369 4805 generic.go:334] "Generic (PLEG): container finished" podID="e1211b72-d818-49f8-88a1-4f371a660930" containerID="249fc24c61ca87c693a097dcc1f1f4ff1e2316bd4cab15254b2f2988a6330181" exitCode=0 Dec 03 14:49:55 crc kubenswrapper[4805]: I1203 14:49:55.460484 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-64css" event={"ID":"e1211b72-d818-49f8-88a1-4f371a660930","Type":"ContainerDied","Data":"249fc24c61ca87c693a097dcc1f1f4ff1e2316bd4cab15254b2f2988a6330181"} Dec 03 14:49:55 crc kubenswrapper[4805]: I1203 14:49:55.470447 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-fkwj5" event={"ID":"7b48ed82-1c33-4361-9c24-7ac4a3ff8715","Type":"ContainerStarted","Data":"434394ed82d82a94690cc46bda8cae90792b9424908dc9ee99505d9b2d4070ee"} Dec 03 14:49:55 crc kubenswrapper[4805]: I1203 14:49:55.474002 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-n2pkd" event={"ID":"dcceed7d-71aa-432f-94b3-94ce65599157","Type":"ContainerStarted","Data":"a4241d2d9d223b2e571d10a25626916349525f6d050d92f102e055ec0df63d79"} Dec 03 14:49:55 crc kubenswrapper[4805]: I1203 14:49:55.474043 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-n2pkd" event={"ID":"dcceed7d-71aa-432f-94b3-94ce65599157","Type":"ContainerStarted","Data":"4f5ad3a7634b6148b208698721aaa1a1bc201d630d888a8394e0d89956b3365d"} Dec 03 14:49:56 crc kubenswrapper[4805]: I1203 14:49:56.486870 4805 generic.go:334] "Generic (PLEG): container finished" podID="7b48ed82-1c33-4361-9c24-7ac4a3ff8715" containerID="434394ed82d82a94690cc46bda8cae90792b9424908dc9ee99505d9b2d4070ee" exitCode=0 Dec 03 14:49:56 crc kubenswrapper[4805]: I1203 14:49:56.486982 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-fkwj5" event={"ID":"7b48ed82-1c33-4361-9c24-7ac4a3ff8715","Type":"ContainerDied","Data":"434394ed82d82a94690cc46bda8cae90792b9424908dc9ee99505d9b2d4070ee"} Dec 03 14:49:56 crc kubenswrapper[4805]: I1203 14:49:56.493236 4805 generic.go:334] "Generic (PLEG): container finished" podID="dcceed7d-71aa-432f-94b3-94ce65599157" containerID="a4241d2d9d223b2e571d10a25626916349525f6d050d92f102e055ec0df63d79" exitCode=0 Dec 03 14:49:56 crc kubenswrapper[4805]: I1203 14:49:56.493347 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-n2pkd" event={"ID":"dcceed7d-71aa-432f-94b3-94ce65599157","Type":"ContainerDied","Data":"a4241d2d9d223b2e571d10a25626916349525f6d050d92f102e055ec0df63d79"} Dec 03 14:49:56 crc kubenswrapper[4805]: I1203 14:49:56.503508 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-64css" event={"ID":"e1211b72-d818-49f8-88a1-4f371a660930","Type":"ContainerStarted","Data":"6fdd0bf6e87760bbb90b4e09d8e28b17438ac6e310c817a1ac480fbcdd2642c3"} Dec 03 14:49:56 crc kubenswrapper[4805]: I1203 14:49:56.566702 4805 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-64css" podStartSLOduration=3.037186293 podStartE2EDuration="5.56668111s" podCreationTimestamp="2025-12-03 14:49:51 +0000 UTC" firstStartedPulling="2025-12-03 14:49:53.384938546 +0000 UTC m=+2423.047855469" lastFinishedPulling="2025-12-03 14:49:55.914433363 +0000 UTC m=+2425.577350286" observedRunningTime="2025-12-03 14:49:56.556593857 +0000 UTC m=+2426.219510780" watchObservedRunningTime="2025-12-03 14:49:56.56668111 +0000 UTC m=+2426.229598033" Dec 03 14:49:57 crc kubenswrapper[4805]: I1203 14:49:57.693901 4805 scope.go:117] "RemoveContainer" containerID="fec9f6923538948ef3e55903bb8e108d1f94095a493c6b5117232e21f41ab624" Dec 03 14:49:57 crc kubenswrapper[4805]: E1203 14:49:57.694189 4805 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-gskh4_openshift-machine-config-operator(6098937f-e3f6-45e8-a647-4994a79cd711)\"" pod="openshift-machine-config-operator/machine-config-daemon-gskh4" podUID="6098937f-e3f6-45e8-a647-4994a79cd711" Dec 03 14:49:59 crc kubenswrapper[4805]: I1203 14:49:59.542440 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-fkwj5" event={"ID":"7b48ed82-1c33-4361-9c24-7ac4a3ff8715","Type":"ContainerStarted","Data":"3c36a225ea185572dd923d7651b2ba7b5417aede1f0d94a81f918bac0251f2dc"} Dec 03 14:50:00 crc kubenswrapper[4805]: I1203 14:50:00.561456 4805 generic.go:334] "Generic (PLEG): container finished" podID="dcceed7d-71aa-432f-94b3-94ce65599157" containerID="c613f21aa697fdbbca70292713ee8e799539aa36bcc69bff368ed9da7857f39e" exitCode=0 Dec 03 14:50:00 crc kubenswrapper[4805]: I1203 14:50:00.561556 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-n2pkd" event={"ID":"dcceed7d-71aa-432f-94b3-94ce65599157","Type":"ContainerDied","Data":"c613f21aa697fdbbca70292713ee8e799539aa36bcc69bff368ed9da7857f39e"} Dec 03 14:50:00 crc kubenswrapper[4805]: I1203 14:50:00.584028 4805 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-fkwj5" podStartSLOduration=4.53264495 podStartE2EDuration="7.584007049s" podCreationTimestamp="2025-12-03 14:49:53 +0000 UTC" firstStartedPulling="2025-12-03 14:49:54.448652799 +0000 UTC m=+2424.111569742" lastFinishedPulling="2025-12-03 14:49:57.500014908 +0000 UTC m=+2427.162931841" observedRunningTime="2025-12-03 14:49:59.565269952 +0000 UTC m=+2429.228186895" watchObservedRunningTime="2025-12-03 14:50:00.584007049 +0000 UTC m=+2430.246923972" Dec 03 14:50:02 crc kubenswrapper[4805]: I1203 14:50:02.344762 4805 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-64css" Dec 03 14:50:02 crc kubenswrapper[4805]: I1203 14:50:02.345405 4805 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-64css" Dec 03 14:50:02 crc kubenswrapper[4805]: I1203 14:50:02.398600 4805 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-64css" Dec 03 14:50:02 crc kubenswrapper[4805]: I1203 14:50:02.588476 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-n2pkd" event={"ID":"dcceed7d-71aa-432f-94b3-94ce65599157","Type":"ContainerStarted","Data":"eea5bec2ab8cc6e0297401179d333a03753a51e91980dcb7111c319360d8f6d8"} Dec 03 14:50:02 crc kubenswrapper[4805]: I1203 14:50:02.613162 4805 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-n2pkd" podStartSLOduration=3.704353263 podStartE2EDuration="8.613143187s" podCreationTimestamp="2025-12-03 14:49:54 +0000 UTC" firstStartedPulling="2025-12-03 14:49:56.499769181 +0000 UTC m=+2426.162686114" lastFinishedPulling="2025-12-03 14:50:01.408559085 +0000 UTC m=+2431.071476038" observedRunningTime="2025-12-03 14:50:02.606060605 +0000 UTC m=+2432.268977528" watchObservedRunningTime="2025-12-03 14:50:02.613143187 +0000 UTC m=+2432.276060100" Dec 03 14:50:02 crc kubenswrapper[4805]: I1203 14:50:02.635790 4805 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-64css" Dec 03 14:50:03 crc kubenswrapper[4805]: I1203 14:50:03.722404 4805 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-fkwj5" Dec 03 14:50:03 crc kubenswrapper[4805]: I1203 14:50:03.722470 4805 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-fkwj5" Dec 03 14:50:04 crc kubenswrapper[4805]: I1203 14:50:04.756106 4805 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-n2pkd" Dec 03 14:50:04 crc kubenswrapper[4805]: I1203 14:50:04.757819 4805 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-n2pkd" Dec 03 14:50:04 crc kubenswrapper[4805]: I1203 14:50:04.771209 4805 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/certified-operators-fkwj5" podUID="7b48ed82-1c33-4361-9c24-7ac4a3ff8715" containerName="registry-server" probeResult="failure" output=< Dec 03 14:50:04 crc kubenswrapper[4805]: timeout: failed to connect service ":50051" within 1s Dec 03 14:50:04 crc kubenswrapper[4805]: > Dec 03 14:50:04 crc kubenswrapper[4805]: I1203 14:50:04.777489 4805 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-64css"] Dec 03 14:50:04 crc kubenswrapper[4805]: I1203 14:50:04.777763 4805 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-64css" podUID="e1211b72-d818-49f8-88a1-4f371a660930" containerName="registry-server" containerID="cri-o://6fdd0bf6e87760bbb90b4e09d8e28b17438ac6e310c817a1ac480fbcdd2642c3" gracePeriod=2 Dec 03 14:50:04 crc kubenswrapper[4805]: I1203 14:50:04.838780 4805 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-n2pkd" Dec 03 14:50:05 crc kubenswrapper[4805]: I1203 14:50:05.262894 4805 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-64css" Dec 03 14:50:05 crc kubenswrapper[4805]: I1203 14:50:05.444332 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6h5fl\" (UniqueName: \"kubernetes.io/projected/e1211b72-d818-49f8-88a1-4f371a660930-kube-api-access-6h5fl\") pod \"e1211b72-d818-49f8-88a1-4f371a660930\" (UID: \"e1211b72-d818-49f8-88a1-4f371a660930\") " Dec 03 14:50:05 crc kubenswrapper[4805]: I1203 14:50:05.444378 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e1211b72-d818-49f8-88a1-4f371a660930-catalog-content\") pod \"e1211b72-d818-49f8-88a1-4f371a660930\" (UID: \"e1211b72-d818-49f8-88a1-4f371a660930\") " Dec 03 14:50:05 crc kubenswrapper[4805]: I1203 14:50:05.444477 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e1211b72-d818-49f8-88a1-4f371a660930-utilities\") pod \"e1211b72-d818-49f8-88a1-4f371a660930\" (UID: \"e1211b72-d818-49f8-88a1-4f371a660930\") " Dec 03 14:50:05 crc kubenswrapper[4805]: I1203 14:50:05.445750 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e1211b72-d818-49f8-88a1-4f371a660930-utilities" (OuterVolumeSpecName: "utilities") pod "e1211b72-d818-49f8-88a1-4f371a660930" (UID: "e1211b72-d818-49f8-88a1-4f371a660930"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 14:50:05 crc kubenswrapper[4805]: I1203 14:50:05.455752 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e1211b72-d818-49f8-88a1-4f371a660930-kube-api-access-6h5fl" (OuterVolumeSpecName: "kube-api-access-6h5fl") pod "e1211b72-d818-49f8-88a1-4f371a660930" (UID: "e1211b72-d818-49f8-88a1-4f371a660930"). InnerVolumeSpecName "kube-api-access-6h5fl". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 14:50:05 crc kubenswrapper[4805]: I1203 14:50:05.546712 4805 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6h5fl\" (UniqueName: \"kubernetes.io/projected/e1211b72-d818-49f8-88a1-4f371a660930-kube-api-access-6h5fl\") on node \"crc\" DevicePath \"\"" Dec 03 14:50:05 crc kubenswrapper[4805]: I1203 14:50:05.546744 4805 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e1211b72-d818-49f8-88a1-4f371a660930-utilities\") on node \"crc\" DevicePath \"\"" Dec 03 14:50:05 crc kubenswrapper[4805]: I1203 14:50:05.572312 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e1211b72-d818-49f8-88a1-4f371a660930-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "e1211b72-d818-49f8-88a1-4f371a660930" (UID: "e1211b72-d818-49f8-88a1-4f371a660930"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 14:50:05 crc kubenswrapper[4805]: I1203 14:50:05.623486 4805 generic.go:334] "Generic (PLEG): container finished" podID="e1211b72-d818-49f8-88a1-4f371a660930" containerID="6fdd0bf6e87760bbb90b4e09d8e28b17438ac6e310c817a1ac480fbcdd2642c3" exitCode=0 Dec 03 14:50:05 crc kubenswrapper[4805]: I1203 14:50:05.623526 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-64css" event={"ID":"e1211b72-d818-49f8-88a1-4f371a660930","Type":"ContainerDied","Data":"6fdd0bf6e87760bbb90b4e09d8e28b17438ac6e310c817a1ac480fbcdd2642c3"} Dec 03 14:50:05 crc kubenswrapper[4805]: I1203 14:50:05.623583 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-64css" event={"ID":"e1211b72-d818-49f8-88a1-4f371a660930","Type":"ContainerDied","Data":"cc660ef338ffebbf20a37491dffbb058ac8aac823e34699acdb49d60d2a19d04"} Dec 03 14:50:05 crc kubenswrapper[4805]: I1203 14:50:05.623606 4805 scope.go:117] "RemoveContainer" containerID="6fdd0bf6e87760bbb90b4e09d8e28b17438ac6e310c817a1ac480fbcdd2642c3" Dec 03 14:50:05 crc kubenswrapper[4805]: I1203 14:50:05.624360 4805 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-64css" Dec 03 14:50:05 crc kubenswrapper[4805]: I1203 14:50:05.649000 4805 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e1211b72-d818-49f8-88a1-4f371a660930-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 03 14:50:05 crc kubenswrapper[4805]: I1203 14:50:05.652366 4805 scope.go:117] "RemoveContainer" containerID="249fc24c61ca87c693a097dcc1f1f4ff1e2316bd4cab15254b2f2988a6330181" Dec 03 14:50:05 crc kubenswrapper[4805]: I1203 14:50:05.682347 4805 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-64css"] Dec 03 14:50:05 crc kubenswrapper[4805]: I1203 14:50:05.686333 4805 scope.go:117] "RemoveContainer" containerID="119b5a60716eb65e3c2c1d10ce2cda5e341c71c0a9442c88e4a2076747fd9e6e" Dec 03 14:50:05 crc kubenswrapper[4805]: I1203 14:50:05.690950 4805 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-64css"] Dec 03 14:50:05 crc kubenswrapper[4805]: I1203 14:50:05.724730 4805 scope.go:117] "RemoveContainer" containerID="6fdd0bf6e87760bbb90b4e09d8e28b17438ac6e310c817a1ac480fbcdd2642c3" Dec 03 14:50:05 crc kubenswrapper[4805]: E1203 14:50:05.725230 4805 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6fdd0bf6e87760bbb90b4e09d8e28b17438ac6e310c817a1ac480fbcdd2642c3\": container with ID starting with 6fdd0bf6e87760bbb90b4e09d8e28b17438ac6e310c817a1ac480fbcdd2642c3 not found: ID does not exist" containerID="6fdd0bf6e87760bbb90b4e09d8e28b17438ac6e310c817a1ac480fbcdd2642c3" Dec 03 14:50:05 crc kubenswrapper[4805]: I1203 14:50:05.725276 4805 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6fdd0bf6e87760bbb90b4e09d8e28b17438ac6e310c817a1ac480fbcdd2642c3"} err="failed to get container status \"6fdd0bf6e87760bbb90b4e09d8e28b17438ac6e310c817a1ac480fbcdd2642c3\": rpc error: code = NotFound desc = could not find container \"6fdd0bf6e87760bbb90b4e09d8e28b17438ac6e310c817a1ac480fbcdd2642c3\": container with ID starting with 6fdd0bf6e87760bbb90b4e09d8e28b17438ac6e310c817a1ac480fbcdd2642c3 not found: ID does not exist" Dec 03 14:50:05 crc kubenswrapper[4805]: I1203 14:50:05.725297 4805 scope.go:117] "RemoveContainer" containerID="249fc24c61ca87c693a097dcc1f1f4ff1e2316bd4cab15254b2f2988a6330181" Dec 03 14:50:05 crc kubenswrapper[4805]: E1203 14:50:05.725765 4805 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"249fc24c61ca87c693a097dcc1f1f4ff1e2316bd4cab15254b2f2988a6330181\": container with ID starting with 249fc24c61ca87c693a097dcc1f1f4ff1e2316bd4cab15254b2f2988a6330181 not found: ID does not exist" containerID="249fc24c61ca87c693a097dcc1f1f4ff1e2316bd4cab15254b2f2988a6330181" Dec 03 14:50:05 crc kubenswrapper[4805]: I1203 14:50:05.725783 4805 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"249fc24c61ca87c693a097dcc1f1f4ff1e2316bd4cab15254b2f2988a6330181"} err="failed to get container status \"249fc24c61ca87c693a097dcc1f1f4ff1e2316bd4cab15254b2f2988a6330181\": rpc error: code = NotFound desc = could not find container \"249fc24c61ca87c693a097dcc1f1f4ff1e2316bd4cab15254b2f2988a6330181\": container with ID starting with 249fc24c61ca87c693a097dcc1f1f4ff1e2316bd4cab15254b2f2988a6330181 not found: ID does not exist" Dec 03 14:50:05 crc kubenswrapper[4805]: I1203 14:50:05.725809 4805 scope.go:117] "RemoveContainer" containerID="119b5a60716eb65e3c2c1d10ce2cda5e341c71c0a9442c88e4a2076747fd9e6e" Dec 03 14:50:05 crc kubenswrapper[4805]: E1203 14:50:05.726104 4805 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"119b5a60716eb65e3c2c1d10ce2cda5e341c71c0a9442c88e4a2076747fd9e6e\": container with ID starting with 119b5a60716eb65e3c2c1d10ce2cda5e341c71c0a9442c88e4a2076747fd9e6e not found: ID does not exist" containerID="119b5a60716eb65e3c2c1d10ce2cda5e341c71c0a9442c88e4a2076747fd9e6e" Dec 03 14:50:05 crc kubenswrapper[4805]: I1203 14:50:05.726137 4805 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"119b5a60716eb65e3c2c1d10ce2cda5e341c71c0a9442c88e4a2076747fd9e6e"} err="failed to get container status \"119b5a60716eb65e3c2c1d10ce2cda5e341c71c0a9442c88e4a2076747fd9e6e\": rpc error: code = NotFound desc = could not find container \"119b5a60716eb65e3c2c1d10ce2cda5e341c71c0a9442c88e4a2076747fd9e6e\": container with ID starting with 119b5a60716eb65e3c2c1d10ce2cda5e341c71c0a9442c88e4a2076747fd9e6e not found: ID does not exist" Dec 03 14:50:06 crc kubenswrapper[4805]: I1203 14:50:06.713064 4805 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e1211b72-d818-49f8-88a1-4f371a660930" path="/var/lib/kubelet/pods/e1211b72-d818-49f8-88a1-4f371a660930/volumes" Dec 03 14:50:08 crc kubenswrapper[4805]: I1203 14:50:08.695730 4805 scope.go:117] "RemoveContainer" containerID="fec9f6923538948ef3e55903bb8e108d1f94095a493c6b5117232e21f41ab624" Dec 03 14:50:08 crc kubenswrapper[4805]: E1203 14:50:08.696748 4805 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-gskh4_openshift-machine-config-operator(6098937f-e3f6-45e8-a647-4994a79cd711)\"" pod="openshift-machine-config-operator/machine-config-daemon-gskh4" podUID="6098937f-e3f6-45e8-a647-4994a79cd711" Dec 03 14:50:13 crc kubenswrapper[4805]: I1203 14:50:13.811402 4805 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-fkwj5" Dec 03 14:50:13 crc kubenswrapper[4805]: I1203 14:50:13.873895 4805 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-fkwj5" Dec 03 14:50:14 crc kubenswrapper[4805]: I1203 14:50:14.070508 4805 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-fkwj5"] Dec 03 14:50:14 crc kubenswrapper[4805]: I1203 14:50:14.813474 4805 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-n2pkd" Dec 03 14:50:15 crc kubenswrapper[4805]: I1203 14:50:15.744776 4805 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-fkwj5" podUID="7b48ed82-1c33-4361-9c24-7ac4a3ff8715" containerName="registry-server" containerID="cri-o://3c36a225ea185572dd923d7651b2ba7b5417aede1f0d94a81f918bac0251f2dc" gracePeriod=2 Dec 03 14:50:16 crc kubenswrapper[4805]: I1203 14:50:16.243724 4805 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-fkwj5" Dec 03 14:50:16 crc kubenswrapper[4805]: I1203 14:50:16.289040 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zlnmw\" (UniqueName: \"kubernetes.io/projected/7b48ed82-1c33-4361-9c24-7ac4a3ff8715-kube-api-access-zlnmw\") pod \"7b48ed82-1c33-4361-9c24-7ac4a3ff8715\" (UID: \"7b48ed82-1c33-4361-9c24-7ac4a3ff8715\") " Dec 03 14:50:16 crc kubenswrapper[4805]: I1203 14:50:16.289103 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7b48ed82-1c33-4361-9c24-7ac4a3ff8715-catalog-content\") pod \"7b48ed82-1c33-4361-9c24-7ac4a3ff8715\" (UID: \"7b48ed82-1c33-4361-9c24-7ac4a3ff8715\") " Dec 03 14:50:16 crc kubenswrapper[4805]: I1203 14:50:16.289175 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7b48ed82-1c33-4361-9c24-7ac4a3ff8715-utilities\") pod \"7b48ed82-1c33-4361-9c24-7ac4a3ff8715\" (UID: \"7b48ed82-1c33-4361-9c24-7ac4a3ff8715\") " Dec 03 14:50:16 crc kubenswrapper[4805]: I1203 14:50:16.290262 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7b48ed82-1c33-4361-9c24-7ac4a3ff8715-utilities" (OuterVolumeSpecName: "utilities") pod "7b48ed82-1c33-4361-9c24-7ac4a3ff8715" (UID: "7b48ed82-1c33-4361-9c24-7ac4a3ff8715"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 14:50:16 crc kubenswrapper[4805]: I1203 14:50:16.294538 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7b48ed82-1c33-4361-9c24-7ac4a3ff8715-kube-api-access-zlnmw" (OuterVolumeSpecName: "kube-api-access-zlnmw") pod "7b48ed82-1c33-4361-9c24-7ac4a3ff8715" (UID: "7b48ed82-1c33-4361-9c24-7ac4a3ff8715"). InnerVolumeSpecName "kube-api-access-zlnmw". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 14:50:16 crc kubenswrapper[4805]: I1203 14:50:16.356665 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7b48ed82-1c33-4361-9c24-7ac4a3ff8715-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "7b48ed82-1c33-4361-9c24-7ac4a3ff8715" (UID: "7b48ed82-1c33-4361-9c24-7ac4a3ff8715"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 14:50:16 crc kubenswrapper[4805]: I1203 14:50:16.391383 4805 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7b48ed82-1c33-4361-9c24-7ac4a3ff8715-utilities\") on node \"crc\" DevicePath \"\"" Dec 03 14:50:16 crc kubenswrapper[4805]: I1203 14:50:16.391430 4805 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zlnmw\" (UniqueName: \"kubernetes.io/projected/7b48ed82-1c33-4361-9c24-7ac4a3ff8715-kube-api-access-zlnmw\") on node \"crc\" DevicePath \"\"" Dec 03 14:50:16 crc kubenswrapper[4805]: I1203 14:50:16.391452 4805 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7b48ed82-1c33-4361-9c24-7ac4a3ff8715-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 03 14:50:16 crc kubenswrapper[4805]: I1203 14:50:16.471816 4805 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-n2pkd"] Dec 03 14:50:16 crc kubenswrapper[4805]: I1203 14:50:16.472531 4805 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-n2pkd" podUID="dcceed7d-71aa-432f-94b3-94ce65599157" containerName="registry-server" containerID="cri-o://eea5bec2ab8cc6e0297401179d333a03753a51e91980dcb7111c319360d8f6d8" gracePeriod=2 Dec 03 14:50:16 crc kubenswrapper[4805]: I1203 14:50:16.762811 4805 generic.go:334] "Generic (PLEG): container finished" podID="dcceed7d-71aa-432f-94b3-94ce65599157" containerID="eea5bec2ab8cc6e0297401179d333a03753a51e91980dcb7111c319360d8f6d8" exitCode=0 Dec 03 14:50:16 crc kubenswrapper[4805]: I1203 14:50:16.762952 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-n2pkd" event={"ID":"dcceed7d-71aa-432f-94b3-94ce65599157","Type":"ContainerDied","Data":"eea5bec2ab8cc6e0297401179d333a03753a51e91980dcb7111c319360d8f6d8"} Dec 03 14:50:16 crc kubenswrapper[4805]: I1203 14:50:16.765466 4805 generic.go:334] "Generic (PLEG): container finished" podID="7b48ed82-1c33-4361-9c24-7ac4a3ff8715" containerID="3c36a225ea185572dd923d7651b2ba7b5417aede1f0d94a81f918bac0251f2dc" exitCode=0 Dec 03 14:50:16 crc kubenswrapper[4805]: I1203 14:50:16.765505 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-fkwj5" event={"ID":"7b48ed82-1c33-4361-9c24-7ac4a3ff8715","Type":"ContainerDied","Data":"3c36a225ea185572dd923d7651b2ba7b5417aede1f0d94a81f918bac0251f2dc"} Dec 03 14:50:16 crc kubenswrapper[4805]: I1203 14:50:16.765526 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-fkwj5" event={"ID":"7b48ed82-1c33-4361-9c24-7ac4a3ff8715","Type":"ContainerDied","Data":"0d0c3997e74fa7c67e49c7a19d58ad0e6eb68e070761cf6f27b78af146685c62"} Dec 03 14:50:16 crc kubenswrapper[4805]: I1203 14:50:16.765545 4805 scope.go:117] "RemoveContainer" containerID="3c36a225ea185572dd923d7651b2ba7b5417aede1f0d94a81f918bac0251f2dc" Dec 03 14:50:16 crc kubenswrapper[4805]: I1203 14:50:16.765697 4805 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-fkwj5" Dec 03 14:50:16 crc kubenswrapper[4805]: I1203 14:50:16.798276 4805 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-fkwj5"] Dec 03 14:50:16 crc kubenswrapper[4805]: I1203 14:50:16.805906 4805 scope.go:117] "RemoveContainer" containerID="434394ed82d82a94690cc46bda8cae90792b9424908dc9ee99505d9b2d4070ee" Dec 03 14:50:16 crc kubenswrapper[4805]: I1203 14:50:16.810513 4805 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-fkwj5"] Dec 03 14:50:16 crc kubenswrapper[4805]: I1203 14:50:16.833176 4805 scope.go:117] "RemoveContainer" containerID="16fb611feed7a9e6917e4609f89a801fef9f9c130c6ce0fa6adca48860d97915" Dec 03 14:50:16 crc kubenswrapper[4805]: I1203 14:50:16.863944 4805 scope.go:117] "RemoveContainer" containerID="3c36a225ea185572dd923d7651b2ba7b5417aede1f0d94a81f918bac0251f2dc" Dec 03 14:50:16 crc kubenswrapper[4805]: E1203 14:50:16.864515 4805 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3c36a225ea185572dd923d7651b2ba7b5417aede1f0d94a81f918bac0251f2dc\": container with ID starting with 3c36a225ea185572dd923d7651b2ba7b5417aede1f0d94a81f918bac0251f2dc not found: ID does not exist" containerID="3c36a225ea185572dd923d7651b2ba7b5417aede1f0d94a81f918bac0251f2dc" Dec 03 14:50:16 crc kubenswrapper[4805]: I1203 14:50:16.864547 4805 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3c36a225ea185572dd923d7651b2ba7b5417aede1f0d94a81f918bac0251f2dc"} err="failed to get container status \"3c36a225ea185572dd923d7651b2ba7b5417aede1f0d94a81f918bac0251f2dc\": rpc error: code = NotFound desc = could not find container \"3c36a225ea185572dd923d7651b2ba7b5417aede1f0d94a81f918bac0251f2dc\": container with ID starting with 3c36a225ea185572dd923d7651b2ba7b5417aede1f0d94a81f918bac0251f2dc not found: ID does not exist" Dec 03 14:50:16 crc kubenswrapper[4805]: I1203 14:50:16.864574 4805 scope.go:117] "RemoveContainer" containerID="434394ed82d82a94690cc46bda8cae90792b9424908dc9ee99505d9b2d4070ee" Dec 03 14:50:16 crc kubenswrapper[4805]: E1203 14:50:16.865093 4805 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"434394ed82d82a94690cc46bda8cae90792b9424908dc9ee99505d9b2d4070ee\": container with ID starting with 434394ed82d82a94690cc46bda8cae90792b9424908dc9ee99505d9b2d4070ee not found: ID does not exist" containerID="434394ed82d82a94690cc46bda8cae90792b9424908dc9ee99505d9b2d4070ee" Dec 03 14:50:16 crc kubenswrapper[4805]: I1203 14:50:16.865118 4805 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"434394ed82d82a94690cc46bda8cae90792b9424908dc9ee99505d9b2d4070ee"} err="failed to get container status \"434394ed82d82a94690cc46bda8cae90792b9424908dc9ee99505d9b2d4070ee\": rpc error: code = NotFound desc = could not find container \"434394ed82d82a94690cc46bda8cae90792b9424908dc9ee99505d9b2d4070ee\": container with ID starting with 434394ed82d82a94690cc46bda8cae90792b9424908dc9ee99505d9b2d4070ee not found: ID does not exist" Dec 03 14:50:16 crc kubenswrapper[4805]: I1203 14:50:16.865132 4805 scope.go:117] "RemoveContainer" containerID="16fb611feed7a9e6917e4609f89a801fef9f9c130c6ce0fa6adca48860d97915" Dec 03 14:50:16 crc kubenswrapper[4805]: E1203 14:50:16.865461 4805 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"16fb611feed7a9e6917e4609f89a801fef9f9c130c6ce0fa6adca48860d97915\": container with ID starting with 16fb611feed7a9e6917e4609f89a801fef9f9c130c6ce0fa6adca48860d97915 not found: ID does not exist" containerID="16fb611feed7a9e6917e4609f89a801fef9f9c130c6ce0fa6adca48860d97915" Dec 03 14:50:16 crc kubenswrapper[4805]: I1203 14:50:16.865489 4805 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"16fb611feed7a9e6917e4609f89a801fef9f9c130c6ce0fa6adca48860d97915"} err="failed to get container status \"16fb611feed7a9e6917e4609f89a801fef9f9c130c6ce0fa6adca48860d97915\": rpc error: code = NotFound desc = could not find container \"16fb611feed7a9e6917e4609f89a801fef9f9c130c6ce0fa6adca48860d97915\": container with ID starting with 16fb611feed7a9e6917e4609f89a801fef9f9c130c6ce0fa6adca48860d97915 not found: ID does not exist" Dec 03 14:50:16 crc kubenswrapper[4805]: I1203 14:50:16.990073 4805 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-n2pkd" Dec 03 14:50:17 crc kubenswrapper[4805]: I1203 14:50:17.109441 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-m6sqm\" (UniqueName: \"kubernetes.io/projected/dcceed7d-71aa-432f-94b3-94ce65599157-kube-api-access-m6sqm\") pod \"dcceed7d-71aa-432f-94b3-94ce65599157\" (UID: \"dcceed7d-71aa-432f-94b3-94ce65599157\") " Dec 03 14:50:17 crc kubenswrapper[4805]: I1203 14:50:17.109569 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/dcceed7d-71aa-432f-94b3-94ce65599157-utilities\") pod \"dcceed7d-71aa-432f-94b3-94ce65599157\" (UID: \"dcceed7d-71aa-432f-94b3-94ce65599157\") " Dec 03 14:50:17 crc kubenswrapper[4805]: I1203 14:50:17.109944 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/dcceed7d-71aa-432f-94b3-94ce65599157-catalog-content\") pod \"dcceed7d-71aa-432f-94b3-94ce65599157\" (UID: \"dcceed7d-71aa-432f-94b3-94ce65599157\") " Dec 03 14:50:17 crc kubenswrapper[4805]: I1203 14:50:17.110769 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/dcceed7d-71aa-432f-94b3-94ce65599157-utilities" (OuterVolumeSpecName: "utilities") pod "dcceed7d-71aa-432f-94b3-94ce65599157" (UID: "dcceed7d-71aa-432f-94b3-94ce65599157"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 14:50:17 crc kubenswrapper[4805]: I1203 14:50:17.121446 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/dcceed7d-71aa-432f-94b3-94ce65599157-kube-api-access-m6sqm" (OuterVolumeSpecName: "kube-api-access-m6sqm") pod "dcceed7d-71aa-432f-94b3-94ce65599157" (UID: "dcceed7d-71aa-432f-94b3-94ce65599157"). InnerVolumeSpecName "kube-api-access-m6sqm". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 14:50:17 crc kubenswrapper[4805]: I1203 14:50:17.188114 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/dcceed7d-71aa-432f-94b3-94ce65599157-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "dcceed7d-71aa-432f-94b3-94ce65599157" (UID: "dcceed7d-71aa-432f-94b3-94ce65599157"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 14:50:17 crc kubenswrapper[4805]: I1203 14:50:17.211954 4805 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/dcceed7d-71aa-432f-94b3-94ce65599157-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 03 14:50:17 crc kubenswrapper[4805]: I1203 14:50:17.212248 4805 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-m6sqm\" (UniqueName: \"kubernetes.io/projected/dcceed7d-71aa-432f-94b3-94ce65599157-kube-api-access-m6sqm\") on node \"crc\" DevicePath \"\"" Dec 03 14:50:17 crc kubenswrapper[4805]: I1203 14:50:17.212340 4805 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/dcceed7d-71aa-432f-94b3-94ce65599157-utilities\") on node \"crc\" DevicePath \"\"" Dec 03 14:50:17 crc kubenswrapper[4805]: I1203 14:50:17.781954 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-n2pkd" event={"ID":"dcceed7d-71aa-432f-94b3-94ce65599157","Type":"ContainerDied","Data":"4f5ad3a7634b6148b208698721aaa1a1bc201d630d888a8394e0d89956b3365d"} Dec 03 14:50:17 crc kubenswrapper[4805]: I1203 14:50:17.782035 4805 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-n2pkd" Dec 03 14:50:17 crc kubenswrapper[4805]: I1203 14:50:17.782062 4805 scope.go:117] "RemoveContainer" containerID="eea5bec2ab8cc6e0297401179d333a03753a51e91980dcb7111c319360d8f6d8" Dec 03 14:50:17 crc kubenswrapper[4805]: I1203 14:50:17.811270 4805 scope.go:117] "RemoveContainer" containerID="c613f21aa697fdbbca70292713ee8e799539aa36bcc69bff368ed9da7857f39e" Dec 03 14:50:17 crc kubenswrapper[4805]: I1203 14:50:17.837262 4805 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-n2pkd"] Dec 03 14:50:17 crc kubenswrapper[4805]: I1203 14:50:17.848200 4805 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-n2pkd"] Dec 03 14:50:17 crc kubenswrapper[4805]: I1203 14:50:17.861158 4805 scope.go:117] "RemoveContainer" containerID="a4241d2d9d223b2e571d10a25626916349525f6d050d92f102e055ec0df63d79" Dec 03 14:50:18 crc kubenswrapper[4805]: I1203 14:50:18.706481 4805 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7b48ed82-1c33-4361-9c24-7ac4a3ff8715" path="/var/lib/kubelet/pods/7b48ed82-1c33-4361-9c24-7ac4a3ff8715/volumes" Dec 03 14:50:18 crc kubenswrapper[4805]: I1203 14:50:18.707347 4805 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="dcceed7d-71aa-432f-94b3-94ce65599157" path="/var/lib/kubelet/pods/dcceed7d-71aa-432f-94b3-94ce65599157/volumes" Dec 03 14:50:19 crc kubenswrapper[4805]: I1203 14:50:19.695384 4805 scope.go:117] "RemoveContainer" containerID="fec9f6923538948ef3e55903bb8e108d1f94095a493c6b5117232e21f41ab624" Dec 03 14:50:19 crc kubenswrapper[4805]: E1203 14:50:19.695731 4805 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-gskh4_openshift-machine-config-operator(6098937f-e3f6-45e8-a647-4994a79cd711)\"" pod="openshift-machine-config-operator/machine-config-daemon-gskh4" podUID="6098937f-e3f6-45e8-a647-4994a79cd711" Dec 03 14:50:34 crc kubenswrapper[4805]: I1203 14:50:34.694906 4805 scope.go:117] "RemoveContainer" containerID="fec9f6923538948ef3e55903bb8e108d1f94095a493c6b5117232e21f41ab624" Dec 03 14:50:34 crc kubenswrapper[4805]: E1203 14:50:34.695711 4805 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-gskh4_openshift-machine-config-operator(6098937f-e3f6-45e8-a647-4994a79cd711)\"" pod="openshift-machine-config-operator/machine-config-daemon-gskh4" podUID="6098937f-e3f6-45e8-a647-4994a79cd711" Dec 03 14:50:47 crc kubenswrapper[4805]: I1203 14:50:47.695452 4805 scope.go:117] "RemoveContainer" containerID="fec9f6923538948ef3e55903bb8e108d1f94095a493c6b5117232e21f41ab624" Dec 03 14:50:47 crc kubenswrapper[4805]: E1203 14:50:47.696654 4805 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-gskh4_openshift-machine-config-operator(6098937f-e3f6-45e8-a647-4994a79cd711)\"" pod="openshift-machine-config-operator/machine-config-daemon-gskh4" podUID="6098937f-e3f6-45e8-a647-4994a79cd711" Dec 03 14:50:58 crc kubenswrapper[4805]: I1203 14:50:58.694698 4805 scope.go:117] "RemoveContainer" containerID="fec9f6923538948ef3e55903bb8e108d1f94095a493c6b5117232e21f41ab624" Dec 03 14:50:58 crc kubenswrapper[4805]: E1203 14:50:58.695675 4805 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-gskh4_openshift-machine-config-operator(6098937f-e3f6-45e8-a647-4994a79cd711)\"" pod="openshift-machine-config-operator/machine-config-daemon-gskh4" podUID="6098937f-e3f6-45e8-a647-4994a79cd711" Dec 03 14:51:12 crc kubenswrapper[4805]: I1203 14:51:12.695433 4805 scope.go:117] "RemoveContainer" containerID="fec9f6923538948ef3e55903bb8e108d1f94095a493c6b5117232e21f41ab624" Dec 03 14:51:12 crc kubenswrapper[4805]: E1203 14:51:12.696408 4805 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-gskh4_openshift-machine-config-operator(6098937f-e3f6-45e8-a647-4994a79cd711)\"" pod="openshift-machine-config-operator/machine-config-daemon-gskh4" podUID="6098937f-e3f6-45e8-a647-4994a79cd711" Dec 03 14:51:24 crc kubenswrapper[4805]: I1203 14:51:24.695773 4805 scope.go:117] "RemoveContainer" containerID="fec9f6923538948ef3e55903bb8e108d1f94095a493c6b5117232e21f41ab624" Dec 03 14:51:24 crc kubenswrapper[4805]: E1203 14:51:24.697133 4805 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-gskh4_openshift-machine-config-operator(6098937f-e3f6-45e8-a647-4994a79cd711)\"" pod="openshift-machine-config-operator/machine-config-daemon-gskh4" podUID="6098937f-e3f6-45e8-a647-4994a79cd711" Dec 03 14:51:38 crc kubenswrapper[4805]: I1203 14:51:38.694978 4805 scope.go:117] "RemoveContainer" containerID="fec9f6923538948ef3e55903bb8e108d1f94095a493c6b5117232e21f41ab624" Dec 03 14:51:38 crc kubenswrapper[4805]: E1203 14:51:38.696093 4805 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-gskh4_openshift-machine-config-operator(6098937f-e3f6-45e8-a647-4994a79cd711)\"" pod="openshift-machine-config-operator/machine-config-daemon-gskh4" podUID="6098937f-e3f6-45e8-a647-4994a79cd711" Dec 03 14:51:52 crc kubenswrapper[4805]: I1203 14:51:52.696820 4805 scope.go:117] "RemoveContainer" containerID="fec9f6923538948ef3e55903bb8e108d1f94095a493c6b5117232e21f41ab624" Dec 03 14:51:52 crc kubenswrapper[4805]: E1203 14:51:52.697997 4805 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-gskh4_openshift-machine-config-operator(6098937f-e3f6-45e8-a647-4994a79cd711)\"" pod="openshift-machine-config-operator/machine-config-daemon-gskh4" podUID="6098937f-e3f6-45e8-a647-4994a79cd711" Dec 03 14:52:05 crc kubenswrapper[4805]: I1203 14:52:05.694429 4805 scope.go:117] "RemoveContainer" containerID="fec9f6923538948ef3e55903bb8e108d1f94095a493c6b5117232e21f41ab624" Dec 03 14:52:05 crc kubenswrapper[4805]: E1203 14:52:05.695157 4805 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-gskh4_openshift-machine-config-operator(6098937f-e3f6-45e8-a647-4994a79cd711)\"" pod="openshift-machine-config-operator/machine-config-daemon-gskh4" podUID="6098937f-e3f6-45e8-a647-4994a79cd711" Dec 03 14:52:17 crc kubenswrapper[4805]: I1203 14:52:17.695645 4805 scope.go:117] "RemoveContainer" containerID="fec9f6923538948ef3e55903bb8e108d1f94095a493c6b5117232e21f41ab624" Dec 03 14:52:17 crc kubenswrapper[4805]: E1203 14:52:17.696748 4805 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-gskh4_openshift-machine-config-operator(6098937f-e3f6-45e8-a647-4994a79cd711)\"" pod="openshift-machine-config-operator/machine-config-daemon-gskh4" podUID="6098937f-e3f6-45e8-a647-4994a79cd711" Dec 03 14:52:30 crc kubenswrapper[4805]: I1203 14:52:30.695173 4805 scope.go:117] "RemoveContainer" containerID="fec9f6923538948ef3e55903bb8e108d1f94095a493c6b5117232e21f41ab624" Dec 03 14:52:30 crc kubenswrapper[4805]: E1203 14:52:30.696280 4805 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-gskh4_openshift-machine-config-operator(6098937f-e3f6-45e8-a647-4994a79cd711)\"" pod="openshift-machine-config-operator/machine-config-daemon-gskh4" podUID="6098937f-e3f6-45e8-a647-4994a79cd711" Dec 03 14:52:41 crc kubenswrapper[4805]: I1203 14:52:41.694617 4805 scope.go:117] "RemoveContainer" containerID="fec9f6923538948ef3e55903bb8e108d1f94095a493c6b5117232e21f41ab624" Dec 03 14:52:41 crc kubenswrapper[4805]: E1203 14:52:41.695442 4805 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-gskh4_openshift-machine-config-operator(6098937f-e3f6-45e8-a647-4994a79cd711)\"" pod="openshift-machine-config-operator/machine-config-daemon-gskh4" podUID="6098937f-e3f6-45e8-a647-4994a79cd711" Dec 03 14:52:54 crc kubenswrapper[4805]: I1203 14:52:54.695991 4805 scope.go:117] "RemoveContainer" containerID="fec9f6923538948ef3e55903bb8e108d1f94095a493c6b5117232e21f41ab624" Dec 03 14:52:54 crc kubenswrapper[4805]: E1203 14:52:54.697254 4805 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-gskh4_openshift-machine-config-operator(6098937f-e3f6-45e8-a647-4994a79cd711)\"" pod="openshift-machine-config-operator/machine-config-daemon-gskh4" podUID="6098937f-e3f6-45e8-a647-4994a79cd711" Dec 03 14:53:07 crc kubenswrapper[4805]: I1203 14:53:07.694934 4805 scope.go:117] "RemoveContainer" containerID="fec9f6923538948ef3e55903bb8e108d1f94095a493c6b5117232e21f41ab624" Dec 03 14:53:07 crc kubenswrapper[4805]: E1203 14:53:07.696005 4805 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-gskh4_openshift-machine-config-operator(6098937f-e3f6-45e8-a647-4994a79cd711)\"" pod="openshift-machine-config-operator/machine-config-daemon-gskh4" podUID="6098937f-e3f6-45e8-a647-4994a79cd711" Dec 03 14:53:18 crc kubenswrapper[4805]: I1203 14:53:18.695048 4805 scope.go:117] "RemoveContainer" containerID="fec9f6923538948ef3e55903bb8e108d1f94095a493c6b5117232e21f41ab624" Dec 03 14:53:18 crc kubenswrapper[4805]: E1203 14:53:18.696202 4805 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-gskh4_openshift-machine-config-operator(6098937f-e3f6-45e8-a647-4994a79cd711)\"" pod="openshift-machine-config-operator/machine-config-daemon-gskh4" podUID="6098937f-e3f6-45e8-a647-4994a79cd711" Dec 03 14:53:18 crc kubenswrapper[4805]: I1203 14:53:18.747583 4805 generic.go:334] "Generic (PLEG): container finished" podID="1776968b-6613-4b93-bb53-a91c3b6c3d2c" containerID="b388e6b00c345fb123ed444bc2e736f841bc1e5e7980d411065a9f38a5783437" exitCode=0 Dec 03 14:53:18 crc kubenswrapper[4805]: I1203 14:53:18.747637 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-lmlms" event={"ID":"1776968b-6613-4b93-bb53-a91c3b6c3d2c","Type":"ContainerDied","Data":"b388e6b00c345fb123ed444bc2e736f841bc1e5e7980d411065a9f38a5783437"} Dec 03 14:53:20 crc kubenswrapper[4805]: I1203 14:53:20.244758 4805 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-lmlms" Dec 03 14:53:20 crc kubenswrapper[4805]: I1203 14:53:20.396338 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-szhk8\" (UniqueName: \"kubernetes.io/projected/1776968b-6613-4b93-bb53-a91c3b6c3d2c-kube-api-access-szhk8\") pod \"1776968b-6613-4b93-bb53-a91c3b6c3d2c\" (UID: \"1776968b-6613-4b93-bb53-a91c3b6c3d2c\") " Dec 03 14:53:20 crc kubenswrapper[4805]: I1203 14:53:20.396573 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"libvirt-secret-0\" (UniqueName: \"kubernetes.io/secret/1776968b-6613-4b93-bb53-a91c3b6c3d2c-libvirt-secret-0\") pod \"1776968b-6613-4b93-bb53-a91c3b6c3d2c\" (UID: \"1776968b-6613-4b93-bb53-a91c3b6c3d2c\") " Dec 03 14:53:20 crc kubenswrapper[4805]: I1203 14:53:20.396690 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/1776968b-6613-4b93-bb53-a91c3b6c3d2c-ssh-key\") pod \"1776968b-6613-4b93-bb53-a91c3b6c3d2c\" (UID: \"1776968b-6613-4b93-bb53-a91c3b6c3d2c\") " Dec 03 14:53:20 crc kubenswrapper[4805]: I1203 14:53:20.396774 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/1776968b-6613-4b93-bb53-a91c3b6c3d2c-inventory\") pod \"1776968b-6613-4b93-bb53-a91c3b6c3d2c\" (UID: \"1776968b-6613-4b93-bb53-a91c3b6c3d2c\") " Dec 03 14:53:20 crc kubenswrapper[4805]: I1203 14:53:20.396825 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1776968b-6613-4b93-bb53-a91c3b6c3d2c-libvirt-combined-ca-bundle\") pod \"1776968b-6613-4b93-bb53-a91c3b6c3d2c\" (UID: \"1776968b-6613-4b93-bb53-a91c3b6c3d2c\") " Dec 03 14:53:20 crc kubenswrapper[4805]: I1203 14:53:20.402608 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1776968b-6613-4b93-bb53-a91c3b6c3d2c-libvirt-combined-ca-bundle" (OuterVolumeSpecName: "libvirt-combined-ca-bundle") pod "1776968b-6613-4b93-bb53-a91c3b6c3d2c" (UID: "1776968b-6613-4b93-bb53-a91c3b6c3d2c"). InnerVolumeSpecName "libvirt-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 14:53:20 crc kubenswrapper[4805]: I1203 14:53:20.403069 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1776968b-6613-4b93-bb53-a91c3b6c3d2c-kube-api-access-szhk8" (OuterVolumeSpecName: "kube-api-access-szhk8") pod "1776968b-6613-4b93-bb53-a91c3b6c3d2c" (UID: "1776968b-6613-4b93-bb53-a91c3b6c3d2c"). InnerVolumeSpecName "kube-api-access-szhk8". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 14:53:20 crc kubenswrapper[4805]: I1203 14:53:20.423758 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1776968b-6613-4b93-bb53-a91c3b6c3d2c-libvirt-secret-0" (OuterVolumeSpecName: "libvirt-secret-0") pod "1776968b-6613-4b93-bb53-a91c3b6c3d2c" (UID: "1776968b-6613-4b93-bb53-a91c3b6c3d2c"). InnerVolumeSpecName "libvirt-secret-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 14:53:20 crc kubenswrapper[4805]: I1203 14:53:20.446314 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1776968b-6613-4b93-bb53-a91c3b6c3d2c-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "1776968b-6613-4b93-bb53-a91c3b6c3d2c" (UID: "1776968b-6613-4b93-bb53-a91c3b6c3d2c"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 14:53:20 crc kubenswrapper[4805]: I1203 14:53:20.454399 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1776968b-6613-4b93-bb53-a91c3b6c3d2c-inventory" (OuterVolumeSpecName: "inventory") pod "1776968b-6613-4b93-bb53-a91c3b6c3d2c" (UID: "1776968b-6613-4b93-bb53-a91c3b6c3d2c"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 14:53:20 crc kubenswrapper[4805]: I1203 14:53:20.498967 4805 reconciler_common.go:293] "Volume detached for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1776968b-6613-4b93-bb53-a91c3b6c3d2c-libvirt-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 14:53:20 crc kubenswrapper[4805]: I1203 14:53:20.499000 4805 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-szhk8\" (UniqueName: \"kubernetes.io/projected/1776968b-6613-4b93-bb53-a91c3b6c3d2c-kube-api-access-szhk8\") on node \"crc\" DevicePath \"\"" Dec 03 14:53:20 crc kubenswrapper[4805]: I1203 14:53:20.499012 4805 reconciler_common.go:293] "Volume detached for volume \"libvirt-secret-0\" (UniqueName: \"kubernetes.io/secret/1776968b-6613-4b93-bb53-a91c3b6c3d2c-libvirt-secret-0\") on node \"crc\" DevicePath \"\"" Dec 03 14:53:20 crc kubenswrapper[4805]: I1203 14:53:20.499020 4805 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/1776968b-6613-4b93-bb53-a91c3b6c3d2c-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 03 14:53:20 crc kubenswrapper[4805]: I1203 14:53:20.499030 4805 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/1776968b-6613-4b93-bb53-a91c3b6c3d2c-inventory\") on node \"crc\" DevicePath \"\"" Dec 03 14:53:20 crc kubenswrapper[4805]: I1203 14:53:20.769651 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-lmlms" event={"ID":"1776968b-6613-4b93-bb53-a91c3b6c3d2c","Type":"ContainerDied","Data":"84a9a8ddff972f2cded1372841d93a4d2004ba1dd9d9d7ae1e81848f9043f078"} Dec 03 14:53:20 crc kubenswrapper[4805]: I1203 14:53:20.769704 4805 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="84a9a8ddff972f2cded1372841d93a4d2004ba1dd9d9d7ae1e81848f9043f078" Dec 03 14:53:20 crc kubenswrapper[4805]: I1203 14:53:20.769767 4805 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-lmlms" Dec 03 14:53:20 crc kubenswrapper[4805]: I1203 14:53:20.890450 4805 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-edpm-deployment-openstack-edpm-ipam-75h75"] Dec 03 14:53:20 crc kubenswrapper[4805]: E1203 14:53:20.890833 4805 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="dcceed7d-71aa-432f-94b3-94ce65599157" containerName="registry-server" Dec 03 14:53:20 crc kubenswrapper[4805]: I1203 14:53:20.890975 4805 state_mem.go:107] "Deleted CPUSet assignment" podUID="dcceed7d-71aa-432f-94b3-94ce65599157" containerName="registry-server" Dec 03 14:53:20 crc kubenswrapper[4805]: E1203 14:53:20.890997 4805 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1776968b-6613-4b93-bb53-a91c3b6c3d2c" containerName="libvirt-edpm-deployment-openstack-edpm-ipam" Dec 03 14:53:20 crc kubenswrapper[4805]: I1203 14:53:20.891003 4805 state_mem.go:107] "Deleted CPUSet assignment" podUID="1776968b-6613-4b93-bb53-a91c3b6c3d2c" containerName="libvirt-edpm-deployment-openstack-edpm-ipam" Dec 03 14:53:20 crc kubenswrapper[4805]: E1203 14:53:20.891015 4805 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e1211b72-d818-49f8-88a1-4f371a660930" containerName="registry-server" Dec 03 14:53:20 crc kubenswrapper[4805]: I1203 14:53:20.891022 4805 state_mem.go:107] "Deleted CPUSet assignment" podUID="e1211b72-d818-49f8-88a1-4f371a660930" containerName="registry-server" Dec 03 14:53:20 crc kubenswrapper[4805]: E1203 14:53:20.891036 4805 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e1211b72-d818-49f8-88a1-4f371a660930" containerName="extract-content" Dec 03 14:53:20 crc kubenswrapper[4805]: I1203 14:53:20.891042 4805 state_mem.go:107] "Deleted CPUSet assignment" podUID="e1211b72-d818-49f8-88a1-4f371a660930" containerName="extract-content" Dec 03 14:53:20 crc kubenswrapper[4805]: E1203 14:53:20.891056 4805 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7b48ed82-1c33-4361-9c24-7ac4a3ff8715" containerName="extract-utilities" Dec 03 14:53:20 crc kubenswrapper[4805]: I1203 14:53:20.891063 4805 state_mem.go:107] "Deleted CPUSet assignment" podUID="7b48ed82-1c33-4361-9c24-7ac4a3ff8715" containerName="extract-utilities" Dec 03 14:53:20 crc kubenswrapper[4805]: E1203 14:53:20.891085 4805 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7b48ed82-1c33-4361-9c24-7ac4a3ff8715" containerName="registry-server" Dec 03 14:53:20 crc kubenswrapper[4805]: I1203 14:53:20.891093 4805 state_mem.go:107] "Deleted CPUSet assignment" podUID="7b48ed82-1c33-4361-9c24-7ac4a3ff8715" containerName="registry-server" Dec 03 14:53:20 crc kubenswrapper[4805]: E1203 14:53:20.891104 4805 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7b48ed82-1c33-4361-9c24-7ac4a3ff8715" containerName="extract-content" Dec 03 14:53:20 crc kubenswrapper[4805]: I1203 14:53:20.891111 4805 state_mem.go:107] "Deleted CPUSet assignment" podUID="7b48ed82-1c33-4361-9c24-7ac4a3ff8715" containerName="extract-content" Dec 03 14:53:20 crc kubenswrapper[4805]: E1203 14:53:20.891127 4805 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="dcceed7d-71aa-432f-94b3-94ce65599157" containerName="extract-content" Dec 03 14:53:20 crc kubenswrapper[4805]: I1203 14:53:20.891133 4805 state_mem.go:107] "Deleted CPUSet assignment" podUID="dcceed7d-71aa-432f-94b3-94ce65599157" containerName="extract-content" Dec 03 14:53:20 crc kubenswrapper[4805]: E1203 14:53:20.891149 4805 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="dcceed7d-71aa-432f-94b3-94ce65599157" containerName="extract-utilities" Dec 03 14:53:20 crc kubenswrapper[4805]: I1203 14:53:20.891155 4805 state_mem.go:107] "Deleted CPUSet assignment" podUID="dcceed7d-71aa-432f-94b3-94ce65599157" containerName="extract-utilities" Dec 03 14:53:20 crc kubenswrapper[4805]: E1203 14:53:20.891162 4805 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e1211b72-d818-49f8-88a1-4f371a660930" containerName="extract-utilities" Dec 03 14:53:20 crc kubenswrapper[4805]: I1203 14:53:20.891168 4805 state_mem.go:107] "Deleted CPUSet assignment" podUID="e1211b72-d818-49f8-88a1-4f371a660930" containerName="extract-utilities" Dec 03 14:53:20 crc kubenswrapper[4805]: I1203 14:53:20.891345 4805 memory_manager.go:354] "RemoveStaleState removing state" podUID="1776968b-6613-4b93-bb53-a91c3b6c3d2c" containerName="libvirt-edpm-deployment-openstack-edpm-ipam" Dec 03 14:53:20 crc kubenswrapper[4805]: I1203 14:53:20.891361 4805 memory_manager.go:354] "RemoveStaleState removing state" podUID="7b48ed82-1c33-4361-9c24-7ac4a3ff8715" containerName="registry-server" Dec 03 14:53:20 crc kubenswrapper[4805]: I1203 14:53:20.891371 4805 memory_manager.go:354] "RemoveStaleState removing state" podUID="dcceed7d-71aa-432f-94b3-94ce65599157" containerName="registry-server" Dec 03 14:53:20 crc kubenswrapper[4805]: I1203 14:53:20.891393 4805 memory_manager.go:354] "RemoveStaleState removing state" podUID="e1211b72-d818-49f8-88a1-4f371a660930" containerName="registry-server" Dec 03 14:53:20 crc kubenswrapper[4805]: I1203 14:53:20.892664 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-75h75" Dec 03 14:53:20 crc kubenswrapper[4805]: I1203 14:53:20.905161 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-compute-config" Dec 03 14:53:20 crc kubenswrapper[4805]: I1203 14:53:20.905243 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Dec 03 14:53:20 crc kubenswrapper[4805]: I1203 14:53:20.905258 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-migration-ssh-key" Dec 03 14:53:20 crc kubenswrapper[4805]: I1203 14:53:20.905382 4805 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 03 14:53:20 crc kubenswrapper[4805]: I1203 14:53:20.907749 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-n7jt4" Dec 03 14:53:20 crc kubenswrapper[4805]: I1203 14:53:20.907935 4805 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"nova-extra-config" Dec 03 14:53:20 crc kubenswrapper[4805]: I1203 14:53:20.908078 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Dec 03 14:53:20 crc kubenswrapper[4805]: I1203 14:53:20.911696 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-edpm-deployment-openstack-edpm-ipam-75h75"] Dec 03 14:53:21 crc kubenswrapper[4805]: I1203 14:53:21.010903 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dkm9v\" (UniqueName: \"kubernetes.io/projected/7e18e1db-7529-412b-bf01-17cb78e689b9-kube-api-access-dkm9v\") pod \"nova-edpm-deployment-openstack-edpm-ipam-75h75\" (UID: \"7e18e1db-7529-412b-bf01-17cb78e689b9\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-75h75" Dec 03 14:53:21 crc kubenswrapper[4805]: I1203 14:53:21.010960 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-extra-config-0\" (UniqueName: \"kubernetes.io/configmap/7e18e1db-7529-412b-bf01-17cb78e689b9-nova-extra-config-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-75h75\" (UID: \"7e18e1db-7529-412b-bf01-17cb78e689b9\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-75h75" Dec 03 14:53:21 crc kubenswrapper[4805]: I1203 14:53:21.011003 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/7e18e1db-7529-412b-bf01-17cb78e689b9-nova-cell1-compute-config-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-75h75\" (UID: \"7e18e1db-7529-412b-bf01-17cb78e689b9\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-75h75" Dec 03 14:53:21 crc kubenswrapper[4805]: I1203 14:53:21.011025 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/7e18e1db-7529-412b-bf01-17cb78e689b9-ssh-key\") pod \"nova-edpm-deployment-openstack-edpm-ipam-75h75\" (UID: \"7e18e1db-7529-412b-bf01-17cb78e689b9\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-75h75" Dec 03 14:53:21 crc kubenswrapper[4805]: I1203 14:53:21.011067 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/7e18e1db-7529-412b-bf01-17cb78e689b9-nova-migration-ssh-key-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-75h75\" (UID: \"7e18e1db-7529-412b-bf01-17cb78e689b9\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-75h75" Dec 03 14:53:21 crc kubenswrapper[4805]: I1203 14:53:21.011105 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7e18e1db-7529-412b-bf01-17cb78e689b9-nova-combined-ca-bundle\") pod \"nova-edpm-deployment-openstack-edpm-ipam-75h75\" (UID: \"7e18e1db-7529-412b-bf01-17cb78e689b9\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-75h75" Dec 03 14:53:21 crc kubenswrapper[4805]: I1203 14:53:21.011188 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/7e18e1db-7529-412b-bf01-17cb78e689b9-nova-migration-ssh-key-1\") pod \"nova-edpm-deployment-openstack-edpm-ipam-75h75\" (UID: \"7e18e1db-7529-412b-bf01-17cb78e689b9\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-75h75" Dec 03 14:53:21 crc kubenswrapper[4805]: I1203 14:53:21.011254 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/7e18e1db-7529-412b-bf01-17cb78e689b9-inventory\") pod \"nova-edpm-deployment-openstack-edpm-ipam-75h75\" (UID: \"7e18e1db-7529-412b-bf01-17cb78e689b9\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-75h75" Dec 03 14:53:21 crc kubenswrapper[4805]: I1203 14:53:21.011314 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/7e18e1db-7529-412b-bf01-17cb78e689b9-nova-cell1-compute-config-1\") pod \"nova-edpm-deployment-openstack-edpm-ipam-75h75\" (UID: \"7e18e1db-7529-412b-bf01-17cb78e689b9\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-75h75" Dec 03 14:53:21 crc kubenswrapper[4805]: I1203 14:53:21.113388 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/7e18e1db-7529-412b-bf01-17cb78e689b9-nova-cell1-compute-config-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-75h75\" (UID: \"7e18e1db-7529-412b-bf01-17cb78e689b9\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-75h75" Dec 03 14:53:21 crc kubenswrapper[4805]: I1203 14:53:21.113435 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/7e18e1db-7529-412b-bf01-17cb78e689b9-ssh-key\") pod \"nova-edpm-deployment-openstack-edpm-ipam-75h75\" (UID: \"7e18e1db-7529-412b-bf01-17cb78e689b9\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-75h75" Dec 03 14:53:21 crc kubenswrapper[4805]: I1203 14:53:21.113464 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/7e18e1db-7529-412b-bf01-17cb78e689b9-nova-migration-ssh-key-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-75h75\" (UID: \"7e18e1db-7529-412b-bf01-17cb78e689b9\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-75h75" Dec 03 14:53:21 crc kubenswrapper[4805]: I1203 14:53:21.113497 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7e18e1db-7529-412b-bf01-17cb78e689b9-nova-combined-ca-bundle\") pod \"nova-edpm-deployment-openstack-edpm-ipam-75h75\" (UID: \"7e18e1db-7529-412b-bf01-17cb78e689b9\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-75h75" Dec 03 14:53:21 crc kubenswrapper[4805]: I1203 14:53:21.113555 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/7e18e1db-7529-412b-bf01-17cb78e689b9-nova-migration-ssh-key-1\") pod \"nova-edpm-deployment-openstack-edpm-ipam-75h75\" (UID: \"7e18e1db-7529-412b-bf01-17cb78e689b9\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-75h75" Dec 03 14:53:21 crc kubenswrapper[4805]: I1203 14:53:21.113616 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/7e18e1db-7529-412b-bf01-17cb78e689b9-inventory\") pod \"nova-edpm-deployment-openstack-edpm-ipam-75h75\" (UID: \"7e18e1db-7529-412b-bf01-17cb78e689b9\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-75h75" Dec 03 14:53:21 crc kubenswrapper[4805]: I1203 14:53:21.113670 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/7e18e1db-7529-412b-bf01-17cb78e689b9-nova-cell1-compute-config-1\") pod \"nova-edpm-deployment-openstack-edpm-ipam-75h75\" (UID: \"7e18e1db-7529-412b-bf01-17cb78e689b9\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-75h75" Dec 03 14:53:21 crc kubenswrapper[4805]: I1203 14:53:21.113711 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dkm9v\" (UniqueName: \"kubernetes.io/projected/7e18e1db-7529-412b-bf01-17cb78e689b9-kube-api-access-dkm9v\") pod \"nova-edpm-deployment-openstack-edpm-ipam-75h75\" (UID: \"7e18e1db-7529-412b-bf01-17cb78e689b9\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-75h75" Dec 03 14:53:21 crc kubenswrapper[4805]: I1203 14:53:21.113739 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-extra-config-0\" (UniqueName: \"kubernetes.io/configmap/7e18e1db-7529-412b-bf01-17cb78e689b9-nova-extra-config-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-75h75\" (UID: \"7e18e1db-7529-412b-bf01-17cb78e689b9\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-75h75" Dec 03 14:53:21 crc kubenswrapper[4805]: I1203 14:53:21.114903 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-extra-config-0\" (UniqueName: \"kubernetes.io/configmap/7e18e1db-7529-412b-bf01-17cb78e689b9-nova-extra-config-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-75h75\" (UID: \"7e18e1db-7529-412b-bf01-17cb78e689b9\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-75h75" Dec 03 14:53:21 crc kubenswrapper[4805]: I1203 14:53:21.117665 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/7e18e1db-7529-412b-bf01-17cb78e689b9-nova-cell1-compute-config-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-75h75\" (UID: \"7e18e1db-7529-412b-bf01-17cb78e689b9\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-75h75" Dec 03 14:53:21 crc kubenswrapper[4805]: I1203 14:53:21.118273 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/7e18e1db-7529-412b-bf01-17cb78e689b9-ssh-key\") pod \"nova-edpm-deployment-openstack-edpm-ipam-75h75\" (UID: \"7e18e1db-7529-412b-bf01-17cb78e689b9\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-75h75" Dec 03 14:53:21 crc kubenswrapper[4805]: I1203 14:53:21.118334 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/7e18e1db-7529-412b-bf01-17cb78e689b9-nova-migration-ssh-key-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-75h75\" (UID: \"7e18e1db-7529-412b-bf01-17cb78e689b9\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-75h75" Dec 03 14:53:21 crc kubenswrapper[4805]: I1203 14:53:21.118606 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/7e18e1db-7529-412b-bf01-17cb78e689b9-inventory\") pod \"nova-edpm-deployment-openstack-edpm-ipam-75h75\" (UID: \"7e18e1db-7529-412b-bf01-17cb78e689b9\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-75h75" Dec 03 14:53:21 crc kubenswrapper[4805]: I1203 14:53:21.119107 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/7e18e1db-7529-412b-bf01-17cb78e689b9-nova-migration-ssh-key-1\") pod \"nova-edpm-deployment-openstack-edpm-ipam-75h75\" (UID: \"7e18e1db-7529-412b-bf01-17cb78e689b9\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-75h75" Dec 03 14:53:21 crc kubenswrapper[4805]: I1203 14:53:21.120181 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/7e18e1db-7529-412b-bf01-17cb78e689b9-nova-cell1-compute-config-1\") pod \"nova-edpm-deployment-openstack-edpm-ipam-75h75\" (UID: \"7e18e1db-7529-412b-bf01-17cb78e689b9\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-75h75" Dec 03 14:53:21 crc kubenswrapper[4805]: I1203 14:53:21.120305 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7e18e1db-7529-412b-bf01-17cb78e689b9-nova-combined-ca-bundle\") pod \"nova-edpm-deployment-openstack-edpm-ipam-75h75\" (UID: \"7e18e1db-7529-412b-bf01-17cb78e689b9\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-75h75" Dec 03 14:53:21 crc kubenswrapper[4805]: I1203 14:53:21.130524 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dkm9v\" (UniqueName: \"kubernetes.io/projected/7e18e1db-7529-412b-bf01-17cb78e689b9-kube-api-access-dkm9v\") pod \"nova-edpm-deployment-openstack-edpm-ipam-75h75\" (UID: \"7e18e1db-7529-412b-bf01-17cb78e689b9\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-75h75" Dec 03 14:53:21 crc kubenswrapper[4805]: I1203 14:53:21.226638 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-75h75" Dec 03 14:53:21 crc kubenswrapper[4805]: I1203 14:53:21.781805 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-edpm-deployment-openstack-edpm-ipam-75h75"] Dec 03 14:53:22 crc kubenswrapper[4805]: I1203 14:53:22.818716 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-75h75" event={"ID":"7e18e1db-7529-412b-bf01-17cb78e689b9","Type":"ContainerStarted","Data":"7616d8fd673eac4e5972a3f8f8dfaa218baf877f969a4730c5a8e8a90b3e0945"} Dec 03 14:53:22 crc kubenswrapper[4805]: I1203 14:53:22.819253 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-75h75" event={"ID":"7e18e1db-7529-412b-bf01-17cb78e689b9","Type":"ContainerStarted","Data":"3ce5c1d9cf5b363bb9ab6a2d4a4fa6f549a685868733dadbd47c509c002f9d87"} Dec 03 14:53:22 crc kubenswrapper[4805]: I1203 14:53:22.844749 4805 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-75h75" podStartSLOduration=2.404410984 podStartE2EDuration="2.84473447s" podCreationTimestamp="2025-12-03 14:53:20 +0000 UTC" firstStartedPulling="2025-12-03 14:53:21.8003365 +0000 UTC m=+2631.463253423" lastFinishedPulling="2025-12-03 14:53:22.240659986 +0000 UTC m=+2631.903576909" observedRunningTime="2025-12-03 14:53:22.839746425 +0000 UTC m=+2632.502663348" watchObservedRunningTime="2025-12-03 14:53:22.84473447 +0000 UTC m=+2632.507651393" Dec 03 14:53:32 crc kubenswrapper[4805]: I1203 14:53:32.696241 4805 scope.go:117] "RemoveContainer" containerID="fec9f6923538948ef3e55903bb8e108d1f94095a493c6b5117232e21f41ab624" Dec 03 14:53:32 crc kubenswrapper[4805]: E1203 14:53:32.697436 4805 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-gskh4_openshift-machine-config-operator(6098937f-e3f6-45e8-a647-4994a79cd711)\"" pod="openshift-machine-config-operator/machine-config-daemon-gskh4" podUID="6098937f-e3f6-45e8-a647-4994a79cd711" Dec 03 14:53:44 crc kubenswrapper[4805]: I1203 14:53:44.694737 4805 scope.go:117] "RemoveContainer" containerID="fec9f6923538948ef3e55903bb8e108d1f94095a493c6b5117232e21f41ab624" Dec 03 14:53:44 crc kubenswrapper[4805]: E1203 14:53:44.695660 4805 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-gskh4_openshift-machine-config-operator(6098937f-e3f6-45e8-a647-4994a79cd711)\"" pod="openshift-machine-config-operator/machine-config-daemon-gskh4" podUID="6098937f-e3f6-45e8-a647-4994a79cd711" Dec 03 14:53:55 crc kubenswrapper[4805]: I1203 14:53:55.695294 4805 scope.go:117] "RemoveContainer" containerID="fec9f6923538948ef3e55903bb8e108d1f94095a493c6b5117232e21f41ab624" Dec 03 14:53:55 crc kubenswrapper[4805]: E1203 14:53:55.696191 4805 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-gskh4_openshift-machine-config-operator(6098937f-e3f6-45e8-a647-4994a79cd711)\"" pod="openshift-machine-config-operator/machine-config-daemon-gskh4" podUID="6098937f-e3f6-45e8-a647-4994a79cd711" Dec 03 14:54:07 crc kubenswrapper[4805]: I1203 14:54:07.694730 4805 scope.go:117] "RemoveContainer" containerID="fec9f6923538948ef3e55903bb8e108d1f94095a493c6b5117232e21f41ab624" Dec 03 14:54:07 crc kubenswrapper[4805]: E1203 14:54:07.695683 4805 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-gskh4_openshift-machine-config-operator(6098937f-e3f6-45e8-a647-4994a79cd711)\"" pod="openshift-machine-config-operator/machine-config-daemon-gskh4" podUID="6098937f-e3f6-45e8-a647-4994a79cd711" Dec 03 14:54:18 crc kubenswrapper[4805]: I1203 14:54:18.694831 4805 scope.go:117] "RemoveContainer" containerID="fec9f6923538948ef3e55903bb8e108d1f94095a493c6b5117232e21f41ab624" Dec 03 14:54:18 crc kubenswrapper[4805]: E1203 14:54:18.695638 4805 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-gskh4_openshift-machine-config-operator(6098937f-e3f6-45e8-a647-4994a79cd711)\"" pod="openshift-machine-config-operator/machine-config-daemon-gskh4" podUID="6098937f-e3f6-45e8-a647-4994a79cd711" Dec 03 14:54:31 crc kubenswrapper[4805]: I1203 14:54:31.694710 4805 scope.go:117] "RemoveContainer" containerID="fec9f6923538948ef3e55903bb8e108d1f94095a493c6b5117232e21f41ab624" Dec 03 14:54:31 crc kubenswrapper[4805]: E1203 14:54:31.696827 4805 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-gskh4_openshift-machine-config-operator(6098937f-e3f6-45e8-a647-4994a79cd711)\"" pod="openshift-machine-config-operator/machine-config-daemon-gskh4" podUID="6098937f-e3f6-45e8-a647-4994a79cd711" Dec 03 14:54:43 crc kubenswrapper[4805]: I1203 14:54:43.695393 4805 scope.go:117] "RemoveContainer" containerID="fec9f6923538948ef3e55903bb8e108d1f94095a493c6b5117232e21f41ab624" Dec 03 14:54:43 crc kubenswrapper[4805]: E1203 14:54:43.696315 4805 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-gskh4_openshift-machine-config-operator(6098937f-e3f6-45e8-a647-4994a79cd711)\"" pod="openshift-machine-config-operator/machine-config-daemon-gskh4" podUID="6098937f-e3f6-45e8-a647-4994a79cd711" Dec 03 14:54:54 crc kubenswrapper[4805]: I1203 14:54:54.695098 4805 scope.go:117] "RemoveContainer" containerID="fec9f6923538948ef3e55903bb8e108d1f94095a493c6b5117232e21f41ab624" Dec 03 14:54:55 crc kubenswrapper[4805]: I1203 14:54:55.814278 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-gskh4" event={"ID":"6098937f-e3f6-45e8-a647-4994a79cd711","Type":"ContainerStarted","Data":"2e3f74f03f0ba2719087d9ca75ea7670c41a3ad116d1af23fbb1aeb3c76a97d3"} Dec 03 14:55:14 crc kubenswrapper[4805]: I1203 14:55:14.617086 4805 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-s2cxk"] Dec 03 14:55:14 crc kubenswrapper[4805]: I1203 14:55:14.620440 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-s2cxk" Dec 03 14:55:14 crc kubenswrapper[4805]: I1203 14:55:14.634440 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-s2cxk"] Dec 03 14:55:14 crc kubenswrapper[4805]: I1203 14:55:14.751018 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/dc26e67f-c323-4f89-b53c-d091c0a54e25-catalog-content\") pod \"redhat-marketplace-s2cxk\" (UID: \"dc26e67f-c323-4f89-b53c-d091c0a54e25\") " pod="openshift-marketplace/redhat-marketplace-s2cxk" Dec 03 14:55:14 crc kubenswrapper[4805]: I1203 14:55:14.751202 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-99n2x\" (UniqueName: \"kubernetes.io/projected/dc26e67f-c323-4f89-b53c-d091c0a54e25-kube-api-access-99n2x\") pod \"redhat-marketplace-s2cxk\" (UID: \"dc26e67f-c323-4f89-b53c-d091c0a54e25\") " pod="openshift-marketplace/redhat-marketplace-s2cxk" Dec 03 14:55:14 crc kubenswrapper[4805]: I1203 14:55:14.751680 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/dc26e67f-c323-4f89-b53c-d091c0a54e25-utilities\") pod \"redhat-marketplace-s2cxk\" (UID: \"dc26e67f-c323-4f89-b53c-d091c0a54e25\") " pod="openshift-marketplace/redhat-marketplace-s2cxk" Dec 03 14:55:14 crc kubenswrapper[4805]: I1203 14:55:14.853444 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/dc26e67f-c323-4f89-b53c-d091c0a54e25-utilities\") pod \"redhat-marketplace-s2cxk\" (UID: \"dc26e67f-c323-4f89-b53c-d091c0a54e25\") " pod="openshift-marketplace/redhat-marketplace-s2cxk" Dec 03 14:55:14 crc kubenswrapper[4805]: I1203 14:55:14.854149 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/dc26e67f-c323-4f89-b53c-d091c0a54e25-utilities\") pod \"redhat-marketplace-s2cxk\" (UID: \"dc26e67f-c323-4f89-b53c-d091c0a54e25\") " pod="openshift-marketplace/redhat-marketplace-s2cxk" Dec 03 14:55:14 crc kubenswrapper[4805]: I1203 14:55:14.854532 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/dc26e67f-c323-4f89-b53c-d091c0a54e25-catalog-content\") pod \"redhat-marketplace-s2cxk\" (UID: \"dc26e67f-c323-4f89-b53c-d091c0a54e25\") " pod="openshift-marketplace/redhat-marketplace-s2cxk" Dec 03 14:55:14 crc kubenswrapper[4805]: I1203 14:55:14.854695 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-99n2x\" (UniqueName: \"kubernetes.io/projected/dc26e67f-c323-4f89-b53c-d091c0a54e25-kube-api-access-99n2x\") pod \"redhat-marketplace-s2cxk\" (UID: \"dc26e67f-c323-4f89-b53c-d091c0a54e25\") " pod="openshift-marketplace/redhat-marketplace-s2cxk" Dec 03 14:55:14 crc kubenswrapper[4805]: I1203 14:55:14.854971 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/dc26e67f-c323-4f89-b53c-d091c0a54e25-catalog-content\") pod \"redhat-marketplace-s2cxk\" (UID: \"dc26e67f-c323-4f89-b53c-d091c0a54e25\") " pod="openshift-marketplace/redhat-marketplace-s2cxk" Dec 03 14:55:14 crc kubenswrapper[4805]: I1203 14:55:14.883247 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-99n2x\" (UniqueName: \"kubernetes.io/projected/dc26e67f-c323-4f89-b53c-d091c0a54e25-kube-api-access-99n2x\") pod \"redhat-marketplace-s2cxk\" (UID: \"dc26e67f-c323-4f89-b53c-d091c0a54e25\") " pod="openshift-marketplace/redhat-marketplace-s2cxk" Dec 03 14:55:14 crc kubenswrapper[4805]: I1203 14:55:14.947248 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-s2cxk" Dec 03 14:55:15 crc kubenswrapper[4805]: I1203 14:55:15.419094 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-s2cxk"] Dec 03 14:55:15 crc kubenswrapper[4805]: W1203 14:55:15.422299 4805 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-poddc26e67f_c323_4f89_b53c_d091c0a54e25.slice/crio-e029b5bffde13cd1138c46089e6948de2eaab5f7927033f012d362349f1da606 WatchSource:0}: Error finding container e029b5bffde13cd1138c46089e6948de2eaab5f7927033f012d362349f1da606: Status 404 returned error can't find the container with id e029b5bffde13cd1138c46089e6948de2eaab5f7927033f012d362349f1da606 Dec 03 14:55:16 crc kubenswrapper[4805]: I1203 14:55:16.072128 4805 generic.go:334] "Generic (PLEG): container finished" podID="dc26e67f-c323-4f89-b53c-d091c0a54e25" containerID="5fdd05f82f6170f4fbfee6d9d82a427d4f5b136ec58b24671c155d7dfb3d8bc2" exitCode=0 Dec 03 14:55:16 crc kubenswrapper[4805]: I1203 14:55:16.072200 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-s2cxk" event={"ID":"dc26e67f-c323-4f89-b53c-d091c0a54e25","Type":"ContainerDied","Data":"5fdd05f82f6170f4fbfee6d9d82a427d4f5b136ec58b24671c155d7dfb3d8bc2"} Dec 03 14:55:16 crc kubenswrapper[4805]: I1203 14:55:16.072430 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-s2cxk" event={"ID":"dc26e67f-c323-4f89-b53c-d091c0a54e25","Type":"ContainerStarted","Data":"e029b5bffde13cd1138c46089e6948de2eaab5f7927033f012d362349f1da606"} Dec 03 14:55:16 crc kubenswrapper[4805]: I1203 14:55:16.074456 4805 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 03 14:55:17 crc kubenswrapper[4805]: I1203 14:55:17.083249 4805 generic.go:334] "Generic (PLEG): container finished" podID="dc26e67f-c323-4f89-b53c-d091c0a54e25" containerID="7b65a7fb32ae6359f625bd3dc417454dacf006a5251bdb3c5cfbb131802aa77b" exitCode=0 Dec 03 14:55:17 crc kubenswrapper[4805]: I1203 14:55:17.083322 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-s2cxk" event={"ID":"dc26e67f-c323-4f89-b53c-d091c0a54e25","Type":"ContainerDied","Data":"7b65a7fb32ae6359f625bd3dc417454dacf006a5251bdb3c5cfbb131802aa77b"} Dec 03 14:55:18 crc kubenswrapper[4805]: I1203 14:55:18.094530 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-s2cxk" event={"ID":"dc26e67f-c323-4f89-b53c-d091c0a54e25","Type":"ContainerStarted","Data":"2aebf37755b90dfbbef4d6c5d258785e3c4ecea3627a9cdd07b97748cd2c2d75"} Dec 03 14:55:18 crc kubenswrapper[4805]: I1203 14:55:18.121033 4805 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-s2cxk" podStartSLOduration=2.7108745819999998 podStartE2EDuration="4.121014732s" podCreationTimestamp="2025-12-03 14:55:14 +0000 UTC" firstStartedPulling="2025-12-03 14:55:16.074212246 +0000 UTC m=+2745.737129159" lastFinishedPulling="2025-12-03 14:55:17.484352386 +0000 UTC m=+2747.147269309" observedRunningTime="2025-12-03 14:55:18.113209941 +0000 UTC m=+2747.776126874" watchObservedRunningTime="2025-12-03 14:55:18.121014732 +0000 UTC m=+2747.783931655" Dec 03 14:55:24 crc kubenswrapper[4805]: I1203 14:55:24.947754 4805 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-s2cxk" Dec 03 14:55:24 crc kubenswrapper[4805]: I1203 14:55:24.948980 4805 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-s2cxk" Dec 03 14:55:25 crc kubenswrapper[4805]: I1203 14:55:25.025300 4805 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-s2cxk" Dec 03 14:55:25 crc kubenswrapper[4805]: I1203 14:55:25.256709 4805 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-s2cxk" Dec 03 14:55:25 crc kubenswrapper[4805]: I1203 14:55:25.318339 4805 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-s2cxk"] Dec 03 14:55:27 crc kubenswrapper[4805]: I1203 14:55:27.180689 4805 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-s2cxk" podUID="dc26e67f-c323-4f89-b53c-d091c0a54e25" containerName="registry-server" containerID="cri-o://2aebf37755b90dfbbef4d6c5d258785e3c4ecea3627a9cdd07b97748cd2c2d75" gracePeriod=2 Dec 03 14:55:27 crc kubenswrapper[4805]: I1203 14:55:27.750086 4805 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-s2cxk" Dec 03 14:55:27 crc kubenswrapper[4805]: I1203 14:55:27.942601 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/dc26e67f-c323-4f89-b53c-d091c0a54e25-catalog-content\") pod \"dc26e67f-c323-4f89-b53c-d091c0a54e25\" (UID: \"dc26e67f-c323-4f89-b53c-d091c0a54e25\") " Dec 03 14:55:27 crc kubenswrapper[4805]: I1203 14:55:27.943090 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/dc26e67f-c323-4f89-b53c-d091c0a54e25-utilities\") pod \"dc26e67f-c323-4f89-b53c-d091c0a54e25\" (UID: \"dc26e67f-c323-4f89-b53c-d091c0a54e25\") " Dec 03 14:55:27 crc kubenswrapper[4805]: I1203 14:55:27.943136 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-99n2x\" (UniqueName: \"kubernetes.io/projected/dc26e67f-c323-4f89-b53c-d091c0a54e25-kube-api-access-99n2x\") pod \"dc26e67f-c323-4f89-b53c-d091c0a54e25\" (UID: \"dc26e67f-c323-4f89-b53c-d091c0a54e25\") " Dec 03 14:55:27 crc kubenswrapper[4805]: I1203 14:55:27.943795 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/dc26e67f-c323-4f89-b53c-d091c0a54e25-utilities" (OuterVolumeSpecName: "utilities") pod "dc26e67f-c323-4f89-b53c-d091c0a54e25" (UID: "dc26e67f-c323-4f89-b53c-d091c0a54e25"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 14:55:27 crc kubenswrapper[4805]: I1203 14:55:27.952044 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/dc26e67f-c323-4f89-b53c-d091c0a54e25-kube-api-access-99n2x" (OuterVolumeSpecName: "kube-api-access-99n2x") pod "dc26e67f-c323-4f89-b53c-d091c0a54e25" (UID: "dc26e67f-c323-4f89-b53c-d091c0a54e25"). InnerVolumeSpecName "kube-api-access-99n2x". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 14:55:27 crc kubenswrapper[4805]: I1203 14:55:27.964685 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/dc26e67f-c323-4f89-b53c-d091c0a54e25-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "dc26e67f-c323-4f89-b53c-d091c0a54e25" (UID: "dc26e67f-c323-4f89-b53c-d091c0a54e25"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 14:55:28 crc kubenswrapper[4805]: I1203 14:55:28.044915 4805 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/dc26e67f-c323-4f89-b53c-d091c0a54e25-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 03 14:55:28 crc kubenswrapper[4805]: I1203 14:55:28.044950 4805 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/dc26e67f-c323-4f89-b53c-d091c0a54e25-utilities\") on node \"crc\" DevicePath \"\"" Dec 03 14:55:28 crc kubenswrapper[4805]: I1203 14:55:28.044966 4805 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-99n2x\" (UniqueName: \"kubernetes.io/projected/dc26e67f-c323-4f89-b53c-d091c0a54e25-kube-api-access-99n2x\") on node \"crc\" DevicePath \"\"" Dec 03 14:55:28 crc kubenswrapper[4805]: I1203 14:55:28.198280 4805 generic.go:334] "Generic (PLEG): container finished" podID="dc26e67f-c323-4f89-b53c-d091c0a54e25" containerID="2aebf37755b90dfbbef4d6c5d258785e3c4ecea3627a9cdd07b97748cd2c2d75" exitCode=0 Dec 03 14:55:28 crc kubenswrapper[4805]: I1203 14:55:28.198343 4805 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-s2cxk" Dec 03 14:55:28 crc kubenswrapper[4805]: I1203 14:55:28.198328 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-s2cxk" event={"ID":"dc26e67f-c323-4f89-b53c-d091c0a54e25","Type":"ContainerDied","Data":"2aebf37755b90dfbbef4d6c5d258785e3c4ecea3627a9cdd07b97748cd2c2d75"} Dec 03 14:55:28 crc kubenswrapper[4805]: I1203 14:55:28.198514 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-s2cxk" event={"ID":"dc26e67f-c323-4f89-b53c-d091c0a54e25","Type":"ContainerDied","Data":"e029b5bffde13cd1138c46089e6948de2eaab5f7927033f012d362349f1da606"} Dec 03 14:55:28 crc kubenswrapper[4805]: I1203 14:55:28.198548 4805 scope.go:117] "RemoveContainer" containerID="2aebf37755b90dfbbef4d6c5d258785e3c4ecea3627a9cdd07b97748cd2c2d75" Dec 03 14:55:28 crc kubenswrapper[4805]: I1203 14:55:28.239307 4805 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-s2cxk"] Dec 03 14:55:28 crc kubenswrapper[4805]: I1203 14:55:28.248417 4805 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-s2cxk"] Dec 03 14:55:28 crc kubenswrapper[4805]: I1203 14:55:28.251960 4805 scope.go:117] "RemoveContainer" containerID="7b65a7fb32ae6359f625bd3dc417454dacf006a5251bdb3c5cfbb131802aa77b" Dec 03 14:55:28 crc kubenswrapper[4805]: I1203 14:55:28.275264 4805 scope.go:117] "RemoveContainer" containerID="5fdd05f82f6170f4fbfee6d9d82a427d4f5b136ec58b24671c155d7dfb3d8bc2" Dec 03 14:55:28 crc kubenswrapper[4805]: I1203 14:55:28.331374 4805 scope.go:117] "RemoveContainer" containerID="2aebf37755b90dfbbef4d6c5d258785e3c4ecea3627a9cdd07b97748cd2c2d75" Dec 03 14:55:28 crc kubenswrapper[4805]: E1203 14:55:28.331773 4805 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2aebf37755b90dfbbef4d6c5d258785e3c4ecea3627a9cdd07b97748cd2c2d75\": container with ID starting with 2aebf37755b90dfbbef4d6c5d258785e3c4ecea3627a9cdd07b97748cd2c2d75 not found: ID does not exist" containerID="2aebf37755b90dfbbef4d6c5d258785e3c4ecea3627a9cdd07b97748cd2c2d75" Dec 03 14:55:28 crc kubenswrapper[4805]: I1203 14:55:28.331805 4805 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2aebf37755b90dfbbef4d6c5d258785e3c4ecea3627a9cdd07b97748cd2c2d75"} err="failed to get container status \"2aebf37755b90dfbbef4d6c5d258785e3c4ecea3627a9cdd07b97748cd2c2d75\": rpc error: code = NotFound desc = could not find container \"2aebf37755b90dfbbef4d6c5d258785e3c4ecea3627a9cdd07b97748cd2c2d75\": container with ID starting with 2aebf37755b90dfbbef4d6c5d258785e3c4ecea3627a9cdd07b97748cd2c2d75 not found: ID does not exist" Dec 03 14:55:28 crc kubenswrapper[4805]: I1203 14:55:28.331826 4805 scope.go:117] "RemoveContainer" containerID="7b65a7fb32ae6359f625bd3dc417454dacf006a5251bdb3c5cfbb131802aa77b" Dec 03 14:55:28 crc kubenswrapper[4805]: E1203 14:55:28.332127 4805 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7b65a7fb32ae6359f625bd3dc417454dacf006a5251bdb3c5cfbb131802aa77b\": container with ID starting with 7b65a7fb32ae6359f625bd3dc417454dacf006a5251bdb3c5cfbb131802aa77b not found: ID does not exist" containerID="7b65a7fb32ae6359f625bd3dc417454dacf006a5251bdb3c5cfbb131802aa77b" Dec 03 14:55:28 crc kubenswrapper[4805]: I1203 14:55:28.332155 4805 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7b65a7fb32ae6359f625bd3dc417454dacf006a5251bdb3c5cfbb131802aa77b"} err="failed to get container status \"7b65a7fb32ae6359f625bd3dc417454dacf006a5251bdb3c5cfbb131802aa77b\": rpc error: code = NotFound desc = could not find container \"7b65a7fb32ae6359f625bd3dc417454dacf006a5251bdb3c5cfbb131802aa77b\": container with ID starting with 7b65a7fb32ae6359f625bd3dc417454dacf006a5251bdb3c5cfbb131802aa77b not found: ID does not exist" Dec 03 14:55:28 crc kubenswrapper[4805]: I1203 14:55:28.332176 4805 scope.go:117] "RemoveContainer" containerID="5fdd05f82f6170f4fbfee6d9d82a427d4f5b136ec58b24671c155d7dfb3d8bc2" Dec 03 14:55:28 crc kubenswrapper[4805]: E1203 14:55:28.332439 4805 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5fdd05f82f6170f4fbfee6d9d82a427d4f5b136ec58b24671c155d7dfb3d8bc2\": container with ID starting with 5fdd05f82f6170f4fbfee6d9d82a427d4f5b136ec58b24671c155d7dfb3d8bc2 not found: ID does not exist" containerID="5fdd05f82f6170f4fbfee6d9d82a427d4f5b136ec58b24671c155d7dfb3d8bc2" Dec 03 14:55:28 crc kubenswrapper[4805]: I1203 14:55:28.332461 4805 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5fdd05f82f6170f4fbfee6d9d82a427d4f5b136ec58b24671c155d7dfb3d8bc2"} err="failed to get container status \"5fdd05f82f6170f4fbfee6d9d82a427d4f5b136ec58b24671c155d7dfb3d8bc2\": rpc error: code = NotFound desc = could not find container \"5fdd05f82f6170f4fbfee6d9d82a427d4f5b136ec58b24671c155d7dfb3d8bc2\": container with ID starting with 5fdd05f82f6170f4fbfee6d9d82a427d4f5b136ec58b24671c155d7dfb3d8bc2 not found: ID does not exist" Dec 03 14:55:28 crc kubenswrapper[4805]: I1203 14:55:28.707966 4805 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="dc26e67f-c323-4f89-b53c-d091c0a54e25" path="/var/lib/kubelet/pods/dc26e67f-c323-4f89-b53c-d091c0a54e25/volumes" Dec 03 14:56:13 crc kubenswrapper[4805]: I1203 14:56:13.730667 4805 generic.go:334] "Generic (PLEG): container finished" podID="7e18e1db-7529-412b-bf01-17cb78e689b9" containerID="7616d8fd673eac4e5972a3f8f8dfaa218baf877f969a4730c5a8e8a90b3e0945" exitCode=0 Dec 03 14:56:13 crc kubenswrapper[4805]: I1203 14:56:13.730770 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-75h75" event={"ID":"7e18e1db-7529-412b-bf01-17cb78e689b9","Type":"ContainerDied","Data":"7616d8fd673eac4e5972a3f8f8dfaa218baf877f969a4730c5a8e8a90b3e0945"} Dec 03 14:56:15 crc kubenswrapper[4805]: I1203 14:56:15.156148 4805 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-75h75" Dec 03 14:56:15 crc kubenswrapper[4805]: I1203 14:56:15.333355 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/7e18e1db-7529-412b-bf01-17cb78e689b9-nova-cell1-compute-config-0\") pod \"7e18e1db-7529-412b-bf01-17cb78e689b9\" (UID: \"7e18e1db-7529-412b-bf01-17cb78e689b9\") " Dec 03 14:56:15 crc kubenswrapper[4805]: I1203 14:56:15.333452 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7e18e1db-7529-412b-bf01-17cb78e689b9-nova-combined-ca-bundle\") pod \"7e18e1db-7529-412b-bf01-17cb78e689b9\" (UID: \"7e18e1db-7529-412b-bf01-17cb78e689b9\") " Dec 03 14:56:15 crc kubenswrapper[4805]: I1203 14:56:15.333516 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/7e18e1db-7529-412b-bf01-17cb78e689b9-nova-migration-ssh-key-0\") pod \"7e18e1db-7529-412b-bf01-17cb78e689b9\" (UID: \"7e18e1db-7529-412b-bf01-17cb78e689b9\") " Dec 03 14:56:15 crc kubenswrapper[4805]: I1203 14:56:15.333579 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-extra-config-0\" (UniqueName: \"kubernetes.io/configmap/7e18e1db-7529-412b-bf01-17cb78e689b9-nova-extra-config-0\") pod \"7e18e1db-7529-412b-bf01-17cb78e689b9\" (UID: \"7e18e1db-7529-412b-bf01-17cb78e689b9\") " Dec 03 14:56:15 crc kubenswrapper[4805]: I1203 14:56:15.333623 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/7e18e1db-7529-412b-bf01-17cb78e689b9-nova-migration-ssh-key-1\") pod \"7e18e1db-7529-412b-bf01-17cb78e689b9\" (UID: \"7e18e1db-7529-412b-bf01-17cb78e689b9\") " Dec 03 14:56:15 crc kubenswrapper[4805]: I1203 14:56:15.333672 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/7e18e1db-7529-412b-bf01-17cb78e689b9-ssh-key\") pod \"7e18e1db-7529-412b-bf01-17cb78e689b9\" (UID: \"7e18e1db-7529-412b-bf01-17cb78e689b9\") " Dec 03 14:56:15 crc kubenswrapper[4805]: I1203 14:56:15.333732 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/7e18e1db-7529-412b-bf01-17cb78e689b9-inventory\") pod \"7e18e1db-7529-412b-bf01-17cb78e689b9\" (UID: \"7e18e1db-7529-412b-bf01-17cb78e689b9\") " Dec 03 14:56:15 crc kubenswrapper[4805]: I1203 14:56:15.333781 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dkm9v\" (UniqueName: \"kubernetes.io/projected/7e18e1db-7529-412b-bf01-17cb78e689b9-kube-api-access-dkm9v\") pod \"7e18e1db-7529-412b-bf01-17cb78e689b9\" (UID: \"7e18e1db-7529-412b-bf01-17cb78e689b9\") " Dec 03 14:56:15 crc kubenswrapper[4805]: I1203 14:56:15.333856 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/7e18e1db-7529-412b-bf01-17cb78e689b9-nova-cell1-compute-config-1\") pod \"7e18e1db-7529-412b-bf01-17cb78e689b9\" (UID: \"7e18e1db-7529-412b-bf01-17cb78e689b9\") " Dec 03 14:56:15 crc kubenswrapper[4805]: I1203 14:56:15.341580 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7e18e1db-7529-412b-bf01-17cb78e689b9-nova-combined-ca-bundle" (OuterVolumeSpecName: "nova-combined-ca-bundle") pod "7e18e1db-7529-412b-bf01-17cb78e689b9" (UID: "7e18e1db-7529-412b-bf01-17cb78e689b9"). InnerVolumeSpecName "nova-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 14:56:15 crc kubenswrapper[4805]: I1203 14:56:15.341690 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7e18e1db-7529-412b-bf01-17cb78e689b9-kube-api-access-dkm9v" (OuterVolumeSpecName: "kube-api-access-dkm9v") pod "7e18e1db-7529-412b-bf01-17cb78e689b9" (UID: "7e18e1db-7529-412b-bf01-17cb78e689b9"). InnerVolumeSpecName "kube-api-access-dkm9v". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 14:56:15 crc kubenswrapper[4805]: I1203 14:56:15.363553 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7e18e1db-7529-412b-bf01-17cb78e689b9-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "7e18e1db-7529-412b-bf01-17cb78e689b9" (UID: "7e18e1db-7529-412b-bf01-17cb78e689b9"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 14:56:15 crc kubenswrapper[4805]: I1203 14:56:15.364170 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7e18e1db-7529-412b-bf01-17cb78e689b9-nova-cell1-compute-config-1" (OuterVolumeSpecName: "nova-cell1-compute-config-1") pod "7e18e1db-7529-412b-bf01-17cb78e689b9" (UID: "7e18e1db-7529-412b-bf01-17cb78e689b9"). InnerVolumeSpecName "nova-cell1-compute-config-1". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 14:56:15 crc kubenswrapper[4805]: I1203 14:56:15.376716 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7e18e1db-7529-412b-bf01-17cb78e689b9-nova-cell1-compute-config-0" (OuterVolumeSpecName: "nova-cell1-compute-config-0") pod "7e18e1db-7529-412b-bf01-17cb78e689b9" (UID: "7e18e1db-7529-412b-bf01-17cb78e689b9"). InnerVolumeSpecName "nova-cell1-compute-config-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 14:56:15 crc kubenswrapper[4805]: I1203 14:56:15.377540 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7e18e1db-7529-412b-bf01-17cb78e689b9-inventory" (OuterVolumeSpecName: "inventory") pod "7e18e1db-7529-412b-bf01-17cb78e689b9" (UID: "7e18e1db-7529-412b-bf01-17cb78e689b9"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 14:56:15 crc kubenswrapper[4805]: I1203 14:56:15.383329 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7e18e1db-7529-412b-bf01-17cb78e689b9-nova-extra-config-0" (OuterVolumeSpecName: "nova-extra-config-0") pod "7e18e1db-7529-412b-bf01-17cb78e689b9" (UID: "7e18e1db-7529-412b-bf01-17cb78e689b9"). InnerVolumeSpecName "nova-extra-config-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 14:56:15 crc kubenswrapper[4805]: I1203 14:56:15.388922 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7e18e1db-7529-412b-bf01-17cb78e689b9-nova-migration-ssh-key-1" (OuterVolumeSpecName: "nova-migration-ssh-key-1") pod "7e18e1db-7529-412b-bf01-17cb78e689b9" (UID: "7e18e1db-7529-412b-bf01-17cb78e689b9"). InnerVolumeSpecName "nova-migration-ssh-key-1". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 14:56:15 crc kubenswrapper[4805]: I1203 14:56:15.393728 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7e18e1db-7529-412b-bf01-17cb78e689b9-nova-migration-ssh-key-0" (OuterVolumeSpecName: "nova-migration-ssh-key-0") pod "7e18e1db-7529-412b-bf01-17cb78e689b9" (UID: "7e18e1db-7529-412b-bf01-17cb78e689b9"). InnerVolumeSpecName "nova-migration-ssh-key-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 14:56:15 crc kubenswrapper[4805]: I1203 14:56:15.437304 4805 reconciler_common.go:293] "Volume detached for volume \"nova-extra-config-0\" (UniqueName: \"kubernetes.io/configmap/7e18e1db-7529-412b-bf01-17cb78e689b9-nova-extra-config-0\") on node \"crc\" DevicePath \"\"" Dec 03 14:56:15 crc kubenswrapper[4805]: I1203 14:56:15.437357 4805 reconciler_common.go:293] "Volume detached for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/7e18e1db-7529-412b-bf01-17cb78e689b9-nova-migration-ssh-key-1\") on node \"crc\" DevicePath \"\"" Dec 03 14:56:15 crc kubenswrapper[4805]: I1203 14:56:15.437371 4805 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/7e18e1db-7529-412b-bf01-17cb78e689b9-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 03 14:56:15 crc kubenswrapper[4805]: I1203 14:56:15.437388 4805 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/7e18e1db-7529-412b-bf01-17cb78e689b9-inventory\") on node \"crc\" DevicePath \"\"" Dec 03 14:56:15 crc kubenswrapper[4805]: I1203 14:56:15.437400 4805 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dkm9v\" (UniqueName: \"kubernetes.io/projected/7e18e1db-7529-412b-bf01-17cb78e689b9-kube-api-access-dkm9v\") on node \"crc\" DevicePath \"\"" Dec 03 14:56:15 crc kubenswrapper[4805]: I1203 14:56:15.437415 4805 reconciler_common.go:293] "Volume detached for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/7e18e1db-7529-412b-bf01-17cb78e689b9-nova-cell1-compute-config-1\") on node \"crc\" DevicePath \"\"" Dec 03 14:56:15 crc kubenswrapper[4805]: I1203 14:56:15.437431 4805 reconciler_common.go:293] "Volume detached for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/7e18e1db-7529-412b-bf01-17cb78e689b9-nova-cell1-compute-config-0\") on node \"crc\" DevicePath \"\"" Dec 03 14:56:15 crc kubenswrapper[4805]: I1203 14:56:15.437443 4805 reconciler_common.go:293] "Volume detached for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7e18e1db-7529-412b-bf01-17cb78e689b9-nova-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 14:56:15 crc kubenswrapper[4805]: I1203 14:56:15.437455 4805 reconciler_common.go:293] "Volume detached for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/7e18e1db-7529-412b-bf01-17cb78e689b9-nova-migration-ssh-key-0\") on node \"crc\" DevicePath \"\"" Dec 03 14:56:15 crc kubenswrapper[4805]: I1203 14:56:15.751810 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-75h75" event={"ID":"7e18e1db-7529-412b-bf01-17cb78e689b9","Type":"ContainerDied","Data":"3ce5c1d9cf5b363bb9ab6a2d4a4fa6f549a685868733dadbd47c509c002f9d87"} Dec 03 14:56:15 crc kubenswrapper[4805]: I1203 14:56:15.751868 4805 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="3ce5c1d9cf5b363bb9ab6a2d4a4fa6f549a685868733dadbd47c509c002f9d87" Dec 03 14:56:15 crc kubenswrapper[4805]: I1203 14:56:15.751899 4805 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-75h75" Dec 03 14:56:15 crc kubenswrapper[4805]: I1203 14:56:15.880517 4805 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/telemetry-edpm-deployment-openstack-edpm-ipam-ctxx4"] Dec 03 14:56:15 crc kubenswrapper[4805]: E1203 14:56:15.880882 4805 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="dc26e67f-c323-4f89-b53c-d091c0a54e25" containerName="extract-content" Dec 03 14:56:15 crc kubenswrapper[4805]: I1203 14:56:15.880893 4805 state_mem.go:107] "Deleted CPUSet assignment" podUID="dc26e67f-c323-4f89-b53c-d091c0a54e25" containerName="extract-content" Dec 03 14:56:15 crc kubenswrapper[4805]: E1203 14:56:15.880907 4805 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="dc26e67f-c323-4f89-b53c-d091c0a54e25" containerName="registry-server" Dec 03 14:56:15 crc kubenswrapper[4805]: I1203 14:56:15.880913 4805 state_mem.go:107] "Deleted CPUSet assignment" podUID="dc26e67f-c323-4f89-b53c-d091c0a54e25" containerName="registry-server" Dec 03 14:56:15 crc kubenswrapper[4805]: E1203 14:56:15.880925 4805 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="dc26e67f-c323-4f89-b53c-d091c0a54e25" containerName="extract-utilities" Dec 03 14:56:15 crc kubenswrapper[4805]: I1203 14:56:15.880931 4805 state_mem.go:107] "Deleted CPUSet assignment" podUID="dc26e67f-c323-4f89-b53c-d091c0a54e25" containerName="extract-utilities" Dec 03 14:56:15 crc kubenswrapper[4805]: E1203 14:56:15.880948 4805 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7e18e1db-7529-412b-bf01-17cb78e689b9" containerName="nova-edpm-deployment-openstack-edpm-ipam" Dec 03 14:56:15 crc kubenswrapper[4805]: I1203 14:56:15.880954 4805 state_mem.go:107] "Deleted CPUSet assignment" podUID="7e18e1db-7529-412b-bf01-17cb78e689b9" containerName="nova-edpm-deployment-openstack-edpm-ipam" Dec 03 14:56:15 crc kubenswrapper[4805]: I1203 14:56:15.881129 4805 memory_manager.go:354] "RemoveStaleState removing state" podUID="dc26e67f-c323-4f89-b53c-d091c0a54e25" containerName="registry-server" Dec 03 14:56:15 crc kubenswrapper[4805]: I1203 14:56:15.881143 4805 memory_manager.go:354] "RemoveStaleState removing state" podUID="7e18e1db-7529-412b-bf01-17cb78e689b9" containerName="nova-edpm-deployment-openstack-edpm-ipam" Dec 03 14:56:15 crc kubenswrapper[4805]: I1203 14:56:15.881708 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-ctxx4" Dec 03 14:56:15 crc kubenswrapper[4805]: I1203 14:56:15.886979 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-n7jt4" Dec 03 14:56:15 crc kubenswrapper[4805]: I1203 14:56:15.887518 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Dec 03 14:56:15 crc kubenswrapper[4805]: I1203 14:56:15.887951 4805 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 03 14:56:15 crc kubenswrapper[4805]: I1203 14:56:15.888095 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Dec 03 14:56:15 crc kubenswrapper[4805]: I1203 14:56:15.888162 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-compute-config-data" Dec 03 14:56:15 crc kubenswrapper[4805]: I1203 14:56:15.914874 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/telemetry-edpm-deployment-openstack-edpm-ipam-ctxx4"] Dec 03 14:56:16 crc kubenswrapper[4805]: I1203 14:56:16.049714 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-compute-config-data-0\" (UniqueName: \"kubernetes.io/secret/4c03480a-7dce-48dd-9e5d-1aea5c540a72-ceilometer-compute-config-data-0\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-ctxx4\" (UID: \"4c03480a-7dce-48dd-9e5d-1aea5c540a72\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-ctxx4" Dec 03 14:56:16 crc kubenswrapper[4805]: I1203 14:56:16.050060 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-compute-config-data-1\" (UniqueName: \"kubernetes.io/secret/4c03480a-7dce-48dd-9e5d-1aea5c540a72-ceilometer-compute-config-data-1\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-ctxx4\" (UID: \"4c03480a-7dce-48dd-9e5d-1aea5c540a72\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-ctxx4" Dec 03 14:56:16 crc kubenswrapper[4805]: I1203 14:56:16.050293 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-compute-config-data-2\" (UniqueName: \"kubernetes.io/secret/4c03480a-7dce-48dd-9e5d-1aea5c540a72-ceilometer-compute-config-data-2\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-ctxx4\" (UID: \"4c03480a-7dce-48dd-9e5d-1aea5c540a72\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-ctxx4" Dec 03 14:56:16 crc kubenswrapper[4805]: I1203 14:56:16.050368 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-srx8j\" (UniqueName: \"kubernetes.io/projected/4c03480a-7dce-48dd-9e5d-1aea5c540a72-kube-api-access-srx8j\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-ctxx4\" (UID: \"4c03480a-7dce-48dd-9e5d-1aea5c540a72\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-ctxx4" Dec 03 14:56:16 crc kubenswrapper[4805]: I1203 14:56:16.050507 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/4c03480a-7dce-48dd-9e5d-1aea5c540a72-inventory\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-ctxx4\" (UID: \"4c03480a-7dce-48dd-9e5d-1aea5c540a72\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-ctxx4" Dec 03 14:56:16 crc kubenswrapper[4805]: I1203 14:56:16.050658 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4c03480a-7dce-48dd-9e5d-1aea5c540a72-telemetry-combined-ca-bundle\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-ctxx4\" (UID: \"4c03480a-7dce-48dd-9e5d-1aea5c540a72\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-ctxx4" Dec 03 14:56:16 crc kubenswrapper[4805]: I1203 14:56:16.050965 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/4c03480a-7dce-48dd-9e5d-1aea5c540a72-ssh-key\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-ctxx4\" (UID: \"4c03480a-7dce-48dd-9e5d-1aea5c540a72\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-ctxx4" Dec 03 14:56:16 crc kubenswrapper[4805]: I1203 14:56:16.153576 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-compute-config-data-1\" (UniqueName: \"kubernetes.io/secret/4c03480a-7dce-48dd-9e5d-1aea5c540a72-ceilometer-compute-config-data-1\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-ctxx4\" (UID: \"4c03480a-7dce-48dd-9e5d-1aea5c540a72\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-ctxx4" Dec 03 14:56:16 crc kubenswrapper[4805]: I1203 14:56:16.154269 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-compute-config-data-2\" (UniqueName: \"kubernetes.io/secret/4c03480a-7dce-48dd-9e5d-1aea5c540a72-ceilometer-compute-config-data-2\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-ctxx4\" (UID: \"4c03480a-7dce-48dd-9e5d-1aea5c540a72\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-ctxx4" Dec 03 14:56:16 crc kubenswrapper[4805]: I1203 14:56:16.154941 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-srx8j\" (UniqueName: \"kubernetes.io/projected/4c03480a-7dce-48dd-9e5d-1aea5c540a72-kube-api-access-srx8j\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-ctxx4\" (UID: \"4c03480a-7dce-48dd-9e5d-1aea5c540a72\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-ctxx4" Dec 03 14:56:16 crc kubenswrapper[4805]: I1203 14:56:16.155527 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/4c03480a-7dce-48dd-9e5d-1aea5c540a72-inventory\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-ctxx4\" (UID: \"4c03480a-7dce-48dd-9e5d-1aea5c540a72\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-ctxx4" Dec 03 14:56:16 crc kubenswrapper[4805]: I1203 14:56:16.156047 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4c03480a-7dce-48dd-9e5d-1aea5c540a72-telemetry-combined-ca-bundle\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-ctxx4\" (UID: \"4c03480a-7dce-48dd-9e5d-1aea5c540a72\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-ctxx4" Dec 03 14:56:16 crc kubenswrapper[4805]: I1203 14:56:16.156362 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/4c03480a-7dce-48dd-9e5d-1aea5c540a72-ssh-key\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-ctxx4\" (UID: \"4c03480a-7dce-48dd-9e5d-1aea5c540a72\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-ctxx4" Dec 03 14:56:16 crc kubenswrapper[4805]: I1203 14:56:16.158455 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-compute-config-data-0\" (UniqueName: \"kubernetes.io/secret/4c03480a-7dce-48dd-9e5d-1aea5c540a72-ceilometer-compute-config-data-0\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-ctxx4\" (UID: \"4c03480a-7dce-48dd-9e5d-1aea5c540a72\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-ctxx4" Dec 03 14:56:16 crc kubenswrapper[4805]: I1203 14:56:16.160991 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-compute-config-data-2\" (UniqueName: \"kubernetes.io/secret/4c03480a-7dce-48dd-9e5d-1aea5c540a72-ceilometer-compute-config-data-2\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-ctxx4\" (UID: \"4c03480a-7dce-48dd-9e5d-1aea5c540a72\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-ctxx4" Dec 03 14:56:16 crc kubenswrapper[4805]: I1203 14:56:16.162559 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/4c03480a-7dce-48dd-9e5d-1aea5c540a72-inventory\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-ctxx4\" (UID: \"4c03480a-7dce-48dd-9e5d-1aea5c540a72\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-ctxx4" Dec 03 14:56:16 crc kubenswrapper[4805]: I1203 14:56:16.163467 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/4c03480a-7dce-48dd-9e5d-1aea5c540a72-ssh-key\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-ctxx4\" (UID: \"4c03480a-7dce-48dd-9e5d-1aea5c540a72\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-ctxx4" Dec 03 14:56:16 crc kubenswrapper[4805]: I1203 14:56:16.164131 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-compute-config-data-1\" (UniqueName: \"kubernetes.io/secret/4c03480a-7dce-48dd-9e5d-1aea5c540a72-ceilometer-compute-config-data-1\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-ctxx4\" (UID: \"4c03480a-7dce-48dd-9e5d-1aea5c540a72\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-ctxx4" Dec 03 14:56:16 crc kubenswrapper[4805]: I1203 14:56:16.164502 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4c03480a-7dce-48dd-9e5d-1aea5c540a72-telemetry-combined-ca-bundle\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-ctxx4\" (UID: \"4c03480a-7dce-48dd-9e5d-1aea5c540a72\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-ctxx4" Dec 03 14:56:16 crc kubenswrapper[4805]: I1203 14:56:16.166529 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-compute-config-data-0\" (UniqueName: \"kubernetes.io/secret/4c03480a-7dce-48dd-9e5d-1aea5c540a72-ceilometer-compute-config-data-0\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-ctxx4\" (UID: \"4c03480a-7dce-48dd-9e5d-1aea5c540a72\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-ctxx4" Dec 03 14:56:16 crc kubenswrapper[4805]: I1203 14:56:16.188790 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-srx8j\" (UniqueName: \"kubernetes.io/projected/4c03480a-7dce-48dd-9e5d-1aea5c540a72-kube-api-access-srx8j\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-ctxx4\" (UID: \"4c03480a-7dce-48dd-9e5d-1aea5c540a72\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-ctxx4" Dec 03 14:56:16 crc kubenswrapper[4805]: I1203 14:56:16.209375 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-ctxx4" Dec 03 14:56:16 crc kubenswrapper[4805]: I1203 14:56:16.661541 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/telemetry-edpm-deployment-openstack-edpm-ipam-ctxx4"] Dec 03 14:56:16 crc kubenswrapper[4805]: W1203 14:56:16.664341 4805 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod4c03480a_7dce_48dd_9e5d_1aea5c540a72.slice/crio-a652664dcf298145a97c05dd5025032308cb44d6fb5044f2c4a7d61809b04b27 WatchSource:0}: Error finding container a652664dcf298145a97c05dd5025032308cb44d6fb5044f2c4a7d61809b04b27: Status 404 returned error can't find the container with id a652664dcf298145a97c05dd5025032308cb44d6fb5044f2c4a7d61809b04b27 Dec 03 14:56:16 crc kubenswrapper[4805]: I1203 14:56:16.760606 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-ctxx4" event={"ID":"4c03480a-7dce-48dd-9e5d-1aea5c540a72","Type":"ContainerStarted","Data":"a652664dcf298145a97c05dd5025032308cb44d6fb5044f2c4a7d61809b04b27"} Dec 03 14:56:17 crc kubenswrapper[4805]: I1203 14:56:17.774267 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-ctxx4" event={"ID":"4c03480a-7dce-48dd-9e5d-1aea5c540a72","Type":"ContainerStarted","Data":"dbfa93f8a8ba3e02449235d4cccb1374046d9b0b1116baceedd2856b014fe16c"} Dec 03 14:56:17 crc kubenswrapper[4805]: I1203 14:56:17.814158 4805 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-ctxx4" podStartSLOduration=2.363034421 podStartE2EDuration="2.8141404s" podCreationTimestamp="2025-12-03 14:56:15 +0000 UTC" firstStartedPulling="2025-12-03 14:56:16.66771722 +0000 UTC m=+2806.330634143" lastFinishedPulling="2025-12-03 14:56:17.118823199 +0000 UTC m=+2806.781740122" observedRunningTime="2025-12-03 14:56:17.802070514 +0000 UTC m=+2807.464987427" watchObservedRunningTime="2025-12-03 14:56:17.8141404 +0000 UTC m=+2807.477057323" Dec 03 14:57:13 crc kubenswrapper[4805]: I1203 14:57:13.916853 4805 patch_prober.go:28] interesting pod/machine-config-daemon-gskh4 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 03 14:57:13 crc kubenswrapper[4805]: I1203 14:57:13.917345 4805 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-gskh4" podUID="6098937f-e3f6-45e8-a647-4994a79cd711" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 03 14:57:43 crc kubenswrapper[4805]: I1203 14:57:43.916901 4805 patch_prober.go:28] interesting pod/machine-config-daemon-gskh4 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 03 14:57:43 crc kubenswrapper[4805]: I1203 14:57:43.919035 4805 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-gskh4" podUID="6098937f-e3f6-45e8-a647-4994a79cd711" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 03 14:58:13 crc kubenswrapper[4805]: I1203 14:58:13.917132 4805 patch_prober.go:28] interesting pod/machine-config-daemon-gskh4 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 03 14:58:13 crc kubenswrapper[4805]: I1203 14:58:13.917955 4805 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-gskh4" podUID="6098937f-e3f6-45e8-a647-4994a79cd711" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 03 14:58:13 crc kubenswrapper[4805]: I1203 14:58:13.918009 4805 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-gskh4" Dec 03 14:58:13 crc kubenswrapper[4805]: I1203 14:58:13.918831 4805 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"2e3f74f03f0ba2719087d9ca75ea7670c41a3ad116d1af23fbb1aeb3c76a97d3"} pod="openshift-machine-config-operator/machine-config-daemon-gskh4" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 03 14:58:13 crc kubenswrapper[4805]: I1203 14:58:13.918907 4805 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-gskh4" podUID="6098937f-e3f6-45e8-a647-4994a79cd711" containerName="machine-config-daemon" containerID="cri-o://2e3f74f03f0ba2719087d9ca75ea7670c41a3ad116d1af23fbb1aeb3c76a97d3" gracePeriod=600 Dec 03 14:58:15 crc kubenswrapper[4805]: I1203 14:58:15.065959 4805 generic.go:334] "Generic (PLEG): container finished" podID="6098937f-e3f6-45e8-a647-4994a79cd711" containerID="2e3f74f03f0ba2719087d9ca75ea7670c41a3ad116d1af23fbb1aeb3c76a97d3" exitCode=0 Dec 03 14:58:15 crc kubenswrapper[4805]: I1203 14:58:15.066020 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-gskh4" event={"ID":"6098937f-e3f6-45e8-a647-4994a79cd711","Type":"ContainerDied","Data":"2e3f74f03f0ba2719087d9ca75ea7670c41a3ad116d1af23fbb1aeb3c76a97d3"} Dec 03 14:58:15 crc kubenswrapper[4805]: I1203 14:58:15.066743 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-gskh4" event={"ID":"6098937f-e3f6-45e8-a647-4994a79cd711","Type":"ContainerStarted","Data":"13e0da4d7fc8f4085be601d845d688938c16ea7941e52ca511e39a37d6e0ba4e"} Dec 03 14:58:15 crc kubenswrapper[4805]: I1203 14:58:15.066774 4805 scope.go:117] "RemoveContainer" containerID="fec9f6923538948ef3e55903bb8e108d1f94095a493c6b5117232e21f41ab624" Dec 03 14:58:46 crc kubenswrapper[4805]: I1203 14:58:46.442427 4805 generic.go:334] "Generic (PLEG): container finished" podID="4c03480a-7dce-48dd-9e5d-1aea5c540a72" containerID="dbfa93f8a8ba3e02449235d4cccb1374046d9b0b1116baceedd2856b014fe16c" exitCode=0 Dec 03 14:58:46 crc kubenswrapper[4805]: I1203 14:58:46.442515 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-ctxx4" event={"ID":"4c03480a-7dce-48dd-9e5d-1aea5c540a72","Type":"ContainerDied","Data":"dbfa93f8a8ba3e02449235d4cccb1374046d9b0b1116baceedd2856b014fe16c"} Dec 03 14:58:47 crc kubenswrapper[4805]: I1203 14:58:47.990831 4805 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-ctxx4" Dec 03 14:58:48 crc kubenswrapper[4805]: I1203 14:58:48.183156 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-compute-config-data-1\" (UniqueName: \"kubernetes.io/secret/4c03480a-7dce-48dd-9e5d-1aea5c540a72-ceilometer-compute-config-data-1\") pod \"4c03480a-7dce-48dd-9e5d-1aea5c540a72\" (UID: \"4c03480a-7dce-48dd-9e5d-1aea5c540a72\") " Dec 03 14:58:48 crc kubenswrapper[4805]: I1203 14:58:48.183246 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4c03480a-7dce-48dd-9e5d-1aea5c540a72-telemetry-combined-ca-bundle\") pod \"4c03480a-7dce-48dd-9e5d-1aea5c540a72\" (UID: \"4c03480a-7dce-48dd-9e5d-1aea5c540a72\") " Dec 03 14:58:48 crc kubenswrapper[4805]: I1203 14:58:48.183312 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-compute-config-data-0\" (UniqueName: \"kubernetes.io/secret/4c03480a-7dce-48dd-9e5d-1aea5c540a72-ceilometer-compute-config-data-0\") pod \"4c03480a-7dce-48dd-9e5d-1aea5c540a72\" (UID: \"4c03480a-7dce-48dd-9e5d-1aea5c540a72\") " Dec 03 14:58:48 crc kubenswrapper[4805]: I1203 14:58:48.183366 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/4c03480a-7dce-48dd-9e5d-1aea5c540a72-ssh-key\") pod \"4c03480a-7dce-48dd-9e5d-1aea5c540a72\" (UID: \"4c03480a-7dce-48dd-9e5d-1aea5c540a72\") " Dec 03 14:58:48 crc kubenswrapper[4805]: I1203 14:58:48.183597 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-compute-config-data-2\" (UniqueName: \"kubernetes.io/secret/4c03480a-7dce-48dd-9e5d-1aea5c540a72-ceilometer-compute-config-data-2\") pod \"4c03480a-7dce-48dd-9e5d-1aea5c540a72\" (UID: \"4c03480a-7dce-48dd-9e5d-1aea5c540a72\") " Dec 03 14:58:48 crc kubenswrapper[4805]: I1203 14:58:48.183927 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/4c03480a-7dce-48dd-9e5d-1aea5c540a72-inventory\") pod \"4c03480a-7dce-48dd-9e5d-1aea5c540a72\" (UID: \"4c03480a-7dce-48dd-9e5d-1aea5c540a72\") " Dec 03 14:58:48 crc kubenswrapper[4805]: I1203 14:58:48.184572 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-srx8j\" (UniqueName: \"kubernetes.io/projected/4c03480a-7dce-48dd-9e5d-1aea5c540a72-kube-api-access-srx8j\") pod \"4c03480a-7dce-48dd-9e5d-1aea5c540a72\" (UID: \"4c03480a-7dce-48dd-9e5d-1aea5c540a72\") " Dec 03 14:58:48 crc kubenswrapper[4805]: I1203 14:58:48.193098 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4c03480a-7dce-48dd-9e5d-1aea5c540a72-kube-api-access-srx8j" (OuterVolumeSpecName: "kube-api-access-srx8j") pod "4c03480a-7dce-48dd-9e5d-1aea5c540a72" (UID: "4c03480a-7dce-48dd-9e5d-1aea5c540a72"). InnerVolumeSpecName "kube-api-access-srx8j". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 14:58:48 crc kubenswrapper[4805]: I1203 14:58:48.193120 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4c03480a-7dce-48dd-9e5d-1aea5c540a72-telemetry-combined-ca-bundle" (OuterVolumeSpecName: "telemetry-combined-ca-bundle") pod "4c03480a-7dce-48dd-9e5d-1aea5c540a72" (UID: "4c03480a-7dce-48dd-9e5d-1aea5c540a72"). InnerVolumeSpecName "telemetry-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 14:58:48 crc kubenswrapper[4805]: I1203 14:58:48.216566 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4c03480a-7dce-48dd-9e5d-1aea5c540a72-ceilometer-compute-config-data-1" (OuterVolumeSpecName: "ceilometer-compute-config-data-1") pod "4c03480a-7dce-48dd-9e5d-1aea5c540a72" (UID: "4c03480a-7dce-48dd-9e5d-1aea5c540a72"). InnerVolumeSpecName "ceilometer-compute-config-data-1". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 14:58:48 crc kubenswrapper[4805]: I1203 14:58:48.224002 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4c03480a-7dce-48dd-9e5d-1aea5c540a72-ceilometer-compute-config-data-2" (OuterVolumeSpecName: "ceilometer-compute-config-data-2") pod "4c03480a-7dce-48dd-9e5d-1aea5c540a72" (UID: "4c03480a-7dce-48dd-9e5d-1aea5c540a72"). InnerVolumeSpecName "ceilometer-compute-config-data-2". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 14:58:48 crc kubenswrapper[4805]: I1203 14:58:48.224467 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4c03480a-7dce-48dd-9e5d-1aea5c540a72-ceilometer-compute-config-data-0" (OuterVolumeSpecName: "ceilometer-compute-config-data-0") pod "4c03480a-7dce-48dd-9e5d-1aea5c540a72" (UID: "4c03480a-7dce-48dd-9e5d-1aea5c540a72"). InnerVolumeSpecName "ceilometer-compute-config-data-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 14:58:48 crc kubenswrapper[4805]: I1203 14:58:48.224994 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4c03480a-7dce-48dd-9e5d-1aea5c540a72-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "4c03480a-7dce-48dd-9e5d-1aea5c540a72" (UID: "4c03480a-7dce-48dd-9e5d-1aea5c540a72"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 14:58:48 crc kubenswrapper[4805]: I1203 14:58:48.239009 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4c03480a-7dce-48dd-9e5d-1aea5c540a72-inventory" (OuterVolumeSpecName: "inventory") pod "4c03480a-7dce-48dd-9e5d-1aea5c540a72" (UID: "4c03480a-7dce-48dd-9e5d-1aea5c540a72"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 14:58:48 crc kubenswrapper[4805]: I1203 14:58:48.288303 4805 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/4c03480a-7dce-48dd-9e5d-1aea5c540a72-inventory\") on node \"crc\" DevicePath \"\"" Dec 03 14:58:48 crc kubenswrapper[4805]: I1203 14:58:48.288346 4805 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-srx8j\" (UniqueName: \"kubernetes.io/projected/4c03480a-7dce-48dd-9e5d-1aea5c540a72-kube-api-access-srx8j\") on node \"crc\" DevicePath \"\"" Dec 03 14:58:48 crc kubenswrapper[4805]: I1203 14:58:48.288364 4805 reconciler_common.go:293] "Volume detached for volume \"ceilometer-compute-config-data-1\" (UniqueName: \"kubernetes.io/secret/4c03480a-7dce-48dd-9e5d-1aea5c540a72-ceilometer-compute-config-data-1\") on node \"crc\" DevicePath \"\"" Dec 03 14:58:48 crc kubenswrapper[4805]: I1203 14:58:48.288393 4805 reconciler_common.go:293] "Volume detached for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4c03480a-7dce-48dd-9e5d-1aea5c540a72-telemetry-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 14:58:48 crc kubenswrapper[4805]: I1203 14:58:48.288406 4805 reconciler_common.go:293] "Volume detached for volume \"ceilometer-compute-config-data-0\" (UniqueName: \"kubernetes.io/secret/4c03480a-7dce-48dd-9e5d-1aea5c540a72-ceilometer-compute-config-data-0\") on node \"crc\" DevicePath \"\"" Dec 03 14:58:48 crc kubenswrapper[4805]: I1203 14:58:48.288421 4805 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/4c03480a-7dce-48dd-9e5d-1aea5c540a72-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 03 14:58:48 crc kubenswrapper[4805]: I1203 14:58:48.288433 4805 reconciler_common.go:293] "Volume detached for volume \"ceilometer-compute-config-data-2\" (UniqueName: \"kubernetes.io/secret/4c03480a-7dce-48dd-9e5d-1aea5c540a72-ceilometer-compute-config-data-2\") on node \"crc\" DevicePath \"\"" Dec 03 14:58:48 crc kubenswrapper[4805]: I1203 14:58:48.543988 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-ctxx4" event={"ID":"4c03480a-7dce-48dd-9e5d-1aea5c540a72","Type":"ContainerDied","Data":"a652664dcf298145a97c05dd5025032308cb44d6fb5044f2c4a7d61809b04b27"} Dec 03 14:58:48 crc kubenswrapper[4805]: I1203 14:58:48.544051 4805 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="a652664dcf298145a97c05dd5025032308cb44d6fb5044f2c4a7d61809b04b27" Dec 03 14:58:48 crc kubenswrapper[4805]: I1203 14:58:48.544225 4805 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-ctxx4" Dec 03 15:00:00 crc kubenswrapper[4805]: I1203 15:00:00.156587 4805 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29412900-pjrk4"] Dec 03 15:00:00 crc kubenswrapper[4805]: E1203 15:00:00.157870 4805 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4c03480a-7dce-48dd-9e5d-1aea5c540a72" containerName="telemetry-edpm-deployment-openstack-edpm-ipam" Dec 03 15:00:00 crc kubenswrapper[4805]: I1203 15:00:00.157891 4805 state_mem.go:107] "Deleted CPUSet assignment" podUID="4c03480a-7dce-48dd-9e5d-1aea5c540a72" containerName="telemetry-edpm-deployment-openstack-edpm-ipam" Dec 03 15:00:00 crc kubenswrapper[4805]: I1203 15:00:00.158172 4805 memory_manager.go:354] "RemoveStaleState removing state" podUID="4c03480a-7dce-48dd-9e5d-1aea5c540a72" containerName="telemetry-edpm-deployment-openstack-edpm-ipam" Dec 03 15:00:00 crc kubenswrapper[4805]: I1203 15:00:00.158998 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29412900-pjrk4" Dec 03 15:00:00 crc kubenswrapper[4805]: I1203 15:00:00.161855 4805 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Dec 03 15:00:00 crc kubenswrapper[4805]: I1203 15:00:00.161950 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Dec 03 15:00:00 crc kubenswrapper[4805]: I1203 15:00:00.167167 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29412900-pjrk4"] Dec 03 15:00:00 crc kubenswrapper[4805]: I1203 15:00:00.175147 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/055eaef9-82c4-4ad0-ab32-9eeca92fecb4-config-volume\") pod \"collect-profiles-29412900-pjrk4\" (UID: \"055eaef9-82c4-4ad0-ab32-9eeca92fecb4\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29412900-pjrk4" Dec 03 15:00:00 crc kubenswrapper[4805]: I1203 15:00:00.175237 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/055eaef9-82c4-4ad0-ab32-9eeca92fecb4-secret-volume\") pod \"collect-profiles-29412900-pjrk4\" (UID: \"055eaef9-82c4-4ad0-ab32-9eeca92fecb4\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29412900-pjrk4" Dec 03 15:00:00 crc kubenswrapper[4805]: I1203 15:00:00.175281 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vhxcr\" (UniqueName: \"kubernetes.io/projected/055eaef9-82c4-4ad0-ab32-9eeca92fecb4-kube-api-access-vhxcr\") pod \"collect-profiles-29412900-pjrk4\" (UID: \"055eaef9-82c4-4ad0-ab32-9eeca92fecb4\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29412900-pjrk4" Dec 03 15:00:00 crc kubenswrapper[4805]: I1203 15:00:00.277181 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/055eaef9-82c4-4ad0-ab32-9eeca92fecb4-config-volume\") pod \"collect-profiles-29412900-pjrk4\" (UID: \"055eaef9-82c4-4ad0-ab32-9eeca92fecb4\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29412900-pjrk4" Dec 03 15:00:00 crc kubenswrapper[4805]: I1203 15:00:00.277285 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/055eaef9-82c4-4ad0-ab32-9eeca92fecb4-secret-volume\") pod \"collect-profiles-29412900-pjrk4\" (UID: \"055eaef9-82c4-4ad0-ab32-9eeca92fecb4\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29412900-pjrk4" Dec 03 15:00:00 crc kubenswrapper[4805]: I1203 15:00:00.277331 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vhxcr\" (UniqueName: \"kubernetes.io/projected/055eaef9-82c4-4ad0-ab32-9eeca92fecb4-kube-api-access-vhxcr\") pod \"collect-profiles-29412900-pjrk4\" (UID: \"055eaef9-82c4-4ad0-ab32-9eeca92fecb4\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29412900-pjrk4" Dec 03 15:00:00 crc kubenswrapper[4805]: I1203 15:00:00.278192 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/055eaef9-82c4-4ad0-ab32-9eeca92fecb4-config-volume\") pod \"collect-profiles-29412900-pjrk4\" (UID: \"055eaef9-82c4-4ad0-ab32-9eeca92fecb4\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29412900-pjrk4" Dec 03 15:00:00 crc kubenswrapper[4805]: I1203 15:00:00.286047 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/055eaef9-82c4-4ad0-ab32-9eeca92fecb4-secret-volume\") pod \"collect-profiles-29412900-pjrk4\" (UID: \"055eaef9-82c4-4ad0-ab32-9eeca92fecb4\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29412900-pjrk4" Dec 03 15:00:00 crc kubenswrapper[4805]: I1203 15:00:00.308430 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vhxcr\" (UniqueName: \"kubernetes.io/projected/055eaef9-82c4-4ad0-ab32-9eeca92fecb4-kube-api-access-vhxcr\") pod \"collect-profiles-29412900-pjrk4\" (UID: \"055eaef9-82c4-4ad0-ab32-9eeca92fecb4\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29412900-pjrk4" Dec 03 15:00:00 crc kubenswrapper[4805]: I1203 15:00:00.520567 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29412900-pjrk4" Dec 03 15:00:01 crc kubenswrapper[4805]: I1203 15:00:01.005223 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29412900-pjrk4"] Dec 03 15:00:01 crc kubenswrapper[4805]: W1203 15:00:01.010388 4805 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod055eaef9_82c4_4ad0_ab32_9eeca92fecb4.slice/crio-8f89ae60934965d9f7f7469a1b23af4bc1c94e8fed7a7bee383e24ce92a34029 WatchSource:0}: Error finding container 8f89ae60934965d9f7f7469a1b23af4bc1c94e8fed7a7bee383e24ce92a34029: Status 404 returned error can't find the container with id 8f89ae60934965d9f7f7469a1b23af4bc1c94e8fed7a7bee383e24ce92a34029 Dec 03 15:00:01 crc kubenswrapper[4805]: I1203 15:00:01.333986 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29412900-pjrk4" event={"ID":"055eaef9-82c4-4ad0-ab32-9eeca92fecb4","Type":"ContainerStarted","Data":"3fbbd682b518332ba42ec97240b127bc0757c7b09ae7e98fbc614d845414251a"} Dec 03 15:00:01 crc kubenswrapper[4805]: I1203 15:00:01.334496 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29412900-pjrk4" event={"ID":"055eaef9-82c4-4ad0-ab32-9eeca92fecb4","Type":"ContainerStarted","Data":"8f89ae60934965d9f7f7469a1b23af4bc1c94e8fed7a7bee383e24ce92a34029"} Dec 03 15:00:01 crc kubenswrapper[4805]: I1203 15:00:01.365903 4805 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/collect-profiles-29412900-pjrk4" podStartSLOduration=1.365878764 podStartE2EDuration="1.365878764s" podCreationTimestamp="2025-12-03 15:00:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 15:00:01.361343332 +0000 UTC m=+3031.024260265" watchObservedRunningTime="2025-12-03 15:00:01.365878764 +0000 UTC m=+3031.028795687" Dec 03 15:00:02 crc kubenswrapper[4805]: I1203 15:00:02.347620 4805 generic.go:334] "Generic (PLEG): container finished" podID="055eaef9-82c4-4ad0-ab32-9eeca92fecb4" containerID="3fbbd682b518332ba42ec97240b127bc0757c7b09ae7e98fbc614d845414251a" exitCode=0 Dec 03 15:00:02 crc kubenswrapper[4805]: I1203 15:00:02.347742 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29412900-pjrk4" event={"ID":"055eaef9-82c4-4ad0-ab32-9eeca92fecb4","Type":"ContainerDied","Data":"3fbbd682b518332ba42ec97240b127bc0757c7b09ae7e98fbc614d845414251a"} Dec 03 15:00:03 crc kubenswrapper[4805]: I1203 15:00:03.721113 4805 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29412900-pjrk4" Dec 03 15:00:03 crc kubenswrapper[4805]: I1203 15:00:03.747520 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/055eaef9-82c4-4ad0-ab32-9eeca92fecb4-secret-volume\") pod \"055eaef9-82c4-4ad0-ab32-9eeca92fecb4\" (UID: \"055eaef9-82c4-4ad0-ab32-9eeca92fecb4\") " Dec 03 15:00:03 crc kubenswrapper[4805]: I1203 15:00:03.747642 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vhxcr\" (UniqueName: \"kubernetes.io/projected/055eaef9-82c4-4ad0-ab32-9eeca92fecb4-kube-api-access-vhxcr\") pod \"055eaef9-82c4-4ad0-ab32-9eeca92fecb4\" (UID: \"055eaef9-82c4-4ad0-ab32-9eeca92fecb4\") " Dec 03 15:00:03 crc kubenswrapper[4805]: I1203 15:00:03.747693 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/055eaef9-82c4-4ad0-ab32-9eeca92fecb4-config-volume\") pod \"055eaef9-82c4-4ad0-ab32-9eeca92fecb4\" (UID: \"055eaef9-82c4-4ad0-ab32-9eeca92fecb4\") " Dec 03 15:00:03 crc kubenswrapper[4805]: I1203 15:00:03.748310 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/055eaef9-82c4-4ad0-ab32-9eeca92fecb4-config-volume" (OuterVolumeSpecName: "config-volume") pod "055eaef9-82c4-4ad0-ab32-9eeca92fecb4" (UID: "055eaef9-82c4-4ad0-ab32-9eeca92fecb4"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 15:00:03 crc kubenswrapper[4805]: I1203 15:00:03.748823 4805 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/055eaef9-82c4-4ad0-ab32-9eeca92fecb4-config-volume\") on node \"crc\" DevicePath \"\"" Dec 03 15:00:03 crc kubenswrapper[4805]: I1203 15:00:03.753861 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/055eaef9-82c4-4ad0-ab32-9eeca92fecb4-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "055eaef9-82c4-4ad0-ab32-9eeca92fecb4" (UID: "055eaef9-82c4-4ad0-ab32-9eeca92fecb4"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 15:00:03 crc kubenswrapper[4805]: I1203 15:00:03.767083 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/055eaef9-82c4-4ad0-ab32-9eeca92fecb4-kube-api-access-vhxcr" (OuterVolumeSpecName: "kube-api-access-vhxcr") pod "055eaef9-82c4-4ad0-ab32-9eeca92fecb4" (UID: "055eaef9-82c4-4ad0-ab32-9eeca92fecb4"). InnerVolumeSpecName "kube-api-access-vhxcr". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 15:00:03 crc kubenswrapper[4805]: I1203 15:00:03.850813 4805 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/055eaef9-82c4-4ad0-ab32-9eeca92fecb4-secret-volume\") on node \"crc\" DevicePath \"\"" Dec 03 15:00:03 crc kubenswrapper[4805]: I1203 15:00:03.850889 4805 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vhxcr\" (UniqueName: \"kubernetes.io/projected/055eaef9-82c4-4ad0-ab32-9eeca92fecb4-kube-api-access-vhxcr\") on node \"crc\" DevicePath \"\"" Dec 03 15:00:04 crc kubenswrapper[4805]: I1203 15:00:04.369709 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29412900-pjrk4" event={"ID":"055eaef9-82c4-4ad0-ab32-9eeca92fecb4","Type":"ContainerDied","Data":"8f89ae60934965d9f7f7469a1b23af4bc1c94e8fed7a7bee383e24ce92a34029"} Dec 03 15:00:04 crc kubenswrapper[4805]: I1203 15:00:04.369751 4805 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="8f89ae60934965d9f7f7469a1b23af4bc1c94e8fed7a7bee383e24ce92a34029" Dec 03 15:00:04 crc kubenswrapper[4805]: I1203 15:00:04.369794 4805 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29412900-pjrk4" Dec 03 15:00:04 crc kubenswrapper[4805]: I1203 15:00:04.449822 4805 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29412855-8rxc9"] Dec 03 15:00:04 crc kubenswrapper[4805]: I1203 15:00:04.458782 4805 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29412855-8rxc9"] Dec 03 15:00:04 crc kubenswrapper[4805]: I1203 15:00:04.706615 4805 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7c461d93-d743-4981-b160-48e927a4d443" path="/var/lib/kubelet/pods/7c461d93-d743-4981-b160-48e927a4d443/volumes" Dec 03 15:00:38 crc kubenswrapper[4805]: I1203 15:00:38.828995 4805 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-wdgtm"] Dec 03 15:00:38 crc kubenswrapper[4805]: E1203 15:00:38.830800 4805 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="055eaef9-82c4-4ad0-ab32-9eeca92fecb4" containerName="collect-profiles" Dec 03 15:00:38 crc kubenswrapper[4805]: I1203 15:00:38.830833 4805 state_mem.go:107] "Deleted CPUSet assignment" podUID="055eaef9-82c4-4ad0-ab32-9eeca92fecb4" containerName="collect-profiles" Dec 03 15:00:38 crc kubenswrapper[4805]: I1203 15:00:38.831446 4805 memory_manager.go:354] "RemoveStaleState removing state" podUID="055eaef9-82c4-4ad0-ab32-9eeca92fecb4" containerName="collect-profiles" Dec 03 15:00:38 crc kubenswrapper[4805]: I1203 15:00:38.834957 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-wdgtm" Dec 03 15:00:38 crc kubenswrapper[4805]: I1203 15:00:38.851109 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-wdgtm"] Dec 03 15:00:38 crc kubenswrapper[4805]: I1203 15:00:38.964776 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2t2jv\" (UniqueName: \"kubernetes.io/projected/524c89c7-bfc2-4a89-8805-2415d2024237-kube-api-access-2t2jv\") pod \"redhat-operators-wdgtm\" (UID: \"524c89c7-bfc2-4a89-8805-2415d2024237\") " pod="openshift-marketplace/redhat-operators-wdgtm" Dec 03 15:00:38 crc kubenswrapper[4805]: I1203 15:00:38.965131 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/524c89c7-bfc2-4a89-8805-2415d2024237-utilities\") pod \"redhat-operators-wdgtm\" (UID: \"524c89c7-bfc2-4a89-8805-2415d2024237\") " pod="openshift-marketplace/redhat-operators-wdgtm" Dec 03 15:00:38 crc kubenswrapper[4805]: I1203 15:00:38.965220 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/524c89c7-bfc2-4a89-8805-2415d2024237-catalog-content\") pod \"redhat-operators-wdgtm\" (UID: \"524c89c7-bfc2-4a89-8805-2415d2024237\") " pod="openshift-marketplace/redhat-operators-wdgtm" Dec 03 15:00:39 crc kubenswrapper[4805]: I1203 15:00:39.066863 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/524c89c7-bfc2-4a89-8805-2415d2024237-catalog-content\") pod \"redhat-operators-wdgtm\" (UID: \"524c89c7-bfc2-4a89-8805-2415d2024237\") " pod="openshift-marketplace/redhat-operators-wdgtm" Dec 03 15:00:39 crc kubenswrapper[4805]: I1203 15:00:39.067037 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2t2jv\" (UniqueName: \"kubernetes.io/projected/524c89c7-bfc2-4a89-8805-2415d2024237-kube-api-access-2t2jv\") pod \"redhat-operators-wdgtm\" (UID: \"524c89c7-bfc2-4a89-8805-2415d2024237\") " pod="openshift-marketplace/redhat-operators-wdgtm" Dec 03 15:00:39 crc kubenswrapper[4805]: I1203 15:00:39.067079 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/524c89c7-bfc2-4a89-8805-2415d2024237-utilities\") pod \"redhat-operators-wdgtm\" (UID: \"524c89c7-bfc2-4a89-8805-2415d2024237\") " pod="openshift-marketplace/redhat-operators-wdgtm" Dec 03 15:00:39 crc kubenswrapper[4805]: I1203 15:00:39.067494 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/524c89c7-bfc2-4a89-8805-2415d2024237-catalog-content\") pod \"redhat-operators-wdgtm\" (UID: \"524c89c7-bfc2-4a89-8805-2415d2024237\") " pod="openshift-marketplace/redhat-operators-wdgtm" Dec 03 15:00:39 crc kubenswrapper[4805]: I1203 15:00:39.067596 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/524c89c7-bfc2-4a89-8805-2415d2024237-utilities\") pod \"redhat-operators-wdgtm\" (UID: \"524c89c7-bfc2-4a89-8805-2415d2024237\") " pod="openshift-marketplace/redhat-operators-wdgtm" Dec 03 15:00:39 crc kubenswrapper[4805]: I1203 15:00:39.084949 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2t2jv\" (UniqueName: \"kubernetes.io/projected/524c89c7-bfc2-4a89-8805-2415d2024237-kube-api-access-2t2jv\") pod \"redhat-operators-wdgtm\" (UID: \"524c89c7-bfc2-4a89-8805-2415d2024237\") " pod="openshift-marketplace/redhat-operators-wdgtm" Dec 03 15:00:39 crc kubenswrapper[4805]: I1203 15:00:39.166251 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-wdgtm" Dec 03 15:00:39 crc kubenswrapper[4805]: I1203 15:00:39.663211 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-wdgtm"] Dec 03 15:00:39 crc kubenswrapper[4805]: I1203 15:00:39.737435 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-wdgtm" event={"ID":"524c89c7-bfc2-4a89-8805-2415d2024237","Type":"ContainerStarted","Data":"9b0eab6df98664b855501539c518a5e55ba40c53e0eca00d53401f816879424e"} Dec 03 15:00:40 crc kubenswrapper[4805]: I1203 15:00:40.748096 4805 generic.go:334] "Generic (PLEG): container finished" podID="524c89c7-bfc2-4a89-8805-2415d2024237" containerID="e8ad258dfcb33d46740d53e05b9119296ff0b0fb60c15cde4ae58a2020569ebd" exitCode=0 Dec 03 15:00:40 crc kubenswrapper[4805]: I1203 15:00:40.748151 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-wdgtm" event={"ID":"524c89c7-bfc2-4a89-8805-2415d2024237","Type":"ContainerDied","Data":"e8ad258dfcb33d46740d53e05b9119296ff0b0fb60c15cde4ae58a2020569ebd"} Dec 03 15:00:40 crc kubenswrapper[4805]: I1203 15:00:40.750350 4805 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 03 15:00:41 crc kubenswrapper[4805]: I1203 15:00:41.758798 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-wdgtm" event={"ID":"524c89c7-bfc2-4a89-8805-2415d2024237","Type":"ContainerStarted","Data":"ce5c886e6ad57aa48f61f5f331f104a2171b03b2dd5653aa6c2921279ecb4c58"} Dec 03 15:00:42 crc kubenswrapper[4805]: I1203 15:00:42.781298 4805 generic.go:334] "Generic (PLEG): container finished" podID="524c89c7-bfc2-4a89-8805-2415d2024237" containerID="ce5c886e6ad57aa48f61f5f331f104a2171b03b2dd5653aa6c2921279ecb4c58" exitCode=0 Dec 03 15:00:42 crc kubenswrapper[4805]: I1203 15:00:42.781416 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-wdgtm" event={"ID":"524c89c7-bfc2-4a89-8805-2415d2024237","Type":"ContainerDied","Data":"ce5c886e6ad57aa48f61f5f331f104a2171b03b2dd5653aa6c2921279ecb4c58"} Dec 03 15:00:43 crc kubenswrapper[4805]: I1203 15:00:43.798712 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-wdgtm" event={"ID":"524c89c7-bfc2-4a89-8805-2415d2024237","Type":"ContainerStarted","Data":"eb362c76079d36369def3f50836650193cecba1cd0b80a1661f7d6bcaf9e57c8"} Dec 03 15:00:43 crc kubenswrapper[4805]: I1203 15:00:43.828254 4805 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-wdgtm" podStartSLOduration=3.38912174 podStartE2EDuration="5.828236186s" podCreationTimestamp="2025-12-03 15:00:38 +0000 UTC" firstStartedPulling="2025-12-03 15:00:40.750134104 +0000 UTC m=+3070.413051027" lastFinishedPulling="2025-12-03 15:00:43.18924854 +0000 UTC m=+3072.852165473" observedRunningTime="2025-12-03 15:00:43.81836558 +0000 UTC m=+3073.481282513" watchObservedRunningTime="2025-12-03 15:00:43.828236186 +0000 UTC m=+3073.491153119" Dec 03 15:00:43 crc kubenswrapper[4805]: I1203 15:00:43.917036 4805 patch_prober.go:28] interesting pod/machine-config-daemon-gskh4 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 03 15:00:43 crc kubenswrapper[4805]: I1203 15:00:43.917487 4805 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-gskh4" podUID="6098937f-e3f6-45e8-a647-4994a79cd711" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 03 15:00:45 crc kubenswrapper[4805]: I1203 15:00:45.318599 4805 scope.go:117] "RemoveContainer" containerID="b5b408b815391ac337fc70c2d75fd6ceee87e7854c2a7696867fcf21ebaf87cb" Dec 03 15:00:49 crc kubenswrapper[4805]: I1203 15:00:49.166464 4805 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-wdgtm" Dec 03 15:00:49 crc kubenswrapper[4805]: I1203 15:00:49.167116 4805 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-wdgtm" Dec 03 15:00:50 crc kubenswrapper[4805]: I1203 15:00:50.217683 4805 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-wdgtm" podUID="524c89c7-bfc2-4a89-8805-2415d2024237" containerName="registry-server" probeResult="failure" output=< Dec 03 15:00:50 crc kubenswrapper[4805]: timeout: failed to connect service ":50051" within 1s Dec 03 15:00:50 crc kubenswrapper[4805]: > Dec 03 15:00:59 crc kubenswrapper[4805]: I1203 15:00:59.230906 4805 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-wdgtm" Dec 03 15:00:59 crc kubenswrapper[4805]: I1203 15:00:59.285427 4805 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-wdgtm" Dec 03 15:00:59 crc kubenswrapper[4805]: I1203 15:00:59.480809 4805 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-wdgtm"] Dec 03 15:01:00 crc kubenswrapper[4805]: I1203 15:01:00.155204 4805 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-cron-29412901-wgktp"] Dec 03 15:01:00 crc kubenswrapper[4805]: I1203 15:01:00.157700 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-cron-29412901-wgktp" Dec 03 15:01:00 crc kubenswrapper[4805]: I1203 15:01:00.171027 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-cron-29412901-wgktp"] Dec 03 15:01:00 crc kubenswrapper[4805]: I1203 15:01:00.312978 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ae49c111-f5d4-4884-857c-d03296d16312-config-data\") pod \"keystone-cron-29412901-wgktp\" (UID: \"ae49c111-f5d4-4884-857c-d03296d16312\") " pod="openstack/keystone-cron-29412901-wgktp" Dec 03 15:01:00 crc kubenswrapper[4805]: I1203 15:01:00.313106 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ae49c111-f5d4-4884-857c-d03296d16312-combined-ca-bundle\") pod \"keystone-cron-29412901-wgktp\" (UID: \"ae49c111-f5d4-4884-857c-d03296d16312\") " pod="openstack/keystone-cron-29412901-wgktp" Dec 03 15:01:00 crc kubenswrapper[4805]: I1203 15:01:00.313377 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nmv69\" (UniqueName: \"kubernetes.io/projected/ae49c111-f5d4-4884-857c-d03296d16312-kube-api-access-nmv69\") pod \"keystone-cron-29412901-wgktp\" (UID: \"ae49c111-f5d4-4884-857c-d03296d16312\") " pod="openstack/keystone-cron-29412901-wgktp" Dec 03 15:01:00 crc kubenswrapper[4805]: I1203 15:01:00.313456 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/ae49c111-f5d4-4884-857c-d03296d16312-fernet-keys\") pod \"keystone-cron-29412901-wgktp\" (UID: \"ae49c111-f5d4-4884-857c-d03296d16312\") " pod="openstack/keystone-cron-29412901-wgktp" Dec 03 15:01:00 crc kubenswrapper[4805]: I1203 15:01:00.416021 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nmv69\" (UniqueName: \"kubernetes.io/projected/ae49c111-f5d4-4884-857c-d03296d16312-kube-api-access-nmv69\") pod \"keystone-cron-29412901-wgktp\" (UID: \"ae49c111-f5d4-4884-857c-d03296d16312\") " pod="openstack/keystone-cron-29412901-wgktp" Dec 03 15:01:00 crc kubenswrapper[4805]: I1203 15:01:00.416157 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/ae49c111-f5d4-4884-857c-d03296d16312-fernet-keys\") pod \"keystone-cron-29412901-wgktp\" (UID: \"ae49c111-f5d4-4884-857c-d03296d16312\") " pod="openstack/keystone-cron-29412901-wgktp" Dec 03 15:01:00 crc kubenswrapper[4805]: I1203 15:01:00.416267 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ae49c111-f5d4-4884-857c-d03296d16312-config-data\") pod \"keystone-cron-29412901-wgktp\" (UID: \"ae49c111-f5d4-4884-857c-d03296d16312\") " pod="openstack/keystone-cron-29412901-wgktp" Dec 03 15:01:00 crc kubenswrapper[4805]: I1203 15:01:00.416371 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ae49c111-f5d4-4884-857c-d03296d16312-combined-ca-bundle\") pod \"keystone-cron-29412901-wgktp\" (UID: \"ae49c111-f5d4-4884-857c-d03296d16312\") " pod="openstack/keystone-cron-29412901-wgktp" Dec 03 15:01:00 crc kubenswrapper[4805]: I1203 15:01:00.425173 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ae49c111-f5d4-4884-857c-d03296d16312-config-data\") pod \"keystone-cron-29412901-wgktp\" (UID: \"ae49c111-f5d4-4884-857c-d03296d16312\") " pod="openstack/keystone-cron-29412901-wgktp" Dec 03 15:01:00 crc kubenswrapper[4805]: I1203 15:01:00.426488 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/ae49c111-f5d4-4884-857c-d03296d16312-fernet-keys\") pod \"keystone-cron-29412901-wgktp\" (UID: \"ae49c111-f5d4-4884-857c-d03296d16312\") " pod="openstack/keystone-cron-29412901-wgktp" Dec 03 15:01:00 crc kubenswrapper[4805]: I1203 15:01:00.427459 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ae49c111-f5d4-4884-857c-d03296d16312-combined-ca-bundle\") pod \"keystone-cron-29412901-wgktp\" (UID: \"ae49c111-f5d4-4884-857c-d03296d16312\") " pod="openstack/keystone-cron-29412901-wgktp" Dec 03 15:01:00 crc kubenswrapper[4805]: I1203 15:01:00.447301 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nmv69\" (UniqueName: \"kubernetes.io/projected/ae49c111-f5d4-4884-857c-d03296d16312-kube-api-access-nmv69\") pod \"keystone-cron-29412901-wgktp\" (UID: \"ae49c111-f5d4-4884-857c-d03296d16312\") " pod="openstack/keystone-cron-29412901-wgktp" Dec 03 15:01:00 crc kubenswrapper[4805]: I1203 15:01:00.481049 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-cron-29412901-wgktp" Dec 03 15:01:00 crc kubenswrapper[4805]: I1203 15:01:00.948338 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-cron-29412901-wgktp"] Dec 03 15:01:00 crc kubenswrapper[4805]: W1203 15:01:00.951603 4805 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podae49c111_f5d4_4884_857c_d03296d16312.slice/crio-f3f8985017de69c69e37c1e266ce1c8847681b107ca7c27c58814f57470f0569 WatchSource:0}: Error finding container f3f8985017de69c69e37c1e266ce1c8847681b107ca7c27c58814f57470f0569: Status 404 returned error can't find the container with id f3f8985017de69c69e37c1e266ce1c8847681b107ca7c27c58814f57470f0569 Dec 03 15:01:00 crc kubenswrapper[4805]: I1203 15:01:00.979192 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-cron-29412901-wgktp" event={"ID":"ae49c111-f5d4-4884-857c-d03296d16312","Type":"ContainerStarted","Data":"f3f8985017de69c69e37c1e266ce1c8847681b107ca7c27c58814f57470f0569"} Dec 03 15:01:00 crc kubenswrapper[4805]: I1203 15:01:00.979398 4805 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-wdgtm" podUID="524c89c7-bfc2-4a89-8805-2415d2024237" containerName="registry-server" containerID="cri-o://eb362c76079d36369def3f50836650193cecba1cd0b80a1661f7d6bcaf9e57c8" gracePeriod=2 Dec 03 15:01:01 crc kubenswrapper[4805]: I1203 15:01:01.340045 4805 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-wdgtm" Dec 03 15:01:01 crc kubenswrapper[4805]: I1203 15:01:01.539558 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/524c89c7-bfc2-4a89-8805-2415d2024237-catalog-content\") pod \"524c89c7-bfc2-4a89-8805-2415d2024237\" (UID: \"524c89c7-bfc2-4a89-8805-2415d2024237\") " Dec 03 15:01:01 crc kubenswrapper[4805]: I1203 15:01:01.539680 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2t2jv\" (UniqueName: \"kubernetes.io/projected/524c89c7-bfc2-4a89-8805-2415d2024237-kube-api-access-2t2jv\") pod \"524c89c7-bfc2-4a89-8805-2415d2024237\" (UID: \"524c89c7-bfc2-4a89-8805-2415d2024237\") " Dec 03 15:01:01 crc kubenswrapper[4805]: I1203 15:01:01.539740 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/524c89c7-bfc2-4a89-8805-2415d2024237-utilities\") pod \"524c89c7-bfc2-4a89-8805-2415d2024237\" (UID: \"524c89c7-bfc2-4a89-8805-2415d2024237\") " Dec 03 15:01:01 crc kubenswrapper[4805]: I1203 15:01:01.541027 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/524c89c7-bfc2-4a89-8805-2415d2024237-utilities" (OuterVolumeSpecName: "utilities") pod "524c89c7-bfc2-4a89-8805-2415d2024237" (UID: "524c89c7-bfc2-4a89-8805-2415d2024237"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 15:01:01 crc kubenswrapper[4805]: I1203 15:01:01.550953 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/524c89c7-bfc2-4a89-8805-2415d2024237-kube-api-access-2t2jv" (OuterVolumeSpecName: "kube-api-access-2t2jv") pod "524c89c7-bfc2-4a89-8805-2415d2024237" (UID: "524c89c7-bfc2-4a89-8805-2415d2024237"). InnerVolumeSpecName "kube-api-access-2t2jv". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 15:01:01 crc kubenswrapper[4805]: I1203 15:01:01.641962 4805 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2t2jv\" (UniqueName: \"kubernetes.io/projected/524c89c7-bfc2-4a89-8805-2415d2024237-kube-api-access-2t2jv\") on node \"crc\" DevicePath \"\"" Dec 03 15:01:01 crc kubenswrapper[4805]: I1203 15:01:01.641994 4805 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/524c89c7-bfc2-4a89-8805-2415d2024237-utilities\") on node \"crc\" DevicePath \"\"" Dec 03 15:01:01 crc kubenswrapper[4805]: I1203 15:01:01.650832 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/524c89c7-bfc2-4a89-8805-2415d2024237-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "524c89c7-bfc2-4a89-8805-2415d2024237" (UID: "524c89c7-bfc2-4a89-8805-2415d2024237"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 15:01:01 crc kubenswrapper[4805]: I1203 15:01:01.744085 4805 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/524c89c7-bfc2-4a89-8805-2415d2024237-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 03 15:01:01 crc kubenswrapper[4805]: I1203 15:01:01.988494 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-cron-29412901-wgktp" event={"ID":"ae49c111-f5d4-4884-857c-d03296d16312","Type":"ContainerStarted","Data":"2bad62b13479c45ae0eac61e80c795326c66328f9d9d9e595d35aef1f824ecc5"} Dec 03 15:01:01 crc kubenswrapper[4805]: I1203 15:01:01.991111 4805 generic.go:334] "Generic (PLEG): container finished" podID="524c89c7-bfc2-4a89-8805-2415d2024237" containerID="eb362c76079d36369def3f50836650193cecba1cd0b80a1661f7d6bcaf9e57c8" exitCode=0 Dec 03 15:01:01 crc kubenswrapper[4805]: I1203 15:01:01.991133 4805 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-wdgtm" Dec 03 15:01:01 crc kubenswrapper[4805]: I1203 15:01:01.991147 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-wdgtm" event={"ID":"524c89c7-bfc2-4a89-8805-2415d2024237","Type":"ContainerDied","Data":"eb362c76079d36369def3f50836650193cecba1cd0b80a1661f7d6bcaf9e57c8"} Dec 03 15:01:01 crc kubenswrapper[4805]: I1203 15:01:01.991892 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-wdgtm" event={"ID":"524c89c7-bfc2-4a89-8805-2415d2024237","Type":"ContainerDied","Data":"9b0eab6df98664b855501539c518a5e55ba40c53e0eca00d53401f816879424e"} Dec 03 15:01:01 crc kubenswrapper[4805]: I1203 15:01:01.991912 4805 scope.go:117] "RemoveContainer" containerID="eb362c76079d36369def3f50836650193cecba1cd0b80a1661f7d6bcaf9e57c8" Dec 03 15:01:02 crc kubenswrapper[4805]: I1203 15:01:02.010764 4805 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-cron-29412901-wgktp" podStartSLOduration=2.010744299 podStartE2EDuration="2.010744299s" podCreationTimestamp="2025-12-03 15:01:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 15:01:02.006192807 +0000 UTC m=+3091.669109730" watchObservedRunningTime="2025-12-03 15:01:02.010744299 +0000 UTC m=+3091.673661222" Dec 03 15:01:02 crc kubenswrapper[4805]: I1203 15:01:02.022389 4805 scope.go:117] "RemoveContainer" containerID="ce5c886e6ad57aa48f61f5f331f104a2171b03b2dd5653aa6c2921279ecb4c58" Dec 03 15:01:02 crc kubenswrapper[4805]: I1203 15:01:02.052153 4805 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-wdgtm"] Dec 03 15:01:02 crc kubenswrapper[4805]: I1203 15:01:02.056923 4805 scope.go:117] "RemoveContainer" containerID="e8ad258dfcb33d46740d53e05b9119296ff0b0fb60c15cde4ae58a2020569ebd" Dec 03 15:01:02 crc kubenswrapper[4805]: I1203 15:01:02.059853 4805 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-wdgtm"] Dec 03 15:01:02 crc kubenswrapper[4805]: I1203 15:01:02.107190 4805 scope.go:117] "RemoveContainer" containerID="eb362c76079d36369def3f50836650193cecba1cd0b80a1661f7d6bcaf9e57c8" Dec 03 15:01:02 crc kubenswrapper[4805]: E1203 15:01:02.107760 4805 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"eb362c76079d36369def3f50836650193cecba1cd0b80a1661f7d6bcaf9e57c8\": container with ID starting with eb362c76079d36369def3f50836650193cecba1cd0b80a1661f7d6bcaf9e57c8 not found: ID does not exist" containerID="eb362c76079d36369def3f50836650193cecba1cd0b80a1661f7d6bcaf9e57c8" Dec 03 15:01:02 crc kubenswrapper[4805]: I1203 15:01:02.107811 4805 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"eb362c76079d36369def3f50836650193cecba1cd0b80a1661f7d6bcaf9e57c8"} err="failed to get container status \"eb362c76079d36369def3f50836650193cecba1cd0b80a1661f7d6bcaf9e57c8\": rpc error: code = NotFound desc = could not find container \"eb362c76079d36369def3f50836650193cecba1cd0b80a1661f7d6bcaf9e57c8\": container with ID starting with eb362c76079d36369def3f50836650193cecba1cd0b80a1661f7d6bcaf9e57c8 not found: ID does not exist" Dec 03 15:01:02 crc kubenswrapper[4805]: I1203 15:01:02.107861 4805 scope.go:117] "RemoveContainer" containerID="ce5c886e6ad57aa48f61f5f331f104a2171b03b2dd5653aa6c2921279ecb4c58" Dec 03 15:01:02 crc kubenswrapper[4805]: E1203 15:01:02.108435 4805 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ce5c886e6ad57aa48f61f5f331f104a2171b03b2dd5653aa6c2921279ecb4c58\": container with ID starting with ce5c886e6ad57aa48f61f5f331f104a2171b03b2dd5653aa6c2921279ecb4c58 not found: ID does not exist" containerID="ce5c886e6ad57aa48f61f5f331f104a2171b03b2dd5653aa6c2921279ecb4c58" Dec 03 15:01:02 crc kubenswrapper[4805]: I1203 15:01:02.108511 4805 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ce5c886e6ad57aa48f61f5f331f104a2171b03b2dd5653aa6c2921279ecb4c58"} err="failed to get container status \"ce5c886e6ad57aa48f61f5f331f104a2171b03b2dd5653aa6c2921279ecb4c58\": rpc error: code = NotFound desc = could not find container \"ce5c886e6ad57aa48f61f5f331f104a2171b03b2dd5653aa6c2921279ecb4c58\": container with ID starting with ce5c886e6ad57aa48f61f5f331f104a2171b03b2dd5653aa6c2921279ecb4c58 not found: ID does not exist" Dec 03 15:01:02 crc kubenswrapper[4805]: I1203 15:01:02.108556 4805 scope.go:117] "RemoveContainer" containerID="e8ad258dfcb33d46740d53e05b9119296ff0b0fb60c15cde4ae58a2020569ebd" Dec 03 15:01:02 crc kubenswrapper[4805]: E1203 15:01:02.109107 4805 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e8ad258dfcb33d46740d53e05b9119296ff0b0fb60c15cde4ae58a2020569ebd\": container with ID starting with e8ad258dfcb33d46740d53e05b9119296ff0b0fb60c15cde4ae58a2020569ebd not found: ID does not exist" containerID="e8ad258dfcb33d46740d53e05b9119296ff0b0fb60c15cde4ae58a2020569ebd" Dec 03 15:01:02 crc kubenswrapper[4805]: I1203 15:01:02.109146 4805 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e8ad258dfcb33d46740d53e05b9119296ff0b0fb60c15cde4ae58a2020569ebd"} err="failed to get container status \"e8ad258dfcb33d46740d53e05b9119296ff0b0fb60c15cde4ae58a2020569ebd\": rpc error: code = NotFound desc = could not find container \"e8ad258dfcb33d46740d53e05b9119296ff0b0fb60c15cde4ae58a2020569ebd\": container with ID starting with e8ad258dfcb33d46740d53e05b9119296ff0b0fb60c15cde4ae58a2020569ebd not found: ID does not exist" Dec 03 15:01:02 crc kubenswrapper[4805]: I1203 15:01:02.708405 4805 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="524c89c7-bfc2-4a89-8805-2415d2024237" path="/var/lib/kubelet/pods/524c89c7-bfc2-4a89-8805-2415d2024237/volumes" Dec 03 15:01:04 crc kubenswrapper[4805]: I1203 15:01:04.032996 4805 generic.go:334] "Generic (PLEG): container finished" podID="ae49c111-f5d4-4884-857c-d03296d16312" containerID="2bad62b13479c45ae0eac61e80c795326c66328f9d9d9e595d35aef1f824ecc5" exitCode=0 Dec 03 15:01:04 crc kubenswrapper[4805]: I1203 15:01:04.033121 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-cron-29412901-wgktp" event={"ID":"ae49c111-f5d4-4884-857c-d03296d16312","Type":"ContainerDied","Data":"2bad62b13479c45ae0eac61e80c795326c66328f9d9d9e595d35aef1f824ecc5"} Dec 03 15:01:05 crc kubenswrapper[4805]: I1203 15:01:05.411591 4805 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-cron-29412901-wgktp" Dec 03 15:01:05 crc kubenswrapper[4805]: I1203 15:01:05.520827 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/ae49c111-f5d4-4884-857c-d03296d16312-fernet-keys\") pod \"ae49c111-f5d4-4884-857c-d03296d16312\" (UID: \"ae49c111-f5d4-4884-857c-d03296d16312\") " Dec 03 15:01:05 crc kubenswrapper[4805]: I1203 15:01:05.521271 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ae49c111-f5d4-4884-857c-d03296d16312-combined-ca-bundle\") pod \"ae49c111-f5d4-4884-857c-d03296d16312\" (UID: \"ae49c111-f5d4-4884-857c-d03296d16312\") " Dec 03 15:01:05 crc kubenswrapper[4805]: I1203 15:01:05.521309 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ae49c111-f5d4-4884-857c-d03296d16312-config-data\") pod \"ae49c111-f5d4-4884-857c-d03296d16312\" (UID: \"ae49c111-f5d4-4884-857c-d03296d16312\") " Dec 03 15:01:05 crc kubenswrapper[4805]: I1203 15:01:05.521360 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nmv69\" (UniqueName: \"kubernetes.io/projected/ae49c111-f5d4-4884-857c-d03296d16312-kube-api-access-nmv69\") pod \"ae49c111-f5d4-4884-857c-d03296d16312\" (UID: \"ae49c111-f5d4-4884-857c-d03296d16312\") " Dec 03 15:01:05 crc kubenswrapper[4805]: I1203 15:01:05.531780 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ae49c111-f5d4-4884-857c-d03296d16312-kube-api-access-nmv69" (OuterVolumeSpecName: "kube-api-access-nmv69") pod "ae49c111-f5d4-4884-857c-d03296d16312" (UID: "ae49c111-f5d4-4884-857c-d03296d16312"). InnerVolumeSpecName "kube-api-access-nmv69". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 15:01:05 crc kubenswrapper[4805]: I1203 15:01:05.534226 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ae49c111-f5d4-4884-857c-d03296d16312-fernet-keys" (OuterVolumeSpecName: "fernet-keys") pod "ae49c111-f5d4-4884-857c-d03296d16312" (UID: "ae49c111-f5d4-4884-857c-d03296d16312"). InnerVolumeSpecName "fernet-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 15:01:05 crc kubenswrapper[4805]: I1203 15:01:05.558991 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ae49c111-f5d4-4884-857c-d03296d16312-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "ae49c111-f5d4-4884-857c-d03296d16312" (UID: "ae49c111-f5d4-4884-857c-d03296d16312"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 15:01:05 crc kubenswrapper[4805]: I1203 15:01:05.596764 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ae49c111-f5d4-4884-857c-d03296d16312-config-data" (OuterVolumeSpecName: "config-data") pod "ae49c111-f5d4-4884-857c-d03296d16312" (UID: "ae49c111-f5d4-4884-857c-d03296d16312"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 15:01:05 crc kubenswrapper[4805]: I1203 15:01:05.623669 4805 reconciler_common.go:293] "Volume detached for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/ae49c111-f5d4-4884-857c-d03296d16312-fernet-keys\") on node \"crc\" DevicePath \"\"" Dec 03 15:01:05 crc kubenswrapper[4805]: I1203 15:01:05.623713 4805 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ae49c111-f5d4-4884-857c-d03296d16312-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 15:01:05 crc kubenswrapper[4805]: I1203 15:01:05.623727 4805 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ae49c111-f5d4-4884-857c-d03296d16312-config-data\") on node \"crc\" DevicePath \"\"" Dec 03 15:01:05 crc kubenswrapper[4805]: I1203 15:01:05.623740 4805 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nmv69\" (UniqueName: \"kubernetes.io/projected/ae49c111-f5d4-4884-857c-d03296d16312-kube-api-access-nmv69\") on node \"crc\" DevicePath \"\"" Dec 03 15:01:06 crc kubenswrapper[4805]: I1203 15:01:06.056927 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-cron-29412901-wgktp" event={"ID":"ae49c111-f5d4-4884-857c-d03296d16312","Type":"ContainerDied","Data":"f3f8985017de69c69e37c1e266ce1c8847681b107ca7c27c58814f57470f0569"} Dec 03 15:01:06 crc kubenswrapper[4805]: I1203 15:01:06.056976 4805 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="f3f8985017de69c69e37c1e266ce1c8847681b107ca7c27c58814f57470f0569" Dec 03 15:01:06 crc kubenswrapper[4805]: I1203 15:01:06.057009 4805 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-cron-29412901-wgktp" Dec 03 15:01:13 crc kubenswrapper[4805]: I1203 15:01:13.917737 4805 patch_prober.go:28] interesting pod/machine-config-daemon-gskh4 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 03 15:01:13 crc kubenswrapper[4805]: I1203 15:01:13.918330 4805 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-gskh4" podUID="6098937f-e3f6-45e8-a647-4994a79cd711" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 03 15:01:25 crc kubenswrapper[4805]: I1203 15:01:25.534625 4805 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-r9rwr"] Dec 03 15:01:25 crc kubenswrapper[4805]: E1203 15:01:25.535580 4805 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="524c89c7-bfc2-4a89-8805-2415d2024237" containerName="registry-server" Dec 03 15:01:25 crc kubenswrapper[4805]: I1203 15:01:25.535595 4805 state_mem.go:107] "Deleted CPUSet assignment" podUID="524c89c7-bfc2-4a89-8805-2415d2024237" containerName="registry-server" Dec 03 15:01:25 crc kubenswrapper[4805]: E1203 15:01:25.535622 4805 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="524c89c7-bfc2-4a89-8805-2415d2024237" containerName="extract-content" Dec 03 15:01:25 crc kubenswrapper[4805]: I1203 15:01:25.535630 4805 state_mem.go:107] "Deleted CPUSet assignment" podUID="524c89c7-bfc2-4a89-8805-2415d2024237" containerName="extract-content" Dec 03 15:01:25 crc kubenswrapper[4805]: E1203 15:01:25.535646 4805 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="524c89c7-bfc2-4a89-8805-2415d2024237" containerName="extract-utilities" Dec 03 15:01:25 crc kubenswrapper[4805]: I1203 15:01:25.535653 4805 state_mem.go:107] "Deleted CPUSet assignment" podUID="524c89c7-bfc2-4a89-8805-2415d2024237" containerName="extract-utilities" Dec 03 15:01:25 crc kubenswrapper[4805]: E1203 15:01:25.535667 4805 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ae49c111-f5d4-4884-857c-d03296d16312" containerName="keystone-cron" Dec 03 15:01:25 crc kubenswrapper[4805]: I1203 15:01:25.535673 4805 state_mem.go:107] "Deleted CPUSet assignment" podUID="ae49c111-f5d4-4884-857c-d03296d16312" containerName="keystone-cron" Dec 03 15:01:25 crc kubenswrapper[4805]: I1203 15:01:25.535919 4805 memory_manager.go:354] "RemoveStaleState removing state" podUID="524c89c7-bfc2-4a89-8805-2415d2024237" containerName="registry-server" Dec 03 15:01:25 crc kubenswrapper[4805]: I1203 15:01:25.535950 4805 memory_manager.go:354] "RemoveStaleState removing state" podUID="ae49c111-f5d4-4884-857c-d03296d16312" containerName="keystone-cron" Dec 03 15:01:25 crc kubenswrapper[4805]: I1203 15:01:25.538029 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-r9rwr" Dec 03 15:01:25 crc kubenswrapper[4805]: I1203 15:01:25.566444 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-r9rwr"] Dec 03 15:01:25 crc kubenswrapper[4805]: I1203 15:01:25.623976 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8b1d1e18-a325-4512-a3a7-48b21944317f-utilities\") pod \"certified-operators-r9rwr\" (UID: \"8b1d1e18-a325-4512-a3a7-48b21944317f\") " pod="openshift-marketplace/certified-operators-r9rwr" Dec 03 15:01:25 crc kubenswrapper[4805]: I1203 15:01:25.624066 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8w8kj\" (UniqueName: \"kubernetes.io/projected/8b1d1e18-a325-4512-a3a7-48b21944317f-kube-api-access-8w8kj\") pod \"certified-operators-r9rwr\" (UID: \"8b1d1e18-a325-4512-a3a7-48b21944317f\") " pod="openshift-marketplace/certified-operators-r9rwr" Dec 03 15:01:25 crc kubenswrapper[4805]: I1203 15:01:25.624240 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8b1d1e18-a325-4512-a3a7-48b21944317f-catalog-content\") pod \"certified-operators-r9rwr\" (UID: \"8b1d1e18-a325-4512-a3a7-48b21944317f\") " pod="openshift-marketplace/certified-operators-r9rwr" Dec 03 15:01:25 crc kubenswrapper[4805]: I1203 15:01:25.726230 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8w8kj\" (UniqueName: \"kubernetes.io/projected/8b1d1e18-a325-4512-a3a7-48b21944317f-kube-api-access-8w8kj\") pod \"certified-operators-r9rwr\" (UID: \"8b1d1e18-a325-4512-a3a7-48b21944317f\") " pod="openshift-marketplace/certified-operators-r9rwr" Dec 03 15:01:25 crc kubenswrapper[4805]: I1203 15:01:25.726460 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8b1d1e18-a325-4512-a3a7-48b21944317f-catalog-content\") pod \"certified-operators-r9rwr\" (UID: \"8b1d1e18-a325-4512-a3a7-48b21944317f\") " pod="openshift-marketplace/certified-operators-r9rwr" Dec 03 15:01:25 crc kubenswrapper[4805]: I1203 15:01:25.726554 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8b1d1e18-a325-4512-a3a7-48b21944317f-utilities\") pod \"certified-operators-r9rwr\" (UID: \"8b1d1e18-a325-4512-a3a7-48b21944317f\") " pod="openshift-marketplace/certified-operators-r9rwr" Dec 03 15:01:25 crc kubenswrapper[4805]: I1203 15:01:25.727804 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8b1d1e18-a325-4512-a3a7-48b21944317f-utilities\") pod \"certified-operators-r9rwr\" (UID: \"8b1d1e18-a325-4512-a3a7-48b21944317f\") " pod="openshift-marketplace/certified-operators-r9rwr" Dec 03 15:01:25 crc kubenswrapper[4805]: I1203 15:01:25.727891 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8b1d1e18-a325-4512-a3a7-48b21944317f-catalog-content\") pod \"certified-operators-r9rwr\" (UID: \"8b1d1e18-a325-4512-a3a7-48b21944317f\") " pod="openshift-marketplace/certified-operators-r9rwr" Dec 03 15:01:25 crc kubenswrapper[4805]: I1203 15:01:25.755888 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8w8kj\" (UniqueName: \"kubernetes.io/projected/8b1d1e18-a325-4512-a3a7-48b21944317f-kube-api-access-8w8kj\") pod \"certified-operators-r9rwr\" (UID: \"8b1d1e18-a325-4512-a3a7-48b21944317f\") " pod="openshift-marketplace/certified-operators-r9rwr" Dec 03 15:01:25 crc kubenswrapper[4805]: I1203 15:01:25.872144 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-r9rwr" Dec 03 15:01:26 crc kubenswrapper[4805]: I1203 15:01:26.370365 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-r9rwr"] Dec 03 15:01:26 crc kubenswrapper[4805]: W1203 15:01:26.385031 4805 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod8b1d1e18_a325_4512_a3a7_48b21944317f.slice/crio-fbd0cfd7d0edac39f2ff8f162ef664f3744ae3b4858a96d7c368f5e8e3b09efd WatchSource:0}: Error finding container fbd0cfd7d0edac39f2ff8f162ef664f3744ae3b4858a96d7c368f5e8e3b09efd: Status 404 returned error can't find the container with id fbd0cfd7d0edac39f2ff8f162ef664f3744ae3b4858a96d7c368f5e8e3b09efd Dec 03 15:01:26 crc kubenswrapper[4805]: I1203 15:01:26.798442 4805 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_telemetry-operator-controller-manager-cc9d8f87b-qfmmx_8cbfcd07-e136-4bfd-b3c6-6615c911d440/manager/0.log" Dec 03 15:01:27 crc kubenswrapper[4805]: I1203 15:01:27.252516 4805 generic.go:334] "Generic (PLEG): container finished" podID="8b1d1e18-a325-4512-a3a7-48b21944317f" containerID="da1d10a688034fb07bec11aa970653b7d9535e3cf6ba9e895a9876a6ba6660fd" exitCode=0 Dec 03 15:01:27 crc kubenswrapper[4805]: I1203 15:01:27.252997 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-r9rwr" event={"ID":"8b1d1e18-a325-4512-a3a7-48b21944317f","Type":"ContainerDied","Data":"da1d10a688034fb07bec11aa970653b7d9535e3cf6ba9e895a9876a6ba6660fd"} Dec 03 15:01:27 crc kubenswrapper[4805]: I1203 15:01:27.253037 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-r9rwr" event={"ID":"8b1d1e18-a325-4512-a3a7-48b21944317f","Type":"ContainerStarted","Data":"fbd0cfd7d0edac39f2ff8f162ef664f3744ae3b4858a96d7c368f5e8e3b09efd"} Dec 03 15:01:28 crc kubenswrapper[4805]: I1203 15:01:28.510286 4805 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/openstackclient"] Dec 03 15:01:28 crc kubenswrapper[4805]: I1203 15:01:28.510929 4805 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/openstackclient" podUID="29c53624-ef4b-43ab-8922-181a78908d24" containerName="openstackclient" containerID="cri-o://26262b91874fa88789d54d9ed21dcbd2e4122b6fb279633aaa0a8892e3a82995" gracePeriod=2 Dec 03 15:01:28 crc kubenswrapper[4805]: I1203 15:01:28.526999 4805 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/openstackclient"] Dec 03 15:01:28 crc kubenswrapper[4805]: I1203 15:01:28.561153 4805 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/openstackclient"] Dec 03 15:01:28 crc kubenswrapper[4805]: E1203 15:01:28.561560 4805 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="29c53624-ef4b-43ab-8922-181a78908d24" containerName="openstackclient" Dec 03 15:01:28 crc kubenswrapper[4805]: I1203 15:01:28.561578 4805 state_mem.go:107] "Deleted CPUSet assignment" podUID="29c53624-ef4b-43ab-8922-181a78908d24" containerName="openstackclient" Dec 03 15:01:28 crc kubenswrapper[4805]: I1203 15:01:28.561786 4805 memory_manager.go:354] "RemoveStaleState removing state" podUID="29c53624-ef4b-43ab-8922-181a78908d24" containerName="openstackclient" Dec 03 15:01:28 crc kubenswrapper[4805]: I1203 15:01:28.562414 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Dec 03 15:01:28 crc kubenswrapper[4805]: I1203 15:01:28.566714 4805 status_manager.go:861] "Pod was deleted and then recreated, skipping status update" pod="openstack/openstackclient" oldPodUID="29c53624-ef4b-43ab-8922-181a78908d24" podUID="8524cabe-a1dd-46c5-98d1-73c121150721" Dec 03 15:01:28 crc kubenswrapper[4805]: I1203 15:01:28.575804 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstackclient"] Dec 03 15:01:28 crc kubenswrapper[4805]: I1203 15:01:28.640270 4805 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/openstackclient"] Dec 03 15:01:28 crc kubenswrapper[4805]: E1203 15:01:28.641003 4805 pod_workers.go:1301] "Error syncing pod, skipping" err="unmounted volumes=[combined-ca-bundle kube-api-access-977nw openstack-config openstack-config-secret], unattached volumes=[], failed to process volumes=[combined-ca-bundle kube-api-access-977nw openstack-config openstack-config-secret]: context canceled" pod="openstack/openstackclient" podUID="8524cabe-a1dd-46c5-98d1-73c121150721" Dec 03 15:01:28 crc kubenswrapper[4805]: I1203 15:01:28.654546 4805 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/openstackclient"] Dec 03 15:01:28 crc kubenswrapper[4805]: I1203 15:01:28.664660 4805 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/openstackclient"] Dec 03 15:01:28 crc kubenswrapper[4805]: I1203 15:01:28.666041 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Dec 03 15:01:28 crc kubenswrapper[4805]: I1203 15:01:28.676101 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstackclient"] Dec 03 15:01:28 crc kubenswrapper[4805]: I1203 15:01:28.705880 4805 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8524cabe-a1dd-46c5-98d1-73c121150721" path="/var/lib/kubelet/pods/8524cabe-a1dd-46c5-98d1-73c121150721/volumes" Dec 03 15:01:28 crc kubenswrapper[4805]: I1203 15:01:28.786987 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8a387e8e-801f-4724-85a6-fbd066455197-combined-ca-bundle\") pod \"openstackclient\" (UID: \"8a387e8e-801f-4724-85a6-fbd066455197\") " pod="openstack/openstackclient" Dec 03 15:01:28 crc kubenswrapper[4805]: I1203 15:01:28.787059 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/8a387e8e-801f-4724-85a6-fbd066455197-openstack-config\") pod \"openstackclient\" (UID: \"8a387e8e-801f-4724-85a6-fbd066455197\") " pod="openstack/openstackclient" Dec 03 15:01:28 crc kubenswrapper[4805]: I1203 15:01:28.787584 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/8a387e8e-801f-4724-85a6-fbd066455197-openstack-config-secret\") pod \"openstackclient\" (UID: \"8a387e8e-801f-4724-85a6-fbd066455197\") " pod="openstack/openstackclient" Dec 03 15:01:28 crc kubenswrapper[4805]: I1203 15:01:28.787730 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ds9p7\" (UniqueName: \"kubernetes.io/projected/8a387e8e-801f-4724-85a6-fbd066455197-kube-api-access-ds9p7\") pod \"openstackclient\" (UID: \"8a387e8e-801f-4724-85a6-fbd066455197\") " pod="openstack/openstackclient" Dec 03 15:01:28 crc kubenswrapper[4805]: I1203 15:01:28.890024 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/8a387e8e-801f-4724-85a6-fbd066455197-openstack-config-secret\") pod \"openstackclient\" (UID: \"8a387e8e-801f-4724-85a6-fbd066455197\") " pod="openstack/openstackclient" Dec 03 15:01:28 crc kubenswrapper[4805]: I1203 15:01:28.890113 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ds9p7\" (UniqueName: \"kubernetes.io/projected/8a387e8e-801f-4724-85a6-fbd066455197-kube-api-access-ds9p7\") pod \"openstackclient\" (UID: \"8a387e8e-801f-4724-85a6-fbd066455197\") " pod="openstack/openstackclient" Dec 03 15:01:28 crc kubenswrapper[4805]: I1203 15:01:28.890194 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8a387e8e-801f-4724-85a6-fbd066455197-combined-ca-bundle\") pod \"openstackclient\" (UID: \"8a387e8e-801f-4724-85a6-fbd066455197\") " pod="openstack/openstackclient" Dec 03 15:01:28 crc kubenswrapper[4805]: I1203 15:01:28.890231 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/8a387e8e-801f-4724-85a6-fbd066455197-openstack-config\") pod \"openstackclient\" (UID: \"8a387e8e-801f-4724-85a6-fbd066455197\") " pod="openstack/openstackclient" Dec 03 15:01:28 crc kubenswrapper[4805]: I1203 15:01:28.891183 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/8a387e8e-801f-4724-85a6-fbd066455197-openstack-config\") pod \"openstackclient\" (UID: \"8a387e8e-801f-4724-85a6-fbd066455197\") " pod="openstack/openstackclient" Dec 03 15:01:28 crc kubenswrapper[4805]: I1203 15:01:28.898545 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8a387e8e-801f-4724-85a6-fbd066455197-combined-ca-bundle\") pod \"openstackclient\" (UID: \"8a387e8e-801f-4724-85a6-fbd066455197\") " pod="openstack/openstackclient" Dec 03 15:01:28 crc kubenswrapper[4805]: I1203 15:01:28.906752 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/8a387e8e-801f-4724-85a6-fbd066455197-openstack-config-secret\") pod \"openstackclient\" (UID: \"8a387e8e-801f-4724-85a6-fbd066455197\") " pod="openstack/openstackclient" Dec 03 15:01:28 crc kubenswrapper[4805]: I1203 15:01:28.909624 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ds9p7\" (UniqueName: \"kubernetes.io/projected/8a387e8e-801f-4724-85a6-fbd066455197-kube-api-access-ds9p7\") pod \"openstackclient\" (UID: \"8a387e8e-801f-4724-85a6-fbd066455197\") " pod="openstack/openstackclient" Dec 03 15:01:29 crc kubenswrapper[4805]: I1203 15:01:29.026012 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Dec 03 15:01:29 crc kubenswrapper[4805]: I1203 15:01:29.276073 4805 generic.go:334] "Generic (PLEG): container finished" podID="8b1d1e18-a325-4512-a3a7-48b21944317f" containerID="723e22ceaaa9d5708c3e9b1d42b7dad6f527afeebd185e8071d4a3c74d8463e4" exitCode=0 Dec 03 15:01:29 crc kubenswrapper[4805]: I1203 15:01:29.276130 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-r9rwr" event={"ID":"8b1d1e18-a325-4512-a3a7-48b21944317f","Type":"ContainerDied","Data":"723e22ceaaa9d5708c3e9b1d42b7dad6f527afeebd185e8071d4a3c74d8463e4"} Dec 03 15:01:29 crc kubenswrapper[4805]: I1203 15:01:29.276388 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Dec 03 15:01:29 crc kubenswrapper[4805]: I1203 15:01:29.288169 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Dec 03 15:01:29 crc kubenswrapper[4805]: I1203 15:01:29.298653 4805 status_manager.go:861] "Pod was deleted and then recreated, skipping status update" pod="openstack/openstackclient" oldPodUID="8524cabe-a1dd-46c5-98d1-73c121150721" podUID="8a387e8e-801f-4724-85a6-fbd066455197" Dec 03 15:01:29 crc kubenswrapper[4805]: I1203 15:01:29.522826 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstackclient"] Dec 03 15:01:29 crc kubenswrapper[4805]: W1203 15:01:29.524516 4805 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod8a387e8e_801f_4724_85a6_fbd066455197.slice/crio-2c4e55d62df5867ccb7179830ee841daafb5b5b9910065fda72a1d17f6142b2e WatchSource:0}: Error finding container 2c4e55d62df5867ccb7179830ee841daafb5b5b9910065fda72a1d17f6142b2e: Status 404 returned error can't find the container with id 2c4e55d62df5867ccb7179830ee841daafb5b5b9910065fda72a1d17f6142b2e Dec 03 15:01:29 crc kubenswrapper[4805]: I1203 15:01:29.694231 4805 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/aodh-db-create-9xnk5"] Dec 03 15:01:29 crc kubenswrapper[4805]: I1203 15:01:29.695464 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-db-create-9xnk5" Dec 03 15:01:29 crc kubenswrapper[4805]: I1203 15:01:29.701743 4805 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/aodh-624d-account-create-update-xd9kw"] Dec 03 15:01:29 crc kubenswrapper[4805]: I1203 15:01:29.705834 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-624d-account-create-update-xd9kw" Dec 03 15:01:29 crc kubenswrapper[4805]: I1203 15:01:29.710402 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"aodh-db-secret" Dec 03 15:01:29 crc kubenswrapper[4805]: I1203 15:01:29.769796 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/aodh-db-create-9xnk5"] Dec 03 15:01:29 crc kubenswrapper[4805]: I1203 15:01:29.809196 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/aodh-624d-account-create-update-xd9kw"] Dec 03 15:01:29 crc kubenswrapper[4805]: I1203 15:01:29.810440 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qh724\" (UniqueName: \"kubernetes.io/projected/ff3965d5-1e07-45d0-acb0-f1f25716821f-kube-api-access-qh724\") pod \"aodh-db-create-9xnk5\" (UID: \"ff3965d5-1e07-45d0-acb0-f1f25716821f\") " pod="openstack/aodh-db-create-9xnk5" Dec 03 15:01:29 crc kubenswrapper[4805]: I1203 15:01:29.810491 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5wwmk\" (UniqueName: \"kubernetes.io/projected/c3131917-28b3-4dd5-a575-7517f266b36c-kube-api-access-5wwmk\") pod \"aodh-624d-account-create-update-xd9kw\" (UID: \"c3131917-28b3-4dd5-a575-7517f266b36c\") " pod="openstack/aodh-624d-account-create-update-xd9kw" Dec 03 15:01:29 crc kubenswrapper[4805]: I1203 15:01:29.810596 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/c3131917-28b3-4dd5-a575-7517f266b36c-operator-scripts\") pod \"aodh-624d-account-create-update-xd9kw\" (UID: \"c3131917-28b3-4dd5-a575-7517f266b36c\") " pod="openstack/aodh-624d-account-create-update-xd9kw" Dec 03 15:01:29 crc kubenswrapper[4805]: I1203 15:01:29.810650 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/ff3965d5-1e07-45d0-acb0-f1f25716821f-operator-scripts\") pod \"aodh-db-create-9xnk5\" (UID: \"ff3965d5-1e07-45d0-acb0-f1f25716821f\") " pod="openstack/aodh-db-create-9xnk5" Dec 03 15:01:29 crc kubenswrapper[4805]: I1203 15:01:29.912279 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qh724\" (UniqueName: \"kubernetes.io/projected/ff3965d5-1e07-45d0-acb0-f1f25716821f-kube-api-access-qh724\") pod \"aodh-db-create-9xnk5\" (UID: \"ff3965d5-1e07-45d0-acb0-f1f25716821f\") " pod="openstack/aodh-db-create-9xnk5" Dec 03 15:01:29 crc kubenswrapper[4805]: I1203 15:01:29.912338 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5wwmk\" (UniqueName: \"kubernetes.io/projected/c3131917-28b3-4dd5-a575-7517f266b36c-kube-api-access-5wwmk\") pod \"aodh-624d-account-create-update-xd9kw\" (UID: \"c3131917-28b3-4dd5-a575-7517f266b36c\") " pod="openstack/aodh-624d-account-create-update-xd9kw" Dec 03 15:01:29 crc kubenswrapper[4805]: I1203 15:01:29.912446 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/c3131917-28b3-4dd5-a575-7517f266b36c-operator-scripts\") pod \"aodh-624d-account-create-update-xd9kw\" (UID: \"c3131917-28b3-4dd5-a575-7517f266b36c\") " pod="openstack/aodh-624d-account-create-update-xd9kw" Dec 03 15:01:29 crc kubenswrapper[4805]: I1203 15:01:29.912513 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/ff3965d5-1e07-45d0-acb0-f1f25716821f-operator-scripts\") pod \"aodh-db-create-9xnk5\" (UID: \"ff3965d5-1e07-45d0-acb0-f1f25716821f\") " pod="openstack/aodh-db-create-9xnk5" Dec 03 15:01:29 crc kubenswrapper[4805]: I1203 15:01:29.913396 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/ff3965d5-1e07-45d0-acb0-f1f25716821f-operator-scripts\") pod \"aodh-db-create-9xnk5\" (UID: \"ff3965d5-1e07-45d0-acb0-f1f25716821f\") " pod="openstack/aodh-db-create-9xnk5" Dec 03 15:01:29 crc kubenswrapper[4805]: I1203 15:01:29.913608 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/c3131917-28b3-4dd5-a575-7517f266b36c-operator-scripts\") pod \"aodh-624d-account-create-update-xd9kw\" (UID: \"c3131917-28b3-4dd5-a575-7517f266b36c\") " pod="openstack/aodh-624d-account-create-update-xd9kw" Dec 03 15:01:29 crc kubenswrapper[4805]: I1203 15:01:29.933585 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5wwmk\" (UniqueName: \"kubernetes.io/projected/c3131917-28b3-4dd5-a575-7517f266b36c-kube-api-access-5wwmk\") pod \"aodh-624d-account-create-update-xd9kw\" (UID: \"c3131917-28b3-4dd5-a575-7517f266b36c\") " pod="openstack/aodh-624d-account-create-update-xd9kw" Dec 03 15:01:29 crc kubenswrapper[4805]: I1203 15:01:29.934453 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qh724\" (UniqueName: \"kubernetes.io/projected/ff3965d5-1e07-45d0-acb0-f1f25716821f-kube-api-access-qh724\") pod \"aodh-db-create-9xnk5\" (UID: \"ff3965d5-1e07-45d0-acb0-f1f25716821f\") " pod="openstack/aodh-db-create-9xnk5" Dec 03 15:01:30 crc kubenswrapper[4805]: I1203 15:01:30.041266 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-db-create-9xnk5" Dec 03 15:01:30 crc kubenswrapper[4805]: I1203 15:01:30.082221 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-624d-account-create-update-xd9kw" Dec 03 15:01:30 crc kubenswrapper[4805]: I1203 15:01:30.294464 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-r9rwr" event={"ID":"8b1d1e18-a325-4512-a3a7-48b21944317f","Type":"ContainerStarted","Data":"b5fcc2bed14aeb90beb03d11f1c1e67a07a6a265bcf42921ec359b37d1e799df"} Dec 03 15:01:30 crc kubenswrapper[4805]: I1203 15:01:30.295975 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Dec 03 15:01:30 crc kubenswrapper[4805]: I1203 15:01:30.296611 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstackclient" event={"ID":"8a387e8e-801f-4724-85a6-fbd066455197","Type":"ContainerStarted","Data":"d2436082da13f8316ca1b0be86c6c49eb13c4416ceecc8b65b3e8a3615ba5f90"} Dec 03 15:01:30 crc kubenswrapper[4805]: I1203 15:01:30.296648 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstackclient" event={"ID":"8a387e8e-801f-4724-85a6-fbd066455197","Type":"ContainerStarted","Data":"2c4e55d62df5867ccb7179830ee841daafb5b5b9910065fda72a1d17f6142b2e"} Dec 03 15:01:30 crc kubenswrapper[4805]: I1203 15:01:30.328122 4805 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-r9rwr" podStartSLOduration=2.8590855939999997 podStartE2EDuration="5.328102556s" podCreationTimestamp="2025-12-03 15:01:25 +0000 UTC" firstStartedPulling="2025-12-03 15:01:27.255258526 +0000 UTC m=+3116.918175449" lastFinishedPulling="2025-12-03 15:01:29.724275488 +0000 UTC m=+3119.387192411" observedRunningTime="2025-12-03 15:01:30.323466622 +0000 UTC m=+3119.986383575" watchObservedRunningTime="2025-12-03 15:01:30.328102556 +0000 UTC m=+3119.991019479" Dec 03 15:01:30 crc kubenswrapper[4805]: I1203 15:01:30.345012 4805 status_manager.go:861] "Pod was deleted and then recreated, skipping status update" pod="openstack/openstackclient" oldPodUID="8524cabe-a1dd-46c5-98d1-73c121150721" podUID="8a387e8e-801f-4724-85a6-fbd066455197" Dec 03 15:01:30 crc kubenswrapper[4805]: I1203 15:01:30.345198 4805 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/openstackclient" podStartSLOduration=2.345187317 podStartE2EDuration="2.345187317s" podCreationTimestamp="2025-12-03 15:01:28 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 15:01:30.342826124 +0000 UTC m=+3120.005743047" watchObservedRunningTime="2025-12-03 15:01:30.345187317 +0000 UTC m=+3120.008104260" Dec 03 15:01:30 crc kubenswrapper[4805]: I1203 15:01:30.571129 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/aodh-db-create-9xnk5"] Dec 03 15:01:30 crc kubenswrapper[4805]: I1203 15:01:30.681207 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/aodh-624d-account-create-update-xd9kw"] Dec 03 15:01:30 crc kubenswrapper[4805]: I1203 15:01:30.736932 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"aodh-db-secret" Dec 03 15:01:30 crc kubenswrapper[4805]: I1203 15:01:30.926067 4805 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Dec 03 15:01:30 crc kubenswrapper[4805]: I1203 15:01:30.928560 4805 status_manager.go:861] "Pod was deleted and then recreated, skipping status update" pod="openstack/openstackclient" oldPodUID="29c53624-ef4b-43ab-8922-181a78908d24" podUID="8a387e8e-801f-4724-85a6-fbd066455197" Dec 03 15:01:31 crc kubenswrapper[4805]: I1203 15:01:31.064987 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9pll7\" (UniqueName: \"kubernetes.io/projected/29c53624-ef4b-43ab-8922-181a78908d24-kube-api-access-9pll7\") pod \"29c53624-ef4b-43ab-8922-181a78908d24\" (UID: \"29c53624-ef4b-43ab-8922-181a78908d24\") " Dec 03 15:01:31 crc kubenswrapper[4805]: I1203 15:01:31.065094 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/29c53624-ef4b-43ab-8922-181a78908d24-openstack-config-secret\") pod \"29c53624-ef4b-43ab-8922-181a78908d24\" (UID: \"29c53624-ef4b-43ab-8922-181a78908d24\") " Dec 03 15:01:31 crc kubenswrapper[4805]: I1203 15:01:31.065126 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/29c53624-ef4b-43ab-8922-181a78908d24-combined-ca-bundle\") pod \"29c53624-ef4b-43ab-8922-181a78908d24\" (UID: \"29c53624-ef4b-43ab-8922-181a78908d24\") " Dec 03 15:01:31 crc kubenswrapper[4805]: I1203 15:01:31.065204 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/29c53624-ef4b-43ab-8922-181a78908d24-openstack-config\") pod \"29c53624-ef4b-43ab-8922-181a78908d24\" (UID: \"29c53624-ef4b-43ab-8922-181a78908d24\") " Dec 03 15:01:31 crc kubenswrapper[4805]: I1203 15:01:31.070999 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/29c53624-ef4b-43ab-8922-181a78908d24-kube-api-access-9pll7" (OuterVolumeSpecName: "kube-api-access-9pll7") pod "29c53624-ef4b-43ab-8922-181a78908d24" (UID: "29c53624-ef4b-43ab-8922-181a78908d24"). InnerVolumeSpecName "kube-api-access-9pll7". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 15:01:31 crc kubenswrapper[4805]: I1203 15:01:31.097557 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/29c53624-ef4b-43ab-8922-181a78908d24-openstack-config" (OuterVolumeSpecName: "openstack-config") pod "29c53624-ef4b-43ab-8922-181a78908d24" (UID: "29c53624-ef4b-43ab-8922-181a78908d24"). InnerVolumeSpecName "openstack-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 15:01:31 crc kubenswrapper[4805]: I1203 15:01:31.117351 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/29c53624-ef4b-43ab-8922-181a78908d24-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "29c53624-ef4b-43ab-8922-181a78908d24" (UID: "29c53624-ef4b-43ab-8922-181a78908d24"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 15:01:31 crc kubenswrapper[4805]: I1203 15:01:31.135225 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/29c53624-ef4b-43ab-8922-181a78908d24-openstack-config-secret" (OuterVolumeSpecName: "openstack-config-secret") pod "29c53624-ef4b-43ab-8922-181a78908d24" (UID: "29c53624-ef4b-43ab-8922-181a78908d24"). InnerVolumeSpecName "openstack-config-secret". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 15:01:31 crc kubenswrapper[4805]: I1203 15:01:31.167634 4805 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9pll7\" (UniqueName: \"kubernetes.io/projected/29c53624-ef4b-43ab-8922-181a78908d24-kube-api-access-9pll7\") on node \"crc\" DevicePath \"\"" Dec 03 15:01:31 crc kubenswrapper[4805]: I1203 15:01:31.167670 4805 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/29c53624-ef4b-43ab-8922-181a78908d24-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 15:01:31 crc kubenswrapper[4805]: I1203 15:01:31.167681 4805 reconciler_common.go:293] "Volume detached for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/29c53624-ef4b-43ab-8922-181a78908d24-openstack-config-secret\") on node \"crc\" DevicePath \"\"" Dec 03 15:01:31 crc kubenswrapper[4805]: I1203 15:01:31.167690 4805 reconciler_common.go:293] "Volume detached for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/29c53624-ef4b-43ab-8922-181a78908d24-openstack-config\") on node \"crc\" DevicePath \"\"" Dec 03 15:01:31 crc kubenswrapper[4805]: I1203 15:01:31.305742 4805 generic.go:334] "Generic (PLEG): container finished" podID="c3131917-28b3-4dd5-a575-7517f266b36c" containerID="802e74b1d3b9b5ea1ce841167f986d4201dcce31e8b8090f7bee58a18f2709b9" exitCode=0 Dec 03 15:01:31 crc kubenswrapper[4805]: I1203 15:01:31.305787 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-624d-account-create-update-xd9kw" event={"ID":"c3131917-28b3-4dd5-a575-7517f266b36c","Type":"ContainerDied","Data":"802e74b1d3b9b5ea1ce841167f986d4201dcce31e8b8090f7bee58a18f2709b9"} Dec 03 15:01:31 crc kubenswrapper[4805]: I1203 15:01:31.305825 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-624d-account-create-update-xd9kw" event={"ID":"c3131917-28b3-4dd5-a575-7517f266b36c","Type":"ContainerStarted","Data":"caacaa7c90459a8f83bbdb43e50f7ecf448d602af5cd306d0d8cc7c84509a798"} Dec 03 15:01:31 crc kubenswrapper[4805]: I1203 15:01:31.307635 4805 generic.go:334] "Generic (PLEG): container finished" podID="ff3965d5-1e07-45d0-acb0-f1f25716821f" containerID="8265f7bd78602bdcbae0a347221e8ffbb0b0c660a3740f8759fbec420e1b17eb" exitCode=0 Dec 03 15:01:31 crc kubenswrapper[4805]: I1203 15:01:31.307732 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-db-create-9xnk5" event={"ID":"ff3965d5-1e07-45d0-acb0-f1f25716821f","Type":"ContainerDied","Data":"8265f7bd78602bdcbae0a347221e8ffbb0b0c660a3740f8759fbec420e1b17eb"} Dec 03 15:01:31 crc kubenswrapper[4805]: I1203 15:01:31.307785 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-db-create-9xnk5" event={"ID":"ff3965d5-1e07-45d0-acb0-f1f25716821f","Type":"ContainerStarted","Data":"86e87e4fc18b682af62b341a513f72d247578922b807c2921809d7e0bcd69eec"} Dec 03 15:01:31 crc kubenswrapper[4805]: I1203 15:01:31.309948 4805 generic.go:334] "Generic (PLEG): container finished" podID="29c53624-ef4b-43ab-8922-181a78908d24" containerID="26262b91874fa88789d54d9ed21dcbd2e4122b6fb279633aaa0a8892e3a82995" exitCode=137 Dec 03 15:01:31 crc kubenswrapper[4805]: I1203 15:01:31.310222 4805 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Dec 03 15:01:31 crc kubenswrapper[4805]: I1203 15:01:31.310244 4805 scope.go:117] "RemoveContainer" containerID="26262b91874fa88789d54d9ed21dcbd2e4122b6fb279633aaa0a8892e3a82995" Dec 03 15:01:31 crc kubenswrapper[4805]: I1203 15:01:31.343343 4805 status_manager.go:861] "Pod was deleted and then recreated, skipping status update" pod="openstack/openstackclient" oldPodUID="29c53624-ef4b-43ab-8922-181a78908d24" podUID="8a387e8e-801f-4724-85a6-fbd066455197" Dec 03 15:01:31 crc kubenswrapper[4805]: I1203 15:01:31.347939 4805 scope.go:117] "RemoveContainer" containerID="26262b91874fa88789d54d9ed21dcbd2e4122b6fb279633aaa0a8892e3a82995" Dec 03 15:01:31 crc kubenswrapper[4805]: E1203 15:01:31.348399 4805 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"26262b91874fa88789d54d9ed21dcbd2e4122b6fb279633aaa0a8892e3a82995\": container with ID starting with 26262b91874fa88789d54d9ed21dcbd2e4122b6fb279633aaa0a8892e3a82995 not found: ID does not exist" containerID="26262b91874fa88789d54d9ed21dcbd2e4122b6fb279633aaa0a8892e3a82995" Dec 03 15:01:31 crc kubenswrapper[4805]: I1203 15:01:31.348430 4805 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"26262b91874fa88789d54d9ed21dcbd2e4122b6fb279633aaa0a8892e3a82995"} err="failed to get container status \"26262b91874fa88789d54d9ed21dcbd2e4122b6fb279633aaa0a8892e3a82995\": rpc error: code = NotFound desc = could not find container \"26262b91874fa88789d54d9ed21dcbd2e4122b6fb279633aaa0a8892e3a82995\": container with ID starting with 26262b91874fa88789d54d9ed21dcbd2e4122b6fb279633aaa0a8892e3a82995 not found: ID does not exist" Dec 03 15:01:32 crc kubenswrapper[4805]: I1203 15:01:32.704858 4805 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="29c53624-ef4b-43ab-8922-181a78908d24" path="/var/lib/kubelet/pods/29c53624-ef4b-43ab-8922-181a78908d24/volumes" Dec 03 15:01:32 crc kubenswrapper[4805]: I1203 15:01:32.736966 4805 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-db-create-9xnk5" Dec 03 15:01:32 crc kubenswrapper[4805]: I1203 15:01:32.747526 4805 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-624d-account-create-update-xd9kw" Dec 03 15:01:32 crc kubenswrapper[4805]: I1203 15:01:32.810363 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5wwmk\" (UniqueName: \"kubernetes.io/projected/c3131917-28b3-4dd5-a575-7517f266b36c-kube-api-access-5wwmk\") pod \"c3131917-28b3-4dd5-a575-7517f266b36c\" (UID: \"c3131917-28b3-4dd5-a575-7517f266b36c\") " Dec 03 15:01:32 crc kubenswrapper[4805]: I1203 15:01:32.810451 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qh724\" (UniqueName: \"kubernetes.io/projected/ff3965d5-1e07-45d0-acb0-f1f25716821f-kube-api-access-qh724\") pod \"ff3965d5-1e07-45d0-acb0-f1f25716821f\" (UID: \"ff3965d5-1e07-45d0-acb0-f1f25716821f\") " Dec 03 15:01:32 crc kubenswrapper[4805]: I1203 15:01:32.810519 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/c3131917-28b3-4dd5-a575-7517f266b36c-operator-scripts\") pod \"c3131917-28b3-4dd5-a575-7517f266b36c\" (UID: \"c3131917-28b3-4dd5-a575-7517f266b36c\") " Dec 03 15:01:32 crc kubenswrapper[4805]: I1203 15:01:32.810545 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/ff3965d5-1e07-45d0-acb0-f1f25716821f-operator-scripts\") pod \"ff3965d5-1e07-45d0-acb0-f1f25716821f\" (UID: \"ff3965d5-1e07-45d0-acb0-f1f25716821f\") " Dec 03 15:01:32 crc kubenswrapper[4805]: I1203 15:01:32.811412 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ff3965d5-1e07-45d0-acb0-f1f25716821f-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "ff3965d5-1e07-45d0-acb0-f1f25716821f" (UID: "ff3965d5-1e07-45d0-acb0-f1f25716821f"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 15:01:32 crc kubenswrapper[4805]: I1203 15:01:32.811609 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c3131917-28b3-4dd5-a575-7517f266b36c-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "c3131917-28b3-4dd5-a575-7517f266b36c" (UID: "c3131917-28b3-4dd5-a575-7517f266b36c"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 15:01:32 crc kubenswrapper[4805]: I1203 15:01:32.815311 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c3131917-28b3-4dd5-a575-7517f266b36c-kube-api-access-5wwmk" (OuterVolumeSpecName: "kube-api-access-5wwmk") pod "c3131917-28b3-4dd5-a575-7517f266b36c" (UID: "c3131917-28b3-4dd5-a575-7517f266b36c"). InnerVolumeSpecName "kube-api-access-5wwmk". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 15:01:32 crc kubenswrapper[4805]: I1203 15:01:32.815922 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ff3965d5-1e07-45d0-acb0-f1f25716821f-kube-api-access-qh724" (OuterVolumeSpecName: "kube-api-access-qh724") pod "ff3965d5-1e07-45d0-acb0-f1f25716821f" (UID: "ff3965d5-1e07-45d0-acb0-f1f25716821f"). InnerVolumeSpecName "kube-api-access-qh724". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 15:01:32 crc kubenswrapper[4805]: I1203 15:01:32.912918 4805 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5wwmk\" (UniqueName: \"kubernetes.io/projected/c3131917-28b3-4dd5-a575-7517f266b36c-kube-api-access-5wwmk\") on node \"crc\" DevicePath \"\"" Dec 03 15:01:32 crc kubenswrapper[4805]: I1203 15:01:32.912968 4805 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qh724\" (UniqueName: \"kubernetes.io/projected/ff3965d5-1e07-45d0-acb0-f1f25716821f-kube-api-access-qh724\") on node \"crc\" DevicePath \"\"" Dec 03 15:01:32 crc kubenswrapper[4805]: I1203 15:01:32.912988 4805 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/c3131917-28b3-4dd5-a575-7517f266b36c-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 03 15:01:32 crc kubenswrapper[4805]: I1203 15:01:32.913006 4805 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/ff3965d5-1e07-45d0-acb0-f1f25716821f-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 03 15:01:33 crc kubenswrapper[4805]: I1203 15:01:33.330737 4805 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-db-create-9xnk5" Dec 03 15:01:33 crc kubenswrapper[4805]: I1203 15:01:33.330732 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-db-create-9xnk5" event={"ID":"ff3965d5-1e07-45d0-acb0-f1f25716821f","Type":"ContainerDied","Data":"86e87e4fc18b682af62b341a513f72d247578922b807c2921809d7e0bcd69eec"} Dec 03 15:01:33 crc kubenswrapper[4805]: I1203 15:01:33.330878 4805 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="86e87e4fc18b682af62b341a513f72d247578922b807c2921809d7e0bcd69eec" Dec 03 15:01:33 crc kubenswrapper[4805]: I1203 15:01:33.332602 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-624d-account-create-update-xd9kw" event={"ID":"c3131917-28b3-4dd5-a575-7517f266b36c","Type":"ContainerDied","Data":"caacaa7c90459a8f83bbdb43e50f7ecf448d602af5cd306d0d8cc7c84509a798"} Dec 03 15:01:33 crc kubenswrapper[4805]: I1203 15:01:33.332627 4805 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="caacaa7c90459a8f83bbdb43e50f7ecf448d602af5cd306d0d8cc7c84509a798" Dec 03 15:01:33 crc kubenswrapper[4805]: I1203 15:01:33.332644 4805 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-624d-account-create-update-xd9kw" Dec 03 15:01:35 crc kubenswrapper[4805]: I1203 15:01:35.179272 4805 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/aodh-db-sync-znm4q"] Dec 03 15:01:35 crc kubenswrapper[4805]: E1203 15:01:35.180660 4805 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ff3965d5-1e07-45d0-acb0-f1f25716821f" containerName="mariadb-database-create" Dec 03 15:01:35 crc kubenswrapper[4805]: I1203 15:01:35.180685 4805 state_mem.go:107] "Deleted CPUSet assignment" podUID="ff3965d5-1e07-45d0-acb0-f1f25716821f" containerName="mariadb-database-create" Dec 03 15:01:35 crc kubenswrapper[4805]: E1203 15:01:35.180702 4805 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c3131917-28b3-4dd5-a575-7517f266b36c" containerName="mariadb-account-create-update" Dec 03 15:01:35 crc kubenswrapper[4805]: I1203 15:01:35.180711 4805 state_mem.go:107] "Deleted CPUSet assignment" podUID="c3131917-28b3-4dd5-a575-7517f266b36c" containerName="mariadb-account-create-update" Dec 03 15:01:35 crc kubenswrapper[4805]: I1203 15:01:35.180973 4805 memory_manager.go:354] "RemoveStaleState removing state" podUID="ff3965d5-1e07-45d0-acb0-f1f25716821f" containerName="mariadb-database-create" Dec 03 15:01:35 crc kubenswrapper[4805]: I1203 15:01:35.181006 4805 memory_manager.go:354] "RemoveStaleState removing state" podUID="c3131917-28b3-4dd5-a575-7517f266b36c" containerName="mariadb-account-create-update" Dec 03 15:01:35 crc kubenswrapper[4805]: I1203 15:01:35.181770 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-db-sync-znm4q" Dec 03 15:01:35 crc kubenswrapper[4805]: I1203 15:01:35.184422 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"osp-secret" Dec 03 15:01:35 crc kubenswrapper[4805]: I1203 15:01:35.185036 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"telemetry-autoscaling-dockercfg-wc8kz" Dec 03 15:01:35 crc kubenswrapper[4805]: I1203 15:01:35.188140 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"aodh-config-data" Dec 03 15:01:35 crc kubenswrapper[4805]: I1203 15:01:35.188188 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"aodh-scripts" Dec 03 15:01:35 crc kubenswrapper[4805]: I1203 15:01:35.189271 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/aodh-db-sync-znm4q"] Dec 03 15:01:35 crc kubenswrapper[4805]: I1203 15:01:35.276064 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/07c5b863-2051-46da-b63b-87210297c6cb-config-data\") pod \"aodh-db-sync-znm4q\" (UID: \"07c5b863-2051-46da-b63b-87210297c6cb\") " pod="openstack/aodh-db-sync-znm4q" Dec 03 15:01:35 crc kubenswrapper[4805]: I1203 15:01:35.276161 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-n4fbx\" (UniqueName: \"kubernetes.io/projected/07c5b863-2051-46da-b63b-87210297c6cb-kube-api-access-n4fbx\") pod \"aodh-db-sync-znm4q\" (UID: \"07c5b863-2051-46da-b63b-87210297c6cb\") " pod="openstack/aodh-db-sync-znm4q" Dec 03 15:01:35 crc kubenswrapper[4805]: I1203 15:01:35.276212 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/07c5b863-2051-46da-b63b-87210297c6cb-combined-ca-bundle\") pod \"aodh-db-sync-znm4q\" (UID: \"07c5b863-2051-46da-b63b-87210297c6cb\") " pod="openstack/aodh-db-sync-znm4q" Dec 03 15:01:35 crc kubenswrapper[4805]: I1203 15:01:35.276246 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/07c5b863-2051-46da-b63b-87210297c6cb-scripts\") pod \"aodh-db-sync-znm4q\" (UID: \"07c5b863-2051-46da-b63b-87210297c6cb\") " pod="openstack/aodh-db-sync-znm4q" Dec 03 15:01:35 crc kubenswrapper[4805]: I1203 15:01:35.377645 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-n4fbx\" (UniqueName: \"kubernetes.io/projected/07c5b863-2051-46da-b63b-87210297c6cb-kube-api-access-n4fbx\") pod \"aodh-db-sync-znm4q\" (UID: \"07c5b863-2051-46da-b63b-87210297c6cb\") " pod="openstack/aodh-db-sync-znm4q" Dec 03 15:01:35 crc kubenswrapper[4805]: I1203 15:01:35.377722 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/07c5b863-2051-46da-b63b-87210297c6cb-combined-ca-bundle\") pod \"aodh-db-sync-znm4q\" (UID: \"07c5b863-2051-46da-b63b-87210297c6cb\") " pod="openstack/aodh-db-sync-znm4q" Dec 03 15:01:35 crc kubenswrapper[4805]: I1203 15:01:35.377756 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/07c5b863-2051-46da-b63b-87210297c6cb-scripts\") pod \"aodh-db-sync-znm4q\" (UID: \"07c5b863-2051-46da-b63b-87210297c6cb\") " pod="openstack/aodh-db-sync-znm4q" Dec 03 15:01:35 crc kubenswrapper[4805]: I1203 15:01:35.377819 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/07c5b863-2051-46da-b63b-87210297c6cb-config-data\") pod \"aodh-db-sync-znm4q\" (UID: \"07c5b863-2051-46da-b63b-87210297c6cb\") " pod="openstack/aodh-db-sync-znm4q" Dec 03 15:01:35 crc kubenswrapper[4805]: I1203 15:01:35.387831 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/07c5b863-2051-46da-b63b-87210297c6cb-scripts\") pod \"aodh-db-sync-znm4q\" (UID: \"07c5b863-2051-46da-b63b-87210297c6cb\") " pod="openstack/aodh-db-sync-znm4q" Dec 03 15:01:35 crc kubenswrapper[4805]: I1203 15:01:35.399740 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/07c5b863-2051-46da-b63b-87210297c6cb-config-data\") pod \"aodh-db-sync-znm4q\" (UID: \"07c5b863-2051-46da-b63b-87210297c6cb\") " pod="openstack/aodh-db-sync-znm4q" Dec 03 15:01:35 crc kubenswrapper[4805]: I1203 15:01:35.400030 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/07c5b863-2051-46da-b63b-87210297c6cb-combined-ca-bundle\") pod \"aodh-db-sync-znm4q\" (UID: \"07c5b863-2051-46da-b63b-87210297c6cb\") " pod="openstack/aodh-db-sync-znm4q" Dec 03 15:01:35 crc kubenswrapper[4805]: I1203 15:01:35.405581 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-n4fbx\" (UniqueName: \"kubernetes.io/projected/07c5b863-2051-46da-b63b-87210297c6cb-kube-api-access-n4fbx\") pod \"aodh-db-sync-znm4q\" (UID: \"07c5b863-2051-46da-b63b-87210297c6cb\") " pod="openstack/aodh-db-sync-znm4q" Dec 03 15:01:35 crc kubenswrapper[4805]: I1203 15:01:35.516069 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-db-sync-znm4q" Dec 03 15:01:35 crc kubenswrapper[4805]: I1203 15:01:35.872655 4805 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-r9rwr" Dec 03 15:01:35 crc kubenswrapper[4805]: I1203 15:01:35.873056 4805 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-r9rwr" Dec 03 15:01:35 crc kubenswrapper[4805]: I1203 15:01:35.927625 4805 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-r9rwr" Dec 03 15:01:35 crc kubenswrapper[4805]: I1203 15:01:35.985480 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/aodh-db-sync-znm4q"] Dec 03 15:01:36 crc kubenswrapper[4805]: I1203 15:01:36.361617 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-db-sync-znm4q" event={"ID":"07c5b863-2051-46da-b63b-87210297c6cb","Type":"ContainerStarted","Data":"37b43f37ab0a1aebb1cabb07b93a414ae67eefe503881c5e206fef56a4fcad7a"} Dec 03 15:01:36 crc kubenswrapper[4805]: I1203 15:01:36.418396 4805 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-r9rwr" Dec 03 15:01:36 crc kubenswrapper[4805]: I1203 15:01:36.463210 4805 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-r9rwr"] Dec 03 15:01:38 crc kubenswrapper[4805]: I1203 15:01:38.383981 4805 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-r9rwr" podUID="8b1d1e18-a325-4512-a3a7-48b21944317f" containerName="registry-server" containerID="cri-o://b5fcc2bed14aeb90beb03d11f1c1e67a07a6a265bcf42921ec359b37d1e799df" gracePeriod=2 Dec 03 15:01:39 crc kubenswrapper[4805]: I1203 15:01:39.400373 4805 generic.go:334] "Generic (PLEG): container finished" podID="8b1d1e18-a325-4512-a3a7-48b21944317f" containerID="b5fcc2bed14aeb90beb03d11f1c1e67a07a6a265bcf42921ec359b37d1e799df" exitCode=0 Dec 03 15:01:39 crc kubenswrapper[4805]: I1203 15:01:39.400461 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-r9rwr" event={"ID":"8b1d1e18-a325-4512-a3a7-48b21944317f","Type":"ContainerDied","Data":"b5fcc2bed14aeb90beb03d11f1c1e67a07a6a265bcf42921ec359b37d1e799df"} Dec 03 15:01:40 crc kubenswrapper[4805]: I1203 15:01:40.364635 4805 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-r9rwr" Dec 03 15:01:40 crc kubenswrapper[4805]: I1203 15:01:40.414727 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-db-sync-znm4q" event={"ID":"07c5b863-2051-46da-b63b-87210297c6cb","Type":"ContainerStarted","Data":"0746524baeba3bd134675117e4c1d81496a866fffa6dd8f2fc7b3675e9635340"} Dec 03 15:01:40 crc kubenswrapper[4805]: I1203 15:01:40.418343 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-r9rwr" event={"ID":"8b1d1e18-a325-4512-a3a7-48b21944317f","Type":"ContainerDied","Data":"fbd0cfd7d0edac39f2ff8f162ef664f3744ae3b4858a96d7c368f5e8e3b09efd"} Dec 03 15:01:40 crc kubenswrapper[4805]: I1203 15:01:40.418402 4805 scope.go:117] "RemoveContainer" containerID="b5fcc2bed14aeb90beb03d11f1c1e67a07a6a265bcf42921ec359b37d1e799df" Dec 03 15:01:40 crc kubenswrapper[4805]: I1203 15:01:40.418514 4805 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-r9rwr" Dec 03 15:01:40 crc kubenswrapper[4805]: I1203 15:01:40.459467 4805 scope.go:117] "RemoveContainer" containerID="723e22ceaaa9d5708c3e9b1d42b7dad6f527afeebd185e8071d4a3c74d8463e4" Dec 03 15:01:40 crc kubenswrapper[4805]: I1203 15:01:40.488939 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8w8kj\" (UniqueName: \"kubernetes.io/projected/8b1d1e18-a325-4512-a3a7-48b21944317f-kube-api-access-8w8kj\") pod \"8b1d1e18-a325-4512-a3a7-48b21944317f\" (UID: \"8b1d1e18-a325-4512-a3a7-48b21944317f\") " Dec 03 15:01:40 crc kubenswrapper[4805]: I1203 15:01:40.489094 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8b1d1e18-a325-4512-a3a7-48b21944317f-catalog-content\") pod \"8b1d1e18-a325-4512-a3a7-48b21944317f\" (UID: \"8b1d1e18-a325-4512-a3a7-48b21944317f\") " Dec 03 15:01:40 crc kubenswrapper[4805]: I1203 15:01:40.489150 4805 scope.go:117] "RemoveContainer" containerID="da1d10a688034fb07bec11aa970653b7d9535e3cf6ba9e895a9876a6ba6660fd" Dec 03 15:01:40 crc kubenswrapper[4805]: I1203 15:01:40.489156 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8b1d1e18-a325-4512-a3a7-48b21944317f-utilities\") pod \"8b1d1e18-a325-4512-a3a7-48b21944317f\" (UID: \"8b1d1e18-a325-4512-a3a7-48b21944317f\") " Dec 03 15:01:40 crc kubenswrapper[4805]: I1203 15:01:40.489832 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8b1d1e18-a325-4512-a3a7-48b21944317f-utilities" (OuterVolumeSpecName: "utilities") pod "8b1d1e18-a325-4512-a3a7-48b21944317f" (UID: "8b1d1e18-a325-4512-a3a7-48b21944317f"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 15:01:40 crc kubenswrapper[4805]: I1203 15:01:40.490230 4805 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8b1d1e18-a325-4512-a3a7-48b21944317f-utilities\") on node \"crc\" DevicePath \"\"" Dec 03 15:01:40 crc kubenswrapper[4805]: I1203 15:01:40.496014 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8b1d1e18-a325-4512-a3a7-48b21944317f-kube-api-access-8w8kj" (OuterVolumeSpecName: "kube-api-access-8w8kj") pod "8b1d1e18-a325-4512-a3a7-48b21944317f" (UID: "8b1d1e18-a325-4512-a3a7-48b21944317f"). InnerVolumeSpecName "kube-api-access-8w8kj". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 15:01:40 crc kubenswrapper[4805]: I1203 15:01:40.538805 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8b1d1e18-a325-4512-a3a7-48b21944317f-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "8b1d1e18-a325-4512-a3a7-48b21944317f" (UID: "8b1d1e18-a325-4512-a3a7-48b21944317f"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 15:01:40 crc kubenswrapper[4805]: I1203 15:01:40.591504 4805 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8w8kj\" (UniqueName: \"kubernetes.io/projected/8b1d1e18-a325-4512-a3a7-48b21944317f-kube-api-access-8w8kj\") on node \"crc\" DevicePath \"\"" Dec 03 15:01:40 crc kubenswrapper[4805]: I1203 15:01:40.591538 4805 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8b1d1e18-a325-4512-a3a7-48b21944317f-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 03 15:01:40 crc kubenswrapper[4805]: E1203 15:01:40.600646 4805 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod29c53624_ef4b_43ab_8922_181a78908d24.slice/crio-conmon-26262b91874fa88789d54d9ed21dcbd2e4122b6fb279633aaa0a8892e3a82995.scope\": RecentStats: unable to find data in memory cache]" Dec 03 15:01:40 crc kubenswrapper[4805]: I1203 15:01:40.762703 4805 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/aodh-db-sync-znm4q" podStartSLOduration=1.653585687 podStartE2EDuration="5.762676793s" podCreationTimestamp="2025-12-03 15:01:35 +0000 UTC" firstStartedPulling="2025-12-03 15:01:35.999279417 +0000 UTC m=+3125.662196340" lastFinishedPulling="2025-12-03 15:01:40.108370523 +0000 UTC m=+3129.771287446" observedRunningTime="2025-12-03 15:01:40.455701447 +0000 UTC m=+3130.118618380" watchObservedRunningTime="2025-12-03 15:01:40.762676793 +0000 UTC m=+3130.425593726" Dec 03 15:01:40 crc kubenswrapper[4805]: I1203 15:01:40.769296 4805 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-r9rwr"] Dec 03 15:01:40 crc kubenswrapper[4805]: I1203 15:01:40.777592 4805 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-r9rwr"] Dec 03 15:01:42 crc kubenswrapper[4805]: I1203 15:01:42.706702 4805 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8b1d1e18-a325-4512-a3a7-48b21944317f" path="/var/lib/kubelet/pods/8b1d1e18-a325-4512-a3a7-48b21944317f/volumes" Dec 03 15:01:43 crc kubenswrapper[4805]: I1203 15:01:43.482890 4805 generic.go:334] "Generic (PLEG): container finished" podID="07c5b863-2051-46da-b63b-87210297c6cb" containerID="0746524baeba3bd134675117e4c1d81496a866fffa6dd8f2fc7b3675e9635340" exitCode=0 Dec 03 15:01:43 crc kubenswrapper[4805]: I1203 15:01:43.482952 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-db-sync-znm4q" event={"ID":"07c5b863-2051-46da-b63b-87210297c6cb","Type":"ContainerDied","Data":"0746524baeba3bd134675117e4c1d81496a866fffa6dd8f2fc7b3675e9635340"} Dec 03 15:01:43 crc kubenswrapper[4805]: I1203 15:01:43.917910 4805 patch_prober.go:28] interesting pod/machine-config-daemon-gskh4 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 03 15:01:43 crc kubenswrapper[4805]: I1203 15:01:43.917973 4805 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-gskh4" podUID="6098937f-e3f6-45e8-a647-4994a79cd711" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 03 15:01:43 crc kubenswrapper[4805]: I1203 15:01:43.918015 4805 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-gskh4" Dec 03 15:01:43 crc kubenswrapper[4805]: I1203 15:01:43.918728 4805 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"13e0da4d7fc8f4085be601d845d688938c16ea7941e52ca511e39a37d6e0ba4e"} pod="openshift-machine-config-operator/machine-config-daemon-gskh4" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 03 15:01:43 crc kubenswrapper[4805]: I1203 15:01:43.918776 4805 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-gskh4" podUID="6098937f-e3f6-45e8-a647-4994a79cd711" containerName="machine-config-daemon" containerID="cri-o://13e0da4d7fc8f4085be601d845d688938c16ea7941e52ca511e39a37d6e0ba4e" gracePeriod=600 Dec 03 15:01:44 crc kubenswrapper[4805]: E1203 15:01:44.045610 4805 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-gskh4_openshift-machine-config-operator(6098937f-e3f6-45e8-a647-4994a79cd711)\"" pod="openshift-machine-config-operator/machine-config-daemon-gskh4" podUID="6098937f-e3f6-45e8-a647-4994a79cd711" Dec 03 15:01:44 crc kubenswrapper[4805]: I1203 15:01:44.494530 4805 generic.go:334] "Generic (PLEG): container finished" podID="6098937f-e3f6-45e8-a647-4994a79cd711" containerID="13e0da4d7fc8f4085be601d845d688938c16ea7941e52ca511e39a37d6e0ba4e" exitCode=0 Dec 03 15:01:44 crc kubenswrapper[4805]: I1203 15:01:44.494615 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-gskh4" event={"ID":"6098937f-e3f6-45e8-a647-4994a79cd711","Type":"ContainerDied","Data":"13e0da4d7fc8f4085be601d845d688938c16ea7941e52ca511e39a37d6e0ba4e"} Dec 03 15:01:44 crc kubenswrapper[4805]: I1203 15:01:44.495004 4805 scope.go:117] "RemoveContainer" containerID="2e3f74f03f0ba2719087d9ca75ea7670c41a3ad116d1af23fbb1aeb3c76a97d3" Dec 03 15:01:44 crc kubenswrapper[4805]: I1203 15:01:44.495905 4805 scope.go:117] "RemoveContainer" containerID="13e0da4d7fc8f4085be601d845d688938c16ea7941e52ca511e39a37d6e0ba4e" Dec 03 15:01:44 crc kubenswrapper[4805]: E1203 15:01:44.496257 4805 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-gskh4_openshift-machine-config-operator(6098937f-e3f6-45e8-a647-4994a79cd711)\"" pod="openshift-machine-config-operator/machine-config-daemon-gskh4" podUID="6098937f-e3f6-45e8-a647-4994a79cd711" Dec 03 15:01:44 crc kubenswrapper[4805]: I1203 15:01:44.867931 4805 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-db-sync-znm4q" Dec 03 15:01:44 crc kubenswrapper[4805]: I1203 15:01:44.980926 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/07c5b863-2051-46da-b63b-87210297c6cb-config-data\") pod \"07c5b863-2051-46da-b63b-87210297c6cb\" (UID: \"07c5b863-2051-46da-b63b-87210297c6cb\") " Dec 03 15:01:44 crc kubenswrapper[4805]: I1203 15:01:44.980990 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/07c5b863-2051-46da-b63b-87210297c6cb-scripts\") pod \"07c5b863-2051-46da-b63b-87210297c6cb\" (UID: \"07c5b863-2051-46da-b63b-87210297c6cb\") " Dec 03 15:01:44 crc kubenswrapper[4805]: I1203 15:01:44.981053 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/07c5b863-2051-46da-b63b-87210297c6cb-combined-ca-bundle\") pod \"07c5b863-2051-46da-b63b-87210297c6cb\" (UID: \"07c5b863-2051-46da-b63b-87210297c6cb\") " Dec 03 15:01:44 crc kubenswrapper[4805]: I1203 15:01:44.981147 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-n4fbx\" (UniqueName: \"kubernetes.io/projected/07c5b863-2051-46da-b63b-87210297c6cb-kube-api-access-n4fbx\") pod \"07c5b863-2051-46da-b63b-87210297c6cb\" (UID: \"07c5b863-2051-46da-b63b-87210297c6cb\") " Dec 03 15:01:44 crc kubenswrapper[4805]: I1203 15:01:44.987190 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/07c5b863-2051-46da-b63b-87210297c6cb-kube-api-access-n4fbx" (OuterVolumeSpecName: "kube-api-access-n4fbx") pod "07c5b863-2051-46da-b63b-87210297c6cb" (UID: "07c5b863-2051-46da-b63b-87210297c6cb"). InnerVolumeSpecName "kube-api-access-n4fbx". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 15:01:44 crc kubenswrapper[4805]: I1203 15:01:44.987423 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/07c5b863-2051-46da-b63b-87210297c6cb-scripts" (OuterVolumeSpecName: "scripts") pod "07c5b863-2051-46da-b63b-87210297c6cb" (UID: "07c5b863-2051-46da-b63b-87210297c6cb"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 15:01:45 crc kubenswrapper[4805]: I1203 15:01:45.012248 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/07c5b863-2051-46da-b63b-87210297c6cb-config-data" (OuterVolumeSpecName: "config-data") pod "07c5b863-2051-46da-b63b-87210297c6cb" (UID: "07c5b863-2051-46da-b63b-87210297c6cb"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 15:01:45 crc kubenswrapper[4805]: I1203 15:01:45.026722 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/07c5b863-2051-46da-b63b-87210297c6cb-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "07c5b863-2051-46da-b63b-87210297c6cb" (UID: "07c5b863-2051-46da-b63b-87210297c6cb"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 15:01:45 crc kubenswrapper[4805]: I1203 15:01:45.083855 4805 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/07c5b863-2051-46da-b63b-87210297c6cb-config-data\") on node \"crc\" DevicePath \"\"" Dec 03 15:01:45 crc kubenswrapper[4805]: I1203 15:01:45.083889 4805 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/07c5b863-2051-46da-b63b-87210297c6cb-scripts\") on node \"crc\" DevicePath \"\"" Dec 03 15:01:45 crc kubenswrapper[4805]: I1203 15:01:45.083898 4805 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/07c5b863-2051-46da-b63b-87210297c6cb-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 15:01:45 crc kubenswrapper[4805]: I1203 15:01:45.083907 4805 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-n4fbx\" (UniqueName: \"kubernetes.io/projected/07c5b863-2051-46da-b63b-87210297c6cb-kube-api-access-n4fbx\") on node \"crc\" DevicePath \"\"" Dec 03 15:01:45 crc kubenswrapper[4805]: I1203 15:01:45.521812 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-db-sync-znm4q" event={"ID":"07c5b863-2051-46da-b63b-87210297c6cb","Type":"ContainerDied","Data":"37b43f37ab0a1aebb1cabb07b93a414ae67eefe503881c5e206fef56a4fcad7a"} Dec 03 15:01:45 crc kubenswrapper[4805]: I1203 15:01:45.521880 4805 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="37b43f37ab0a1aebb1cabb07b93a414ae67eefe503881c5e206fef56a4fcad7a" Dec 03 15:01:45 crc kubenswrapper[4805]: I1203 15:01:45.521952 4805 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-db-sync-znm4q" Dec 03 15:01:49 crc kubenswrapper[4805]: I1203 15:01:49.691976 4805 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/aodh-0"] Dec 03 15:01:49 crc kubenswrapper[4805]: E1203 15:01:49.692918 4805 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="07c5b863-2051-46da-b63b-87210297c6cb" containerName="aodh-db-sync" Dec 03 15:01:49 crc kubenswrapper[4805]: I1203 15:01:49.692941 4805 state_mem.go:107] "Deleted CPUSet assignment" podUID="07c5b863-2051-46da-b63b-87210297c6cb" containerName="aodh-db-sync" Dec 03 15:01:49 crc kubenswrapper[4805]: E1203 15:01:49.692977 4805 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8b1d1e18-a325-4512-a3a7-48b21944317f" containerName="registry-server" Dec 03 15:01:49 crc kubenswrapper[4805]: I1203 15:01:49.692989 4805 state_mem.go:107] "Deleted CPUSet assignment" podUID="8b1d1e18-a325-4512-a3a7-48b21944317f" containerName="registry-server" Dec 03 15:01:49 crc kubenswrapper[4805]: E1203 15:01:49.693030 4805 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8b1d1e18-a325-4512-a3a7-48b21944317f" containerName="extract-content" Dec 03 15:01:49 crc kubenswrapper[4805]: I1203 15:01:49.693040 4805 state_mem.go:107] "Deleted CPUSet assignment" podUID="8b1d1e18-a325-4512-a3a7-48b21944317f" containerName="extract-content" Dec 03 15:01:49 crc kubenswrapper[4805]: E1203 15:01:49.693072 4805 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8b1d1e18-a325-4512-a3a7-48b21944317f" containerName="extract-utilities" Dec 03 15:01:49 crc kubenswrapper[4805]: I1203 15:01:49.693083 4805 state_mem.go:107] "Deleted CPUSet assignment" podUID="8b1d1e18-a325-4512-a3a7-48b21944317f" containerName="extract-utilities" Dec 03 15:01:49 crc kubenswrapper[4805]: I1203 15:01:49.693591 4805 memory_manager.go:354] "RemoveStaleState removing state" podUID="07c5b863-2051-46da-b63b-87210297c6cb" containerName="aodh-db-sync" Dec 03 15:01:49 crc kubenswrapper[4805]: I1203 15:01:49.693616 4805 memory_manager.go:354] "RemoveStaleState removing state" podUID="8b1d1e18-a325-4512-a3a7-48b21944317f" containerName="registry-server" Dec 03 15:01:49 crc kubenswrapper[4805]: I1203 15:01:49.700816 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-0" Dec 03 15:01:49 crc kubenswrapper[4805]: I1203 15:01:49.706763 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"aodh-config-data" Dec 03 15:01:49 crc kubenswrapper[4805]: I1203 15:01:49.707113 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"telemetry-autoscaling-dockercfg-wc8kz" Dec 03 15:01:49 crc kubenswrapper[4805]: I1203 15:01:49.707319 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"aodh-scripts" Dec 03 15:01:49 crc kubenswrapper[4805]: I1203 15:01:49.707635 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/aodh-0"] Dec 03 15:01:49 crc kubenswrapper[4805]: I1203 15:01:49.783141 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-m5mbv\" (UniqueName: \"kubernetes.io/projected/cfdd6465-a0a8-4dc4-ad55-b367cb27a2b5-kube-api-access-m5mbv\") pod \"aodh-0\" (UID: \"cfdd6465-a0a8-4dc4-ad55-b367cb27a2b5\") " pod="openstack/aodh-0" Dec 03 15:01:49 crc kubenswrapper[4805]: I1203 15:01:49.783231 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cfdd6465-a0a8-4dc4-ad55-b367cb27a2b5-config-data\") pod \"aodh-0\" (UID: \"cfdd6465-a0a8-4dc4-ad55-b367cb27a2b5\") " pod="openstack/aodh-0" Dec 03 15:01:49 crc kubenswrapper[4805]: I1203 15:01:49.783252 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cfdd6465-a0a8-4dc4-ad55-b367cb27a2b5-combined-ca-bundle\") pod \"aodh-0\" (UID: \"cfdd6465-a0a8-4dc4-ad55-b367cb27a2b5\") " pod="openstack/aodh-0" Dec 03 15:01:49 crc kubenswrapper[4805]: I1203 15:01:49.783292 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/cfdd6465-a0a8-4dc4-ad55-b367cb27a2b5-scripts\") pod \"aodh-0\" (UID: \"cfdd6465-a0a8-4dc4-ad55-b367cb27a2b5\") " pod="openstack/aodh-0" Dec 03 15:01:49 crc kubenswrapper[4805]: I1203 15:01:49.884765 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/cfdd6465-a0a8-4dc4-ad55-b367cb27a2b5-scripts\") pod \"aodh-0\" (UID: \"cfdd6465-a0a8-4dc4-ad55-b367cb27a2b5\") " pod="openstack/aodh-0" Dec 03 15:01:49 crc kubenswrapper[4805]: I1203 15:01:49.885006 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-m5mbv\" (UniqueName: \"kubernetes.io/projected/cfdd6465-a0a8-4dc4-ad55-b367cb27a2b5-kube-api-access-m5mbv\") pod \"aodh-0\" (UID: \"cfdd6465-a0a8-4dc4-ad55-b367cb27a2b5\") " pod="openstack/aodh-0" Dec 03 15:01:49 crc kubenswrapper[4805]: I1203 15:01:49.885073 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cfdd6465-a0a8-4dc4-ad55-b367cb27a2b5-config-data\") pod \"aodh-0\" (UID: \"cfdd6465-a0a8-4dc4-ad55-b367cb27a2b5\") " pod="openstack/aodh-0" Dec 03 15:01:49 crc kubenswrapper[4805]: I1203 15:01:49.885097 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cfdd6465-a0a8-4dc4-ad55-b367cb27a2b5-combined-ca-bundle\") pod \"aodh-0\" (UID: \"cfdd6465-a0a8-4dc4-ad55-b367cb27a2b5\") " pod="openstack/aodh-0" Dec 03 15:01:49 crc kubenswrapper[4805]: I1203 15:01:49.892329 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cfdd6465-a0a8-4dc4-ad55-b367cb27a2b5-combined-ca-bundle\") pod \"aodh-0\" (UID: \"cfdd6465-a0a8-4dc4-ad55-b367cb27a2b5\") " pod="openstack/aodh-0" Dec 03 15:01:49 crc kubenswrapper[4805]: I1203 15:01:49.898995 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cfdd6465-a0a8-4dc4-ad55-b367cb27a2b5-config-data\") pod \"aodh-0\" (UID: \"cfdd6465-a0a8-4dc4-ad55-b367cb27a2b5\") " pod="openstack/aodh-0" Dec 03 15:01:49 crc kubenswrapper[4805]: I1203 15:01:49.903210 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/cfdd6465-a0a8-4dc4-ad55-b367cb27a2b5-scripts\") pod \"aodh-0\" (UID: \"cfdd6465-a0a8-4dc4-ad55-b367cb27a2b5\") " pod="openstack/aodh-0" Dec 03 15:01:49 crc kubenswrapper[4805]: I1203 15:01:49.909386 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-m5mbv\" (UniqueName: \"kubernetes.io/projected/cfdd6465-a0a8-4dc4-ad55-b367cb27a2b5-kube-api-access-m5mbv\") pod \"aodh-0\" (UID: \"cfdd6465-a0a8-4dc4-ad55-b367cb27a2b5\") " pod="openstack/aodh-0" Dec 03 15:01:50 crc kubenswrapper[4805]: I1203 15:01:50.029772 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-0" Dec 03 15:01:50 crc kubenswrapper[4805]: I1203 15:01:50.533085 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/aodh-0"] Dec 03 15:01:50 crc kubenswrapper[4805]: W1203 15:01:50.547623 4805 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podcfdd6465_a0a8_4dc4_ad55_b367cb27a2b5.slice/crio-597774ad530dc25a7e74d08fd60651d95945c6fe92c220f04e0a3a414897b673 WatchSource:0}: Error finding container 597774ad530dc25a7e74d08fd60651d95945c6fe92c220f04e0a3a414897b673: Status 404 returned error can't find the container with id 597774ad530dc25a7e74d08fd60651d95945c6fe92c220f04e0a3a414897b673 Dec 03 15:01:50 crc kubenswrapper[4805]: I1203 15:01:50.574147 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-0" event={"ID":"cfdd6465-a0a8-4dc4-ad55-b367cb27a2b5","Type":"ContainerStarted","Data":"597774ad530dc25a7e74d08fd60651d95945c6fe92c220f04e0a3a414897b673"} Dec 03 15:01:50 crc kubenswrapper[4805]: E1203 15:01:50.860495 4805 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod29c53624_ef4b_43ab_8922_181a78908d24.slice/crio-conmon-26262b91874fa88789d54d9ed21dcbd2e4122b6fb279633aaa0a8892e3a82995.scope\": RecentStats: unable to find data in memory cache]" Dec 03 15:01:51 crc kubenswrapper[4805]: I1203 15:01:51.625168 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-0" event={"ID":"cfdd6465-a0a8-4dc4-ad55-b367cb27a2b5","Type":"ContainerStarted","Data":"962645f025d997f376be2e9db91f639ab21b690752350e7f3ad6d738ea6e7456"} Dec 03 15:01:52 crc kubenswrapper[4805]: I1203 15:01:52.176854 4805 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 03 15:01:52 crc kubenswrapper[4805]: I1203 15:01:52.177122 4805 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="aee26d6d-7304-4001-90f8-c0ed886a3cda" containerName="sg-core" containerID="cri-o://f62e2a86bdda5a487b4221e167d77f2dbb49598c46ba7c129af938f19abfce52" gracePeriod=30 Dec 03 15:01:52 crc kubenswrapper[4805]: I1203 15:01:52.177263 4805 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="aee26d6d-7304-4001-90f8-c0ed886a3cda" containerName="ceilometer-notification-agent" containerID="cri-o://5166a2fd0b33128b5411b9ce35f22d267a1f26fddd2c64c87f9709551024044c" gracePeriod=30 Dec 03 15:01:52 crc kubenswrapper[4805]: I1203 15:01:52.177307 4805 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="aee26d6d-7304-4001-90f8-c0ed886a3cda" containerName="ceilometer-central-agent" containerID="cri-o://c5da23817a0fadd49f2754c6b0126b250c6c2cee661c805c0a922ca37d9d5a30" gracePeriod=30 Dec 03 15:01:52 crc kubenswrapper[4805]: I1203 15:01:52.177359 4805 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="aee26d6d-7304-4001-90f8-c0ed886a3cda" containerName="proxy-httpd" containerID="cri-o://2cf776a26e533c761ac54e168d3c4825a871fbc428f01ab40780a9f315b80d05" gracePeriod=30 Dec 03 15:01:52 crc kubenswrapper[4805]: I1203 15:01:52.608884 4805 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/aodh-0"] Dec 03 15:01:52 crc kubenswrapper[4805]: I1203 15:01:52.637299 4805 generic.go:334] "Generic (PLEG): container finished" podID="aee26d6d-7304-4001-90f8-c0ed886a3cda" containerID="c5da23817a0fadd49f2754c6b0126b250c6c2cee661c805c0a922ca37d9d5a30" exitCode=0 Dec 03 15:01:52 crc kubenswrapper[4805]: I1203 15:01:52.637338 4805 generic.go:334] "Generic (PLEG): container finished" podID="aee26d6d-7304-4001-90f8-c0ed886a3cda" containerID="2cf776a26e533c761ac54e168d3c4825a871fbc428f01ab40780a9f315b80d05" exitCode=0 Dec 03 15:01:52 crc kubenswrapper[4805]: I1203 15:01:52.637349 4805 generic.go:334] "Generic (PLEG): container finished" podID="aee26d6d-7304-4001-90f8-c0ed886a3cda" containerID="f62e2a86bdda5a487b4221e167d77f2dbb49598c46ba7c129af938f19abfce52" exitCode=2 Dec 03 15:01:52 crc kubenswrapper[4805]: I1203 15:01:52.637372 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"aee26d6d-7304-4001-90f8-c0ed886a3cda","Type":"ContainerDied","Data":"c5da23817a0fadd49f2754c6b0126b250c6c2cee661c805c0a922ca37d9d5a30"} Dec 03 15:01:52 crc kubenswrapper[4805]: I1203 15:01:52.637407 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"aee26d6d-7304-4001-90f8-c0ed886a3cda","Type":"ContainerDied","Data":"2cf776a26e533c761ac54e168d3c4825a871fbc428f01ab40780a9f315b80d05"} Dec 03 15:01:52 crc kubenswrapper[4805]: I1203 15:01:52.637422 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"aee26d6d-7304-4001-90f8-c0ed886a3cda","Type":"ContainerDied","Data":"f62e2a86bdda5a487b4221e167d77f2dbb49598c46ba7c129af938f19abfce52"} Dec 03 15:01:52 crc kubenswrapper[4805]: I1203 15:01:52.637441 4805 scope.go:117] "RemoveContainer" containerID="73941e0d6546b90dd2830c242d337c821bfefb112e9c263e904f7c35bfea49e4" Dec 03 15:01:53 crc kubenswrapper[4805]: I1203 15:01:53.653125 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-0" event={"ID":"cfdd6465-a0a8-4dc4-ad55-b367cb27a2b5","Type":"ContainerStarted","Data":"3797ba396ee119b5e8cb994e9b2aeef70af1f115ba9a92078848e669283054e6"} Dec 03 15:01:54 crc kubenswrapper[4805]: I1203 15:01:54.668739 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-0" event={"ID":"cfdd6465-a0a8-4dc4-ad55-b367cb27a2b5","Type":"ContainerStarted","Data":"7324ec81fc734ace11a78f898680039c33e91f55796fb2e1b43dbb0722390a4e"} Dec 03 15:01:55 crc kubenswrapper[4805]: I1203 15:01:55.682834 4805 generic.go:334] "Generic (PLEG): container finished" podID="aee26d6d-7304-4001-90f8-c0ed886a3cda" containerID="5166a2fd0b33128b5411b9ce35f22d267a1f26fddd2c64c87f9709551024044c" exitCode=0 Dec 03 15:01:55 crc kubenswrapper[4805]: I1203 15:01:55.683063 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"aee26d6d-7304-4001-90f8-c0ed886a3cda","Type":"ContainerDied","Data":"5166a2fd0b33128b5411b9ce35f22d267a1f26fddd2c64c87f9709551024044c"} Dec 03 15:01:55 crc kubenswrapper[4805]: I1203 15:01:55.683155 4805 scope.go:117] "RemoveContainer" containerID="c6e36473c9af1c0d702a0744c34ad51f28c604114568f3a35c2d7eaf0de6e461" Dec 03 15:01:55 crc kubenswrapper[4805]: I1203 15:01:55.957413 4805 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 03 15:01:56 crc kubenswrapper[4805]: I1203 15:01:56.015368 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/aee26d6d-7304-4001-90f8-c0ed886a3cda-log-httpd\") pod \"aee26d6d-7304-4001-90f8-c0ed886a3cda\" (UID: \"aee26d6d-7304-4001-90f8-c0ed886a3cda\") " Dec 03 15:01:56 crc kubenswrapper[4805]: I1203 15:01:56.015550 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/aee26d6d-7304-4001-90f8-c0ed886a3cda-config-data\") pod \"aee26d6d-7304-4001-90f8-c0ed886a3cda\" (UID: \"aee26d6d-7304-4001-90f8-c0ed886a3cda\") " Dec 03 15:01:56 crc kubenswrapper[4805]: I1203 15:01:56.015715 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hdg4f\" (UniqueName: \"kubernetes.io/projected/aee26d6d-7304-4001-90f8-c0ed886a3cda-kube-api-access-hdg4f\") pod \"aee26d6d-7304-4001-90f8-c0ed886a3cda\" (UID: \"aee26d6d-7304-4001-90f8-c0ed886a3cda\") " Dec 03 15:01:56 crc kubenswrapper[4805]: I1203 15:01:56.015799 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/aee26d6d-7304-4001-90f8-c0ed886a3cda-scripts\") pod \"aee26d6d-7304-4001-90f8-c0ed886a3cda\" (UID: \"aee26d6d-7304-4001-90f8-c0ed886a3cda\") " Dec 03 15:01:56 crc kubenswrapper[4805]: I1203 15:01:56.015994 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/aee26d6d-7304-4001-90f8-c0ed886a3cda-combined-ca-bundle\") pod \"aee26d6d-7304-4001-90f8-c0ed886a3cda\" (UID: \"aee26d6d-7304-4001-90f8-c0ed886a3cda\") " Dec 03 15:01:56 crc kubenswrapper[4805]: I1203 15:01:56.016107 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/aee26d6d-7304-4001-90f8-c0ed886a3cda-run-httpd\") pod \"aee26d6d-7304-4001-90f8-c0ed886a3cda\" (UID: \"aee26d6d-7304-4001-90f8-c0ed886a3cda\") " Dec 03 15:01:56 crc kubenswrapper[4805]: I1203 15:01:56.016217 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/aee26d6d-7304-4001-90f8-c0ed886a3cda-ceilometer-tls-certs\") pod \"aee26d6d-7304-4001-90f8-c0ed886a3cda\" (UID: \"aee26d6d-7304-4001-90f8-c0ed886a3cda\") " Dec 03 15:01:56 crc kubenswrapper[4805]: I1203 15:01:56.016373 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/aee26d6d-7304-4001-90f8-c0ed886a3cda-sg-core-conf-yaml\") pod \"aee26d6d-7304-4001-90f8-c0ed886a3cda\" (UID: \"aee26d6d-7304-4001-90f8-c0ed886a3cda\") " Dec 03 15:01:56 crc kubenswrapper[4805]: I1203 15:01:56.017823 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/aee26d6d-7304-4001-90f8-c0ed886a3cda-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "aee26d6d-7304-4001-90f8-c0ed886a3cda" (UID: "aee26d6d-7304-4001-90f8-c0ed886a3cda"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 15:01:56 crc kubenswrapper[4805]: I1203 15:01:56.018365 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/aee26d6d-7304-4001-90f8-c0ed886a3cda-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "aee26d6d-7304-4001-90f8-c0ed886a3cda" (UID: "aee26d6d-7304-4001-90f8-c0ed886a3cda"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 15:01:56 crc kubenswrapper[4805]: I1203 15:01:56.020211 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/aee26d6d-7304-4001-90f8-c0ed886a3cda-scripts" (OuterVolumeSpecName: "scripts") pod "aee26d6d-7304-4001-90f8-c0ed886a3cda" (UID: "aee26d6d-7304-4001-90f8-c0ed886a3cda"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 15:01:56 crc kubenswrapper[4805]: I1203 15:01:56.023387 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/aee26d6d-7304-4001-90f8-c0ed886a3cda-kube-api-access-hdg4f" (OuterVolumeSpecName: "kube-api-access-hdg4f") pod "aee26d6d-7304-4001-90f8-c0ed886a3cda" (UID: "aee26d6d-7304-4001-90f8-c0ed886a3cda"). InnerVolumeSpecName "kube-api-access-hdg4f". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 15:01:56 crc kubenswrapper[4805]: I1203 15:01:56.028858 4805 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/aee26d6d-7304-4001-90f8-c0ed886a3cda-run-httpd\") on node \"crc\" DevicePath \"\"" Dec 03 15:01:56 crc kubenswrapper[4805]: I1203 15:01:56.029910 4805 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/aee26d6d-7304-4001-90f8-c0ed886a3cda-log-httpd\") on node \"crc\" DevicePath \"\"" Dec 03 15:01:56 crc kubenswrapper[4805]: I1203 15:01:56.030012 4805 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hdg4f\" (UniqueName: \"kubernetes.io/projected/aee26d6d-7304-4001-90f8-c0ed886a3cda-kube-api-access-hdg4f\") on node \"crc\" DevicePath \"\"" Dec 03 15:01:56 crc kubenswrapper[4805]: I1203 15:01:56.034179 4805 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/aee26d6d-7304-4001-90f8-c0ed886a3cda-scripts\") on node \"crc\" DevicePath \"\"" Dec 03 15:01:56 crc kubenswrapper[4805]: I1203 15:01:56.077907 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/aee26d6d-7304-4001-90f8-c0ed886a3cda-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "aee26d6d-7304-4001-90f8-c0ed886a3cda" (UID: "aee26d6d-7304-4001-90f8-c0ed886a3cda"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 15:01:56 crc kubenswrapper[4805]: I1203 15:01:56.099628 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/aee26d6d-7304-4001-90f8-c0ed886a3cda-ceilometer-tls-certs" (OuterVolumeSpecName: "ceilometer-tls-certs") pod "aee26d6d-7304-4001-90f8-c0ed886a3cda" (UID: "aee26d6d-7304-4001-90f8-c0ed886a3cda"). InnerVolumeSpecName "ceilometer-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 15:01:56 crc kubenswrapper[4805]: I1203 15:01:56.115861 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/aee26d6d-7304-4001-90f8-c0ed886a3cda-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "aee26d6d-7304-4001-90f8-c0ed886a3cda" (UID: "aee26d6d-7304-4001-90f8-c0ed886a3cda"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 15:01:56 crc kubenswrapper[4805]: I1203 15:01:56.135576 4805 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/aee26d6d-7304-4001-90f8-c0ed886a3cda-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Dec 03 15:01:56 crc kubenswrapper[4805]: I1203 15:01:56.135607 4805 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/aee26d6d-7304-4001-90f8-c0ed886a3cda-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 15:01:56 crc kubenswrapper[4805]: I1203 15:01:56.135615 4805 reconciler_common.go:293] "Volume detached for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/aee26d6d-7304-4001-90f8-c0ed886a3cda-ceilometer-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 03 15:01:56 crc kubenswrapper[4805]: I1203 15:01:56.152372 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/aee26d6d-7304-4001-90f8-c0ed886a3cda-config-data" (OuterVolumeSpecName: "config-data") pod "aee26d6d-7304-4001-90f8-c0ed886a3cda" (UID: "aee26d6d-7304-4001-90f8-c0ed886a3cda"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 15:01:56 crc kubenswrapper[4805]: I1203 15:01:56.237226 4805 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/aee26d6d-7304-4001-90f8-c0ed886a3cda-config-data\") on node \"crc\" DevicePath \"\"" Dec 03 15:01:56 crc kubenswrapper[4805]: I1203 15:01:56.719752 4805 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 03 15:01:56 crc kubenswrapper[4805]: I1203 15:01:56.727054 4805 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/aodh-0" podUID="cfdd6465-a0a8-4dc4-ad55-b367cb27a2b5" containerName="aodh-api" containerID="cri-o://962645f025d997f376be2e9db91f639ab21b690752350e7f3ad6d738ea6e7456" gracePeriod=30 Dec 03 15:01:56 crc kubenswrapper[4805]: I1203 15:01:56.727187 4805 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/aodh-0" podUID="cfdd6465-a0a8-4dc4-ad55-b367cb27a2b5" containerName="aodh-listener" containerID="cri-o://e05fe671c29954321872dfb71956fb33f0b62d4d8bc62c1bd80b77224e98a33a" gracePeriod=30 Dec 03 15:01:56 crc kubenswrapper[4805]: I1203 15:01:56.727260 4805 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/aodh-0" podUID="cfdd6465-a0a8-4dc4-ad55-b367cb27a2b5" containerName="aodh-notifier" containerID="cri-o://7324ec81fc734ace11a78f898680039c33e91f55796fb2e1b43dbb0722390a4e" gracePeriod=30 Dec 03 15:01:56 crc kubenswrapper[4805]: I1203 15:01:56.727325 4805 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/aodh-0" podUID="cfdd6465-a0a8-4dc4-ad55-b367cb27a2b5" containerName="aodh-evaluator" containerID="cri-o://3797ba396ee119b5e8cb994e9b2aeef70af1f115ba9a92078848e669283054e6" gracePeriod=30 Dec 03 15:01:56 crc kubenswrapper[4805]: I1203 15:01:56.753060 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"aee26d6d-7304-4001-90f8-c0ed886a3cda","Type":"ContainerDied","Data":"b44eea74232b42f321d7fb0358f8f4f1fcb9b6397d3e8156b964488f587b25dc"} Dec 03 15:01:56 crc kubenswrapper[4805]: I1203 15:01:56.753118 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-0" event={"ID":"cfdd6465-a0a8-4dc4-ad55-b367cb27a2b5","Type":"ContainerStarted","Data":"e05fe671c29954321872dfb71956fb33f0b62d4d8bc62c1bd80b77224e98a33a"} Dec 03 15:01:56 crc kubenswrapper[4805]: I1203 15:01:56.753140 4805 scope.go:117] "RemoveContainer" containerID="5166a2fd0b33128b5411b9ce35f22d267a1f26fddd2c64c87f9709551024044c" Dec 03 15:01:56 crc kubenswrapper[4805]: I1203 15:01:56.793313 4805 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 03 15:01:56 crc kubenswrapper[4805]: I1203 15:01:56.812786 4805 scope.go:117] "RemoveContainer" containerID="c5da23817a0fadd49f2754c6b0126b250c6c2cee661c805c0a922ca37d9d5a30" Dec 03 15:01:56 crc kubenswrapper[4805]: I1203 15:01:56.814823 4805 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Dec 03 15:01:56 crc kubenswrapper[4805]: I1203 15:01:56.829982 4805 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Dec 03 15:01:56 crc kubenswrapper[4805]: E1203 15:01:56.830390 4805 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="aee26d6d-7304-4001-90f8-c0ed886a3cda" containerName="sg-core" Dec 03 15:01:56 crc kubenswrapper[4805]: I1203 15:01:56.830407 4805 state_mem.go:107] "Deleted CPUSet assignment" podUID="aee26d6d-7304-4001-90f8-c0ed886a3cda" containerName="sg-core" Dec 03 15:01:56 crc kubenswrapper[4805]: E1203 15:01:56.830425 4805 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="aee26d6d-7304-4001-90f8-c0ed886a3cda" containerName="proxy-httpd" Dec 03 15:01:56 crc kubenswrapper[4805]: I1203 15:01:56.830432 4805 state_mem.go:107] "Deleted CPUSet assignment" podUID="aee26d6d-7304-4001-90f8-c0ed886a3cda" containerName="proxy-httpd" Dec 03 15:01:56 crc kubenswrapper[4805]: E1203 15:01:56.830450 4805 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="aee26d6d-7304-4001-90f8-c0ed886a3cda" containerName="ceilometer-central-agent" Dec 03 15:01:56 crc kubenswrapper[4805]: I1203 15:01:56.830456 4805 state_mem.go:107] "Deleted CPUSet assignment" podUID="aee26d6d-7304-4001-90f8-c0ed886a3cda" containerName="ceilometer-central-agent" Dec 03 15:01:56 crc kubenswrapper[4805]: E1203 15:01:56.830467 4805 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="aee26d6d-7304-4001-90f8-c0ed886a3cda" containerName="ceilometer-notification-agent" Dec 03 15:01:56 crc kubenswrapper[4805]: I1203 15:01:56.830509 4805 state_mem.go:107] "Deleted CPUSet assignment" podUID="aee26d6d-7304-4001-90f8-c0ed886a3cda" containerName="ceilometer-notification-agent" Dec 03 15:01:56 crc kubenswrapper[4805]: E1203 15:01:56.830529 4805 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="aee26d6d-7304-4001-90f8-c0ed886a3cda" containerName="ceilometer-central-agent" Dec 03 15:01:56 crc kubenswrapper[4805]: I1203 15:01:56.830536 4805 state_mem.go:107] "Deleted CPUSet assignment" podUID="aee26d6d-7304-4001-90f8-c0ed886a3cda" containerName="ceilometer-central-agent" Dec 03 15:01:56 crc kubenswrapper[4805]: I1203 15:01:56.830744 4805 memory_manager.go:354] "RemoveStaleState removing state" podUID="aee26d6d-7304-4001-90f8-c0ed886a3cda" containerName="proxy-httpd" Dec 03 15:01:56 crc kubenswrapper[4805]: I1203 15:01:56.830759 4805 memory_manager.go:354] "RemoveStaleState removing state" podUID="aee26d6d-7304-4001-90f8-c0ed886a3cda" containerName="ceilometer-notification-agent" Dec 03 15:01:56 crc kubenswrapper[4805]: I1203 15:01:56.830770 4805 memory_manager.go:354] "RemoveStaleState removing state" podUID="aee26d6d-7304-4001-90f8-c0ed886a3cda" containerName="sg-core" Dec 03 15:01:56 crc kubenswrapper[4805]: I1203 15:01:56.830782 4805 memory_manager.go:354] "RemoveStaleState removing state" podUID="aee26d6d-7304-4001-90f8-c0ed886a3cda" containerName="ceilometer-central-agent" Dec 03 15:01:56 crc kubenswrapper[4805]: I1203 15:01:56.830803 4805 memory_manager.go:354] "RemoveStaleState removing state" podUID="aee26d6d-7304-4001-90f8-c0ed886a3cda" containerName="ceilometer-central-agent" Dec 03 15:01:56 crc kubenswrapper[4805]: E1203 15:01:56.831049 4805 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="aee26d6d-7304-4001-90f8-c0ed886a3cda" containerName="ceilometer-notification-agent" Dec 03 15:01:56 crc kubenswrapper[4805]: I1203 15:01:56.831061 4805 state_mem.go:107] "Deleted CPUSet assignment" podUID="aee26d6d-7304-4001-90f8-c0ed886a3cda" containerName="ceilometer-notification-agent" Dec 03 15:01:56 crc kubenswrapper[4805]: I1203 15:01:56.831028 4805 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/aodh-0" podStartSLOduration=2.419031308 podStartE2EDuration="7.831008609s" podCreationTimestamp="2025-12-03 15:01:49 +0000 UTC" firstStartedPulling="2025-12-03 15:01:50.549822845 +0000 UTC m=+3140.212739758" lastFinishedPulling="2025-12-03 15:01:55.961800136 +0000 UTC m=+3145.624717059" observedRunningTime="2025-12-03 15:01:56.795033489 +0000 UTC m=+3146.457950402" watchObservedRunningTime="2025-12-03 15:01:56.831008609 +0000 UTC m=+3146.493925532" Dec 03 15:01:56 crc kubenswrapper[4805]: I1203 15:01:56.831283 4805 memory_manager.go:354] "RemoveStaleState removing state" podUID="aee26d6d-7304-4001-90f8-c0ed886a3cda" containerName="ceilometer-notification-agent" Dec 03 15:01:56 crc kubenswrapper[4805]: I1203 15:01:56.832756 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 03 15:01:56 crc kubenswrapper[4805]: I1203 15:01:56.835203 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Dec 03 15:01:56 crc kubenswrapper[4805]: I1203 15:01:56.847090 4805 scope.go:117] "RemoveContainer" containerID="2cf776a26e533c761ac54e168d3c4825a871fbc428f01ab40780a9f315b80d05" Dec 03 15:01:56 crc kubenswrapper[4805]: I1203 15:01:56.847436 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Dec 03 15:01:56 crc kubenswrapper[4805]: I1203 15:01:56.847656 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ceilometer-internal-svc" Dec 03 15:01:56 crc kubenswrapper[4805]: I1203 15:01:56.852561 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 03 15:01:56 crc kubenswrapper[4805]: I1203 15:01:56.906250 4805 scope.go:117] "RemoveContainer" containerID="f62e2a86bdda5a487b4221e167d77f2dbb49598c46ba7c129af938f19abfce52" Dec 03 15:01:56 crc kubenswrapper[4805]: I1203 15:01:56.950948 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/df32824d-ca3b-4ee5-88a2-1e4591fafc04-scripts\") pod \"ceilometer-0\" (UID: \"df32824d-ca3b-4ee5-88a2-1e4591fafc04\") " pod="openstack/ceilometer-0" Dec 03 15:01:56 crc kubenswrapper[4805]: I1203 15:01:56.951020 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/df32824d-ca3b-4ee5-88a2-1e4591fafc04-log-httpd\") pod \"ceilometer-0\" (UID: \"df32824d-ca3b-4ee5-88a2-1e4591fafc04\") " pod="openstack/ceilometer-0" Dec 03 15:01:56 crc kubenswrapper[4805]: I1203 15:01:56.951041 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/df32824d-ca3b-4ee5-88a2-1e4591fafc04-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"df32824d-ca3b-4ee5-88a2-1e4591fafc04\") " pod="openstack/ceilometer-0" Dec 03 15:01:56 crc kubenswrapper[4805]: I1203 15:01:56.951101 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-j29wh\" (UniqueName: \"kubernetes.io/projected/df32824d-ca3b-4ee5-88a2-1e4591fafc04-kube-api-access-j29wh\") pod \"ceilometer-0\" (UID: \"df32824d-ca3b-4ee5-88a2-1e4591fafc04\") " pod="openstack/ceilometer-0" Dec 03 15:01:56 crc kubenswrapper[4805]: I1203 15:01:56.951225 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/df32824d-ca3b-4ee5-88a2-1e4591fafc04-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"df32824d-ca3b-4ee5-88a2-1e4591fafc04\") " pod="openstack/ceilometer-0" Dec 03 15:01:56 crc kubenswrapper[4805]: I1203 15:01:56.951265 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/df32824d-ca3b-4ee5-88a2-1e4591fafc04-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"df32824d-ca3b-4ee5-88a2-1e4591fafc04\") " pod="openstack/ceilometer-0" Dec 03 15:01:56 crc kubenswrapper[4805]: I1203 15:01:56.951318 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/df32824d-ca3b-4ee5-88a2-1e4591fafc04-run-httpd\") pod \"ceilometer-0\" (UID: \"df32824d-ca3b-4ee5-88a2-1e4591fafc04\") " pod="openstack/ceilometer-0" Dec 03 15:01:56 crc kubenswrapper[4805]: I1203 15:01:56.951365 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/df32824d-ca3b-4ee5-88a2-1e4591fafc04-config-data\") pod \"ceilometer-0\" (UID: \"df32824d-ca3b-4ee5-88a2-1e4591fafc04\") " pod="openstack/ceilometer-0" Dec 03 15:01:57 crc kubenswrapper[4805]: I1203 15:01:57.053451 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/df32824d-ca3b-4ee5-88a2-1e4591fafc04-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"df32824d-ca3b-4ee5-88a2-1e4591fafc04\") " pod="openstack/ceilometer-0" Dec 03 15:01:57 crc kubenswrapper[4805]: I1203 15:01:57.053502 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/df32824d-ca3b-4ee5-88a2-1e4591fafc04-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"df32824d-ca3b-4ee5-88a2-1e4591fafc04\") " pod="openstack/ceilometer-0" Dec 03 15:01:57 crc kubenswrapper[4805]: I1203 15:01:57.053543 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/df32824d-ca3b-4ee5-88a2-1e4591fafc04-run-httpd\") pod \"ceilometer-0\" (UID: \"df32824d-ca3b-4ee5-88a2-1e4591fafc04\") " pod="openstack/ceilometer-0" Dec 03 15:01:57 crc kubenswrapper[4805]: I1203 15:01:57.053570 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/df32824d-ca3b-4ee5-88a2-1e4591fafc04-config-data\") pod \"ceilometer-0\" (UID: \"df32824d-ca3b-4ee5-88a2-1e4591fafc04\") " pod="openstack/ceilometer-0" Dec 03 15:01:57 crc kubenswrapper[4805]: I1203 15:01:57.053602 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/df32824d-ca3b-4ee5-88a2-1e4591fafc04-scripts\") pod \"ceilometer-0\" (UID: \"df32824d-ca3b-4ee5-88a2-1e4591fafc04\") " pod="openstack/ceilometer-0" Dec 03 15:01:57 crc kubenswrapper[4805]: I1203 15:01:57.053629 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/df32824d-ca3b-4ee5-88a2-1e4591fafc04-log-httpd\") pod \"ceilometer-0\" (UID: \"df32824d-ca3b-4ee5-88a2-1e4591fafc04\") " pod="openstack/ceilometer-0" Dec 03 15:01:57 crc kubenswrapper[4805]: I1203 15:01:57.053642 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/df32824d-ca3b-4ee5-88a2-1e4591fafc04-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"df32824d-ca3b-4ee5-88a2-1e4591fafc04\") " pod="openstack/ceilometer-0" Dec 03 15:01:57 crc kubenswrapper[4805]: I1203 15:01:57.053680 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-j29wh\" (UniqueName: \"kubernetes.io/projected/df32824d-ca3b-4ee5-88a2-1e4591fafc04-kube-api-access-j29wh\") pod \"ceilometer-0\" (UID: \"df32824d-ca3b-4ee5-88a2-1e4591fafc04\") " pod="openstack/ceilometer-0" Dec 03 15:01:57 crc kubenswrapper[4805]: I1203 15:01:57.055061 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/df32824d-ca3b-4ee5-88a2-1e4591fafc04-run-httpd\") pod \"ceilometer-0\" (UID: \"df32824d-ca3b-4ee5-88a2-1e4591fafc04\") " pod="openstack/ceilometer-0" Dec 03 15:01:57 crc kubenswrapper[4805]: I1203 15:01:57.055115 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/df32824d-ca3b-4ee5-88a2-1e4591fafc04-log-httpd\") pod \"ceilometer-0\" (UID: \"df32824d-ca3b-4ee5-88a2-1e4591fafc04\") " pod="openstack/ceilometer-0" Dec 03 15:01:57 crc kubenswrapper[4805]: I1203 15:01:57.060080 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/df32824d-ca3b-4ee5-88a2-1e4591fafc04-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"df32824d-ca3b-4ee5-88a2-1e4591fafc04\") " pod="openstack/ceilometer-0" Dec 03 15:01:57 crc kubenswrapper[4805]: I1203 15:01:57.060259 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/df32824d-ca3b-4ee5-88a2-1e4591fafc04-scripts\") pod \"ceilometer-0\" (UID: \"df32824d-ca3b-4ee5-88a2-1e4591fafc04\") " pod="openstack/ceilometer-0" Dec 03 15:01:57 crc kubenswrapper[4805]: I1203 15:01:57.061095 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/df32824d-ca3b-4ee5-88a2-1e4591fafc04-config-data\") pod \"ceilometer-0\" (UID: \"df32824d-ca3b-4ee5-88a2-1e4591fafc04\") " pod="openstack/ceilometer-0" Dec 03 15:01:57 crc kubenswrapper[4805]: I1203 15:01:57.061747 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/df32824d-ca3b-4ee5-88a2-1e4591fafc04-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"df32824d-ca3b-4ee5-88a2-1e4591fafc04\") " pod="openstack/ceilometer-0" Dec 03 15:01:57 crc kubenswrapper[4805]: I1203 15:01:57.063421 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/df32824d-ca3b-4ee5-88a2-1e4591fafc04-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"df32824d-ca3b-4ee5-88a2-1e4591fafc04\") " pod="openstack/ceilometer-0" Dec 03 15:01:57 crc kubenswrapper[4805]: I1203 15:01:57.072330 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-j29wh\" (UniqueName: \"kubernetes.io/projected/df32824d-ca3b-4ee5-88a2-1e4591fafc04-kube-api-access-j29wh\") pod \"ceilometer-0\" (UID: \"df32824d-ca3b-4ee5-88a2-1e4591fafc04\") " pod="openstack/ceilometer-0" Dec 03 15:01:57 crc kubenswrapper[4805]: I1203 15:01:57.201407 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 03 15:01:57 crc kubenswrapper[4805]: I1203 15:01:57.660288 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 03 15:01:57 crc kubenswrapper[4805]: W1203 15:01:57.661703 4805 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poddf32824d_ca3b_4ee5_88a2_1e4591fafc04.slice/crio-bd06748d76a2c736438e9a017b5bfdf26bc9e609a7d1c28852ef7704927c42b8 WatchSource:0}: Error finding container bd06748d76a2c736438e9a017b5bfdf26bc9e609a7d1c28852ef7704927c42b8: Status 404 returned error can't find the container with id bd06748d76a2c736438e9a017b5bfdf26bc9e609a7d1c28852ef7704927c42b8 Dec 03 15:01:57 crc kubenswrapper[4805]: I1203 15:01:57.694204 4805 scope.go:117] "RemoveContainer" containerID="13e0da4d7fc8f4085be601d845d688938c16ea7941e52ca511e39a37d6e0ba4e" Dec 03 15:01:57 crc kubenswrapper[4805]: E1203 15:01:57.694452 4805 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-gskh4_openshift-machine-config-operator(6098937f-e3f6-45e8-a647-4994a79cd711)\"" pod="openshift-machine-config-operator/machine-config-daemon-gskh4" podUID="6098937f-e3f6-45e8-a647-4994a79cd711" Dec 03 15:01:57 crc kubenswrapper[4805]: I1203 15:01:57.743190 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"df32824d-ca3b-4ee5-88a2-1e4591fafc04","Type":"ContainerStarted","Data":"bd06748d76a2c736438e9a017b5bfdf26bc9e609a7d1c28852ef7704927c42b8"} Dec 03 15:01:57 crc kubenswrapper[4805]: I1203 15:01:57.749054 4805 generic.go:334] "Generic (PLEG): container finished" podID="cfdd6465-a0a8-4dc4-ad55-b367cb27a2b5" containerID="7324ec81fc734ace11a78f898680039c33e91f55796fb2e1b43dbb0722390a4e" exitCode=0 Dec 03 15:01:57 crc kubenswrapper[4805]: I1203 15:01:57.749094 4805 generic.go:334] "Generic (PLEG): container finished" podID="cfdd6465-a0a8-4dc4-ad55-b367cb27a2b5" containerID="3797ba396ee119b5e8cb994e9b2aeef70af1f115ba9a92078848e669283054e6" exitCode=0 Dec 03 15:01:57 crc kubenswrapper[4805]: I1203 15:01:57.749104 4805 generic.go:334] "Generic (PLEG): container finished" podID="cfdd6465-a0a8-4dc4-ad55-b367cb27a2b5" containerID="962645f025d997f376be2e9db91f639ab21b690752350e7f3ad6d738ea6e7456" exitCode=0 Dec 03 15:01:57 crc kubenswrapper[4805]: I1203 15:01:57.749126 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-0" event={"ID":"cfdd6465-a0a8-4dc4-ad55-b367cb27a2b5","Type":"ContainerDied","Data":"7324ec81fc734ace11a78f898680039c33e91f55796fb2e1b43dbb0722390a4e"} Dec 03 15:01:57 crc kubenswrapper[4805]: I1203 15:01:57.749155 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-0" event={"ID":"cfdd6465-a0a8-4dc4-ad55-b367cb27a2b5","Type":"ContainerDied","Data":"3797ba396ee119b5e8cb994e9b2aeef70af1f115ba9a92078848e669283054e6"} Dec 03 15:01:57 crc kubenswrapper[4805]: I1203 15:01:57.749167 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-0" event={"ID":"cfdd6465-a0a8-4dc4-ad55-b367cb27a2b5","Type":"ContainerDied","Data":"962645f025d997f376be2e9db91f639ab21b690752350e7f3ad6d738ea6e7456"} Dec 03 15:01:58 crc kubenswrapper[4805]: I1203 15:01:58.720315 4805 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="aee26d6d-7304-4001-90f8-c0ed886a3cda" path="/var/lib/kubelet/pods/aee26d6d-7304-4001-90f8-c0ed886a3cda/volumes" Dec 03 15:01:58 crc kubenswrapper[4805]: I1203 15:01:58.758617 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"df32824d-ca3b-4ee5-88a2-1e4591fafc04","Type":"ContainerStarted","Data":"11cb003c6b5a127b15dd6a2b524c4938af09d807f62d7e1e83d87a645bdee55c"} Dec 03 15:01:59 crc kubenswrapper[4805]: I1203 15:01:59.769624 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"df32824d-ca3b-4ee5-88a2-1e4591fafc04","Type":"ContainerStarted","Data":"ccd999fdea781411169848d6f7dfdc4fbd314f59ab023e34190e22c410a8344b"} Dec 03 15:02:00 crc kubenswrapper[4805]: I1203 15:02:00.784284 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"df32824d-ca3b-4ee5-88a2-1e4591fafc04","Type":"ContainerStarted","Data":"ef555edb790a8292068b556a918550970d999af4ece129e38e9cbd832fb767f3"} Dec 03 15:02:01 crc kubenswrapper[4805]: E1203 15:02:01.130576 4805 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod29c53624_ef4b_43ab_8922_181a78908d24.slice/crio-conmon-26262b91874fa88789d54d9ed21dcbd2e4122b6fb279633aaa0a8892e3a82995.scope\": RecentStats: unable to find data in memory cache]" Dec 03 15:02:02 crc kubenswrapper[4805]: I1203 15:02:02.804790 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"df32824d-ca3b-4ee5-88a2-1e4591fafc04","Type":"ContainerStarted","Data":"cf8955740e9a9741bfd2e47b5c5ebe9c1d88ccd99e1b3b873bf253a3dc7146b3"} Dec 03 15:02:02 crc kubenswrapper[4805]: I1203 15:02:02.805322 4805 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Dec 03 15:02:02 crc kubenswrapper[4805]: I1203 15:02:02.838830 4805 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=2.612480626 podStartE2EDuration="6.838809893s" podCreationTimestamp="2025-12-03 15:01:56 +0000 UTC" firstStartedPulling="2025-12-03 15:01:57.664605722 +0000 UTC m=+3147.327522645" lastFinishedPulling="2025-12-03 15:02:01.890934999 +0000 UTC m=+3151.553851912" observedRunningTime="2025-12-03 15:02:02.835981776 +0000 UTC m=+3152.498898719" watchObservedRunningTime="2025-12-03 15:02:02.838809893 +0000 UTC m=+3152.501726826" Dec 03 15:02:09 crc kubenswrapper[4805]: I1203 15:02:09.695815 4805 scope.go:117] "RemoveContainer" containerID="13e0da4d7fc8f4085be601d845d688938c16ea7941e52ca511e39a37d6e0ba4e" Dec 03 15:02:09 crc kubenswrapper[4805]: E1203 15:02:09.698432 4805 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-gskh4_openshift-machine-config-operator(6098937f-e3f6-45e8-a647-4994a79cd711)\"" pod="openshift-machine-config-operator/machine-config-daemon-gskh4" podUID="6098937f-e3f6-45e8-a647-4994a79cd711" Dec 03 15:02:11 crc kubenswrapper[4805]: E1203 15:02:11.401650 4805 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod29c53624_ef4b_43ab_8922_181a78908d24.slice/crio-conmon-26262b91874fa88789d54d9ed21dcbd2e4122b6fb279633aaa0a8892e3a82995.scope\": RecentStats: unable to find data in memory cache]" Dec 03 15:02:21 crc kubenswrapper[4805]: E1203 15:02:21.893822 4805 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod29c53624_ef4b_43ab_8922_181a78908d24.slice/crio-conmon-26262b91874fa88789d54d9ed21dcbd2e4122b6fb279633aaa0a8892e3a82995.scope\": RecentStats: unable to find data in memory cache]" Dec 03 15:02:24 crc kubenswrapper[4805]: I1203 15:02:24.695655 4805 scope.go:117] "RemoveContainer" containerID="13e0da4d7fc8f4085be601d845d688938c16ea7941e52ca511e39a37d6e0ba4e" Dec 03 15:02:24 crc kubenswrapper[4805]: E1203 15:02:24.696487 4805 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-gskh4_openshift-machine-config-operator(6098937f-e3f6-45e8-a647-4994a79cd711)\"" pod="openshift-machine-config-operator/machine-config-daemon-gskh4" podUID="6098937f-e3f6-45e8-a647-4994a79cd711" Dec 03 15:02:27 crc kubenswrapper[4805]: I1203 15:02:27.073669 4805 generic.go:334] "Generic (PLEG): container finished" podID="cfdd6465-a0a8-4dc4-ad55-b367cb27a2b5" containerID="e05fe671c29954321872dfb71956fb33f0b62d4d8bc62c1bd80b77224e98a33a" exitCode=137 Dec 03 15:02:27 crc kubenswrapper[4805]: I1203 15:02:27.073767 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-0" event={"ID":"cfdd6465-a0a8-4dc4-ad55-b367cb27a2b5","Type":"ContainerDied","Data":"e05fe671c29954321872dfb71956fb33f0b62d4d8bc62c1bd80b77224e98a33a"} Dec 03 15:02:27 crc kubenswrapper[4805]: I1203 15:02:27.074244 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-0" event={"ID":"cfdd6465-a0a8-4dc4-ad55-b367cb27a2b5","Type":"ContainerDied","Data":"597774ad530dc25a7e74d08fd60651d95945c6fe92c220f04e0a3a414897b673"} Dec 03 15:02:27 crc kubenswrapper[4805]: I1203 15:02:27.074269 4805 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="597774ad530dc25a7e74d08fd60651d95945c6fe92c220f04e0a3a414897b673" Dec 03 15:02:27 crc kubenswrapper[4805]: I1203 15:02:27.139959 4805 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-0" Dec 03 15:02:27 crc kubenswrapper[4805]: I1203 15:02:27.210461 4805 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ceilometer-0" Dec 03 15:02:27 crc kubenswrapper[4805]: I1203 15:02:27.246395 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cfdd6465-a0a8-4dc4-ad55-b367cb27a2b5-config-data\") pod \"cfdd6465-a0a8-4dc4-ad55-b367cb27a2b5\" (UID: \"cfdd6465-a0a8-4dc4-ad55-b367cb27a2b5\") " Dec 03 15:02:27 crc kubenswrapper[4805]: I1203 15:02:27.246475 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-m5mbv\" (UniqueName: \"kubernetes.io/projected/cfdd6465-a0a8-4dc4-ad55-b367cb27a2b5-kube-api-access-m5mbv\") pod \"cfdd6465-a0a8-4dc4-ad55-b367cb27a2b5\" (UID: \"cfdd6465-a0a8-4dc4-ad55-b367cb27a2b5\") " Dec 03 15:02:27 crc kubenswrapper[4805]: I1203 15:02:27.246527 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cfdd6465-a0a8-4dc4-ad55-b367cb27a2b5-combined-ca-bundle\") pod \"cfdd6465-a0a8-4dc4-ad55-b367cb27a2b5\" (UID: \"cfdd6465-a0a8-4dc4-ad55-b367cb27a2b5\") " Dec 03 15:02:27 crc kubenswrapper[4805]: I1203 15:02:27.246645 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/cfdd6465-a0a8-4dc4-ad55-b367cb27a2b5-scripts\") pod \"cfdd6465-a0a8-4dc4-ad55-b367cb27a2b5\" (UID: \"cfdd6465-a0a8-4dc4-ad55-b367cb27a2b5\") " Dec 03 15:02:27 crc kubenswrapper[4805]: I1203 15:02:27.252374 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cfdd6465-a0a8-4dc4-ad55-b367cb27a2b5-kube-api-access-m5mbv" (OuterVolumeSpecName: "kube-api-access-m5mbv") pod "cfdd6465-a0a8-4dc4-ad55-b367cb27a2b5" (UID: "cfdd6465-a0a8-4dc4-ad55-b367cb27a2b5"). InnerVolumeSpecName "kube-api-access-m5mbv". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 15:02:27 crc kubenswrapper[4805]: I1203 15:02:27.253972 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/cfdd6465-a0a8-4dc4-ad55-b367cb27a2b5-scripts" (OuterVolumeSpecName: "scripts") pod "cfdd6465-a0a8-4dc4-ad55-b367cb27a2b5" (UID: "cfdd6465-a0a8-4dc4-ad55-b367cb27a2b5"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 15:02:27 crc kubenswrapper[4805]: I1203 15:02:27.351636 4805 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/cfdd6465-a0a8-4dc4-ad55-b367cb27a2b5-scripts\") on node \"crc\" DevicePath \"\"" Dec 03 15:02:27 crc kubenswrapper[4805]: I1203 15:02:27.356054 4805 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-m5mbv\" (UniqueName: \"kubernetes.io/projected/cfdd6465-a0a8-4dc4-ad55-b367cb27a2b5-kube-api-access-m5mbv\") on node \"crc\" DevicePath \"\"" Dec 03 15:02:27 crc kubenswrapper[4805]: I1203 15:02:27.377096 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/cfdd6465-a0a8-4dc4-ad55-b367cb27a2b5-config-data" (OuterVolumeSpecName: "config-data") pod "cfdd6465-a0a8-4dc4-ad55-b367cb27a2b5" (UID: "cfdd6465-a0a8-4dc4-ad55-b367cb27a2b5"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 15:02:27 crc kubenswrapper[4805]: I1203 15:02:27.382728 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/cfdd6465-a0a8-4dc4-ad55-b367cb27a2b5-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "cfdd6465-a0a8-4dc4-ad55-b367cb27a2b5" (UID: "cfdd6465-a0a8-4dc4-ad55-b367cb27a2b5"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 15:02:27 crc kubenswrapper[4805]: I1203 15:02:27.457521 4805 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cfdd6465-a0a8-4dc4-ad55-b367cb27a2b5-config-data\") on node \"crc\" DevicePath \"\"" Dec 03 15:02:27 crc kubenswrapper[4805]: I1203 15:02:27.457576 4805 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cfdd6465-a0a8-4dc4-ad55-b367cb27a2b5-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 15:02:28 crc kubenswrapper[4805]: I1203 15:02:28.084249 4805 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-0" Dec 03 15:02:28 crc kubenswrapper[4805]: I1203 15:02:28.167516 4805 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/aodh-0"] Dec 03 15:02:28 crc kubenswrapper[4805]: I1203 15:02:28.177550 4805 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/aodh-0"] Dec 03 15:02:28 crc kubenswrapper[4805]: I1203 15:02:28.186499 4805 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/aodh-0"] Dec 03 15:02:28 crc kubenswrapper[4805]: E1203 15:02:28.186907 4805 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cfdd6465-a0a8-4dc4-ad55-b367cb27a2b5" containerName="aodh-notifier" Dec 03 15:02:28 crc kubenswrapper[4805]: I1203 15:02:28.186923 4805 state_mem.go:107] "Deleted CPUSet assignment" podUID="cfdd6465-a0a8-4dc4-ad55-b367cb27a2b5" containerName="aodh-notifier" Dec 03 15:02:28 crc kubenswrapper[4805]: E1203 15:02:28.186954 4805 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cfdd6465-a0a8-4dc4-ad55-b367cb27a2b5" containerName="aodh-listener" Dec 03 15:02:28 crc kubenswrapper[4805]: I1203 15:02:28.186962 4805 state_mem.go:107] "Deleted CPUSet assignment" podUID="cfdd6465-a0a8-4dc4-ad55-b367cb27a2b5" containerName="aodh-listener" Dec 03 15:02:28 crc kubenswrapper[4805]: E1203 15:02:28.186987 4805 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cfdd6465-a0a8-4dc4-ad55-b367cb27a2b5" containerName="aodh-api" Dec 03 15:02:28 crc kubenswrapper[4805]: I1203 15:02:28.186993 4805 state_mem.go:107] "Deleted CPUSet assignment" podUID="cfdd6465-a0a8-4dc4-ad55-b367cb27a2b5" containerName="aodh-api" Dec 03 15:02:28 crc kubenswrapper[4805]: E1203 15:02:28.187009 4805 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cfdd6465-a0a8-4dc4-ad55-b367cb27a2b5" containerName="aodh-evaluator" Dec 03 15:02:28 crc kubenswrapper[4805]: I1203 15:02:28.187015 4805 state_mem.go:107] "Deleted CPUSet assignment" podUID="cfdd6465-a0a8-4dc4-ad55-b367cb27a2b5" containerName="aodh-evaluator" Dec 03 15:02:28 crc kubenswrapper[4805]: I1203 15:02:28.187213 4805 memory_manager.go:354] "RemoveStaleState removing state" podUID="cfdd6465-a0a8-4dc4-ad55-b367cb27a2b5" containerName="aodh-evaluator" Dec 03 15:02:28 crc kubenswrapper[4805]: I1203 15:02:28.187227 4805 memory_manager.go:354] "RemoveStaleState removing state" podUID="cfdd6465-a0a8-4dc4-ad55-b367cb27a2b5" containerName="aodh-api" Dec 03 15:02:28 crc kubenswrapper[4805]: I1203 15:02:28.187247 4805 memory_manager.go:354] "RemoveStaleState removing state" podUID="cfdd6465-a0a8-4dc4-ad55-b367cb27a2b5" containerName="aodh-listener" Dec 03 15:02:28 crc kubenswrapper[4805]: I1203 15:02:28.187260 4805 memory_manager.go:354] "RemoveStaleState removing state" podUID="cfdd6465-a0a8-4dc4-ad55-b367cb27a2b5" containerName="aodh-notifier" Dec 03 15:02:28 crc kubenswrapper[4805]: I1203 15:02:28.188887 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-0" Dec 03 15:02:28 crc kubenswrapper[4805]: I1203 15:02:28.191768 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-aodh-internal-svc" Dec 03 15:02:28 crc kubenswrapper[4805]: I1203 15:02:28.191982 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"telemetry-autoscaling-dockercfg-wc8kz" Dec 03 15:02:28 crc kubenswrapper[4805]: I1203 15:02:28.192113 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"aodh-config-data" Dec 03 15:02:28 crc kubenswrapper[4805]: I1203 15:02:28.192400 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"aodh-scripts" Dec 03 15:02:28 crc kubenswrapper[4805]: I1203 15:02:28.193099 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-aodh-public-svc" Dec 03 15:02:28 crc kubenswrapper[4805]: I1203 15:02:28.197338 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/aodh-0"] Dec 03 15:02:28 crc kubenswrapper[4805]: I1203 15:02:28.375207 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/267ba6dc-f62a-4b5b-a1ee-0743df62f50e-combined-ca-bundle\") pod \"aodh-0\" (UID: \"267ba6dc-f62a-4b5b-a1ee-0743df62f50e\") " pod="openstack/aodh-0" Dec 03 15:02:28 crc kubenswrapper[4805]: I1203 15:02:28.375270 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zlhqt\" (UniqueName: \"kubernetes.io/projected/267ba6dc-f62a-4b5b-a1ee-0743df62f50e-kube-api-access-zlhqt\") pod \"aodh-0\" (UID: \"267ba6dc-f62a-4b5b-a1ee-0743df62f50e\") " pod="openstack/aodh-0" Dec 03 15:02:28 crc kubenswrapper[4805]: I1203 15:02:28.375340 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/267ba6dc-f62a-4b5b-a1ee-0743df62f50e-config-data\") pod \"aodh-0\" (UID: \"267ba6dc-f62a-4b5b-a1ee-0743df62f50e\") " pod="openstack/aodh-0" Dec 03 15:02:28 crc kubenswrapper[4805]: I1203 15:02:28.375379 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/267ba6dc-f62a-4b5b-a1ee-0743df62f50e-public-tls-certs\") pod \"aodh-0\" (UID: \"267ba6dc-f62a-4b5b-a1ee-0743df62f50e\") " pod="openstack/aodh-0" Dec 03 15:02:28 crc kubenswrapper[4805]: I1203 15:02:28.375430 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/267ba6dc-f62a-4b5b-a1ee-0743df62f50e-scripts\") pod \"aodh-0\" (UID: \"267ba6dc-f62a-4b5b-a1ee-0743df62f50e\") " pod="openstack/aodh-0" Dec 03 15:02:28 crc kubenswrapper[4805]: I1203 15:02:28.375451 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/267ba6dc-f62a-4b5b-a1ee-0743df62f50e-internal-tls-certs\") pod \"aodh-0\" (UID: \"267ba6dc-f62a-4b5b-a1ee-0743df62f50e\") " pod="openstack/aodh-0" Dec 03 15:02:28 crc kubenswrapper[4805]: I1203 15:02:28.477623 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/267ba6dc-f62a-4b5b-a1ee-0743df62f50e-config-data\") pod \"aodh-0\" (UID: \"267ba6dc-f62a-4b5b-a1ee-0743df62f50e\") " pod="openstack/aodh-0" Dec 03 15:02:28 crc kubenswrapper[4805]: I1203 15:02:28.477724 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/267ba6dc-f62a-4b5b-a1ee-0743df62f50e-public-tls-certs\") pod \"aodh-0\" (UID: \"267ba6dc-f62a-4b5b-a1ee-0743df62f50e\") " pod="openstack/aodh-0" Dec 03 15:02:28 crc kubenswrapper[4805]: I1203 15:02:28.477809 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/267ba6dc-f62a-4b5b-a1ee-0743df62f50e-scripts\") pod \"aodh-0\" (UID: \"267ba6dc-f62a-4b5b-a1ee-0743df62f50e\") " pod="openstack/aodh-0" Dec 03 15:02:28 crc kubenswrapper[4805]: I1203 15:02:28.477865 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/267ba6dc-f62a-4b5b-a1ee-0743df62f50e-internal-tls-certs\") pod \"aodh-0\" (UID: \"267ba6dc-f62a-4b5b-a1ee-0743df62f50e\") " pod="openstack/aodh-0" Dec 03 15:02:28 crc kubenswrapper[4805]: I1203 15:02:28.478087 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/267ba6dc-f62a-4b5b-a1ee-0743df62f50e-combined-ca-bundle\") pod \"aodh-0\" (UID: \"267ba6dc-f62a-4b5b-a1ee-0743df62f50e\") " pod="openstack/aodh-0" Dec 03 15:02:28 crc kubenswrapper[4805]: I1203 15:02:28.478149 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zlhqt\" (UniqueName: \"kubernetes.io/projected/267ba6dc-f62a-4b5b-a1ee-0743df62f50e-kube-api-access-zlhqt\") pod \"aodh-0\" (UID: \"267ba6dc-f62a-4b5b-a1ee-0743df62f50e\") " pod="openstack/aodh-0" Dec 03 15:02:28 crc kubenswrapper[4805]: I1203 15:02:28.484125 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/267ba6dc-f62a-4b5b-a1ee-0743df62f50e-internal-tls-certs\") pod \"aodh-0\" (UID: \"267ba6dc-f62a-4b5b-a1ee-0743df62f50e\") " pod="openstack/aodh-0" Dec 03 15:02:28 crc kubenswrapper[4805]: I1203 15:02:28.484665 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/267ba6dc-f62a-4b5b-a1ee-0743df62f50e-public-tls-certs\") pod \"aodh-0\" (UID: \"267ba6dc-f62a-4b5b-a1ee-0743df62f50e\") " pod="openstack/aodh-0" Dec 03 15:02:28 crc kubenswrapper[4805]: I1203 15:02:28.485080 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/267ba6dc-f62a-4b5b-a1ee-0743df62f50e-scripts\") pod \"aodh-0\" (UID: \"267ba6dc-f62a-4b5b-a1ee-0743df62f50e\") " pod="openstack/aodh-0" Dec 03 15:02:28 crc kubenswrapper[4805]: I1203 15:02:28.485552 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/267ba6dc-f62a-4b5b-a1ee-0743df62f50e-combined-ca-bundle\") pod \"aodh-0\" (UID: \"267ba6dc-f62a-4b5b-a1ee-0743df62f50e\") " pod="openstack/aodh-0" Dec 03 15:02:28 crc kubenswrapper[4805]: I1203 15:02:28.485711 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/267ba6dc-f62a-4b5b-a1ee-0743df62f50e-config-data\") pod \"aodh-0\" (UID: \"267ba6dc-f62a-4b5b-a1ee-0743df62f50e\") " pod="openstack/aodh-0" Dec 03 15:02:28 crc kubenswrapper[4805]: I1203 15:02:28.498602 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zlhqt\" (UniqueName: \"kubernetes.io/projected/267ba6dc-f62a-4b5b-a1ee-0743df62f50e-kube-api-access-zlhqt\") pod \"aodh-0\" (UID: \"267ba6dc-f62a-4b5b-a1ee-0743df62f50e\") " pod="openstack/aodh-0" Dec 03 15:02:28 crc kubenswrapper[4805]: I1203 15:02:28.507030 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-0" Dec 03 15:02:28 crc kubenswrapper[4805]: I1203 15:02:28.708505 4805 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="cfdd6465-a0a8-4dc4-ad55-b367cb27a2b5" path="/var/lib/kubelet/pods/cfdd6465-a0a8-4dc4-ad55-b367cb27a2b5/volumes" Dec 03 15:02:29 crc kubenswrapper[4805]: I1203 15:02:29.046537 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/aodh-0"] Dec 03 15:02:29 crc kubenswrapper[4805]: I1203 15:02:29.097265 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-0" event={"ID":"267ba6dc-f62a-4b5b-a1ee-0743df62f50e","Type":"ContainerStarted","Data":"9d546eff560853a541cdd332349cbdc606d8eeac10f74991e380a9d298f0b126"} Dec 03 15:02:30 crc kubenswrapper[4805]: I1203 15:02:30.109082 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-0" event={"ID":"267ba6dc-f62a-4b5b-a1ee-0743df62f50e","Type":"ContainerStarted","Data":"23c3e7d8a29458fff84d07a4be9660cc586e135426898c0af6bdef9c908890f2"} Dec 03 15:02:31 crc kubenswrapper[4805]: I1203 15:02:31.123141 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-0" event={"ID":"267ba6dc-f62a-4b5b-a1ee-0743df62f50e","Type":"ContainerStarted","Data":"9e6627c97099047d600501614c8fb93027bc47e3b457b199ae443cb0f41ba8cc"} Dec 03 15:02:32 crc kubenswrapper[4805]: I1203 15:02:32.132705 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-0" event={"ID":"267ba6dc-f62a-4b5b-a1ee-0743df62f50e","Type":"ContainerStarted","Data":"20af8007ca6c082e6b1ef2f9f83d8ec39921523a6c30a7cc1e2799cbba60883b"} Dec 03 15:02:34 crc kubenswrapper[4805]: I1203 15:02:34.161935 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-0" event={"ID":"267ba6dc-f62a-4b5b-a1ee-0743df62f50e","Type":"ContainerStarted","Data":"3d089bb5731d01e0995badef708b81e343d6017ece53c031ba0d369fe28420c5"} Dec 03 15:02:34 crc kubenswrapper[4805]: I1203 15:02:34.203118 4805 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/aodh-0" podStartSLOduration=2.422378612 podStartE2EDuration="6.203089466s" podCreationTimestamp="2025-12-03 15:02:28 +0000 UTC" firstStartedPulling="2025-12-03 15:02:29.055224686 +0000 UTC m=+3178.718141609" lastFinishedPulling="2025-12-03 15:02:32.83593554 +0000 UTC m=+3182.498852463" observedRunningTime="2025-12-03 15:02:34.191381151 +0000 UTC m=+3183.854298084" watchObservedRunningTime="2025-12-03 15:02:34.203089466 +0000 UTC m=+3183.866006419" Dec 03 15:02:39 crc kubenswrapper[4805]: I1203 15:02:39.694828 4805 scope.go:117] "RemoveContainer" containerID="13e0da4d7fc8f4085be601d845d688938c16ea7941e52ca511e39a37d6e0ba4e" Dec 03 15:02:39 crc kubenswrapper[4805]: E1203 15:02:39.695595 4805 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-gskh4_openshift-machine-config-operator(6098937f-e3f6-45e8-a647-4994a79cd711)\"" pod="openshift-machine-config-operator/machine-config-daemon-gskh4" podUID="6098937f-e3f6-45e8-a647-4994a79cd711" Dec 03 15:02:53 crc kubenswrapper[4805]: I1203 15:02:53.695553 4805 scope.go:117] "RemoveContainer" containerID="13e0da4d7fc8f4085be601d845d688938c16ea7941e52ca511e39a37d6e0ba4e" Dec 03 15:02:53 crc kubenswrapper[4805]: E1203 15:02:53.696502 4805 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-gskh4_openshift-machine-config-operator(6098937f-e3f6-45e8-a647-4994a79cd711)\"" pod="openshift-machine-config-operator/machine-config-daemon-gskh4" podUID="6098937f-e3f6-45e8-a647-4994a79cd711" Dec 03 15:03:04 crc kubenswrapper[4805]: I1203 15:03:04.694430 4805 scope.go:117] "RemoveContainer" containerID="13e0da4d7fc8f4085be601d845d688938c16ea7941e52ca511e39a37d6e0ba4e" Dec 03 15:03:04 crc kubenswrapper[4805]: E1203 15:03:04.695412 4805 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-gskh4_openshift-machine-config-operator(6098937f-e3f6-45e8-a647-4994a79cd711)\"" pod="openshift-machine-config-operator/machine-config-daemon-gskh4" podUID="6098937f-e3f6-45e8-a647-4994a79cd711" Dec 03 15:03:15 crc kubenswrapper[4805]: I1203 15:03:15.695353 4805 scope.go:117] "RemoveContainer" containerID="13e0da4d7fc8f4085be601d845d688938c16ea7941e52ca511e39a37d6e0ba4e" Dec 03 15:03:15 crc kubenswrapper[4805]: E1203 15:03:15.696075 4805 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-gskh4_openshift-machine-config-operator(6098937f-e3f6-45e8-a647-4994a79cd711)\"" pod="openshift-machine-config-operator/machine-config-daemon-gskh4" podUID="6098937f-e3f6-45e8-a647-4994a79cd711" Dec 03 15:03:30 crc kubenswrapper[4805]: I1203 15:03:30.700984 4805 scope.go:117] "RemoveContainer" containerID="13e0da4d7fc8f4085be601d845d688938c16ea7941e52ca511e39a37d6e0ba4e" Dec 03 15:03:30 crc kubenswrapper[4805]: E1203 15:03:30.702096 4805 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-gskh4_openshift-machine-config-operator(6098937f-e3f6-45e8-a647-4994a79cd711)\"" pod="openshift-machine-config-operator/machine-config-daemon-gskh4" podUID="6098937f-e3f6-45e8-a647-4994a79cd711" Dec 03 15:03:41 crc kubenswrapper[4805]: I1203 15:03:41.695417 4805 scope.go:117] "RemoveContainer" containerID="13e0da4d7fc8f4085be601d845d688938c16ea7941e52ca511e39a37d6e0ba4e" Dec 03 15:03:41 crc kubenswrapper[4805]: E1203 15:03:41.696464 4805 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-gskh4_openshift-machine-config-operator(6098937f-e3f6-45e8-a647-4994a79cd711)\"" pod="openshift-machine-config-operator/machine-config-daemon-gskh4" podUID="6098937f-e3f6-45e8-a647-4994a79cd711" Dec 03 15:03:52 crc kubenswrapper[4805]: I1203 15:03:52.694797 4805 scope.go:117] "RemoveContainer" containerID="13e0da4d7fc8f4085be601d845d688938c16ea7941e52ca511e39a37d6e0ba4e" Dec 03 15:03:52 crc kubenswrapper[4805]: E1203 15:03:52.695952 4805 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-gskh4_openshift-machine-config-operator(6098937f-e3f6-45e8-a647-4994a79cd711)\"" pod="openshift-machine-config-operator/machine-config-daemon-gskh4" podUID="6098937f-e3f6-45e8-a647-4994a79cd711" Dec 03 15:04:04 crc kubenswrapper[4805]: I1203 15:04:04.183212 4805 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-5hfqf"] Dec 03 15:04:04 crc kubenswrapper[4805]: I1203 15:04:04.186669 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-5hfqf" Dec 03 15:04:04 crc kubenswrapper[4805]: I1203 15:04:04.205051 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-5hfqf"] Dec 03 15:04:04 crc kubenswrapper[4805]: I1203 15:04:04.285676 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xvl94\" (UniqueName: \"kubernetes.io/projected/b2bd5da5-21f7-41eb-88ee-a6852e3d4fca-kube-api-access-xvl94\") pod \"community-operators-5hfqf\" (UID: \"b2bd5da5-21f7-41eb-88ee-a6852e3d4fca\") " pod="openshift-marketplace/community-operators-5hfqf" Dec 03 15:04:04 crc kubenswrapper[4805]: I1203 15:04:04.286094 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b2bd5da5-21f7-41eb-88ee-a6852e3d4fca-utilities\") pod \"community-operators-5hfqf\" (UID: \"b2bd5da5-21f7-41eb-88ee-a6852e3d4fca\") " pod="openshift-marketplace/community-operators-5hfqf" Dec 03 15:04:04 crc kubenswrapper[4805]: I1203 15:04:04.286227 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b2bd5da5-21f7-41eb-88ee-a6852e3d4fca-catalog-content\") pod \"community-operators-5hfqf\" (UID: \"b2bd5da5-21f7-41eb-88ee-a6852e3d4fca\") " pod="openshift-marketplace/community-operators-5hfqf" Dec 03 15:04:04 crc kubenswrapper[4805]: I1203 15:04:04.387212 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xvl94\" (UniqueName: \"kubernetes.io/projected/b2bd5da5-21f7-41eb-88ee-a6852e3d4fca-kube-api-access-xvl94\") pod \"community-operators-5hfqf\" (UID: \"b2bd5da5-21f7-41eb-88ee-a6852e3d4fca\") " pod="openshift-marketplace/community-operators-5hfqf" Dec 03 15:04:04 crc kubenswrapper[4805]: I1203 15:04:04.387305 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b2bd5da5-21f7-41eb-88ee-a6852e3d4fca-utilities\") pod \"community-operators-5hfqf\" (UID: \"b2bd5da5-21f7-41eb-88ee-a6852e3d4fca\") " pod="openshift-marketplace/community-operators-5hfqf" Dec 03 15:04:04 crc kubenswrapper[4805]: I1203 15:04:04.387333 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b2bd5da5-21f7-41eb-88ee-a6852e3d4fca-catalog-content\") pod \"community-operators-5hfqf\" (UID: \"b2bd5da5-21f7-41eb-88ee-a6852e3d4fca\") " pod="openshift-marketplace/community-operators-5hfqf" Dec 03 15:04:04 crc kubenswrapper[4805]: I1203 15:04:04.387853 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b2bd5da5-21f7-41eb-88ee-a6852e3d4fca-utilities\") pod \"community-operators-5hfqf\" (UID: \"b2bd5da5-21f7-41eb-88ee-a6852e3d4fca\") " pod="openshift-marketplace/community-operators-5hfqf" Dec 03 15:04:04 crc kubenswrapper[4805]: I1203 15:04:04.387908 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b2bd5da5-21f7-41eb-88ee-a6852e3d4fca-catalog-content\") pod \"community-operators-5hfqf\" (UID: \"b2bd5da5-21f7-41eb-88ee-a6852e3d4fca\") " pod="openshift-marketplace/community-operators-5hfqf" Dec 03 15:04:04 crc kubenswrapper[4805]: I1203 15:04:04.423303 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xvl94\" (UniqueName: \"kubernetes.io/projected/b2bd5da5-21f7-41eb-88ee-a6852e3d4fca-kube-api-access-xvl94\") pod \"community-operators-5hfqf\" (UID: \"b2bd5da5-21f7-41eb-88ee-a6852e3d4fca\") " pod="openshift-marketplace/community-operators-5hfqf" Dec 03 15:04:04 crc kubenswrapper[4805]: I1203 15:04:04.522084 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-5hfqf" Dec 03 15:04:05 crc kubenswrapper[4805]: I1203 15:04:05.053546 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-5hfqf"] Dec 03 15:04:05 crc kubenswrapper[4805]: I1203 15:04:05.089029 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-5hfqf" event={"ID":"b2bd5da5-21f7-41eb-88ee-a6852e3d4fca","Type":"ContainerStarted","Data":"37d93b04eb88342b248c7fb2755641cfc11a2c39b751d62f446edc5414c7d73b"} Dec 03 15:04:06 crc kubenswrapper[4805]: I1203 15:04:06.103353 4805 generic.go:334] "Generic (PLEG): container finished" podID="b2bd5da5-21f7-41eb-88ee-a6852e3d4fca" containerID="fa7aa743815618a0a7f462a927f07f549ddb5ed812e8d6c5be3b3a30875d5d90" exitCode=0 Dec 03 15:04:06 crc kubenswrapper[4805]: I1203 15:04:06.103408 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-5hfqf" event={"ID":"b2bd5da5-21f7-41eb-88ee-a6852e3d4fca","Type":"ContainerDied","Data":"fa7aa743815618a0a7f462a927f07f549ddb5ed812e8d6c5be3b3a30875d5d90"} Dec 03 15:04:07 crc kubenswrapper[4805]: I1203 15:04:07.117245 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-5hfqf" event={"ID":"b2bd5da5-21f7-41eb-88ee-a6852e3d4fca","Type":"ContainerStarted","Data":"421ee28b6a281f2de3196b2cf0730b373bf55cb11ec1a44b44947a0732252c24"} Dec 03 15:04:07 crc kubenswrapper[4805]: I1203 15:04:07.694084 4805 scope.go:117] "RemoveContainer" containerID="13e0da4d7fc8f4085be601d845d688938c16ea7941e52ca511e39a37d6e0ba4e" Dec 03 15:04:07 crc kubenswrapper[4805]: E1203 15:04:07.694357 4805 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-gskh4_openshift-machine-config-operator(6098937f-e3f6-45e8-a647-4994a79cd711)\"" pod="openshift-machine-config-operator/machine-config-daemon-gskh4" podUID="6098937f-e3f6-45e8-a647-4994a79cd711" Dec 03 15:04:08 crc kubenswrapper[4805]: I1203 15:04:08.130740 4805 generic.go:334] "Generic (PLEG): container finished" podID="b2bd5da5-21f7-41eb-88ee-a6852e3d4fca" containerID="421ee28b6a281f2de3196b2cf0730b373bf55cb11ec1a44b44947a0732252c24" exitCode=0 Dec 03 15:04:08 crc kubenswrapper[4805]: I1203 15:04:08.130783 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-5hfqf" event={"ID":"b2bd5da5-21f7-41eb-88ee-a6852e3d4fca","Type":"ContainerDied","Data":"421ee28b6a281f2de3196b2cf0730b373bf55cb11ec1a44b44947a0732252c24"} Dec 03 15:04:09 crc kubenswrapper[4805]: I1203 15:04:09.143784 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-5hfqf" event={"ID":"b2bd5da5-21f7-41eb-88ee-a6852e3d4fca","Type":"ContainerStarted","Data":"047d8d68cf13c3ef5ae465be2592afa53f863428b08f309c702ec29116f2aeda"} Dec 03 15:04:09 crc kubenswrapper[4805]: I1203 15:04:09.169941 4805 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-5hfqf" podStartSLOduration=2.685246438 podStartE2EDuration="5.16982612s" podCreationTimestamp="2025-12-03 15:04:04 +0000 UTC" firstStartedPulling="2025-12-03 15:04:06.105630302 +0000 UTC m=+3275.768547225" lastFinishedPulling="2025-12-03 15:04:08.590209984 +0000 UTC m=+3278.253126907" observedRunningTime="2025-12-03 15:04:09.160620062 +0000 UTC m=+3278.823537005" watchObservedRunningTime="2025-12-03 15:04:09.16982612 +0000 UTC m=+3278.832743063" Dec 03 15:04:14 crc kubenswrapper[4805]: I1203 15:04:14.522932 4805 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-5hfqf" Dec 03 15:04:14 crc kubenswrapper[4805]: I1203 15:04:14.523544 4805 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-5hfqf" Dec 03 15:04:14 crc kubenswrapper[4805]: I1203 15:04:14.577775 4805 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-5hfqf" Dec 03 15:04:15 crc kubenswrapper[4805]: I1203 15:04:15.411506 4805 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-5hfqf" Dec 03 15:04:16 crc kubenswrapper[4805]: I1203 15:04:16.963529 4805 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-5hfqf"] Dec 03 15:04:17 crc kubenswrapper[4805]: I1203 15:04:17.232226 4805 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-5hfqf" podUID="b2bd5da5-21f7-41eb-88ee-a6852e3d4fca" containerName="registry-server" containerID="cri-o://047d8d68cf13c3ef5ae465be2592afa53f863428b08f309c702ec29116f2aeda" gracePeriod=2 Dec 03 15:04:17 crc kubenswrapper[4805]: I1203 15:04:17.712500 4805 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-5hfqf" Dec 03 15:04:17 crc kubenswrapper[4805]: I1203 15:04:17.840175 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b2bd5da5-21f7-41eb-88ee-a6852e3d4fca-utilities\") pod \"b2bd5da5-21f7-41eb-88ee-a6852e3d4fca\" (UID: \"b2bd5da5-21f7-41eb-88ee-a6852e3d4fca\") " Dec 03 15:04:17 crc kubenswrapper[4805]: I1203 15:04:17.840265 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b2bd5da5-21f7-41eb-88ee-a6852e3d4fca-catalog-content\") pod \"b2bd5da5-21f7-41eb-88ee-a6852e3d4fca\" (UID: \"b2bd5da5-21f7-41eb-88ee-a6852e3d4fca\") " Dec 03 15:04:17 crc kubenswrapper[4805]: I1203 15:04:17.840306 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xvl94\" (UniqueName: \"kubernetes.io/projected/b2bd5da5-21f7-41eb-88ee-a6852e3d4fca-kube-api-access-xvl94\") pod \"b2bd5da5-21f7-41eb-88ee-a6852e3d4fca\" (UID: \"b2bd5da5-21f7-41eb-88ee-a6852e3d4fca\") " Dec 03 15:04:17 crc kubenswrapper[4805]: I1203 15:04:17.841018 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b2bd5da5-21f7-41eb-88ee-a6852e3d4fca-utilities" (OuterVolumeSpecName: "utilities") pod "b2bd5da5-21f7-41eb-88ee-a6852e3d4fca" (UID: "b2bd5da5-21f7-41eb-88ee-a6852e3d4fca"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 15:04:17 crc kubenswrapper[4805]: I1203 15:04:17.841575 4805 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b2bd5da5-21f7-41eb-88ee-a6852e3d4fca-utilities\") on node \"crc\" DevicePath \"\"" Dec 03 15:04:17 crc kubenswrapper[4805]: I1203 15:04:17.846514 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b2bd5da5-21f7-41eb-88ee-a6852e3d4fca-kube-api-access-xvl94" (OuterVolumeSpecName: "kube-api-access-xvl94") pod "b2bd5da5-21f7-41eb-88ee-a6852e3d4fca" (UID: "b2bd5da5-21f7-41eb-88ee-a6852e3d4fca"). InnerVolumeSpecName "kube-api-access-xvl94". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 15:04:17 crc kubenswrapper[4805]: I1203 15:04:17.896274 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b2bd5da5-21f7-41eb-88ee-a6852e3d4fca-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "b2bd5da5-21f7-41eb-88ee-a6852e3d4fca" (UID: "b2bd5da5-21f7-41eb-88ee-a6852e3d4fca"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 15:04:17 crc kubenswrapper[4805]: I1203 15:04:17.943768 4805 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b2bd5da5-21f7-41eb-88ee-a6852e3d4fca-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 03 15:04:17 crc kubenswrapper[4805]: I1203 15:04:17.943988 4805 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xvl94\" (UniqueName: \"kubernetes.io/projected/b2bd5da5-21f7-41eb-88ee-a6852e3d4fca-kube-api-access-xvl94\") on node \"crc\" DevicePath \"\"" Dec 03 15:04:18 crc kubenswrapper[4805]: I1203 15:04:18.243100 4805 generic.go:334] "Generic (PLEG): container finished" podID="b2bd5da5-21f7-41eb-88ee-a6852e3d4fca" containerID="047d8d68cf13c3ef5ae465be2592afa53f863428b08f309c702ec29116f2aeda" exitCode=0 Dec 03 15:04:18 crc kubenswrapper[4805]: I1203 15:04:18.243155 4805 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-5hfqf" Dec 03 15:04:18 crc kubenswrapper[4805]: I1203 15:04:18.243164 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-5hfqf" event={"ID":"b2bd5da5-21f7-41eb-88ee-a6852e3d4fca","Type":"ContainerDied","Data":"047d8d68cf13c3ef5ae465be2592afa53f863428b08f309c702ec29116f2aeda"} Dec 03 15:04:18 crc kubenswrapper[4805]: I1203 15:04:18.243207 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-5hfqf" event={"ID":"b2bd5da5-21f7-41eb-88ee-a6852e3d4fca","Type":"ContainerDied","Data":"37d93b04eb88342b248c7fb2755641cfc11a2c39b751d62f446edc5414c7d73b"} Dec 03 15:04:18 crc kubenswrapper[4805]: I1203 15:04:18.243232 4805 scope.go:117] "RemoveContainer" containerID="047d8d68cf13c3ef5ae465be2592afa53f863428b08f309c702ec29116f2aeda" Dec 03 15:04:18 crc kubenswrapper[4805]: I1203 15:04:18.285393 4805 scope.go:117] "RemoveContainer" containerID="421ee28b6a281f2de3196b2cf0730b373bf55cb11ec1a44b44947a0732252c24" Dec 03 15:04:18 crc kubenswrapper[4805]: I1203 15:04:18.287977 4805 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-5hfqf"] Dec 03 15:04:18 crc kubenswrapper[4805]: I1203 15:04:18.295387 4805 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-5hfqf"] Dec 03 15:04:18 crc kubenswrapper[4805]: I1203 15:04:18.308007 4805 scope.go:117] "RemoveContainer" containerID="fa7aa743815618a0a7f462a927f07f549ddb5ed812e8d6c5be3b3a30875d5d90" Dec 03 15:04:18 crc kubenswrapper[4805]: I1203 15:04:18.353978 4805 scope.go:117] "RemoveContainer" containerID="047d8d68cf13c3ef5ae465be2592afa53f863428b08f309c702ec29116f2aeda" Dec 03 15:04:18 crc kubenswrapper[4805]: E1203 15:04:18.354448 4805 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"047d8d68cf13c3ef5ae465be2592afa53f863428b08f309c702ec29116f2aeda\": container with ID starting with 047d8d68cf13c3ef5ae465be2592afa53f863428b08f309c702ec29116f2aeda not found: ID does not exist" containerID="047d8d68cf13c3ef5ae465be2592afa53f863428b08f309c702ec29116f2aeda" Dec 03 15:04:18 crc kubenswrapper[4805]: I1203 15:04:18.354524 4805 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"047d8d68cf13c3ef5ae465be2592afa53f863428b08f309c702ec29116f2aeda"} err="failed to get container status \"047d8d68cf13c3ef5ae465be2592afa53f863428b08f309c702ec29116f2aeda\": rpc error: code = NotFound desc = could not find container \"047d8d68cf13c3ef5ae465be2592afa53f863428b08f309c702ec29116f2aeda\": container with ID starting with 047d8d68cf13c3ef5ae465be2592afa53f863428b08f309c702ec29116f2aeda not found: ID does not exist" Dec 03 15:04:18 crc kubenswrapper[4805]: I1203 15:04:18.354567 4805 scope.go:117] "RemoveContainer" containerID="421ee28b6a281f2de3196b2cf0730b373bf55cb11ec1a44b44947a0732252c24" Dec 03 15:04:18 crc kubenswrapper[4805]: E1203 15:04:18.355010 4805 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"421ee28b6a281f2de3196b2cf0730b373bf55cb11ec1a44b44947a0732252c24\": container with ID starting with 421ee28b6a281f2de3196b2cf0730b373bf55cb11ec1a44b44947a0732252c24 not found: ID does not exist" containerID="421ee28b6a281f2de3196b2cf0730b373bf55cb11ec1a44b44947a0732252c24" Dec 03 15:04:18 crc kubenswrapper[4805]: I1203 15:04:18.355055 4805 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"421ee28b6a281f2de3196b2cf0730b373bf55cb11ec1a44b44947a0732252c24"} err="failed to get container status \"421ee28b6a281f2de3196b2cf0730b373bf55cb11ec1a44b44947a0732252c24\": rpc error: code = NotFound desc = could not find container \"421ee28b6a281f2de3196b2cf0730b373bf55cb11ec1a44b44947a0732252c24\": container with ID starting with 421ee28b6a281f2de3196b2cf0730b373bf55cb11ec1a44b44947a0732252c24 not found: ID does not exist" Dec 03 15:04:18 crc kubenswrapper[4805]: I1203 15:04:18.355081 4805 scope.go:117] "RemoveContainer" containerID="fa7aa743815618a0a7f462a927f07f549ddb5ed812e8d6c5be3b3a30875d5d90" Dec 03 15:04:18 crc kubenswrapper[4805]: E1203 15:04:18.355498 4805 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"fa7aa743815618a0a7f462a927f07f549ddb5ed812e8d6c5be3b3a30875d5d90\": container with ID starting with fa7aa743815618a0a7f462a927f07f549ddb5ed812e8d6c5be3b3a30875d5d90 not found: ID does not exist" containerID="fa7aa743815618a0a7f462a927f07f549ddb5ed812e8d6c5be3b3a30875d5d90" Dec 03 15:04:18 crc kubenswrapper[4805]: I1203 15:04:18.355533 4805 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"fa7aa743815618a0a7f462a927f07f549ddb5ed812e8d6c5be3b3a30875d5d90"} err="failed to get container status \"fa7aa743815618a0a7f462a927f07f549ddb5ed812e8d6c5be3b3a30875d5d90\": rpc error: code = NotFound desc = could not find container \"fa7aa743815618a0a7f462a927f07f549ddb5ed812e8d6c5be3b3a30875d5d90\": container with ID starting with fa7aa743815618a0a7f462a927f07f549ddb5ed812e8d6c5be3b3a30875d5d90 not found: ID does not exist" Dec 03 15:04:18 crc kubenswrapper[4805]: I1203 15:04:18.709134 4805 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b2bd5da5-21f7-41eb-88ee-a6852e3d4fca" path="/var/lib/kubelet/pods/b2bd5da5-21f7-41eb-88ee-a6852e3d4fca/volumes" Dec 03 15:04:19 crc kubenswrapper[4805]: I1203 15:04:19.695274 4805 scope.go:117] "RemoveContainer" containerID="13e0da4d7fc8f4085be601d845d688938c16ea7941e52ca511e39a37d6e0ba4e" Dec 03 15:04:19 crc kubenswrapper[4805]: E1203 15:04:19.696162 4805 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-gskh4_openshift-machine-config-operator(6098937f-e3f6-45e8-a647-4994a79cd711)\"" pod="openshift-machine-config-operator/machine-config-daemon-gskh4" podUID="6098937f-e3f6-45e8-a647-4994a79cd711" Dec 03 15:04:31 crc kubenswrapper[4805]: I1203 15:04:31.695216 4805 scope.go:117] "RemoveContainer" containerID="13e0da4d7fc8f4085be601d845d688938c16ea7941e52ca511e39a37d6e0ba4e" Dec 03 15:04:31 crc kubenswrapper[4805]: E1203 15:04:31.696132 4805 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-gskh4_openshift-machine-config-operator(6098937f-e3f6-45e8-a647-4994a79cd711)\"" pod="openshift-machine-config-operator/machine-config-daemon-gskh4" podUID="6098937f-e3f6-45e8-a647-4994a79cd711" Dec 03 15:04:42 crc kubenswrapper[4805]: I1203 15:04:42.694309 4805 scope.go:117] "RemoveContainer" containerID="13e0da4d7fc8f4085be601d845d688938c16ea7941e52ca511e39a37d6e0ba4e" Dec 03 15:04:42 crc kubenswrapper[4805]: E1203 15:04:42.695173 4805 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-gskh4_openshift-machine-config-operator(6098937f-e3f6-45e8-a647-4994a79cd711)\"" pod="openshift-machine-config-operator/machine-config-daemon-gskh4" podUID="6098937f-e3f6-45e8-a647-4994a79cd711" Dec 03 15:04:57 crc kubenswrapper[4805]: I1203 15:04:57.695592 4805 scope.go:117] "RemoveContainer" containerID="13e0da4d7fc8f4085be601d845d688938c16ea7941e52ca511e39a37d6e0ba4e" Dec 03 15:04:57 crc kubenswrapper[4805]: E1203 15:04:57.696251 4805 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-gskh4_openshift-machine-config-operator(6098937f-e3f6-45e8-a647-4994a79cd711)\"" pod="openshift-machine-config-operator/machine-config-daemon-gskh4" podUID="6098937f-e3f6-45e8-a647-4994a79cd711" Dec 03 15:05:08 crc kubenswrapper[4805]: I1203 15:05:08.694750 4805 scope.go:117] "RemoveContainer" containerID="13e0da4d7fc8f4085be601d845d688938c16ea7941e52ca511e39a37d6e0ba4e" Dec 03 15:05:08 crc kubenswrapper[4805]: E1203 15:05:08.695881 4805 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-gskh4_openshift-machine-config-operator(6098937f-e3f6-45e8-a647-4994a79cd711)\"" pod="openshift-machine-config-operator/machine-config-daemon-gskh4" podUID="6098937f-e3f6-45e8-a647-4994a79cd711" Dec 03 15:05:21 crc kubenswrapper[4805]: I1203 15:05:21.695264 4805 scope.go:117] "RemoveContainer" containerID="13e0da4d7fc8f4085be601d845d688938c16ea7941e52ca511e39a37d6e0ba4e" Dec 03 15:05:21 crc kubenswrapper[4805]: E1203 15:05:21.697177 4805 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-gskh4_openshift-machine-config-operator(6098937f-e3f6-45e8-a647-4994a79cd711)\"" pod="openshift-machine-config-operator/machine-config-daemon-gskh4" podUID="6098937f-e3f6-45e8-a647-4994a79cd711" Dec 03 15:05:30 crc kubenswrapper[4805]: I1203 15:05:30.106133 4805 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_telemetry-operator-controller-manager-cc9d8f87b-qfmmx_8cbfcd07-e136-4bfd-b3c6-6615c911d440/manager/0.log" Dec 03 15:05:33 crc kubenswrapper[4805]: I1203 15:05:33.699620 4805 scope.go:117] "RemoveContainer" containerID="13e0da4d7fc8f4085be601d845d688938c16ea7941e52ca511e39a37d6e0ba4e" Dec 03 15:05:33 crc kubenswrapper[4805]: E1203 15:05:33.700771 4805 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-gskh4_openshift-machine-config-operator(6098937f-e3f6-45e8-a647-4994a79cd711)\"" pod="openshift-machine-config-operator/machine-config-daemon-gskh4" podUID="6098937f-e3f6-45e8-a647-4994a79cd711" Dec 03 15:05:37 crc kubenswrapper[4805]: I1203 15:05:37.289185 4805 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-blr7n"] Dec 03 15:05:37 crc kubenswrapper[4805]: E1203 15:05:37.289983 4805 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b2bd5da5-21f7-41eb-88ee-a6852e3d4fca" containerName="extract-content" Dec 03 15:05:37 crc kubenswrapper[4805]: I1203 15:05:37.289999 4805 state_mem.go:107] "Deleted CPUSet assignment" podUID="b2bd5da5-21f7-41eb-88ee-a6852e3d4fca" containerName="extract-content" Dec 03 15:05:37 crc kubenswrapper[4805]: E1203 15:05:37.290032 4805 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b2bd5da5-21f7-41eb-88ee-a6852e3d4fca" containerName="extract-utilities" Dec 03 15:05:37 crc kubenswrapper[4805]: I1203 15:05:37.290039 4805 state_mem.go:107] "Deleted CPUSet assignment" podUID="b2bd5da5-21f7-41eb-88ee-a6852e3d4fca" containerName="extract-utilities" Dec 03 15:05:37 crc kubenswrapper[4805]: E1203 15:05:37.290062 4805 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b2bd5da5-21f7-41eb-88ee-a6852e3d4fca" containerName="registry-server" Dec 03 15:05:37 crc kubenswrapper[4805]: I1203 15:05:37.290069 4805 state_mem.go:107] "Deleted CPUSet assignment" podUID="b2bd5da5-21f7-41eb-88ee-a6852e3d4fca" containerName="registry-server" Dec 03 15:05:37 crc kubenswrapper[4805]: I1203 15:05:37.290289 4805 memory_manager.go:354] "RemoveStaleState removing state" podUID="b2bd5da5-21f7-41eb-88ee-a6852e3d4fca" containerName="registry-server" Dec 03 15:05:37 crc kubenswrapper[4805]: I1203 15:05:37.291697 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-blr7n" Dec 03 15:05:37 crc kubenswrapper[4805]: I1203 15:05:37.299452 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-blr7n"] Dec 03 15:05:37 crc kubenswrapper[4805]: I1203 15:05:37.387760 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/daceca8f-f2f6-47dd-9f8f-8e5011f65bc9-catalog-content\") pod \"redhat-marketplace-blr7n\" (UID: \"daceca8f-f2f6-47dd-9f8f-8e5011f65bc9\") " pod="openshift-marketplace/redhat-marketplace-blr7n" Dec 03 15:05:37 crc kubenswrapper[4805]: I1203 15:05:37.387814 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/daceca8f-f2f6-47dd-9f8f-8e5011f65bc9-utilities\") pod \"redhat-marketplace-blr7n\" (UID: \"daceca8f-f2f6-47dd-9f8f-8e5011f65bc9\") " pod="openshift-marketplace/redhat-marketplace-blr7n" Dec 03 15:05:37 crc kubenswrapper[4805]: I1203 15:05:37.388030 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-psrmt\" (UniqueName: \"kubernetes.io/projected/daceca8f-f2f6-47dd-9f8f-8e5011f65bc9-kube-api-access-psrmt\") pod \"redhat-marketplace-blr7n\" (UID: \"daceca8f-f2f6-47dd-9f8f-8e5011f65bc9\") " pod="openshift-marketplace/redhat-marketplace-blr7n" Dec 03 15:05:37 crc kubenswrapper[4805]: I1203 15:05:37.490032 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-psrmt\" (UniqueName: \"kubernetes.io/projected/daceca8f-f2f6-47dd-9f8f-8e5011f65bc9-kube-api-access-psrmt\") pod \"redhat-marketplace-blr7n\" (UID: \"daceca8f-f2f6-47dd-9f8f-8e5011f65bc9\") " pod="openshift-marketplace/redhat-marketplace-blr7n" Dec 03 15:05:37 crc kubenswrapper[4805]: I1203 15:05:37.490211 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/daceca8f-f2f6-47dd-9f8f-8e5011f65bc9-catalog-content\") pod \"redhat-marketplace-blr7n\" (UID: \"daceca8f-f2f6-47dd-9f8f-8e5011f65bc9\") " pod="openshift-marketplace/redhat-marketplace-blr7n" Dec 03 15:05:37 crc kubenswrapper[4805]: I1203 15:05:37.490262 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/daceca8f-f2f6-47dd-9f8f-8e5011f65bc9-utilities\") pod \"redhat-marketplace-blr7n\" (UID: \"daceca8f-f2f6-47dd-9f8f-8e5011f65bc9\") " pod="openshift-marketplace/redhat-marketplace-blr7n" Dec 03 15:05:37 crc kubenswrapper[4805]: I1203 15:05:37.490751 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/daceca8f-f2f6-47dd-9f8f-8e5011f65bc9-catalog-content\") pod \"redhat-marketplace-blr7n\" (UID: \"daceca8f-f2f6-47dd-9f8f-8e5011f65bc9\") " pod="openshift-marketplace/redhat-marketplace-blr7n" Dec 03 15:05:37 crc kubenswrapper[4805]: I1203 15:05:37.490787 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/daceca8f-f2f6-47dd-9f8f-8e5011f65bc9-utilities\") pod \"redhat-marketplace-blr7n\" (UID: \"daceca8f-f2f6-47dd-9f8f-8e5011f65bc9\") " pod="openshift-marketplace/redhat-marketplace-blr7n" Dec 03 15:05:37 crc kubenswrapper[4805]: I1203 15:05:37.521302 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-psrmt\" (UniqueName: \"kubernetes.io/projected/daceca8f-f2f6-47dd-9f8f-8e5011f65bc9-kube-api-access-psrmt\") pod \"redhat-marketplace-blr7n\" (UID: \"daceca8f-f2f6-47dd-9f8f-8e5011f65bc9\") " pod="openshift-marketplace/redhat-marketplace-blr7n" Dec 03 15:05:37 crc kubenswrapper[4805]: I1203 15:05:37.643176 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-blr7n" Dec 03 15:05:38 crc kubenswrapper[4805]: I1203 15:05:38.143495 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-blr7n"] Dec 03 15:05:39 crc kubenswrapper[4805]: I1203 15:05:39.064549 4805 generic.go:334] "Generic (PLEG): container finished" podID="daceca8f-f2f6-47dd-9f8f-8e5011f65bc9" containerID="7bc3991714e5e76bde50e07271eaabaa770255bdb7e5a092646760d44fe40dad" exitCode=0 Dec 03 15:05:39 crc kubenswrapper[4805]: I1203 15:05:39.064606 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-blr7n" event={"ID":"daceca8f-f2f6-47dd-9f8f-8e5011f65bc9","Type":"ContainerDied","Data":"7bc3991714e5e76bde50e07271eaabaa770255bdb7e5a092646760d44fe40dad"} Dec 03 15:05:39 crc kubenswrapper[4805]: I1203 15:05:39.064699 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-blr7n" event={"ID":"daceca8f-f2f6-47dd-9f8f-8e5011f65bc9","Type":"ContainerStarted","Data":"03230a40f901019e6bc96bb468a2009e6fe0b81452be166a8c7c6a4ed9aca751"} Dec 03 15:05:40 crc kubenswrapper[4805]: I1203 15:05:40.075764 4805 generic.go:334] "Generic (PLEG): container finished" podID="daceca8f-f2f6-47dd-9f8f-8e5011f65bc9" containerID="2b631297de8583ab4111db2951af9a465e774c24499a740be1938f64380d1d07" exitCode=0 Dec 03 15:05:40 crc kubenswrapper[4805]: I1203 15:05:40.075894 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-blr7n" event={"ID":"daceca8f-f2f6-47dd-9f8f-8e5011f65bc9","Type":"ContainerDied","Data":"2b631297de8583ab4111db2951af9a465e774c24499a740be1938f64380d1d07"} Dec 03 15:05:41 crc kubenswrapper[4805]: I1203 15:05:41.090990 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-blr7n" event={"ID":"daceca8f-f2f6-47dd-9f8f-8e5011f65bc9","Type":"ContainerStarted","Data":"7d04fcc3c3f11e5ff60aba82589e4940c4555b20e6e050f33c35bd6d9b8eb1d6"} Dec 03 15:05:41 crc kubenswrapper[4805]: I1203 15:05:41.123795 4805 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-blr7n" podStartSLOduration=2.7095997560000002 podStartE2EDuration="4.12377888s" podCreationTimestamp="2025-12-03 15:05:37 +0000 UTC" firstStartedPulling="2025-12-03 15:05:39.067121605 +0000 UTC m=+3368.730038528" lastFinishedPulling="2025-12-03 15:05:40.481300729 +0000 UTC m=+3370.144217652" observedRunningTime="2025-12-03 15:05:41.11637762 +0000 UTC m=+3370.779294543" watchObservedRunningTime="2025-12-03 15:05:41.12377888 +0000 UTC m=+3370.786695803" Dec 03 15:05:46 crc kubenswrapper[4805]: I1203 15:05:46.136028 4805 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210zddkr"] Dec 03 15:05:46 crc kubenswrapper[4805]: I1203 15:05:46.139096 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210zddkr" Dec 03 15:05:46 crc kubenswrapper[4805]: I1203 15:05:46.143056 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"default-dockercfg-vmwhc" Dec 03 15:05:46 crc kubenswrapper[4805]: I1203 15:05:46.152447 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210zddkr"] Dec 03 15:05:46 crc kubenswrapper[4805]: I1203 15:05:46.268698 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lbr8k\" (UniqueName: \"kubernetes.io/projected/95163734-0666-4f76-a68f-722bc3752b38-kube-api-access-lbr8k\") pod \"6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210zddkr\" (UID: \"95163734-0666-4f76-a68f-722bc3752b38\") " pod="openshift-marketplace/6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210zddkr" Dec 03 15:05:46 crc kubenswrapper[4805]: I1203 15:05:46.268766 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/95163734-0666-4f76-a68f-722bc3752b38-util\") pod \"6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210zddkr\" (UID: \"95163734-0666-4f76-a68f-722bc3752b38\") " pod="openshift-marketplace/6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210zddkr" Dec 03 15:05:46 crc kubenswrapper[4805]: I1203 15:05:46.268921 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/95163734-0666-4f76-a68f-722bc3752b38-bundle\") pod \"6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210zddkr\" (UID: \"95163734-0666-4f76-a68f-722bc3752b38\") " pod="openshift-marketplace/6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210zddkr" Dec 03 15:05:46 crc kubenswrapper[4805]: I1203 15:05:46.370931 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lbr8k\" (UniqueName: \"kubernetes.io/projected/95163734-0666-4f76-a68f-722bc3752b38-kube-api-access-lbr8k\") pod \"6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210zddkr\" (UID: \"95163734-0666-4f76-a68f-722bc3752b38\") " pod="openshift-marketplace/6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210zddkr" Dec 03 15:05:46 crc kubenswrapper[4805]: I1203 15:05:46.370998 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/95163734-0666-4f76-a68f-722bc3752b38-util\") pod \"6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210zddkr\" (UID: \"95163734-0666-4f76-a68f-722bc3752b38\") " pod="openshift-marketplace/6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210zddkr" Dec 03 15:05:46 crc kubenswrapper[4805]: I1203 15:05:46.371088 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/95163734-0666-4f76-a68f-722bc3752b38-bundle\") pod \"6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210zddkr\" (UID: \"95163734-0666-4f76-a68f-722bc3752b38\") " pod="openshift-marketplace/6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210zddkr" Dec 03 15:05:46 crc kubenswrapper[4805]: I1203 15:05:46.371819 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/95163734-0666-4f76-a68f-722bc3752b38-bundle\") pod \"6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210zddkr\" (UID: \"95163734-0666-4f76-a68f-722bc3752b38\") " pod="openshift-marketplace/6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210zddkr" Dec 03 15:05:46 crc kubenswrapper[4805]: I1203 15:05:46.371864 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/95163734-0666-4f76-a68f-722bc3752b38-util\") pod \"6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210zddkr\" (UID: \"95163734-0666-4f76-a68f-722bc3752b38\") " pod="openshift-marketplace/6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210zddkr" Dec 03 15:05:46 crc kubenswrapper[4805]: I1203 15:05:46.393450 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lbr8k\" (UniqueName: \"kubernetes.io/projected/95163734-0666-4f76-a68f-722bc3752b38-kube-api-access-lbr8k\") pod \"6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210zddkr\" (UID: \"95163734-0666-4f76-a68f-722bc3752b38\") " pod="openshift-marketplace/6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210zddkr" Dec 03 15:05:46 crc kubenswrapper[4805]: I1203 15:05:46.465691 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210zddkr" Dec 03 15:05:46 crc kubenswrapper[4805]: I1203 15:05:46.977753 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210zddkr"] Dec 03 15:05:47 crc kubenswrapper[4805]: I1203 15:05:47.147974 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210zddkr" event={"ID":"95163734-0666-4f76-a68f-722bc3752b38","Type":"ContainerStarted","Data":"70ff6ca2e06d3ccd422e4617de05ea86bdf5c2ba1ae380cd809e532ee2da02d2"} Dec 03 15:05:47 crc kubenswrapper[4805]: I1203 15:05:47.643513 4805 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-blr7n" Dec 03 15:05:47 crc kubenswrapper[4805]: I1203 15:05:47.643811 4805 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-blr7n" Dec 03 15:05:47 crc kubenswrapper[4805]: I1203 15:05:47.694084 4805 scope.go:117] "RemoveContainer" containerID="13e0da4d7fc8f4085be601d845d688938c16ea7941e52ca511e39a37d6e0ba4e" Dec 03 15:05:47 crc kubenswrapper[4805]: E1203 15:05:47.694327 4805 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-gskh4_openshift-machine-config-operator(6098937f-e3f6-45e8-a647-4994a79cd711)\"" pod="openshift-machine-config-operator/machine-config-daemon-gskh4" podUID="6098937f-e3f6-45e8-a647-4994a79cd711" Dec 03 15:05:47 crc kubenswrapper[4805]: I1203 15:05:47.712399 4805 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-blr7n" Dec 03 15:05:48 crc kubenswrapper[4805]: I1203 15:05:48.159856 4805 generic.go:334] "Generic (PLEG): container finished" podID="95163734-0666-4f76-a68f-722bc3752b38" containerID="f8810b087bf53f9f700c96306deaafde43ce1174210585b5e9c8122b91de0048" exitCode=0 Dec 03 15:05:48 crc kubenswrapper[4805]: I1203 15:05:48.161558 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210zddkr" event={"ID":"95163734-0666-4f76-a68f-722bc3752b38","Type":"ContainerDied","Data":"f8810b087bf53f9f700c96306deaafde43ce1174210585b5e9c8122b91de0048"} Dec 03 15:05:48 crc kubenswrapper[4805]: I1203 15:05:48.163447 4805 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 03 15:05:48 crc kubenswrapper[4805]: I1203 15:05:48.213661 4805 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-blr7n" Dec 03 15:05:50 crc kubenswrapper[4805]: I1203 15:05:50.278900 4805 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-blr7n"] Dec 03 15:05:50 crc kubenswrapper[4805]: I1203 15:05:50.279721 4805 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-blr7n" podUID="daceca8f-f2f6-47dd-9f8f-8e5011f65bc9" containerName="registry-server" containerID="cri-o://7d04fcc3c3f11e5ff60aba82589e4940c4555b20e6e050f33c35bd6d9b8eb1d6" gracePeriod=2 Dec 03 15:05:50 crc kubenswrapper[4805]: I1203 15:05:50.815660 4805 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-blr7n" Dec 03 15:05:50 crc kubenswrapper[4805]: I1203 15:05:50.959945 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-psrmt\" (UniqueName: \"kubernetes.io/projected/daceca8f-f2f6-47dd-9f8f-8e5011f65bc9-kube-api-access-psrmt\") pod \"daceca8f-f2f6-47dd-9f8f-8e5011f65bc9\" (UID: \"daceca8f-f2f6-47dd-9f8f-8e5011f65bc9\") " Dec 03 15:05:50 crc kubenswrapper[4805]: I1203 15:05:50.960347 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/daceca8f-f2f6-47dd-9f8f-8e5011f65bc9-utilities\") pod \"daceca8f-f2f6-47dd-9f8f-8e5011f65bc9\" (UID: \"daceca8f-f2f6-47dd-9f8f-8e5011f65bc9\") " Dec 03 15:05:50 crc kubenswrapper[4805]: I1203 15:05:50.960530 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/daceca8f-f2f6-47dd-9f8f-8e5011f65bc9-catalog-content\") pod \"daceca8f-f2f6-47dd-9f8f-8e5011f65bc9\" (UID: \"daceca8f-f2f6-47dd-9f8f-8e5011f65bc9\") " Dec 03 15:05:50 crc kubenswrapper[4805]: I1203 15:05:50.960979 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/daceca8f-f2f6-47dd-9f8f-8e5011f65bc9-utilities" (OuterVolumeSpecName: "utilities") pod "daceca8f-f2f6-47dd-9f8f-8e5011f65bc9" (UID: "daceca8f-f2f6-47dd-9f8f-8e5011f65bc9"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 15:05:50 crc kubenswrapper[4805]: I1203 15:05:50.961268 4805 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/daceca8f-f2f6-47dd-9f8f-8e5011f65bc9-utilities\") on node \"crc\" DevicePath \"\"" Dec 03 15:05:50 crc kubenswrapper[4805]: I1203 15:05:50.970744 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/daceca8f-f2f6-47dd-9f8f-8e5011f65bc9-kube-api-access-psrmt" (OuterVolumeSpecName: "kube-api-access-psrmt") pod "daceca8f-f2f6-47dd-9f8f-8e5011f65bc9" (UID: "daceca8f-f2f6-47dd-9f8f-8e5011f65bc9"). InnerVolumeSpecName "kube-api-access-psrmt". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 15:05:50 crc kubenswrapper[4805]: I1203 15:05:50.975346 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/daceca8f-f2f6-47dd-9f8f-8e5011f65bc9-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "daceca8f-f2f6-47dd-9f8f-8e5011f65bc9" (UID: "daceca8f-f2f6-47dd-9f8f-8e5011f65bc9"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 15:05:51 crc kubenswrapper[4805]: I1203 15:05:51.062527 4805 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-psrmt\" (UniqueName: \"kubernetes.io/projected/daceca8f-f2f6-47dd-9f8f-8e5011f65bc9-kube-api-access-psrmt\") on node \"crc\" DevicePath \"\"" Dec 03 15:05:51 crc kubenswrapper[4805]: I1203 15:05:51.062560 4805 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/daceca8f-f2f6-47dd-9f8f-8e5011f65bc9-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 03 15:05:51 crc kubenswrapper[4805]: I1203 15:05:51.207069 4805 generic.go:334] "Generic (PLEG): container finished" podID="daceca8f-f2f6-47dd-9f8f-8e5011f65bc9" containerID="7d04fcc3c3f11e5ff60aba82589e4940c4555b20e6e050f33c35bd6d9b8eb1d6" exitCode=0 Dec 03 15:05:51 crc kubenswrapper[4805]: I1203 15:05:51.207104 4805 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-blr7n" Dec 03 15:05:51 crc kubenswrapper[4805]: I1203 15:05:51.207122 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-blr7n" event={"ID":"daceca8f-f2f6-47dd-9f8f-8e5011f65bc9","Type":"ContainerDied","Data":"7d04fcc3c3f11e5ff60aba82589e4940c4555b20e6e050f33c35bd6d9b8eb1d6"} Dec 03 15:05:51 crc kubenswrapper[4805]: I1203 15:05:51.207660 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-blr7n" event={"ID":"daceca8f-f2f6-47dd-9f8f-8e5011f65bc9","Type":"ContainerDied","Data":"03230a40f901019e6bc96bb468a2009e6fe0b81452be166a8c7c6a4ed9aca751"} Dec 03 15:05:51 crc kubenswrapper[4805]: I1203 15:05:51.207719 4805 scope.go:117] "RemoveContainer" containerID="7d04fcc3c3f11e5ff60aba82589e4940c4555b20e6e050f33c35bd6d9b8eb1d6" Dec 03 15:05:51 crc kubenswrapper[4805]: I1203 15:05:51.212708 4805 generic.go:334] "Generic (PLEG): container finished" podID="95163734-0666-4f76-a68f-722bc3752b38" containerID="59f495b5feda108410d67bd513fb81fabb8d83d2cbee7ea5c48500b0c6541323" exitCode=0 Dec 03 15:05:51 crc kubenswrapper[4805]: I1203 15:05:51.212756 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210zddkr" event={"ID":"95163734-0666-4f76-a68f-722bc3752b38","Type":"ContainerDied","Data":"59f495b5feda108410d67bd513fb81fabb8d83d2cbee7ea5c48500b0c6541323"} Dec 03 15:05:51 crc kubenswrapper[4805]: I1203 15:05:51.239266 4805 scope.go:117] "RemoveContainer" containerID="2b631297de8583ab4111db2951af9a465e774c24499a740be1938f64380d1d07" Dec 03 15:05:51 crc kubenswrapper[4805]: I1203 15:05:51.266249 4805 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-blr7n"] Dec 03 15:05:51 crc kubenswrapper[4805]: I1203 15:05:51.275146 4805 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-blr7n"] Dec 03 15:05:51 crc kubenswrapper[4805]: I1203 15:05:51.285907 4805 scope.go:117] "RemoveContainer" containerID="7bc3991714e5e76bde50e07271eaabaa770255bdb7e5a092646760d44fe40dad" Dec 03 15:05:51 crc kubenswrapper[4805]: I1203 15:05:51.328412 4805 scope.go:117] "RemoveContainer" containerID="7d04fcc3c3f11e5ff60aba82589e4940c4555b20e6e050f33c35bd6d9b8eb1d6" Dec 03 15:05:51 crc kubenswrapper[4805]: E1203 15:05:51.328944 4805 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7d04fcc3c3f11e5ff60aba82589e4940c4555b20e6e050f33c35bd6d9b8eb1d6\": container with ID starting with 7d04fcc3c3f11e5ff60aba82589e4940c4555b20e6e050f33c35bd6d9b8eb1d6 not found: ID does not exist" containerID="7d04fcc3c3f11e5ff60aba82589e4940c4555b20e6e050f33c35bd6d9b8eb1d6" Dec 03 15:05:51 crc kubenswrapper[4805]: I1203 15:05:51.328993 4805 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7d04fcc3c3f11e5ff60aba82589e4940c4555b20e6e050f33c35bd6d9b8eb1d6"} err="failed to get container status \"7d04fcc3c3f11e5ff60aba82589e4940c4555b20e6e050f33c35bd6d9b8eb1d6\": rpc error: code = NotFound desc = could not find container \"7d04fcc3c3f11e5ff60aba82589e4940c4555b20e6e050f33c35bd6d9b8eb1d6\": container with ID starting with 7d04fcc3c3f11e5ff60aba82589e4940c4555b20e6e050f33c35bd6d9b8eb1d6 not found: ID does not exist" Dec 03 15:05:51 crc kubenswrapper[4805]: I1203 15:05:51.329021 4805 scope.go:117] "RemoveContainer" containerID="2b631297de8583ab4111db2951af9a465e774c24499a740be1938f64380d1d07" Dec 03 15:05:51 crc kubenswrapper[4805]: E1203 15:05:51.329411 4805 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2b631297de8583ab4111db2951af9a465e774c24499a740be1938f64380d1d07\": container with ID starting with 2b631297de8583ab4111db2951af9a465e774c24499a740be1938f64380d1d07 not found: ID does not exist" containerID="2b631297de8583ab4111db2951af9a465e774c24499a740be1938f64380d1d07" Dec 03 15:05:51 crc kubenswrapper[4805]: I1203 15:05:51.329441 4805 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2b631297de8583ab4111db2951af9a465e774c24499a740be1938f64380d1d07"} err="failed to get container status \"2b631297de8583ab4111db2951af9a465e774c24499a740be1938f64380d1d07\": rpc error: code = NotFound desc = could not find container \"2b631297de8583ab4111db2951af9a465e774c24499a740be1938f64380d1d07\": container with ID starting with 2b631297de8583ab4111db2951af9a465e774c24499a740be1938f64380d1d07 not found: ID does not exist" Dec 03 15:05:51 crc kubenswrapper[4805]: I1203 15:05:51.329460 4805 scope.go:117] "RemoveContainer" containerID="7bc3991714e5e76bde50e07271eaabaa770255bdb7e5a092646760d44fe40dad" Dec 03 15:05:51 crc kubenswrapper[4805]: E1203 15:05:51.329798 4805 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7bc3991714e5e76bde50e07271eaabaa770255bdb7e5a092646760d44fe40dad\": container with ID starting with 7bc3991714e5e76bde50e07271eaabaa770255bdb7e5a092646760d44fe40dad not found: ID does not exist" containerID="7bc3991714e5e76bde50e07271eaabaa770255bdb7e5a092646760d44fe40dad" Dec 03 15:05:51 crc kubenswrapper[4805]: I1203 15:05:51.329873 4805 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7bc3991714e5e76bde50e07271eaabaa770255bdb7e5a092646760d44fe40dad"} err="failed to get container status \"7bc3991714e5e76bde50e07271eaabaa770255bdb7e5a092646760d44fe40dad\": rpc error: code = NotFound desc = could not find container \"7bc3991714e5e76bde50e07271eaabaa770255bdb7e5a092646760d44fe40dad\": container with ID starting with 7bc3991714e5e76bde50e07271eaabaa770255bdb7e5a092646760d44fe40dad not found: ID does not exist" Dec 03 15:05:52 crc kubenswrapper[4805]: I1203 15:05:52.228226 4805 generic.go:334] "Generic (PLEG): container finished" podID="95163734-0666-4f76-a68f-722bc3752b38" containerID="b08bf700741c3f9f0971fe4ea10854cabb85e3b519e7008829948b0040ee05c1" exitCode=0 Dec 03 15:05:52 crc kubenswrapper[4805]: I1203 15:05:52.228296 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210zddkr" event={"ID":"95163734-0666-4f76-a68f-722bc3752b38","Type":"ContainerDied","Data":"b08bf700741c3f9f0971fe4ea10854cabb85e3b519e7008829948b0040ee05c1"} Dec 03 15:05:52 crc kubenswrapper[4805]: I1203 15:05:52.705787 4805 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="daceca8f-f2f6-47dd-9f8f-8e5011f65bc9" path="/var/lib/kubelet/pods/daceca8f-f2f6-47dd-9f8f-8e5011f65bc9/volumes" Dec 03 15:05:53 crc kubenswrapper[4805]: I1203 15:05:53.571553 4805 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210zddkr" Dec 03 15:05:53 crc kubenswrapper[4805]: I1203 15:05:53.609148 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/95163734-0666-4f76-a68f-722bc3752b38-bundle\") pod \"95163734-0666-4f76-a68f-722bc3752b38\" (UID: \"95163734-0666-4f76-a68f-722bc3752b38\") " Dec 03 15:05:53 crc kubenswrapper[4805]: I1203 15:05:53.609247 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/95163734-0666-4f76-a68f-722bc3752b38-util\") pod \"95163734-0666-4f76-a68f-722bc3752b38\" (UID: \"95163734-0666-4f76-a68f-722bc3752b38\") " Dec 03 15:05:53 crc kubenswrapper[4805]: I1203 15:05:53.609374 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lbr8k\" (UniqueName: \"kubernetes.io/projected/95163734-0666-4f76-a68f-722bc3752b38-kube-api-access-lbr8k\") pod \"95163734-0666-4f76-a68f-722bc3752b38\" (UID: \"95163734-0666-4f76-a68f-722bc3752b38\") " Dec 03 15:05:53 crc kubenswrapper[4805]: I1203 15:05:53.612092 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/95163734-0666-4f76-a68f-722bc3752b38-bundle" (OuterVolumeSpecName: "bundle") pod "95163734-0666-4f76-a68f-722bc3752b38" (UID: "95163734-0666-4f76-a68f-722bc3752b38"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 15:05:53 crc kubenswrapper[4805]: I1203 15:05:53.617026 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/95163734-0666-4f76-a68f-722bc3752b38-kube-api-access-lbr8k" (OuterVolumeSpecName: "kube-api-access-lbr8k") pod "95163734-0666-4f76-a68f-722bc3752b38" (UID: "95163734-0666-4f76-a68f-722bc3752b38"). InnerVolumeSpecName "kube-api-access-lbr8k". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 15:05:53 crc kubenswrapper[4805]: I1203 15:05:53.622184 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/95163734-0666-4f76-a68f-722bc3752b38-util" (OuterVolumeSpecName: "util") pod "95163734-0666-4f76-a68f-722bc3752b38" (UID: "95163734-0666-4f76-a68f-722bc3752b38"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 15:05:53 crc kubenswrapper[4805]: I1203 15:05:53.712001 4805 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lbr8k\" (UniqueName: \"kubernetes.io/projected/95163734-0666-4f76-a68f-722bc3752b38-kube-api-access-lbr8k\") on node \"crc\" DevicePath \"\"" Dec 03 15:05:53 crc kubenswrapper[4805]: I1203 15:05:53.712066 4805 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/95163734-0666-4f76-a68f-722bc3752b38-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 15:05:53 crc kubenswrapper[4805]: I1203 15:05:53.712081 4805 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/95163734-0666-4f76-a68f-722bc3752b38-util\") on node \"crc\" DevicePath \"\"" Dec 03 15:05:54 crc kubenswrapper[4805]: I1203 15:05:54.251767 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210zddkr" event={"ID":"95163734-0666-4f76-a68f-722bc3752b38","Type":"ContainerDied","Data":"70ff6ca2e06d3ccd422e4617de05ea86bdf5c2ba1ae380cd809e532ee2da02d2"} Dec 03 15:05:54 crc kubenswrapper[4805]: I1203 15:05:54.251812 4805 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="70ff6ca2e06d3ccd422e4617de05ea86bdf5c2ba1ae380cd809e532ee2da02d2" Dec 03 15:05:54 crc kubenswrapper[4805]: I1203 15:05:54.251913 4805 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210zddkr" Dec 03 15:06:00 crc kubenswrapper[4805]: I1203 15:06:00.707909 4805 scope.go:117] "RemoveContainer" containerID="13e0da4d7fc8f4085be601d845d688938c16ea7941e52ca511e39a37d6e0ba4e" Dec 03 15:06:00 crc kubenswrapper[4805]: E1203 15:06:00.708976 4805 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-gskh4_openshift-machine-config-operator(6098937f-e3f6-45e8-a647-4994a79cd711)\"" pod="openshift-machine-config-operator/machine-config-daemon-gskh4" podUID="6098937f-e3f6-45e8-a647-4994a79cd711" Dec 03 15:06:06 crc kubenswrapper[4805]: I1203 15:06:06.592319 4805 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operators/obo-prometheus-operator-668cf9dfbb-rtvd2"] Dec 03 15:06:06 crc kubenswrapper[4805]: E1203 15:06:06.593452 4805 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="daceca8f-f2f6-47dd-9f8f-8e5011f65bc9" containerName="extract-utilities" Dec 03 15:06:06 crc kubenswrapper[4805]: I1203 15:06:06.593469 4805 state_mem.go:107] "Deleted CPUSet assignment" podUID="daceca8f-f2f6-47dd-9f8f-8e5011f65bc9" containerName="extract-utilities" Dec 03 15:06:06 crc kubenswrapper[4805]: E1203 15:06:06.593485 4805 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="daceca8f-f2f6-47dd-9f8f-8e5011f65bc9" containerName="extract-content" Dec 03 15:06:06 crc kubenswrapper[4805]: I1203 15:06:06.593495 4805 state_mem.go:107] "Deleted CPUSet assignment" podUID="daceca8f-f2f6-47dd-9f8f-8e5011f65bc9" containerName="extract-content" Dec 03 15:06:06 crc kubenswrapper[4805]: E1203 15:06:06.593512 4805 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="95163734-0666-4f76-a68f-722bc3752b38" containerName="extract" Dec 03 15:06:06 crc kubenswrapper[4805]: I1203 15:06:06.593520 4805 state_mem.go:107] "Deleted CPUSet assignment" podUID="95163734-0666-4f76-a68f-722bc3752b38" containerName="extract" Dec 03 15:06:06 crc kubenswrapper[4805]: E1203 15:06:06.593533 4805 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="daceca8f-f2f6-47dd-9f8f-8e5011f65bc9" containerName="registry-server" Dec 03 15:06:06 crc kubenswrapper[4805]: I1203 15:06:06.593540 4805 state_mem.go:107] "Deleted CPUSet assignment" podUID="daceca8f-f2f6-47dd-9f8f-8e5011f65bc9" containerName="registry-server" Dec 03 15:06:06 crc kubenswrapper[4805]: E1203 15:06:06.593557 4805 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="95163734-0666-4f76-a68f-722bc3752b38" containerName="pull" Dec 03 15:06:06 crc kubenswrapper[4805]: I1203 15:06:06.593563 4805 state_mem.go:107] "Deleted CPUSet assignment" podUID="95163734-0666-4f76-a68f-722bc3752b38" containerName="pull" Dec 03 15:06:06 crc kubenswrapper[4805]: E1203 15:06:06.593576 4805 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="95163734-0666-4f76-a68f-722bc3752b38" containerName="util" Dec 03 15:06:06 crc kubenswrapper[4805]: I1203 15:06:06.593585 4805 state_mem.go:107] "Deleted CPUSet assignment" podUID="95163734-0666-4f76-a68f-722bc3752b38" containerName="util" Dec 03 15:06:06 crc kubenswrapper[4805]: I1203 15:06:06.593824 4805 memory_manager.go:354] "RemoveStaleState removing state" podUID="daceca8f-f2f6-47dd-9f8f-8e5011f65bc9" containerName="registry-server" Dec 03 15:06:06 crc kubenswrapper[4805]: I1203 15:06:06.593864 4805 memory_manager.go:354] "RemoveStaleState removing state" podUID="95163734-0666-4f76-a68f-722bc3752b38" containerName="extract" Dec 03 15:06:06 crc kubenswrapper[4805]: I1203 15:06:06.594671 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/obo-prometheus-operator-668cf9dfbb-rtvd2" Dec 03 15:06:06 crc kubenswrapper[4805]: I1203 15:06:06.598754 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operators"/"obo-prometheus-operator-dockercfg-tf78v" Dec 03 15:06:06 crc kubenswrapper[4805]: I1203 15:06:06.599682 4805 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operators"/"kube-root-ca.crt" Dec 03 15:06:06 crc kubenswrapper[4805]: I1203 15:06:06.600478 4805 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operators"/"openshift-service-ca.crt" Dec 03 15:06:06 crc kubenswrapper[4805]: I1203 15:06:06.604670 4805 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operators/obo-prometheus-operator-admission-webhook-59d4564ff5-mh8fd"] Dec 03 15:06:06 crc kubenswrapper[4805]: I1203 15:06:06.606210 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/obo-prometheus-operator-admission-webhook-59d4564ff5-mh8fd" Dec 03 15:06:06 crc kubenswrapper[4805]: I1203 15:06:06.610706 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operators"/"obo-prometheus-operator-admission-webhook-dockercfg-nrrhp" Dec 03 15:06:06 crc kubenswrapper[4805]: I1203 15:06:06.610986 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operators"/"obo-prometheus-operator-admission-webhook-service-cert" Dec 03 15:06:06 crc kubenswrapper[4805]: I1203 15:06:06.616597 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/obo-prometheus-operator-668cf9dfbb-rtvd2"] Dec 03 15:06:06 crc kubenswrapper[4805]: I1203 15:06:06.631677 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/obo-prometheus-operator-admission-webhook-59d4564ff5-mh8fd"] Dec 03 15:06:06 crc kubenswrapper[4805]: I1203 15:06:06.656969 4805 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operators/obo-prometheus-operator-admission-webhook-59d4564ff5-49hbr"] Dec 03 15:06:06 crc kubenswrapper[4805]: I1203 15:06:06.658714 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/obo-prometheus-operator-admission-webhook-59d4564ff5-49hbr" Dec 03 15:06:06 crc kubenswrapper[4805]: I1203 15:06:06.669898 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/obo-prometheus-operator-admission-webhook-59d4564ff5-49hbr"] Dec 03 15:06:06 crc kubenswrapper[4805]: I1203 15:06:06.677746 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7n9xf\" (UniqueName: \"kubernetes.io/projected/063df7dd-86e9-4e2b-a354-9baa6e90ba05-kube-api-access-7n9xf\") pod \"obo-prometheus-operator-668cf9dfbb-rtvd2\" (UID: \"063df7dd-86e9-4e2b-a354-9baa6e90ba05\") " pod="openshift-operators/obo-prometheus-operator-668cf9dfbb-rtvd2" Dec 03 15:06:06 crc kubenswrapper[4805]: I1203 15:06:06.677961 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/21da77ce-bcf7-4ad0-97f8-a0c9d85d13db-webhook-cert\") pod \"obo-prometheus-operator-admission-webhook-59d4564ff5-mh8fd\" (UID: \"21da77ce-bcf7-4ad0-97f8-a0c9d85d13db\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-59d4564ff5-mh8fd" Dec 03 15:06:06 crc kubenswrapper[4805]: I1203 15:06:06.678008 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/21da77ce-bcf7-4ad0-97f8-a0c9d85d13db-apiservice-cert\") pod \"obo-prometheus-operator-admission-webhook-59d4564ff5-mh8fd\" (UID: \"21da77ce-bcf7-4ad0-97f8-a0c9d85d13db\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-59d4564ff5-mh8fd" Dec 03 15:06:06 crc kubenswrapper[4805]: I1203 15:06:06.779487 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7n9xf\" (UniqueName: \"kubernetes.io/projected/063df7dd-86e9-4e2b-a354-9baa6e90ba05-kube-api-access-7n9xf\") pod \"obo-prometheus-operator-668cf9dfbb-rtvd2\" (UID: \"063df7dd-86e9-4e2b-a354-9baa6e90ba05\") " pod="openshift-operators/obo-prometheus-operator-668cf9dfbb-rtvd2" Dec 03 15:06:06 crc kubenswrapper[4805]: I1203 15:06:06.779574 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/94ea81e8-e4cc-4317-a369-b73314c79ab8-webhook-cert\") pod \"obo-prometheus-operator-admission-webhook-59d4564ff5-49hbr\" (UID: \"94ea81e8-e4cc-4317-a369-b73314c79ab8\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-59d4564ff5-49hbr" Dec 03 15:06:06 crc kubenswrapper[4805]: I1203 15:06:06.779647 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/94ea81e8-e4cc-4317-a369-b73314c79ab8-apiservice-cert\") pod \"obo-prometheus-operator-admission-webhook-59d4564ff5-49hbr\" (UID: \"94ea81e8-e4cc-4317-a369-b73314c79ab8\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-59d4564ff5-49hbr" Dec 03 15:06:06 crc kubenswrapper[4805]: I1203 15:06:06.779673 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/21da77ce-bcf7-4ad0-97f8-a0c9d85d13db-webhook-cert\") pod \"obo-prometheus-operator-admission-webhook-59d4564ff5-mh8fd\" (UID: \"21da77ce-bcf7-4ad0-97f8-a0c9d85d13db\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-59d4564ff5-mh8fd" Dec 03 15:06:06 crc kubenswrapper[4805]: I1203 15:06:06.779704 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/21da77ce-bcf7-4ad0-97f8-a0c9d85d13db-apiservice-cert\") pod \"obo-prometheus-operator-admission-webhook-59d4564ff5-mh8fd\" (UID: \"21da77ce-bcf7-4ad0-97f8-a0c9d85d13db\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-59d4564ff5-mh8fd" Dec 03 15:06:06 crc kubenswrapper[4805]: I1203 15:06:06.783026 4805 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operators/observability-operator-d8bb48f5d-m6b9n"] Dec 03 15:06:06 crc kubenswrapper[4805]: I1203 15:06:06.784455 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/observability-operator-d8bb48f5d-m6b9n" Dec 03 15:06:06 crc kubenswrapper[4805]: I1203 15:06:06.789817 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operators"/"observability-operator-tls" Dec 03 15:06:06 crc kubenswrapper[4805]: I1203 15:06:06.790024 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operators"/"observability-operator-sa-dockercfg-dtzkd" Dec 03 15:06:06 crc kubenswrapper[4805]: I1203 15:06:06.800668 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7n9xf\" (UniqueName: \"kubernetes.io/projected/063df7dd-86e9-4e2b-a354-9baa6e90ba05-kube-api-access-7n9xf\") pod \"obo-prometheus-operator-668cf9dfbb-rtvd2\" (UID: \"063df7dd-86e9-4e2b-a354-9baa6e90ba05\") " pod="openshift-operators/obo-prometheus-operator-668cf9dfbb-rtvd2" Dec 03 15:06:06 crc kubenswrapper[4805]: I1203 15:06:06.815797 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/21da77ce-bcf7-4ad0-97f8-a0c9d85d13db-apiservice-cert\") pod \"obo-prometheus-operator-admission-webhook-59d4564ff5-mh8fd\" (UID: \"21da77ce-bcf7-4ad0-97f8-a0c9d85d13db\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-59d4564ff5-mh8fd" Dec 03 15:06:06 crc kubenswrapper[4805]: I1203 15:06:06.819040 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/observability-operator-d8bb48f5d-m6b9n"] Dec 03 15:06:06 crc kubenswrapper[4805]: I1203 15:06:06.821769 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/21da77ce-bcf7-4ad0-97f8-a0c9d85d13db-webhook-cert\") pod \"obo-prometheus-operator-admission-webhook-59d4564ff5-mh8fd\" (UID: \"21da77ce-bcf7-4ad0-97f8-a0c9d85d13db\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-59d4564ff5-mh8fd" Dec 03 15:06:06 crc kubenswrapper[4805]: I1203 15:06:06.881247 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/94ea81e8-e4cc-4317-a369-b73314c79ab8-webhook-cert\") pod \"obo-prometheus-operator-admission-webhook-59d4564ff5-49hbr\" (UID: \"94ea81e8-e4cc-4317-a369-b73314c79ab8\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-59d4564ff5-49hbr" Dec 03 15:06:06 crc kubenswrapper[4805]: I1203 15:06:06.881377 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sdbjk\" (UniqueName: \"kubernetes.io/projected/76c82588-3cbd-4565-abf5-5a291f68160e-kube-api-access-sdbjk\") pod \"observability-operator-d8bb48f5d-m6b9n\" (UID: \"76c82588-3cbd-4565-abf5-5a291f68160e\") " pod="openshift-operators/observability-operator-d8bb48f5d-m6b9n" Dec 03 15:06:06 crc kubenswrapper[4805]: I1203 15:06:06.881424 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/94ea81e8-e4cc-4317-a369-b73314c79ab8-apiservice-cert\") pod \"obo-prometheus-operator-admission-webhook-59d4564ff5-49hbr\" (UID: \"94ea81e8-e4cc-4317-a369-b73314c79ab8\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-59d4564ff5-49hbr" Dec 03 15:06:06 crc kubenswrapper[4805]: I1203 15:06:06.881560 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"observability-operator-tls\" (UniqueName: \"kubernetes.io/secret/76c82588-3cbd-4565-abf5-5a291f68160e-observability-operator-tls\") pod \"observability-operator-d8bb48f5d-m6b9n\" (UID: \"76c82588-3cbd-4565-abf5-5a291f68160e\") " pod="openshift-operators/observability-operator-d8bb48f5d-m6b9n" Dec 03 15:06:06 crc kubenswrapper[4805]: I1203 15:06:06.885025 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/94ea81e8-e4cc-4317-a369-b73314c79ab8-webhook-cert\") pod \"obo-prometheus-operator-admission-webhook-59d4564ff5-49hbr\" (UID: \"94ea81e8-e4cc-4317-a369-b73314c79ab8\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-59d4564ff5-49hbr" Dec 03 15:06:06 crc kubenswrapper[4805]: I1203 15:06:06.890326 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/94ea81e8-e4cc-4317-a369-b73314c79ab8-apiservice-cert\") pod \"obo-prometheus-operator-admission-webhook-59d4564ff5-49hbr\" (UID: \"94ea81e8-e4cc-4317-a369-b73314c79ab8\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-59d4564ff5-49hbr" Dec 03 15:06:06 crc kubenswrapper[4805]: I1203 15:06:06.895653 4805 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operators/perses-operator-5446b9c989-22dsf"] Dec 03 15:06:06 crc kubenswrapper[4805]: I1203 15:06:06.906399 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/perses-operator-5446b9c989-22dsf" Dec 03 15:06:06 crc kubenswrapper[4805]: I1203 15:06:06.910435 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operators"/"perses-operator-dockercfg-mwk68" Dec 03 15:06:06 crc kubenswrapper[4805]: I1203 15:06:06.915716 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/obo-prometheus-operator-668cf9dfbb-rtvd2" Dec 03 15:06:06 crc kubenswrapper[4805]: I1203 15:06:06.916536 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/perses-operator-5446b9c989-22dsf"] Dec 03 15:06:06 crc kubenswrapper[4805]: I1203 15:06:06.926575 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/obo-prometheus-operator-admission-webhook-59d4564ff5-mh8fd" Dec 03 15:06:06 crc kubenswrapper[4805]: I1203 15:06:06.981208 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/obo-prometheus-operator-admission-webhook-59d4564ff5-49hbr" Dec 03 15:06:06 crc kubenswrapper[4805]: I1203 15:06:06.983137 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sdbjk\" (UniqueName: \"kubernetes.io/projected/76c82588-3cbd-4565-abf5-5a291f68160e-kube-api-access-sdbjk\") pod \"observability-operator-d8bb48f5d-m6b9n\" (UID: \"76c82588-3cbd-4565-abf5-5a291f68160e\") " pod="openshift-operators/observability-operator-d8bb48f5d-m6b9n" Dec 03 15:06:06 crc kubenswrapper[4805]: I1203 15:06:06.983276 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openshift-service-ca\" (UniqueName: \"kubernetes.io/configmap/78288c6e-a51d-4870-8530-54415f661eb8-openshift-service-ca\") pod \"perses-operator-5446b9c989-22dsf\" (UID: \"78288c6e-a51d-4870-8530-54415f661eb8\") " pod="openshift-operators/perses-operator-5446b9c989-22dsf" Dec 03 15:06:06 crc kubenswrapper[4805]: I1203 15:06:06.983332 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qmgmb\" (UniqueName: \"kubernetes.io/projected/78288c6e-a51d-4870-8530-54415f661eb8-kube-api-access-qmgmb\") pod \"perses-operator-5446b9c989-22dsf\" (UID: \"78288c6e-a51d-4870-8530-54415f661eb8\") " pod="openshift-operators/perses-operator-5446b9c989-22dsf" Dec 03 15:06:06 crc kubenswrapper[4805]: I1203 15:06:06.983412 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"observability-operator-tls\" (UniqueName: \"kubernetes.io/secret/76c82588-3cbd-4565-abf5-5a291f68160e-observability-operator-tls\") pod \"observability-operator-d8bb48f5d-m6b9n\" (UID: \"76c82588-3cbd-4565-abf5-5a291f68160e\") " pod="openshift-operators/observability-operator-d8bb48f5d-m6b9n" Dec 03 15:06:06 crc kubenswrapper[4805]: I1203 15:06:06.987170 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"observability-operator-tls\" (UniqueName: \"kubernetes.io/secret/76c82588-3cbd-4565-abf5-5a291f68160e-observability-operator-tls\") pod \"observability-operator-d8bb48f5d-m6b9n\" (UID: \"76c82588-3cbd-4565-abf5-5a291f68160e\") " pod="openshift-operators/observability-operator-d8bb48f5d-m6b9n" Dec 03 15:06:07 crc kubenswrapper[4805]: I1203 15:06:06.999798 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sdbjk\" (UniqueName: \"kubernetes.io/projected/76c82588-3cbd-4565-abf5-5a291f68160e-kube-api-access-sdbjk\") pod \"observability-operator-d8bb48f5d-m6b9n\" (UID: \"76c82588-3cbd-4565-abf5-5a291f68160e\") " pod="openshift-operators/observability-operator-d8bb48f5d-m6b9n" Dec 03 15:06:07 crc kubenswrapper[4805]: I1203 15:06:07.085201 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openshift-service-ca\" (UniqueName: \"kubernetes.io/configmap/78288c6e-a51d-4870-8530-54415f661eb8-openshift-service-ca\") pod \"perses-operator-5446b9c989-22dsf\" (UID: \"78288c6e-a51d-4870-8530-54415f661eb8\") " pod="openshift-operators/perses-operator-5446b9c989-22dsf" Dec 03 15:06:07 crc kubenswrapper[4805]: I1203 15:06:07.085556 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qmgmb\" (UniqueName: \"kubernetes.io/projected/78288c6e-a51d-4870-8530-54415f661eb8-kube-api-access-qmgmb\") pod \"perses-operator-5446b9c989-22dsf\" (UID: \"78288c6e-a51d-4870-8530-54415f661eb8\") " pod="openshift-operators/perses-operator-5446b9c989-22dsf" Dec 03 15:06:07 crc kubenswrapper[4805]: I1203 15:06:07.087796 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openshift-service-ca\" (UniqueName: \"kubernetes.io/configmap/78288c6e-a51d-4870-8530-54415f661eb8-openshift-service-ca\") pod \"perses-operator-5446b9c989-22dsf\" (UID: \"78288c6e-a51d-4870-8530-54415f661eb8\") " pod="openshift-operators/perses-operator-5446b9c989-22dsf" Dec 03 15:06:07 crc kubenswrapper[4805]: I1203 15:06:07.121975 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qmgmb\" (UniqueName: \"kubernetes.io/projected/78288c6e-a51d-4870-8530-54415f661eb8-kube-api-access-qmgmb\") pod \"perses-operator-5446b9c989-22dsf\" (UID: \"78288c6e-a51d-4870-8530-54415f661eb8\") " pod="openshift-operators/perses-operator-5446b9c989-22dsf" Dec 03 15:06:07 crc kubenswrapper[4805]: I1203 15:06:07.239961 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/observability-operator-d8bb48f5d-m6b9n" Dec 03 15:06:07 crc kubenswrapper[4805]: I1203 15:06:07.363895 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/perses-operator-5446b9c989-22dsf" Dec 03 15:06:07 crc kubenswrapper[4805]: I1203 15:06:07.506288 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/obo-prometheus-operator-admission-webhook-59d4564ff5-mh8fd"] Dec 03 15:06:07 crc kubenswrapper[4805]: W1203 15:06:07.514358 4805 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod063df7dd_86e9_4e2b_a354_9baa6e90ba05.slice/crio-d0d1e2f324b059d2066ecb0271190b58b574a2f50c60eb8aab204d17b4f606e1 WatchSource:0}: Error finding container d0d1e2f324b059d2066ecb0271190b58b574a2f50c60eb8aab204d17b4f606e1: Status 404 returned error can't find the container with id d0d1e2f324b059d2066ecb0271190b58b574a2f50c60eb8aab204d17b4f606e1 Dec 03 15:06:07 crc kubenswrapper[4805]: I1203 15:06:07.518571 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/obo-prometheus-operator-668cf9dfbb-rtvd2"] Dec 03 15:06:07 crc kubenswrapper[4805]: I1203 15:06:07.665938 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/obo-prometheus-operator-admission-webhook-59d4564ff5-49hbr"] Dec 03 15:06:07 crc kubenswrapper[4805]: I1203 15:06:07.803545 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/observability-operator-d8bb48f5d-m6b9n"] Dec 03 15:06:07 crc kubenswrapper[4805]: W1203 15:06:07.808409 4805 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod76c82588_3cbd_4565_abf5_5a291f68160e.slice/crio-9000b1bbbc99a9e319ba943587d0ccdb56f4e59df7622c7317510d0f5160a259 WatchSource:0}: Error finding container 9000b1bbbc99a9e319ba943587d0ccdb56f4e59df7622c7317510d0f5160a259: Status 404 returned error can't find the container with id 9000b1bbbc99a9e319ba943587d0ccdb56f4e59df7622c7317510d0f5160a259 Dec 03 15:06:07 crc kubenswrapper[4805]: W1203 15:06:07.938565 4805 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod78288c6e_a51d_4870_8530_54415f661eb8.slice/crio-4d1132b6d9428dc4b3d12571fd929a05527ef077dfa5f912219cf6f1a56d8f9a WatchSource:0}: Error finding container 4d1132b6d9428dc4b3d12571fd929a05527ef077dfa5f912219cf6f1a56d8f9a: Status 404 returned error can't find the container with id 4d1132b6d9428dc4b3d12571fd929a05527ef077dfa5f912219cf6f1a56d8f9a Dec 03 15:06:07 crc kubenswrapper[4805]: I1203 15:06:07.942467 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/perses-operator-5446b9c989-22dsf"] Dec 03 15:06:08 crc kubenswrapper[4805]: I1203 15:06:08.467588 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/perses-operator-5446b9c989-22dsf" event={"ID":"78288c6e-a51d-4870-8530-54415f661eb8","Type":"ContainerStarted","Data":"4d1132b6d9428dc4b3d12571fd929a05527ef077dfa5f912219cf6f1a56d8f9a"} Dec 03 15:06:08 crc kubenswrapper[4805]: I1203 15:06:08.469081 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/obo-prometheus-operator-admission-webhook-59d4564ff5-mh8fd" event={"ID":"21da77ce-bcf7-4ad0-97f8-a0c9d85d13db","Type":"ContainerStarted","Data":"f1bf1670c338de4ff6f2dfe08e1dc6651a2ebafd7413a3fa043d42eca27a4e6f"} Dec 03 15:06:08 crc kubenswrapper[4805]: I1203 15:06:08.470091 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/observability-operator-d8bb48f5d-m6b9n" event={"ID":"76c82588-3cbd-4565-abf5-5a291f68160e","Type":"ContainerStarted","Data":"9000b1bbbc99a9e319ba943587d0ccdb56f4e59df7622c7317510d0f5160a259"} Dec 03 15:06:08 crc kubenswrapper[4805]: I1203 15:06:08.471034 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/obo-prometheus-operator-668cf9dfbb-rtvd2" event={"ID":"063df7dd-86e9-4e2b-a354-9baa6e90ba05","Type":"ContainerStarted","Data":"d0d1e2f324b059d2066ecb0271190b58b574a2f50c60eb8aab204d17b4f606e1"} Dec 03 15:06:08 crc kubenswrapper[4805]: I1203 15:06:08.472417 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/obo-prometheus-operator-admission-webhook-59d4564ff5-49hbr" event={"ID":"94ea81e8-e4cc-4317-a369-b73314c79ab8","Type":"ContainerStarted","Data":"38fa93eece2597a37c244463ed44b52cbea4b9aef088167da2ba563f0bd3913b"} Dec 03 15:06:13 crc kubenswrapper[4805]: I1203 15:06:13.694534 4805 scope.go:117] "RemoveContainer" containerID="13e0da4d7fc8f4085be601d845d688938c16ea7941e52ca511e39a37d6e0ba4e" Dec 03 15:06:13 crc kubenswrapper[4805]: E1203 15:06:13.695356 4805 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-gskh4_openshift-machine-config-operator(6098937f-e3f6-45e8-a647-4994a79cd711)\"" pod="openshift-machine-config-operator/machine-config-daemon-gskh4" podUID="6098937f-e3f6-45e8-a647-4994a79cd711" Dec 03 15:06:15 crc kubenswrapper[4805]: I1203 15:06:15.427777 4805 prober.go:107] "Probe failed" probeType="Liveness" pod="metallb-system/speaker-jpk6x" podUID="06e98f2f-39d1-4899-9a18-7cc27e006de9" containerName="speaker" probeResult="failure" output="Get \"http://localhost:29150/metrics\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 03 15:06:15 crc kubenswrapper[4805]: I1203 15:06:15.427804 4805 prober.go:107] "Probe failed" probeType="Readiness" pod="metallb-system/speaker-jpk6x" podUID="06e98f2f-39d1-4899-9a18-7cc27e006de9" containerName="speaker" probeResult="failure" output="Get \"http://localhost:29150/metrics\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 03 15:06:25 crc kubenswrapper[4805]: E1203 15:06:25.598894 4805 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/cluster-observability-operator/obo-prometheus-operator-admission-webhook-rhel9@sha256:43d33f0125e6b990f4a972ac4e952a065d7e72dc1690c6c836963b7341734aec" Dec 03 15:06:25 crc kubenswrapper[4805]: E1203 15:06:25.599551 4805 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:prometheus-operator-admission-webhook,Image:registry.redhat.io/cluster-observability-operator/obo-prometheus-operator-admission-webhook-rhel9@sha256:43d33f0125e6b990f4a972ac4e952a065d7e72dc1690c6c836963b7341734aec,Command:[],Args:[--web.enable-tls=true --web.cert-file=/tmp/k8s-webhook-server/serving-certs/tls.crt --web.key-file=/tmp/k8s-webhook-server/serving-certs/tls.key],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:https,HostPort:0,ContainerPort:8443,Protocol:TCP,HostIP:,},},Env:[]EnvVar{EnvVar{Name:OPERATOR_CONDITION_NAME,Value:cluster-observability-operator.v1.3.0,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{200 -3} {} 200m DecimalSI},memory: {{209715200 0} {} BinarySI},},Requests:ResourceList{cpu: {{50 -3} {} 50m DecimalSI},memory: {{52428800 0} {} 50Mi BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:apiservice-cert,ReadOnly:false,MountPath:/apiserver.local.config/certificates,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:webhook-cert,ReadOnly:false,MountPath:/tmp/k8s-webhook-server/serving-certs,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:*true,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod obo-prometheus-operator-admission-webhook-59d4564ff5-mh8fd_openshift-operators(21da77ce-bcf7-4ad0-97f8-a0c9d85d13db): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Dec 03 15:06:25 crc kubenswrapper[4805]: E1203 15:06:25.601576 4805 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"prometheus-operator-admission-webhook\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-operators/obo-prometheus-operator-admission-webhook-59d4564ff5-mh8fd" podUID="21da77ce-bcf7-4ad0-97f8-a0c9d85d13db" Dec 03 15:06:25 crc kubenswrapper[4805]: E1203 15:06:25.649472 4805 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/cluster-observability-operator/obo-prometheus-operator-admission-webhook-rhel9@sha256:43d33f0125e6b990f4a972ac4e952a065d7e72dc1690c6c836963b7341734aec" Dec 03 15:06:25 crc kubenswrapper[4805]: E1203 15:06:25.649647 4805 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:prometheus-operator-admission-webhook,Image:registry.redhat.io/cluster-observability-operator/obo-prometheus-operator-admission-webhook-rhel9@sha256:43d33f0125e6b990f4a972ac4e952a065d7e72dc1690c6c836963b7341734aec,Command:[],Args:[--web.enable-tls=true --web.cert-file=/tmp/k8s-webhook-server/serving-certs/tls.crt --web.key-file=/tmp/k8s-webhook-server/serving-certs/tls.key],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:https,HostPort:0,ContainerPort:8443,Protocol:TCP,HostIP:,},},Env:[]EnvVar{EnvVar{Name:OPERATOR_CONDITION_NAME,Value:cluster-observability-operator.v1.3.0,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{200 -3} {} 200m DecimalSI},memory: {{209715200 0} {} BinarySI},},Requests:ResourceList{cpu: {{50 -3} {} 50m DecimalSI},memory: {{52428800 0} {} 50Mi BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:apiservice-cert,ReadOnly:false,MountPath:/apiserver.local.config/certificates,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:webhook-cert,ReadOnly:false,MountPath:/tmp/k8s-webhook-server/serving-certs,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:*true,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod obo-prometheus-operator-admission-webhook-59d4564ff5-49hbr_openshift-operators(94ea81e8-e4cc-4317-a369-b73314c79ab8): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Dec 03 15:06:25 crc kubenswrapper[4805]: E1203 15:06:25.653921 4805 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"prometheus-operator-admission-webhook\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-operators/obo-prometheus-operator-admission-webhook-59d4564ff5-49hbr" podUID="94ea81e8-e4cc-4317-a369-b73314c79ab8" Dec 03 15:06:25 crc kubenswrapper[4805]: E1203 15:06:25.694481 4805 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"prometheus-operator-admission-webhook\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/cluster-observability-operator/obo-prometheus-operator-admission-webhook-rhel9@sha256:43d33f0125e6b990f4a972ac4e952a065d7e72dc1690c6c836963b7341734aec\\\"\"" pod="openshift-operators/obo-prometheus-operator-admission-webhook-59d4564ff5-mh8fd" podUID="21da77ce-bcf7-4ad0-97f8-a0c9d85d13db" Dec 03 15:06:25 crc kubenswrapper[4805]: E1203 15:06:25.694691 4805 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"prometheus-operator-admission-webhook\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/cluster-observability-operator/obo-prometheus-operator-admission-webhook-rhel9@sha256:43d33f0125e6b990f4a972ac4e952a065d7e72dc1690c6c836963b7341734aec\\\"\"" pod="openshift-operators/obo-prometheus-operator-admission-webhook-59d4564ff5-49hbr" podUID="94ea81e8-e4cc-4317-a369-b73314c79ab8" Dec 03 15:06:26 crc kubenswrapper[4805]: I1203 15:06:26.698345 4805 scope.go:117] "RemoveContainer" containerID="13e0da4d7fc8f4085be601d845d688938c16ea7941e52ca511e39a37d6e0ba4e" Dec 03 15:06:26 crc kubenswrapper[4805]: E1203 15:06:26.702759 4805 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-gskh4_openshift-machine-config-operator(6098937f-e3f6-45e8-a647-4994a79cd711)\"" pod="openshift-machine-config-operator/machine-config-daemon-gskh4" podUID="6098937f-e3f6-45e8-a647-4994a79cd711" Dec 03 15:06:27 crc kubenswrapper[4805]: E1203 15:06:27.294655 4805 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/cluster-observability-operator/perses-rhel9-operator@sha256:9aec4c328ec43e40481e06ca5808deead74b75c0aacb90e9e72966c3fa14f385" Dec 03 15:06:27 crc kubenswrapper[4805]: E1203 15:06:27.295198 4805 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:perses-operator,Image:registry.redhat.io/cluster-observability-operator/perses-rhel9-operator@sha256:9aec4c328ec43e40481e06ca5808deead74b75c0aacb90e9e72966c3fa14f385,Command:[],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:OPERATOR_CONDITION_NAME,Value:cluster-observability-operator.v1.3.0,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{100 -3} {} 100m DecimalSI},memory: {{134217728 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:openshift-service-ca,ReadOnly:true,MountPath:/ca,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-qmgmb,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000350000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod perses-operator-5446b9c989-22dsf_openshift-operators(78288c6e-a51d-4870-8530-54415f661eb8): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Dec 03 15:06:27 crc kubenswrapper[4805]: E1203 15:06:27.296778 4805 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"perses-operator\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-operators/perses-operator-5446b9c989-22dsf" podUID="78288c6e-a51d-4870-8530-54415f661eb8" Dec 03 15:06:27 crc kubenswrapper[4805]: E1203 15:06:27.734531 4805 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"perses-operator\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/cluster-observability-operator/perses-rhel9-operator@sha256:9aec4c328ec43e40481e06ca5808deead74b75c0aacb90e9e72966c3fa14f385\\\"\"" pod="openshift-operators/perses-operator-5446b9c989-22dsf" podUID="78288c6e-a51d-4870-8530-54415f661eb8" Dec 03 15:06:28 crc kubenswrapper[4805]: I1203 15:06:28.944473 4805 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/aodh-0"] Dec 03 15:06:28 crc kubenswrapper[4805]: I1203 15:06:28.945134 4805 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/aodh-0" podUID="267ba6dc-f62a-4b5b-a1ee-0743df62f50e" containerName="aodh-api" containerID="cri-o://23c3e7d8a29458fff84d07a4be9660cc586e135426898c0af6bdef9c908890f2" gracePeriod=30 Dec 03 15:06:28 crc kubenswrapper[4805]: I1203 15:06:28.945623 4805 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/aodh-0" podUID="267ba6dc-f62a-4b5b-a1ee-0743df62f50e" containerName="aodh-listener" containerID="cri-o://3d089bb5731d01e0995badef708b81e343d6017ece53c031ba0d369fe28420c5" gracePeriod=30 Dec 03 15:06:28 crc kubenswrapper[4805]: I1203 15:06:28.945690 4805 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/aodh-0" podUID="267ba6dc-f62a-4b5b-a1ee-0743df62f50e" containerName="aodh-notifier" containerID="cri-o://20af8007ca6c082e6b1ef2f9f83d8ec39921523a6c30a7cc1e2799cbba60883b" gracePeriod=30 Dec 03 15:06:28 crc kubenswrapper[4805]: I1203 15:06:28.945743 4805 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/aodh-0" podUID="267ba6dc-f62a-4b5b-a1ee-0743df62f50e" containerName="aodh-evaluator" containerID="cri-o://9e6627c97099047d600501614c8fb93027bc47e3b457b199ae443cb0f41ba8cc" gracePeriod=30 Dec 03 15:06:29 crc kubenswrapper[4805]: I1203 15:06:29.765510 4805 generic.go:334] "Generic (PLEG): container finished" podID="267ba6dc-f62a-4b5b-a1ee-0743df62f50e" containerID="9e6627c97099047d600501614c8fb93027bc47e3b457b199ae443cb0f41ba8cc" exitCode=0 Dec 03 15:06:29 crc kubenswrapper[4805]: I1203 15:06:29.765591 4805 generic.go:334] "Generic (PLEG): container finished" podID="267ba6dc-f62a-4b5b-a1ee-0743df62f50e" containerID="23c3e7d8a29458fff84d07a4be9660cc586e135426898c0af6bdef9c908890f2" exitCode=0 Dec 03 15:06:29 crc kubenswrapper[4805]: I1203 15:06:29.765605 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-0" event={"ID":"267ba6dc-f62a-4b5b-a1ee-0743df62f50e","Type":"ContainerDied","Data":"9e6627c97099047d600501614c8fb93027bc47e3b457b199ae443cb0f41ba8cc"} Dec 03 15:06:29 crc kubenswrapper[4805]: I1203 15:06:29.765698 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-0" event={"ID":"267ba6dc-f62a-4b5b-a1ee-0743df62f50e","Type":"ContainerDied","Data":"23c3e7d8a29458fff84d07a4be9660cc586e135426898c0af6bdef9c908890f2"} Dec 03 15:06:30 crc kubenswrapper[4805]: I1203 15:06:30.792132 4805 generic.go:334] "Generic (PLEG): container finished" podID="267ba6dc-f62a-4b5b-a1ee-0743df62f50e" containerID="3d089bb5731d01e0995badef708b81e343d6017ece53c031ba0d369fe28420c5" exitCode=0 Dec 03 15:06:30 crc kubenswrapper[4805]: I1203 15:06:30.792682 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-0" event={"ID":"267ba6dc-f62a-4b5b-a1ee-0743df62f50e","Type":"ContainerDied","Data":"3d089bb5731d01e0995badef708b81e343d6017ece53c031ba0d369fe28420c5"} Dec 03 15:06:30 crc kubenswrapper[4805]: E1203 15:06:30.872459 4805 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/cluster-observability-operator/cluster-observability-rhel9-operator@sha256:ce7d2904f7b238aa37dfe74a0b76bf73629e7a14fa52bf54b0ecf030ca36f1bb" Dec 03 15:06:30 crc kubenswrapper[4805]: E1203 15:06:30.873512 4805 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:operator,Image:registry.redhat.io/cluster-observability-operator/cluster-observability-rhel9-operator@sha256:ce7d2904f7b238aa37dfe74a0b76bf73629e7a14fa52bf54b0ecf030ca36f1bb,Command:[],Args:[--namespace=$(NAMESPACE) --images=perses=$(RELATED_IMAGE_PERSES) --images=alertmanager=$(RELATED_IMAGE_ALERTMANAGER) --images=prometheus=$(RELATED_IMAGE_PROMETHEUS) --images=thanos=$(RELATED_IMAGE_THANOS) --images=ui-dashboards=$(RELATED_IMAGE_CONSOLE_DASHBOARDS_PLUGIN) --images=ui-distributed-tracing=$(RELATED_IMAGE_CONSOLE_DISTRIBUTED_TRACING_PLUGIN) --images=ui-distributed-tracing-pf5=$(RELATED_IMAGE_CONSOLE_DISTRIBUTED_TRACING_PLUGIN_PF5) --images=ui-distributed-tracing-pf4=$(RELATED_IMAGE_CONSOLE_DISTRIBUTED_TRACING_PLUGIN_PF4) --images=ui-logging=$(RELATED_IMAGE_CONSOLE_LOGGING_PLUGIN) --images=ui-logging-pf4=$(RELATED_IMAGE_CONSOLE_LOGGING_PLUGIN_PF4) --images=ui-troubleshooting-panel=$(RELATED_IMAGE_CONSOLE_TROUBLESHOOTING_PANEL_PLUGIN) --images=ui-monitoring=$(RELATED_IMAGE_CONSOLE_MONITORING_PLUGIN) --images=ui-monitoring-pf5=$(RELATED_IMAGE_CONSOLE_MONITORING_PLUGIN_PF5) --images=korrel8r=$(RELATED_IMAGE_KORREL8R) --images=health-analyzer=$(RELATED_IMAGE_CLUSTER_HEALTH_ANALYZER) --openshift.enabled=true],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:NAMESPACE,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:metadata.namespace,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},EnvVar{Name:RELATED_IMAGE_ALERTMANAGER,Value:registry.redhat.io/cluster-observability-operator/alertmanager-rhel9@sha256:e718854a7d6ca8accf0fa72db0eb902e46c44d747ad51dc3f06bba0cefaa3c01,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_PROMETHEUS,Value:registry.redhat.io/cluster-observability-operator/prometheus-rhel9@sha256:17ea20be390a94ab39f5cdd7f0cbc2498046eebcf77fe3dec9aa288d5c2cf46b,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_THANOS,Value:registry.redhat.io/cluster-observability-operator/thanos-rhel9@sha256:d972f4faa5e9c121402d23ed85002f26af48ec36b1b71a7489d677b3913d08b4,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_PERSES,Value:registry.redhat.io/cluster-observability-operator/perses-rhel9@sha256:91531137fc1dcd740e277e0f65e120a0176a16f788c14c27925b61aa0b792ade,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_CONSOLE_DASHBOARDS_PLUGIN,Value:registry.redhat.io/cluster-observability-operator/dashboards-console-plugin-rhel9@sha256:a69da8bbca8a28dd2925f864d51cc31cf761b10532c553095ba40b242ef701cb,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_CONSOLE_DISTRIBUTED_TRACING_PLUGIN,Value:registry.redhat.io/cluster-observability-operator/distributed-tracing-console-plugin-rhel9@sha256:897e1bfad1187062725b54d87107bd0155972257a50d8335dd29e1999b828a4f,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_CONSOLE_DISTRIBUTED_TRACING_PLUGIN_PF5,Value:registry.redhat.io/cluster-observability-operator/distributed-tracing-console-plugin-pf5-rhel9@sha256:95fe5b5746ca8c07ac9217ce2d8ac8e6afad17af210f9d8e0074df1310b209a8,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_CONSOLE_DISTRIBUTED_TRACING_PLUGIN_PF4,Value:registry.redhat.io/cluster-observability-operator/distributed-tracing-console-plugin-pf4-rhel9@sha256:e9d9a89e4d8126a62b1852055482258ee528cac6398dd5d43ebad75ace0f33c9,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_CONSOLE_LOGGING_PLUGIN,Value:registry.redhat.io/cluster-observability-operator/logging-console-plugin-rhel9@sha256:ec684a0645ceb917b019af7ddba68c3533416e356ab0d0320a30e75ca7ebb31b,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_CONSOLE_LOGGING_PLUGIN_PF4,Value:registry.redhat.io/cluster-observability-operator/logging-console-plugin-pf4-rhel9@sha256:3b9693fcde9b3a9494fb04735b1f7cfd0426f10be820fdc3f024175c0d3df1c9,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_CONSOLE_TROUBLESHOOTING_PANEL_PLUGIN,Value:registry.redhat.io/cluster-observability-operator/troubleshooting-panel-console-plugin-rhel9@sha256:580606f194180accc8abba099e17a26dca7522ec6d233fa2fdd40312771703e3,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_CONSOLE_MONITORING_PLUGIN,Value:registry.redhat.io/cluster-observability-operator/monitoring-console-plugin-rhel9@sha256:e03777be39e71701935059cd877603874a13ac94daa73219d4e5e545599d78a9,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_CONSOLE_MONITORING_PLUGIN_PF5,Value:registry.redhat.io/cluster-observability-operator/monitoring-console-plugin-pf5-rhel9@sha256:aa47256193cfd2877853878e1ae97d2ab8b8e5deae62b387cbfad02b284d379c,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_KORREL8R,Value:registry.redhat.io/cluster-observability-operator/korrel8r-rhel9@sha256:c595ff56b2cb85514bf4784db6ddb82e4e657e3e708a7fb695fc4997379a94d4,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_CLUSTER_HEALTH_ANALYZER,Value:registry.redhat.io/cluster-observability-operator/cluster-health-analyzer-rhel9@sha256:45a4ec2a519bcec99e886aa91596d5356a2414a2bd103baaef9fa7838c672eb2,ValueFrom:nil,},EnvVar{Name:OPERATOR_CONDITION_NAME,Value:cluster-observability-operator.v1.3.0,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{400 -3} {} 400m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{100 -3} {} 100m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:observability-operator-tls,ReadOnly:true,MountPath:/etc/tls/private,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-sdbjk,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:0,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:0,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000350000,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod observability-operator-d8bb48f5d-m6b9n_openshift-operators(76c82588-3cbd-4565-abf5-5a291f68160e): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Dec 03 15:06:30 crc kubenswrapper[4805]: E1203 15:06:30.874760 4805 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"operator\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-operators/observability-operator-d8bb48f5d-m6b9n" podUID="76c82588-3cbd-4565-abf5-5a291f68160e" Dec 03 15:06:31 crc kubenswrapper[4805]: I1203 15:06:31.806544 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/obo-prometheus-operator-668cf9dfbb-rtvd2" event={"ID":"063df7dd-86e9-4e2b-a354-9baa6e90ba05","Type":"ContainerStarted","Data":"8ddfc284d8e28b35a373a74da8ee2ba1767e1fbe5867e45444e505763985f6ca"} Dec 03 15:06:31 crc kubenswrapper[4805]: E1203 15:06:31.808811 4805 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"operator\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/cluster-observability-operator/cluster-observability-rhel9-operator@sha256:ce7d2904f7b238aa37dfe74a0b76bf73629e7a14fa52bf54b0ecf030ca36f1bb\\\"\"" pod="openshift-operators/observability-operator-d8bb48f5d-m6b9n" podUID="76c82588-3cbd-4565-abf5-5a291f68160e" Dec 03 15:06:31 crc kubenswrapper[4805]: I1203 15:06:31.850956 4805 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operators/obo-prometheus-operator-668cf9dfbb-rtvd2" podStartSLOduration=2.603259747 podStartE2EDuration="25.850928114s" podCreationTimestamp="2025-12-03 15:06:06 +0000 UTC" firstStartedPulling="2025-12-03 15:06:07.522057005 +0000 UTC m=+3397.184973928" lastFinishedPulling="2025-12-03 15:06:30.769725372 +0000 UTC m=+3420.432642295" observedRunningTime="2025-12-03 15:06:31.849734562 +0000 UTC m=+3421.512651495" watchObservedRunningTime="2025-12-03 15:06:31.850928114 +0000 UTC m=+3421.513845067" Dec 03 15:06:39 crc kubenswrapper[4805]: I1203 15:06:39.545727 4805 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-0" Dec 03 15:06:39 crc kubenswrapper[4805]: I1203 15:06:39.623626 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/267ba6dc-f62a-4b5b-a1ee-0743df62f50e-internal-tls-certs\") pod \"267ba6dc-f62a-4b5b-a1ee-0743df62f50e\" (UID: \"267ba6dc-f62a-4b5b-a1ee-0743df62f50e\") " Dec 03 15:06:39 crc kubenswrapper[4805]: I1203 15:06:39.623711 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/267ba6dc-f62a-4b5b-a1ee-0743df62f50e-scripts\") pod \"267ba6dc-f62a-4b5b-a1ee-0743df62f50e\" (UID: \"267ba6dc-f62a-4b5b-a1ee-0743df62f50e\") " Dec 03 15:06:39 crc kubenswrapper[4805]: I1203 15:06:39.623794 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/267ba6dc-f62a-4b5b-a1ee-0743df62f50e-public-tls-certs\") pod \"267ba6dc-f62a-4b5b-a1ee-0743df62f50e\" (UID: \"267ba6dc-f62a-4b5b-a1ee-0743df62f50e\") " Dec 03 15:06:39 crc kubenswrapper[4805]: I1203 15:06:39.623812 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zlhqt\" (UniqueName: \"kubernetes.io/projected/267ba6dc-f62a-4b5b-a1ee-0743df62f50e-kube-api-access-zlhqt\") pod \"267ba6dc-f62a-4b5b-a1ee-0743df62f50e\" (UID: \"267ba6dc-f62a-4b5b-a1ee-0743df62f50e\") " Dec 03 15:06:39 crc kubenswrapper[4805]: I1203 15:06:39.623928 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/267ba6dc-f62a-4b5b-a1ee-0743df62f50e-config-data\") pod \"267ba6dc-f62a-4b5b-a1ee-0743df62f50e\" (UID: \"267ba6dc-f62a-4b5b-a1ee-0743df62f50e\") " Dec 03 15:06:39 crc kubenswrapper[4805]: I1203 15:06:39.623979 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/267ba6dc-f62a-4b5b-a1ee-0743df62f50e-combined-ca-bundle\") pod \"267ba6dc-f62a-4b5b-a1ee-0743df62f50e\" (UID: \"267ba6dc-f62a-4b5b-a1ee-0743df62f50e\") " Dec 03 15:06:39 crc kubenswrapper[4805]: I1203 15:06:39.630615 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/267ba6dc-f62a-4b5b-a1ee-0743df62f50e-scripts" (OuterVolumeSpecName: "scripts") pod "267ba6dc-f62a-4b5b-a1ee-0743df62f50e" (UID: "267ba6dc-f62a-4b5b-a1ee-0743df62f50e"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 15:06:39 crc kubenswrapper[4805]: I1203 15:06:39.631185 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/267ba6dc-f62a-4b5b-a1ee-0743df62f50e-kube-api-access-zlhqt" (OuterVolumeSpecName: "kube-api-access-zlhqt") pod "267ba6dc-f62a-4b5b-a1ee-0743df62f50e" (UID: "267ba6dc-f62a-4b5b-a1ee-0743df62f50e"). InnerVolumeSpecName "kube-api-access-zlhqt". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 15:06:39 crc kubenswrapper[4805]: I1203 15:06:39.719052 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/267ba6dc-f62a-4b5b-a1ee-0743df62f50e-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "267ba6dc-f62a-4b5b-a1ee-0743df62f50e" (UID: "267ba6dc-f62a-4b5b-a1ee-0743df62f50e"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 15:06:39 crc kubenswrapper[4805]: I1203 15:06:39.726367 4805 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/267ba6dc-f62a-4b5b-a1ee-0743df62f50e-scripts\") on node \"crc\" DevicePath \"\"" Dec 03 15:06:39 crc kubenswrapper[4805]: I1203 15:06:39.726425 4805 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/267ba6dc-f62a-4b5b-a1ee-0743df62f50e-public-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 03 15:06:39 crc kubenswrapper[4805]: I1203 15:06:39.726437 4805 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zlhqt\" (UniqueName: \"kubernetes.io/projected/267ba6dc-f62a-4b5b-a1ee-0743df62f50e-kube-api-access-zlhqt\") on node \"crc\" DevicePath \"\"" Dec 03 15:06:39 crc kubenswrapper[4805]: I1203 15:06:39.762330 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/267ba6dc-f62a-4b5b-a1ee-0743df62f50e-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "267ba6dc-f62a-4b5b-a1ee-0743df62f50e" (UID: "267ba6dc-f62a-4b5b-a1ee-0743df62f50e"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 15:06:39 crc kubenswrapper[4805]: I1203 15:06:39.777952 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/267ba6dc-f62a-4b5b-a1ee-0743df62f50e-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "267ba6dc-f62a-4b5b-a1ee-0743df62f50e" (UID: "267ba6dc-f62a-4b5b-a1ee-0743df62f50e"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 15:06:39 crc kubenswrapper[4805]: I1203 15:06:39.793551 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/267ba6dc-f62a-4b5b-a1ee-0743df62f50e-config-data" (OuterVolumeSpecName: "config-data") pod "267ba6dc-f62a-4b5b-a1ee-0743df62f50e" (UID: "267ba6dc-f62a-4b5b-a1ee-0743df62f50e"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 15:06:39 crc kubenswrapper[4805]: I1203 15:06:39.828635 4805 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/267ba6dc-f62a-4b5b-a1ee-0743df62f50e-config-data\") on node \"crc\" DevicePath \"\"" Dec 03 15:06:39 crc kubenswrapper[4805]: I1203 15:06:39.828663 4805 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/267ba6dc-f62a-4b5b-a1ee-0743df62f50e-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 15:06:39 crc kubenswrapper[4805]: I1203 15:06:39.828673 4805 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/267ba6dc-f62a-4b5b-a1ee-0743df62f50e-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 03 15:06:39 crc kubenswrapper[4805]: I1203 15:06:39.892610 4805 generic.go:334] "Generic (PLEG): container finished" podID="267ba6dc-f62a-4b5b-a1ee-0743df62f50e" containerID="20af8007ca6c082e6b1ef2f9f83d8ec39921523a6c30a7cc1e2799cbba60883b" exitCode=0 Dec 03 15:06:39 crc kubenswrapper[4805]: I1203 15:06:39.892667 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-0" event={"ID":"267ba6dc-f62a-4b5b-a1ee-0743df62f50e","Type":"ContainerDied","Data":"20af8007ca6c082e6b1ef2f9f83d8ec39921523a6c30a7cc1e2799cbba60883b"} Dec 03 15:06:39 crc kubenswrapper[4805]: I1203 15:06:39.892711 4805 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-0" Dec 03 15:06:39 crc kubenswrapper[4805]: I1203 15:06:39.892740 4805 scope.go:117] "RemoveContainer" containerID="3d089bb5731d01e0995badef708b81e343d6017ece53c031ba0d369fe28420c5" Dec 03 15:06:39 crc kubenswrapper[4805]: I1203 15:06:39.892728 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-0" event={"ID":"267ba6dc-f62a-4b5b-a1ee-0743df62f50e","Type":"ContainerDied","Data":"9d546eff560853a541cdd332349cbdc606d8eeac10f74991e380a9d298f0b126"} Dec 03 15:06:39 crc kubenswrapper[4805]: I1203 15:06:39.940242 4805 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/aodh-0"] Dec 03 15:06:39 crc kubenswrapper[4805]: I1203 15:06:39.962937 4805 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/aodh-0"] Dec 03 15:06:39 crc kubenswrapper[4805]: I1203 15:06:39.979892 4805 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/aodh-0"] Dec 03 15:06:39 crc kubenswrapper[4805]: E1203 15:06:39.980327 4805 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="267ba6dc-f62a-4b5b-a1ee-0743df62f50e" containerName="aodh-api" Dec 03 15:06:39 crc kubenswrapper[4805]: I1203 15:06:39.980361 4805 state_mem.go:107] "Deleted CPUSet assignment" podUID="267ba6dc-f62a-4b5b-a1ee-0743df62f50e" containerName="aodh-api" Dec 03 15:06:39 crc kubenswrapper[4805]: E1203 15:06:39.980386 4805 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="267ba6dc-f62a-4b5b-a1ee-0743df62f50e" containerName="aodh-notifier" Dec 03 15:06:39 crc kubenswrapper[4805]: I1203 15:06:39.980392 4805 state_mem.go:107] "Deleted CPUSet assignment" podUID="267ba6dc-f62a-4b5b-a1ee-0743df62f50e" containerName="aodh-notifier" Dec 03 15:06:39 crc kubenswrapper[4805]: E1203 15:06:39.980410 4805 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="267ba6dc-f62a-4b5b-a1ee-0743df62f50e" containerName="aodh-evaluator" Dec 03 15:06:39 crc kubenswrapper[4805]: I1203 15:06:39.980417 4805 state_mem.go:107] "Deleted CPUSet assignment" podUID="267ba6dc-f62a-4b5b-a1ee-0743df62f50e" containerName="aodh-evaluator" Dec 03 15:06:39 crc kubenswrapper[4805]: E1203 15:06:39.980429 4805 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="267ba6dc-f62a-4b5b-a1ee-0743df62f50e" containerName="aodh-listener" Dec 03 15:06:39 crc kubenswrapper[4805]: I1203 15:06:39.980435 4805 state_mem.go:107] "Deleted CPUSet assignment" podUID="267ba6dc-f62a-4b5b-a1ee-0743df62f50e" containerName="aodh-listener" Dec 03 15:06:39 crc kubenswrapper[4805]: I1203 15:06:39.980654 4805 memory_manager.go:354] "RemoveStaleState removing state" podUID="267ba6dc-f62a-4b5b-a1ee-0743df62f50e" containerName="aodh-listener" Dec 03 15:06:39 crc kubenswrapper[4805]: I1203 15:06:39.980668 4805 memory_manager.go:354] "RemoveStaleState removing state" podUID="267ba6dc-f62a-4b5b-a1ee-0743df62f50e" containerName="aodh-notifier" Dec 03 15:06:39 crc kubenswrapper[4805]: I1203 15:06:39.980704 4805 memory_manager.go:354] "RemoveStaleState removing state" podUID="267ba6dc-f62a-4b5b-a1ee-0743df62f50e" containerName="aodh-api" Dec 03 15:06:39 crc kubenswrapper[4805]: I1203 15:06:39.980721 4805 memory_manager.go:354] "RemoveStaleState removing state" podUID="267ba6dc-f62a-4b5b-a1ee-0743df62f50e" containerName="aodh-evaluator" Dec 03 15:06:39 crc kubenswrapper[4805]: I1203 15:06:39.986028 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-0" Dec 03 15:06:39 crc kubenswrapper[4805]: I1203 15:06:39.991360 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"aodh-scripts" Dec 03 15:06:39 crc kubenswrapper[4805]: I1203 15:06:39.991395 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-aodh-public-svc" Dec 03 15:06:39 crc kubenswrapper[4805]: I1203 15:06:39.991693 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-aodh-internal-svc" Dec 03 15:06:39 crc kubenswrapper[4805]: I1203 15:06:39.991882 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"telemetry-autoscaling-dockercfg-wc8kz" Dec 03 15:06:39 crc kubenswrapper[4805]: I1203 15:06:39.991911 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"aodh-config-data" Dec 03 15:06:39 crc kubenswrapper[4805]: I1203 15:06:39.993731 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/aodh-0"] Dec 03 15:06:40 crc kubenswrapper[4805]: I1203 15:06:40.124776 4805 scope.go:117] "RemoveContainer" containerID="20af8007ca6c082e6b1ef2f9f83d8ec39921523a6c30a7cc1e2799cbba60883b" Dec 03 15:06:40 crc kubenswrapper[4805]: I1203 15:06:40.136708 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ef0f7bdc-b48e-41a8-ae3d-9eac4016d70a-combined-ca-bundle\") pod \"aodh-0\" (UID: \"ef0f7bdc-b48e-41a8-ae3d-9eac4016d70a\") " pod="openstack/aodh-0" Dec 03 15:06:40 crc kubenswrapper[4805]: I1203 15:06:40.136906 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/ef0f7bdc-b48e-41a8-ae3d-9eac4016d70a-public-tls-certs\") pod \"aodh-0\" (UID: \"ef0f7bdc-b48e-41a8-ae3d-9eac4016d70a\") " pod="openstack/aodh-0" Dec 03 15:06:40 crc kubenswrapper[4805]: I1203 15:06:40.137053 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/ef0f7bdc-b48e-41a8-ae3d-9eac4016d70a-internal-tls-certs\") pod \"aodh-0\" (UID: \"ef0f7bdc-b48e-41a8-ae3d-9eac4016d70a\") " pod="openstack/aodh-0" Dec 03 15:06:40 crc kubenswrapper[4805]: I1203 15:06:40.137166 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ef0f7bdc-b48e-41a8-ae3d-9eac4016d70a-scripts\") pod \"aodh-0\" (UID: \"ef0f7bdc-b48e-41a8-ae3d-9eac4016d70a\") " pod="openstack/aodh-0" Dec 03 15:06:40 crc kubenswrapper[4805]: I1203 15:06:40.137340 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gmtrk\" (UniqueName: \"kubernetes.io/projected/ef0f7bdc-b48e-41a8-ae3d-9eac4016d70a-kube-api-access-gmtrk\") pod \"aodh-0\" (UID: \"ef0f7bdc-b48e-41a8-ae3d-9eac4016d70a\") " pod="openstack/aodh-0" Dec 03 15:06:40 crc kubenswrapper[4805]: I1203 15:06:40.137487 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ef0f7bdc-b48e-41a8-ae3d-9eac4016d70a-config-data\") pod \"aodh-0\" (UID: \"ef0f7bdc-b48e-41a8-ae3d-9eac4016d70a\") " pod="openstack/aodh-0" Dec 03 15:06:40 crc kubenswrapper[4805]: I1203 15:06:40.205828 4805 scope.go:117] "RemoveContainer" containerID="9e6627c97099047d600501614c8fb93027bc47e3b457b199ae443cb0f41ba8cc" Dec 03 15:06:40 crc kubenswrapper[4805]: I1203 15:06:40.230694 4805 scope.go:117] "RemoveContainer" containerID="23c3e7d8a29458fff84d07a4be9660cc586e135426898c0af6bdef9c908890f2" Dec 03 15:06:40 crc kubenswrapper[4805]: I1203 15:06:40.239310 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gmtrk\" (UniqueName: \"kubernetes.io/projected/ef0f7bdc-b48e-41a8-ae3d-9eac4016d70a-kube-api-access-gmtrk\") pod \"aodh-0\" (UID: \"ef0f7bdc-b48e-41a8-ae3d-9eac4016d70a\") " pod="openstack/aodh-0" Dec 03 15:06:40 crc kubenswrapper[4805]: I1203 15:06:40.239363 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ef0f7bdc-b48e-41a8-ae3d-9eac4016d70a-config-data\") pod \"aodh-0\" (UID: \"ef0f7bdc-b48e-41a8-ae3d-9eac4016d70a\") " pod="openstack/aodh-0" Dec 03 15:06:40 crc kubenswrapper[4805]: I1203 15:06:40.239413 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ef0f7bdc-b48e-41a8-ae3d-9eac4016d70a-combined-ca-bundle\") pod \"aodh-0\" (UID: \"ef0f7bdc-b48e-41a8-ae3d-9eac4016d70a\") " pod="openstack/aodh-0" Dec 03 15:06:40 crc kubenswrapper[4805]: I1203 15:06:40.239436 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/ef0f7bdc-b48e-41a8-ae3d-9eac4016d70a-public-tls-certs\") pod \"aodh-0\" (UID: \"ef0f7bdc-b48e-41a8-ae3d-9eac4016d70a\") " pod="openstack/aodh-0" Dec 03 15:06:40 crc kubenswrapper[4805]: I1203 15:06:40.239472 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/ef0f7bdc-b48e-41a8-ae3d-9eac4016d70a-internal-tls-certs\") pod \"aodh-0\" (UID: \"ef0f7bdc-b48e-41a8-ae3d-9eac4016d70a\") " pod="openstack/aodh-0" Dec 03 15:06:40 crc kubenswrapper[4805]: I1203 15:06:40.239515 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ef0f7bdc-b48e-41a8-ae3d-9eac4016d70a-scripts\") pod \"aodh-0\" (UID: \"ef0f7bdc-b48e-41a8-ae3d-9eac4016d70a\") " pod="openstack/aodh-0" Dec 03 15:06:40 crc kubenswrapper[4805]: I1203 15:06:40.244324 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/ef0f7bdc-b48e-41a8-ae3d-9eac4016d70a-internal-tls-certs\") pod \"aodh-0\" (UID: \"ef0f7bdc-b48e-41a8-ae3d-9eac4016d70a\") " pod="openstack/aodh-0" Dec 03 15:06:40 crc kubenswrapper[4805]: I1203 15:06:40.244330 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ef0f7bdc-b48e-41a8-ae3d-9eac4016d70a-combined-ca-bundle\") pod \"aodh-0\" (UID: \"ef0f7bdc-b48e-41a8-ae3d-9eac4016d70a\") " pod="openstack/aodh-0" Dec 03 15:06:40 crc kubenswrapper[4805]: I1203 15:06:40.244594 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ef0f7bdc-b48e-41a8-ae3d-9eac4016d70a-config-data\") pod \"aodh-0\" (UID: \"ef0f7bdc-b48e-41a8-ae3d-9eac4016d70a\") " pod="openstack/aodh-0" Dec 03 15:06:40 crc kubenswrapper[4805]: I1203 15:06:40.244679 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/ef0f7bdc-b48e-41a8-ae3d-9eac4016d70a-public-tls-certs\") pod \"aodh-0\" (UID: \"ef0f7bdc-b48e-41a8-ae3d-9eac4016d70a\") " pod="openstack/aodh-0" Dec 03 15:06:40 crc kubenswrapper[4805]: I1203 15:06:40.244901 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ef0f7bdc-b48e-41a8-ae3d-9eac4016d70a-scripts\") pod \"aodh-0\" (UID: \"ef0f7bdc-b48e-41a8-ae3d-9eac4016d70a\") " pod="openstack/aodh-0" Dec 03 15:06:40 crc kubenswrapper[4805]: I1203 15:06:40.257086 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gmtrk\" (UniqueName: \"kubernetes.io/projected/ef0f7bdc-b48e-41a8-ae3d-9eac4016d70a-kube-api-access-gmtrk\") pod \"aodh-0\" (UID: \"ef0f7bdc-b48e-41a8-ae3d-9eac4016d70a\") " pod="openstack/aodh-0" Dec 03 15:06:40 crc kubenswrapper[4805]: I1203 15:06:40.274868 4805 scope.go:117] "RemoveContainer" containerID="3d089bb5731d01e0995badef708b81e343d6017ece53c031ba0d369fe28420c5" Dec 03 15:06:40 crc kubenswrapper[4805]: E1203 15:06:40.275358 4805 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3d089bb5731d01e0995badef708b81e343d6017ece53c031ba0d369fe28420c5\": container with ID starting with 3d089bb5731d01e0995badef708b81e343d6017ece53c031ba0d369fe28420c5 not found: ID does not exist" containerID="3d089bb5731d01e0995badef708b81e343d6017ece53c031ba0d369fe28420c5" Dec 03 15:06:40 crc kubenswrapper[4805]: I1203 15:06:40.275465 4805 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3d089bb5731d01e0995badef708b81e343d6017ece53c031ba0d369fe28420c5"} err="failed to get container status \"3d089bb5731d01e0995badef708b81e343d6017ece53c031ba0d369fe28420c5\": rpc error: code = NotFound desc = could not find container \"3d089bb5731d01e0995badef708b81e343d6017ece53c031ba0d369fe28420c5\": container with ID starting with 3d089bb5731d01e0995badef708b81e343d6017ece53c031ba0d369fe28420c5 not found: ID does not exist" Dec 03 15:06:40 crc kubenswrapper[4805]: I1203 15:06:40.275535 4805 scope.go:117] "RemoveContainer" containerID="20af8007ca6c082e6b1ef2f9f83d8ec39921523a6c30a7cc1e2799cbba60883b" Dec 03 15:06:40 crc kubenswrapper[4805]: E1203 15:06:40.276328 4805 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"20af8007ca6c082e6b1ef2f9f83d8ec39921523a6c30a7cc1e2799cbba60883b\": container with ID starting with 20af8007ca6c082e6b1ef2f9f83d8ec39921523a6c30a7cc1e2799cbba60883b not found: ID does not exist" containerID="20af8007ca6c082e6b1ef2f9f83d8ec39921523a6c30a7cc1e2799cbba60883b" Dec 03 15:06:40 crc kubenswrapper[4805]: I1203 15:06:40.276381 4805 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"20af8007ca6c082e6b1ef2f9f83d8ec39921523a6c30a7cc1e2799cbba60883b"} err="failed to get container status \"20af8007ca6c082e6b1ef2f9f83d8ec39921523a6c30a7cc1e2799cbba60883b\": rpc error: code = NotFound desc = could not find container \"20af8007ca6c082e6b1ef2f9f83d8ec39921523a6c30a7cc1e2799cbba60883b\": container with ID starting with 20af8007ca6c082e6b1ef2f9f83d8ec39921523a6c30a7cc1e2799cbba60883b not found: ID does not exist" Dec 03 15:06:40 crc kubenswrapper[4805]: I1203 15:06:40.276405 4805 scope.go:117] "RemoveContainer" containerID="9e6627c97099047d600501614c8fb93027bc47e3b457b199ae443cb0f41ba8cc" Dec 03 15:06:40 crc kubenswrapper[4805]: E1203 15:06:40.277428 4805 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9e6627c97099047d600501614c8fb93027bc47e3b457b199ae443cb0f41ba8cc\": container with ID starting with 9e6627c97099047d600501614c8fb93027bc47e3b457b199ae443cb0f41ba8cc not found: ID does not exist" containerID="9e6627c97099047d600501614c8fb93027bc47e3b457b199ae443cb0f41ba8cc" Dec 03 15:06:40 crc kubenswrapper[4805]: I1203 15:06:40.277510 4805 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9e6627c97099047d600501614c8fb93027bc47e3b457b199ae443cb0f41ba8cc"} err="failed to get container status \"9e6627c97099047d600501614c8fb93027bc47e3b457b199ae443cb0f41ba8cc\": rpc error: code = NotFound desc = could not find container \"9e6627c97099047d600501614c8fb93027bc47e3b457b199ae443cb0f41ba8cc\": container with ID starting with 9e6627c97099047d600501614c8fb93027bc47e3b457b199ae443cb0f41ba8cc not found: ID does not exist" Dec 03 15:06:40 crc kubenswrapper[4805]: I1203 15:06:40.277540 4805 scope.go:117] "RemoveContainer" containerID="23c3e7d8a29458fff84d07a4be9660cc586e135426898c0af6bdef9c908890f2" Dec 03 15:06:40 crc kubenswrapper[4805]: E1203 15:06:40.278032 4805 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"23c3e7d8a29458fff84d07a4be9660cc586e135426898c0af6bdef9c908890f2\": container with ID starting with 23c3e7d8a29458fff84d07a4be9660cc586e135426898c0af6bdef9c908890f2 not found: ID does not exist" containerID="23c3e7d8a29458fff84d07a4be9660cc586e135426898c0af6bdef9c908890f2" Dec 03 15:06:40 crc kubenswrapper[4805]: I1203 15:06:40.278111 4805 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"23c3e7d8a29458fff84d07a4be9660cc586e135426898c0af6bdef9c908890f2"} err="failed to get container status \"23c3e7d8a29458fff84d07a4be9660cc586e135426898c0af6bdef9c908890f2\": rpc error: code = NotFound desc = could not find container \"23c3e7d8a29458fff84d07a4be9660cc586e135426898c0af6bdef9c908890f2\": container with ID starting with 23c3e7d8a29458fff84d07a4be9660cc586e135426898c0af6bdef9c908890f2 not found: ID does not exist" Dec 03 15:06:40 crc kubenswrapper[4805]: I1203 15:06:40.329880 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-0" Dec 03 15:06:40 crc kubenswrapper[4805]: I1203 15:06:40.706409 4805 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="267ba6dc-f62a-4b5b-a1ee-0743df62f50e" path="/var/lib/kubelet/pods/267ba6dc-f62a-4b5b-a1ee-0743df62f50e/volumes" Dec 03 15:06:40 crc kubenswrapper[4805]: I1203 15:06:40.871488 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/aodh-0"] Dec 03 15:06:40 crc kubenswrapper[4805]: W1203 15:06:40.872123 4805 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podef0f7bdc_b48e_41a8_ae3d_9eac4016d70a.slice/crio-ffe45c7453df2f2a0b80ed7599a022e262f35174e810bde7bc2d76fad1256201 WatchSource:0}: Error finding container ffe45c7453df2f2a0b80ed7599a022e262f35174e810bde7bc2d76fad1256201: Status 404 returned error can't find the container with id ffe45c7453df2f2a0b80ed7599a022e262f35174e810bde7bc2d76fad1256201 Dec 03 15:06:40 crc kubenswrapper[4805]: I1203 15:06:40.905192 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/obo-prometheus-operator-admission-webhook-59d4564ff5-mh8fd" event={"ID":"21da77ce-bcf7-4ad0-97f8-a0c9d85d13db","Type":"ContainerStarted","Data":"380bd16b6c160be180f52cbc3af90e5d2c54d6152fa504d8ae113f3f18807369"} Dec 03 15:06:40 crc kubenswrapper[4805]: I1203 15:06:40.907574 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/obo-prometheus-operator-admission-webhook-59d4564ff5-49hbr" event={"ID":"94ea81e8-e4cc-4317-a369-b73314c79ab8","Type":"ContainerStarted","Data":"5871b55994bd4b059e0b3e9803d32c14e9e57af0ae5d19daf8d34eab16ec4ff5"} Dec 03 15:06:40 crc kubenswrapper[4805]: I1203 15:06:40.908934 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-0" event={"ID":"ef0f7bdc-b48e-41a8-ae3d-9eac4016d70a","Type":"ContainerStarted","Data":"ffe45c7453df2f2a0b80ed7599a022e262f35174e810bde7bc2d76fad1256201"} Dec 03 15:06:40 crc kubenswrapper[4805]: I1203 15:06:40.925713 4805 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operators/obo-prometheus-operator-admission-webhook-59d4564ff5-mh8fd" podStartSLOduration=2.231695963 podStartE2EDuration="34.925693693s" podCreationTimestamp="2025-12-03 15:06:06 +0000 UTC" firstStartedPulling="2025-12-03 15:06:07.513067107 +0000 UTC m=+3397.175984030" lastFinishedPulling="2025-12-03 15:06:40.207064837 +0000 UTC m=+3429.869981760" observedRunningTime="2025-12-03 15:06:40.918748409 +0000 UTC m=+3430.581665342" watchObservedRunningTime="2025-12-03 15:06:40.925693693 +0000 UTC m=+3430.588610616" Dec 03 15:06:40 crc kubenswrapper[4805]: I1203 15:06:40.950966 4805 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operators/obo-prometheus-operator-admission-webhook-59d4564ff5-49hbr" podStartSLOduration=2.415353207 podStartE2EDuration="34.950937611s" podCreationTimestamp="2025-12-03 15:06:06 +0000 UTC" firstStartedPulling="2025-12-03 15:06:07.672348226 +0000 UTC m=+3397.335265149" lastFinishedPulling="2025-12-03 15:06:40.20793263 +0000 UTC m=+3429.870849553" observedRunningTime="2025-12-03 15:06:40.945411945 +0000 UTC m=+3430.608328868" watchObservedRunningTime="2025-12-03 15:06:40.950937611 +0000 UTC m=+3430.613854534" Dec 03 15:06:41 crc kubenswrapper[4805]: I1203 15:06:41.695080 4805 scope.go:117] "RemoveContainer" containerID="13e0da4d7fc8f4085be601d845d688938c16ea7941e52ca511e39a37d6e0ba4e" Dec 03 15:06:41 crc kubenswrapper[4805]: E1203 15:06:41.695959 4805 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-gskh4_openshift-machine-config-operator(6098937f-e3f6-45e8-a647-4994a79cd711)\"" pod="openshift-machine-config-operator/machine-config-daemon-gskh4" podUID="6098937f-e3f6-45e8-a647-4994a79cd711" Dec 03 15:06:42 crc kubenswrapper[4805]: I1203 15:06:42.931980 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/perses-operator-5446b9c989-22dsf" event={"ID":"78288c6e-a51d-4870-8530-54415f661eb8","Type":"ContainerStarted","Data":"a9f11773933dc1c1f0a6c5a7ab9ba851eebecc88dab27e72aa685db0e75ec48e"} Dec 03 15:06:42 crc kubenswrapper[4805]: I1203 15:06:42.933434 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-0" event={"ID":"ef0f7bdc-b48e-41a8-ae3d-9eac4016d70a","Type":"ContainerStarted","Data":"ef96a03285bf5bb5faad845eb35c37c353fce53de018d916db07ac65f0cbc9e2"} Dec 03 15:06:42 crc kubenswrapper[4805]: I1203 15:06:42.933531 4805 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operators/perses-operator-5446b9c989-22dsf" Dec 03 15:06:42 crc kubenswrapper[4805]: I1203 15:06:42.955615 4805 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operators/perses-operator-5446b9c989-22dsf" podStartSLOduration=2.880686995 podStartE2EDuration="36.955590475s" podCreationTimestamp="2025-12-03 15:06:06 +0000 UTC" firstStartedPulling="2025-12-03 15:06:07.940250753 +0000 UTC m=+3397.603167676" lastFinishedPulling="2025-12-03 15:06:42.015154233 +0000 UTC m=+3431.678071156" observedRunningTime="2025-12-03 15:06:42.952280197 +0000 UTC m=+3432.615197120" watchObservedRunningTime="2025-12-03 15:06:42.955590475 +0000 UTC m=+3432.618507398" Dec 03 15:06:44 crc kubenswrapper[4805]: I1203 15:06:44.961814 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-0" event={"ID":"ef0f7bdc-b48e-41a8-ae3d-9eac4016d70a","Type":"ContainerStarted","Data":"dfdd2ac28b14d459abf2c4b6a74db213eb06cb0368a0422d75becb5b25e09cb3"} Dec 03 15:06:46 crc kubenswrapper[4805]: I1203 15:06:46.979692 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/observability-operator-d8bb48f5d-m6b9n" event={"ID":"76c82588-3cbd-4565-abf5-5a291f68160e","Type":"ContainerStarted","Data":"d47956b00081219591d5a1f8bce05aef9face39c02cecc7bba756c329d04028f"} Dec 03 15:06:46 crc kubenswrapper[4805]: I1203 15:06:46.982450 4805 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operators/observability-operator-d8bb48f5d-m6b9n" Dec 03 15:06:46 crc kubenswrapper[4805]: I1203 15:06:46.983682 4805 patch_prober.go:28] interesting pod/observability-operator-d8bb48f5d-m6b9n container/operator namespace/openshift-operators: Readiness probe status=failure output="Get \"http://10.217.1.6:8081/healthz\": dial tcp 10.217.1.6:8081: connect: connection refused" start-of-body= Dec 03 15:06:46 crc kubenswrapper[4805]: I1203 15:06:46.983726 4805 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operators/observability-operator-d8bb48f5d-m6b9n" podUID="76c82588-3cbd-4565-abf5-5a291f68160e" containerName="operator" probeResult="failure" output="Get \"http://10.217.1.6:8081/healthz\": dial tcp 10.217.1.6:8081: connect: connection refused" Dec 03 15:06:46 crc kubenswrapper[4805]: I1203 15:06:46.986690 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-0" event={"ID":"ef0f7bdc-b48e-41a8-ae3d-9eac4016d70a","Type":"ContainerStarted","Data":"bf67c1c3e37fd095badea3841d170fb3b8345ca6a40d2106e16d2ccf4ed783e0"} Dec 03 15:06:47 crc kubenswrapper[4805]: I1203 15:06:47.011156 4805 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operators/observability-operator-d8bb48f5d-m6b9n" podStartSLOduration=2.195404612 podStartE2EDuration="41.011135296s" podCreationTimestamp="2025-12-03 15:06:06 +0000 UTC" firstStartedPulling="2025-12-03 15:06:07.810890856 +0000 UTC m=+3397.473807779" lastFinishedPulling="2025-12-03 15:06:46.62662153 +0000 UTC m=+3436.289538463" observedRunningTime="2025-12-03 15:06:47.001555382 +0000 UTC m=+3436.664472325" watchObservedRunningTime="2025-12-03 15:06:47.011135296 +0000 UTC m=+3436.674052209" Dec 03 15:06:47 crc kubenswrapper[4805]: I1203 15:06:47.287888 4805 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operators/observability-operator-d8bb48f5d-m6b9n" Dec 03 15:06:47 crc kubenswrapper[4805]: I1203 15:06:47.370892 4805 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operators/perses-operator-5446b9c989-22dsf" Dec 03 15:06:48 crc kubenswrapper[4805]: I1203 15:06:48.006747 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-0" event={"ID":"ef0f7bdc-b48e-41a8-ae3d-9eac4016d70a","Type":"ContainerStarted","Data":"fb6610f613e5e1b6790967f098610eadabd6b001d42a97c2699e49aa08ac0397"} Dec 03 15:06:48 crc kubenswrapper[4805]: I1203 15:06:48.841031 4805 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/alertmanager-metric-storage-0"] Dec 03 15:06:48 crc kubenswrapper[4805]: I1203 15:06:48.843156 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/alertmanager-metric-storage-0" Dec 03 15:06:48 crc kubenswrapper[4805]: I1203 15:06:48.845135 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"alertmanager-metric-storage-generated" Dec 03 15:06:48 crc kubenswrapper[4805]: I1203 15:06:48.846805 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"alertmanager-metric-storage-cluster-tls-config" Dec 03 15:06:48 crc kubenswrapper[4805]: I1203 15:06:48.847125 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"alertmanager-metric-storage-tls-assets-0" Dec 03 15:06:48 crc kubenswrapper[4805]: I1203 15:06:48.847181 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"metric-storage-alertmanager-dockercfg-mtzw9" Dec 03 15:06:48 crc kubenswrapper[4805]: I1203 15:06:48.847568 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"alertmanager-metric-storage-web-config" Dec 03 15:06:48 crc kubenswrapper[4805]: I1203 15:06:48.870656 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/alertmanager-metric-storage-0"] Dec 03 15:06:48 crc kubenswrapper[4805]: I1203 15:06:48.930620 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/a0412bf9-ae66-4d99-9bb3-ea2fe4637678-tls-assets\") pod \"alertmanager-metric-storage-0\" (UID: \"a0412bf9-ae66-4d99-9bb3-ea2fe4637678\") " pod="openstack/alertmanager-metric-storage-0" Dec 03 15:06:48 crc kubenswrapper[4805]: I1203 15:06:48.930766 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"alertmanager-metric-storage-db\" (UniqueName: \"kubernetes.io/empty-dir/a0412bf9-ae66-4d99-9bb3-ea2fe4637678-alertmanager-metric-storage-db\") pod \"alertmanager-metric-storage-0\" (UID: \"a0412bf9-ae66-4d99-9bb3-ea2fe4637678\") " pod="openstack/alertmanager-metric-storage-0" Dec 03 15:06:48 crc kubenswrapper[4805]: I1203 15:06:48.930925 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/secret/a0412bf9-ae66-4d99-9bb3-ea2fe4637678-config-volume\") pod \"alertmanager-metric-storage-0\" (UID: \"a0412bf9-ae66-4d99-9bb3-ea2fe4637678\") " pod="openstack/alertmanager-metric-storage-0" Dec 03 15:06:48 crc kubenswrapper[4805]: I1203 15:06:48.930986 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/a0412bf9-ae66-4d99-9bb3-ea2fe4637678-config-out\") pod \"alertmanager-metric-storage-0\" (UID: \"a0412bf9-ae66-4d99-9bb3-ea2fe4637678\") " pod="openstack/alertmanager-metric-storage-0" Dec 03 15:06:48 crc kubenswrapper[4805]: I1203 15:06:48.931017 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/a0412bf9-ae66-4d99-9bb3-ea2fe4637678-web-config\") pod \"alertmanager-metric-storage-0\" (UID: \"a0412bf9-ae66-4d99-9bb3-ea2fe4637678\") " pod="openstack/alertmanager-metric-storage-0" Dec 03 15:06:48 crc kubenswrapper[4805]: I1203 15:06:48.931038 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-k4k6f\" (UniqueName: \"kubernetes.io/projected/a0412bf9-ae66-4d99-9bb3-ea2fe4637678-kube-api-access-k4k6f\") pod \"alertmanager-metric-storage-0\" (UID: \"a0412bf9-ae66-4d99-9bb3-ea2fe4637678\") " pod="openstack/alertmanager-metric-storage-0" Dec 03 15:06:48 crc kubenswrapper[4805]: I1203 15:06:48.931084 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cluster-tls-config\" (UniqueName: \"kubernetes.io/secret/a0412bf9-ae66-4d99-9bb3-ea2fe4637678-cluster-tls-config\") pod \"alertmanager-metric-storage-0\" (UID: \"a0412bf9-ae66-4d99-9bb3-ea2fe4637678\") " pod="openstack/alertmanager-metric-storage-0" Dec 03 15:06:49 crc kubenswrapper[4805]: I1203 15:06:49.032321 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/secret/a0412bf9-ae66-4d99-9bb3-ea2fe4637678-config-volume\") pod \"alertmanager-metric-storage-0\" (UID: \"a0412bf9-ae66-4d99-9bb3-ea2fe4637678\") " pod="openstack/alertmanager-metric-storage-0" Dec 03 15:06:49 crc kubenswrapper[4805]: I1203 15:06:49.032384 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/a0412bf9-ae66-4d99-9bb3-ea2fe4637678-config-out\") pod \"alertmanager-metric-storage-0\" (UID: \"a0412bf9-ae66-4d99-9bb3-ea2fe4637678\") " pod="openstack/alertmanager-metric-storage-0" Dec 03 15:06:49 crc kubenswrapper[4805]: I1203 15:06:49.032407 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/a0412bf9-ae66-4d99-9bb3-ea2fe4637678-web-config\") pod \"alertmanager-metric-storage-0\" (UID: \"a0412bf9-ae66-4d99-9bb3-ea2fe4637678\") " pod="openstack/alertmanager-metric-storage-0" Dec 03 15:06:49 crc kubenswrapper[4805]: I1203 15:06:49.032424 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-k4k6f\" (UniqueName: \"kubernetes.io/projected/a0412bf9-ae66-4d99-9bb3-ea2fe4637678-kube-api-access-k4k6f\") pod \"alertmanager-metric-storage-0\" (UID: \"a0412bf9-ae66-4d99-9bb3-ea2fe4637678\") " pod="openstack/alertmanager-metric-storage-0" Dec 03 15:06:49 crc kubenswrapper[4805]: I1203 15:06:49.032457 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cluster-tls-config\" (UniqueName: \"kubernetes.io/secret/a0412bf9-ae66-4d99-9bb3-ea2fe4637678-cluster-tls-config\") pod \"alertmanager-metric-storage-0\" (UID: \"a0412bf9-ae66-4d99-9bb3-ea2fe4637678\") " pod="openstack/alertmanager-metric-storage-0" Dec 03 15:06:49 crc kubenswrapper[4805]: I1203 15:06:49.032491 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/a0412bf9-ae66-4d99-9bb3-ea2fe4637678-tls-assets\") pod \"alertmanager-metric-storage-0\" (UID: \"a0412bf9-ae66-4d99-9bb3-ea2fe4637678\") " pod="openstack/alertmanager-metric-storage-0" Dec 03 15:06:49 crc kubenswrapper[4805]: I1203 15:06:49.032569 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"alertmanager-metric-storage-db\" (UniqueName: \"kubernetes.io/empty-dir/a0412bf9-ae66-4d99-9bb3-ea2fe4637678-alertmanager-metric-storage-db\") pod \"alertmanager-metric-storage-0\" (UID: \"a0412bf9-ae66-4d99-9bb3-ea2fe4637678\") " pod="openstack/alertmanager-metric-storage-0" Dec 03 15:06:49 crc kubenswrapper[4805]: I1203 15:06:49.033050 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"alertmanager-metric-storage-db\" (UniqueName: \"kubernetes.io/empty-dir/a0412bf9-ae66-4d99-9bb3-ea2fe4637678-alertmanager-metric-storage-db\") pod \"alertmanager-metric-storage-0\" (UID: \"a0412bf9-ae66-4d99-9bb3-ea2fe4637678\") " pod="openstack/alertmanager-metric-storage-0" Dec 03 15:06:49 crc kubenswrapper[4805]: I1203 15:06:49.037769 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/secret/a0412bf9-ae66-4d99-9bb3-ea2fe4637678-config-volume\") pod \"alertmanager-metric-storage-0\" (UID: \"a0412bf9-ae66-4d99-9bb3-ea2fe4637678\") " pod="openstack/alertmanager-metric-storage-0" Dec 03 15:06:49 crc kubenswrapper[4805]: I1203 15:06:49.038279 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/a0412bf9-ae66-4d99-9bb3-ea2fe4637678-config-out\") pod \"alertmanager-metric-storage-0\" (UID: \"a0412bf9-ae66-4d99-9bb3-ea2fe4637678\") " pod="openstack/alertmanager-metric-storage-0" Dec 03 15:06:49 crc kubenswrapper[4805]: I1203 15:06:49.038870 4805 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/aodh-0" podStartSLOduration=3.325745249 podStartE2EDuration="10.038831749s" podCreationTimestamp="2025-12-03 15:06:39 +0000 UTC" firstStartedPulling="2025-12-03 15:06:40.873947212 +0000 UTC m=+3430.536864135" lastFinishedPulling="2025-12-03 15:06:47.587033702 +0000 UTC m=+3437.249950635" observedRunningTime="2025-12-03 15:06:49.034104294 +0000 UTC m=+3438.697021217" watchObservedRunningTime="2025-12-03 15:06:49.038831749 +0000 UTC m=+3438.701748672" Dec 03 15:06:49 crc kubenswrapper[4805]: I1203 15:06:49.045496 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/a0412bf9-ae66-4d99-9bb3-ea2fe4637678-web-config\") pod \"alertmanager-metric-storage-0\" (UID: \"a0412bf9-ae66-4d99-9bb3-ea2fe4637678\") " pod="openstack/alertmanager-metric-storage-0" Dec 03 15:06:49 crc kubenswrapper[4805]: I1203 15:06:49.047299 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cluster-tls-config\" (UniqueName: \"kubernetes.io/secret/a0412bf9-ae66-4d99-9bb3-ea2fe4637678-cluster-tls-config\") pod \"alertmanager-metric-storage-0\" (UID: \"a0412bf9-ae66-4d99-9bb3-ea2fe4637678\") " pod="openstack/alertmanager-metric-storage-0" Dec 03 15:06:49 crc kubenswrapper[4805]: I1203 15:06:49.049466 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/a0412bf9-ae66-4d99-9bb3-ea2fe4637678-tls-assets\") pod \"alertmanager-metric-storage-0\" (UID: \"a0412bf9-ae66-4d99-9bb3-ea2fe4637678\") " pod="openstack/alertmanager-metric-storage-0" Dec 03 15:06:49 crc kubenswrapper[4805]: I1203 15:06:49.055757 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-k4k6f\" (UniqueName: \"kubernetes.io/projected/a0412bf9-ae66-4d99-9bb3-ea2fe4637678-kube-api-access-k4k6f\") pod \"alertmanager-metric-storage-0\" (UID: \"a0412bf9-ae66-4d99-9bb3-ea2fe4637678\") " pod="openstack/alertmanager-metric-storage-0" Dec 03 15:06:49 crc kubenswrapper[4805]: I1203 15:06:49.159733 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/alertmanager-metric-storage-0" Dec 03 15:06:49 crc kubenswrapper[4805]: I1203 15:06:49.469500 4805 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/prometheus-metric-storage-0"] Dec 03 15:06:49 crc kubenswrapper[4805]: I1203 15:06:49.473740 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/prometheus-metric-storage-0" Dec 03 15:06:49 crc kubenswrapper[4805]: I1203 15:06:49.476478 4805 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"prometheus-metric-storage-rulefiles-0" Dec 03 15:06:49 crc kubenswrapper[4805]: I1203 15:06:49.476720 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"prometheus-metric-storage-thanos-prometheus-http-client-file" Dec 03 15:06:49 crc kubenswrapper[4805]: I1203 15:06:49.477070 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"prometheus-metric-storage" Dec 03 15:06:49 crc kubenswrapper[4805]: I1203 15:06:49.480150 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"metric-storage-prometheus-dockercfg-kg4lr" Dec 03 15:06:49 crc kubenswrapper[4805]: I1203 15:06:49.480330 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"prometheus-metric-storage-web-config" Dec 03 15:06:49 crc kubenswrapper[4805]: I1203 15:06:49.480447 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"prometheus-metric-storage-tls-assets-0" Dec 03 15:06:49 crc kubenswrapper[4805]: I1203 15:06:49.486990 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/prometheus-metric-storage-0"] Dec 03 15:06:49 crc kubenswrapper[4805]: I1203 15:06:49.553286 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"prometheus-metric-storage-rulefiles-0\" (UniqueName: \"kubernetes.io/configmap/2d49b230-da3b-4b61-bd0f-35408593cb08-prometheus-metric-storage-rulefiles-0\") pod \"prometheus-metric-storage-0\" (UID: \"2d49b230-da3b-4b61-bd0f-35408593cb08\") " pod="openstack/prometheus-metric-storage-0" Dec 03 15:06:49 crc kubenswrapper[4805]: I1203 15:06:49.553357 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"prometheus-metric-storage-0\" (UID: \"2d49b230-da3b-4b61-bd0f-35408593cb08\") " pod="openstack/prometheus-metric-storage-0" Dec 03 15:06:49 crc kubenswrapper[4805]: I1203 15:06:49.553471 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/2d49b230-da3b-4b61-bd0f-35408593cb08-web-config\") pod \"prometheus-metric-storage-0\" (UID: \"2d49b230-da3b-4b61-bd0f-35408593cb08\") " pod="openstack/prometheus-metric-storage-0" Dec 03 15:06:49 crc kubenswrapper[4805]: I1203 15:06:49.553513 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/2d49b230-da3b-4b61-bd0f-35408593cb08-tls-assets\") pod \"prometheus-metric-storage-0\" (UID: \"2d49b230-da3b-4b61-bd0f-35408593cb08\") " pod="openstack/prometheus-metric-storage-0" Dec 03 15:06:49 crc kubenswrapper[4805]: I1203 15:06:49.553543 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/2d49b230-da3b-4b61-bd0f-35408593cb08-config-out\") pod \"prometheus-metric-storage-0\" (UID: \"2d49b230-da3b-4b61-bd0f-35408593cb08\") " pod="openstack/prometheus-metric-storage-0" Dec 03 15:06:49 crc kubenswrapper[4805]: I1203 15:06:49.553583 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/2d49b230-da3b-4b61-bd0f-35408593cb08-config\") pod \"prometheus-metric-storage-0\" (UID: \"2d49b230-da3b-4b61-bd0f-35408593cb08\") " pod="openstack/prometheus-metric-storage-0" Dec 03 15:06:49 crc kubenswrapper[4805]: I1203 15:06:49.553697 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ppf4f\" (UniqueName: \"kubernetes.io/projected/2d49b230-da3b-4b61-bd0f-35408593cb08-kube-api-access-ppf4f\") pod \"prometheus-metric-storage-0\" (UID: \"2d49b230-da3b-4b61-bd0f-35408593cb08\") " pod="openstack/prometheus-metric-storage-0" Dec 03 15:06:49 crc kubenswrapper[4805]: I1203 15:06:49.554049 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"thanos-prometheus-http-client-file\" (UniqueName: \"kubernetes.io/secret/2d49b230-da3b-4b61-bd0f-35408593cb08-thanos-prometheus-http-client-file\") pod \"prometheus-metric-storage-0\" (UID: \"2d49b230-da3b-4b61-bd0f-35408593cb08\") " pod="openstack/prometheus-metric-storage-0" Dec 03 15:06:49 crc kubenswrapper[4805]: I1203 15:06:49.656248 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"thanos-prometheus-http-client-file\" (UniqueName: \"kubernetes.io/secret/2d49b230-da3b-4b61-bd0f-35408593cb08-thanos-prometheus-http-client-file\") pod \"prometheus-metric-storage-0\" (UID: \"2d49b230-da3b-4b61-bd0f-35408593cb08\") " pod="openstack/prometheus-metric-storage-0" Dec 03 15:06:49 crc kubenswrapper[4805]: I1203 15:06:49.656615 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"prometheus-metric-storage-rulefiles-0\" (UniqueName: \"kubernetes.io/configmap/2d49b230-da3b-4b61-bd0f-35408593cb08-prometheus-metric-storage-rulefiles-0\") pod \"prometheus-metric-storage-0\" (UID: \"2d49b230-da3b-4b61-bd0f-35408593cb08\") " pod="openstack/prometheus-metric-storage-0" Dec 03 15:06:49 crc kubenswrapper[4805]: I1203 15:06:49.656656 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"prometheus-metric-storage-0\" (UID: \"2d49b230-da3b-4b61-bd0f-35408593cb08\") " pod="openstack/prometheus-metric-storage-0" Dec 03 15:06:49 crc kubenswrapper[4805]: I1203 15:06:49.656708 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/2d49b230-da3b-4b61-bd0f-35408593cb08-web-config\") pod \"prometheus-metric-storage-0\" (UID: \"2d49b230-da3b-4b61-bd0f-35408593cb08\") " pod="openstack/prometheus-metric-storage-0" Dec 03 15:06:49 crc kubenswrapper[4805]: I1203 15:06:49.656727 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/2d49b230-da3b-4b61-bd0f-35408593cb08-tls-assets\") pod \"prometheus-metric-storage-0\" (UID: \"2d49b230-da3b-4b61-bd0f-35408593cb08\") " pod="openstack/prometheus-metric-storage-0" Dec 03 15:06:49 crc kubenswrapper[4805]: I1203 15:06:49.656746 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/2d49b230-da3b-4b61-bd0f-35408593cb08-config-out\") pod \"prometheus-metric-storage-0\" (UID: \"2d49b230-da3b-4b61-bd0f-35408593cb08\") " pod="openstack/prometheus-metric-storage-0" Dec 03 15:06:49 crc kubenswrapper[4805]: I1203 15:06:49.657033 4805 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"prometheus-metric-storage-0\" (UID: \"2d49b230-da3b-4b61-bd0f-35408593cb08\") device mount path \"/mnt/openstack/pv12\"" pod="openstack/prometheus-metric-storage-0" Dec 03 15:06:49 crc kubenswrapper[4805]: I1203 15:06:49.657290 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/2d49b230-da3b-4b61-bd0f-35408593cb08-config\") pod \"prometheus-metric-storage-0\" (UID: \"2d49b230-da3b-4b61-bd0f-35408593cb08\") " pod="openstack/prometheus-metric-storage-0" Dec 03 15:06:49 crc kubenswrapper[4805]: I1203 15:06:49.657336 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ppf4f\" (UniqueName: \"kubernetes.io/projected/2d49b230-da3b-4b61-bd0f-35408593cb08-kube-api-access-ppf4f\") pod \"prometheus-metric-storage-0\" (UID: \"2d49b230-da3b-4b61-bd0f-35408593cb08\") " pod="openstack/prometheus-metric-storage-0" Dec 03 15:06:49 crc kubenswrapper[4805]: I1203 15:06:49.662214 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"prometheus-metric-storage-rulefiles-0\" (UniqueName: \"kubernetes.io/configmap/2d49b230-da3b-4b61-bd0f-35408593cb08-prometheus-metric-storage-rulefiles-0\") pod \"prometheus-metric-storage-0\" (UID: \"2d49b230-da3b-4b61-bd0f-35408593cb08\") " pod="openstack/prometheus-metric-storage-0" Dec 03 15:06:49 crc kubenswrapper[4805]: I1203 15:06:49.662648 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"thanos-prometheus-http-client-file\" (UniqueName: \"kubernetes.io/secret/2d49b230-da3b-4b61-bd0f-35408593cb08-thanos-prometheus-http-client-file\") pod \"prometheus-metric-storage-0\" (UID: \"2d49b230-da3b-4b61-bd0f-35408593cb08\") " pod="openstack/prometheus-metric-storage-0" Dec 03 15:06:49 crc kubenswrapper[4805]: I1203 15:06:49.662951 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/2d49b230-da3b-4b61-bd0f-35408593cb08-config-out\") pod \"prometheus-metric-storage-0\" (UID: \"2d49b230-da3b-4b61-bd0f-35408593cb08\") " pod="openstack/prometheus-metric-storage-0" Dec 03 15:06:49 crc kubenswrapper[4805]: I1203 15:06:49.668407 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/secret/2d49b230-da3b-4b61-bd0f-35408593cb08-config\") pod \"prometheus-metric-storage-0\" (UID: \"2d49b230-da3b-4b61-bd0f-35408593cb08\") " pod="openstack/prometheus-metric-storage-0" Dec 03 15:06:49 crc kubenswrapper[4805]: I1203 15:06:49.672273 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/2d49b230-da3b-4b61-bd0f-35408593cb08-web-config\") pod \"prometheus-metric-storage-0\" (UID: \"2d49b230-da3b-4b61-bd0f-35408593cb08\") " pod="openstack/prometheus-metric-storage-0" Dec 03 15:06:49 crc kubenswrapper[4805]: I1203 15:06:49.672752 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/2d49b230-da3b-4b61-bd0f-35408593cb08-tls-assets\") pod \"prometheus-metric-storage-0\" (UID: \"2d49b230-da3b-4b61-bd0f-35408593cb08\") " pod="openstack/prometheus-metric-storage-0" Dec 03 15:06:49 crc kubenswrapper[4805]: I1203 15:06:49.678276 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ppf4f\" (UniqueName: \"kubernetes.io/projected/2d49b230-da3b-4b61-bd0f-35408593cb08-kube-api-access-ppf4f\") pod \"prometheus-metric-storage-0\" (UID: \"2d49b230-da3b-4b61-bd0f-35408593cb08\") " pod="openstack/prometheus-metric-storage-0" Dec 03 15:06:49 crc kubenswrapper[4805]: I1203 15:06:49.708831 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/alertmanager-metric-storage-0"] Dec 03 15:06:49 crc kubenswrapper[4805]: I1203 15:06:49.719589 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"prometheus-metric-storage-0\" (UID: \"2d49b230-da3b-4b61-bd0f-35408593cb08\") " pod="openstack/prometheus-metric-storage-0" Dec 03 15:06:49 crc kubenswrapper[4805]: I1203 15:06:49.800405 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/prometheus-metric-storage-0" Dec 03 15:06:50 crc kubenswrapper[4805]: I1203 15:06:50.052025 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/alertmanager-metric-storage-0" event={"ID":"a0412bf9-ae66-4d99-9bb3-ea2fe4637678","Type":"ContainerStarted","Data":"e5de744bac812285e236f0a05333fb8647a54df88d471ef60344a0387e38f76f"} Dec 03 15:06:50 crc kubenswrapper[4805]: I1203 15:06:50.499363 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/prometheus-metric-storage-0"] Dec 03 15:06:51 crc kubenswrapper[4805]: I1203 15:06:51.060754 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"2d49b230-da3b-4b61-bd0f-35408593cb08","Type":"ContainerStarted","Data":"00cb4cbab348fb36d1a0a559de6a935b6e1b30e4966b7970b37c3ac1590fc5a8"} Dec 03 15:06:54 crc kubenswrapper[4805]: I1203 15:06:54.694735 4805 scope.go:117] "RemoveContainer" containerID="13e0da4d7fc8f4085be601d845d688938c16ea7941e52ca511e39a37d6e0ba4e" Dec 03 15:06:58 crc kubenswrapper[4805]: I1203 15:06:58.146799 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-gskh4" event={"ID":"6098937f-e3f6-45e8-a647-4994a79cd711","Type":"ContainerStarted","Data":"e54004d4fb8c135b56b16241c1140f37c36df51d3d40b05d6724caf5e9950f98"} Dec 03 15:07:04 crc kubenswrapper[4805]: I1203 15:07:04.207247 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"2d49b230-da3b-4b61-bd0f-35408593cb08","Type":"ContainerStarted","Data":"71ef45922e649b1ac11a86659ebc27a9c3917cc06c23e145cd0489cda5b944dc"} Dec 03 15:07:04 crc kubenswrapper[4805]: I1203 15:07:04.209664 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/alertmanager-metric-storage-0" event={"ID":"a0412bf9-ae66-4d99-9bb3-ea2fe4637678","Type":"ContainerStarted","Data":"42e542d43d9925cdf44a8f224e627a3788dfbab4502d007ea5cfaf0150e91647"} Dec 03 15:07:10 crc kubenswrapper[4805]: I1203 15:07:10.268476 4805 generic.go:334] "Generic (PLEG): container finished" podID="2d49b230-da3b-4b61-bd0f-35408593cb08" containerID="71ef45922e649b1ac11a86659ebc27a9c3917cc06c23e145cd0489cda5b944dc" exitCode=0 Dec 03 15:07:10 crc kubenswrapper[4805]: I1203 15:07:10.268591 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"2d49b230-da3b-4b61-bd0f-35408593cb08","Type":"ContainerDied","Data":"71ef45922e649b1ac11a86659ebc27a9c3917cc06c23e145cd0489cda5b944dc"} Dec 03 15:07:10 crc kubenswrapper[4805]: I1203 15:07:10.271363 4805 generic.go:334] "Generic (PLEG): container finished" podID="a0412bf9-ae66-4d99-9bb3-ea2fe4637678" containerID="42e542d43d9925cdf44a8f224e627a3788dfbab4502d007ea5cfaf0150e91647" exitCode=0 Dec 03 15:07:10 crc kubenswrapper[4805]: I1203 15:07:10.272533 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/alertmanager-metric-storage-0" event={"ID":"a0412bf9-ae66-4d99-9bb3-ea2fe4637678","Type":"ContainerDied","Data":"42e542d43d9925cdf44a8f224e627a3788dfbab4502d007ea5cfaf0150e91647"} Dec 03 15:07:14 crc kubenswrapper[4805]: I1203 15:07:14.322693 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/alertmanager-metric-storage-0" event={"ID":"a0412bf9-ae66-4d99-9bb3-ea2fe4637678","Type":"ContainerStarted","Data":"0d776608382932e939a07b5928e916eeb9285a367f8451044e0df9c07b5c507d"} Dec 03 15:07:20 crc kubenswrapper[4805]: I1203 15:07:20.422602 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/alertmanager-metric-storage-0" event={"ID":"a0412bf9-ae66-4d99-9bb3-ea2fe4637678","Type":"ContainerStarted","Data":"669939c5581a91786e00dc04682e8da67bdd26c52d3d470031ea4bf3068006a6"} Dec 03 15:07:20 crc kubenswrapper[4805]: I1203 15:07:20.423264 4805 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/alertmanager-metric-storage-0" Dec 03 15:07:20 crc kubenswrapper[4805]: I1203 15:07:20.425403 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"2d49b230-da3b-4b61-bd0f-35408593cb08","Type":"ContainerStarted","Data":"35836fe083be2a43763fe982c9fb1d208fd22a505afe689684ab4ce61acf41c5"} Dec 03 15:07:20 crc kubenswrapper[4805]: I1203 15:07:20.428728 4805 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/alertmanager-metric-storage-0" Dec 03 15:07:20 crc kubenswrapper[4805]: I1203 15:07:20.469697 4805 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/alertmanager-metric-storage-0" podStartSLOduration=9.110284151 podStartE2EDuration="32.469668999s" podCreationTimestamp="2025-12-03 15:06:48 +0000 UTC" firstStartedPulling="2025-12-03 15:06:49.713585064 +0000 UTC m=+3439.376501987" lastFinishedPulling="2025-12-03 15:07:13.072969922 +0000 UTC m=+3462.735886835" observedRunningTime="2025-12-03 15:07:20.455684889 +0000 UTC m=+3470.118601822" watchObservedRunningTime="2025-12-03 15:07:20.469668999 +0000 UTC m=+3470.132585942" Dec 03 15:07:25 crc kubenswrapper[4805]: I1203 15:07:25.492321 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"2d49b230-da3b-4b61-bd0f-35408593cb08","Type":"ContainerStarted","Data":"467fdf6d1cac465bdc39b9b7b49337586881b47a3b21d81736981bc1da42ae23"} Dec 03 15:07:29 crc kubenswrapper[4805]: I1203 15:07:29.534801 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"2d49b230-da3b-4b61-bd0f-35408593cb08","Type":"ContainerStarted","Data":"cf8d8dea50650b296384574cb5114c433e23304443326b7df4f5809103e103cb"} Dec 03 15:07:29 crc kubenswrapper[4805]: I1203 15:07:29.567221 4805 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/prometheus-metric-storage-0" podStartSLOduration=3.232680165 podStartE2EDuration="41.567196273s" podCreationTimestamp="2025-12-03 15:06:48 +0000 UTC" firstStartedPulling="2025-12-03 15:06:50.501502786 +0000 UTC m=+3440.164419709" lastFinishedPulling="2025-12-03 15:07:28.836018894 +0000 UTC m=+3478.498935817" observedRunningTime="2025-12-03 15:07:29.566051793 +0000 UTC m=+3479.228968736" watchObservedRunningTime="2025-12-03 15:07:29.567196273 +0000 UTC m=+3479.230113236" Dec 03 15:07:29 crc kubenswrapper[4805]: I1203 15:07:29.801258 4805 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/prometheus-metric-storage-0" Dec 03 15:07:34 crc kubenswrapper[4805]: I1203 15:07:34.801798 4805 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/prometheus-metric-storage-0" Dec 03 15:07:34 crc kubenswrapper[4805]: I1203 15:07:34.804644 4805 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/prometheus-metric-storage-0" Dec 03 15:07:35 crc kubenswrapper[4805]: I1203 15:07:35.622207 4805 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/prometheus-metric-storage-0" Dec 03 15:07:36 crc kubenswrapper[4805]: I1203 15:07:36.788437 4805 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/openstackclient"] Dec 03 15:07:36 crc kubenswrapper[4805]: I1203 15:07:36.806206 4805 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/openstackclient"] Dec 03 15:07:36 crc kubenswrapper[4805]: I1203 15:07:36.806557 4805 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/openstackclient" podUID="8a387e8e-801f-4724-85a6-fbd066455197" containerName="openstackclient" containerID="cri-o://d2436082da13f8316ca1b0be86c6c49eb13c4416ceecc8b65b3e8a3615ba5f90" gracePeriod=2 Dec 03 15:07:36 crc kubenswrapper[4805]: I1203 15:07:36.815629 4805 status_manager.go:861] "Pod was deleted and then recreated, skipping status update" pod="openstack/openstackclient" oldPodUID="8a387e8e-801f-4724-85a6-fbd066455197" podUID="49d35952-8526-46a9-80f5-b0fe3a359abd" Dec 03 15:07:36 crc kubenswrapper[4805]: I1203 15:07:36.832288 4805 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/openstackclient"] Dec 03 15:07:36 crc kubenswrapper[4805]: E1203 15:07:36.832778 4805 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8a387e8e-801f-4724-85a6-fbd066455197" containerName="openstackclient" Dec 03 15:07:36 crc kubenswrapper[4805]: I1203 15:07:36.832795 4805 state_mem.go:107] "Deleted CPUSet assignment" podUID="8a387e8e-801f-4724-85a6-fbd066455197" containerName="openstackclient" Dec 03 15:07:36 crc kubenswrapper[4805]: I1203 15:07:36.833044 4805 memory_manager.go:354] "RemoveStaleState removing state" podUID="8a387e8e-801f-4724-85a6-fbd066455197" containerName="openstackclient" Dec 03 15:07:36 crc kubenswrapper[4805]: I1203 15:07:36.833695 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Dec 03 15:07:36 crc kubenswrapper[4805]: I1203 15:07:36.857029 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstackclient"] Dec 03 15:07:36 crc kubenswrapper[4805]: I1203 15:07:36.906373 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/49d35952-8526-46a9-80f5-b0fe3a359abd-combined-ca-bundle\") pod \"openstackclient\" (UID: \"49d35952-8526-46a9-80f5-b0fe3a359abd\") " pod="openstack/openstackclient" Dec 03 15:07:36 crc kubenswrapper[4805]: I1203 15:07:36.906473 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/49d35952-8526-46a9-80f5-b0fe3a359abd-openstack-config\") pod \"openstackclient\" (UID: \"49d35952-8526-46a9-80f5-b0fe3a359abd\") " pod="openstack/openstackclient" Dec 03 15:07:36 crc kubenswrapper[4805]: I1203 15:07:36.906494 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cz7z6\" (UniqueName: \"kubernetes.io/projected/49d35952-8526-46a9-80f5-b0fe3a359abd-kube-api-access-cz7z6\") pod \"openstackclient\" (UID: \"49d35952-8526-46a9-80f5-b0fe3a359abd\") " pod="openstack/openstackclient" Dec 03 15:07:36 crc kubenswrapper[4805]: I1203 15:07:36.906536 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/49d35952-8526-46a9-80f5-b0fe3a359abd-openstack-config-secret\") pod \"openstackclient\" (UID: \"49d35952-8526-46a9-80f5-b0fe3a359abd\") " pod="openstack/openstackclient" Dec 03 15:07:37 crc kubenswrapper[4805]: I1203 15:07:37.007561 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/49d35952-8526-46a9-80f5-b0fe3a359abd-openstack-config-secret\") pod \"openstackclient\" (UID: \"49d35952-8526-46a9-80f5-b0fe3a359abd\") " pod="openstack/openstackclient" Dec 03 15:07:37 crc kubenswrapper[4805]: I1203 15:07:37.007699 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/49d35952-8526-46a9-80f5-b0fe3a359abd-combined-ca-bundle\") pod \"openstackclient\" (UID: \"49d35952-8526-46a9-80f5-b0fe3a359abd\") " pod="openstack/openstackclient" Dec 03 15:07:37 crc kubenswrapper[4805]: I1203 15:07:37.007818 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/49d35952-8526-46a9-80f5-b0fe3a359abd-openstack-config\") pod \"openstackclient\" (UID: \"49d35952-8526-46a9-80f5-b0fe3a359abd\") " pod="openstack/openstackclient" Dec 03 15:07:37 crc kubenswrapper[4805]: I1203 15:07:37.007842 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cz7z6\" (UniqueName: \"kubernetes.io/projected/49d35952-8526-46a9-80f5-b0fe3a359abd-kube-api-access-cz7z6\") pod \"openstackclient\" (UID: \"49d35952-8526-46a9-80f5-b0fe3a359abd\") " pod="openstack/openstackclient" Dec 03 15:07:37 crc kubenswrapper[4805]: I1203 15:07:37.008747 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/49d35952-8526-46a9-80f5-b0fe3a359abd-openstack-config\") pod \"openstackclient\" (UID: \"49d35952-8526-46a9-80f5-b0fe3a359abd\") " pod="openstack/openstackclient" Dec 03 15:07:37 crc kubenswrapper[4805]: I1203 15:07:37.014496 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/49d35952-8526-46a9-80f5-b0fe3a359abd-openstack-config-secret\") pod \"openstackclient\" (UID: \"49d35952-8526-46a9-80f5-b0fe3a359abd\") " pod="openstack/openstackclient" Dec 03 15:07:37 crc kubenswrapper[4805]: I1203 15:07:37.023642 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/49d35952-8526-46a9-80f5-b0fe3a359abd-combined-ca-bundle\") pod \"openstackclient\" (UID: \"49d35952-8526-46a9-80f5-b0fe3a359abd\") " pod="openstack/openstackclient" Dec 03 15:07:37 crc kubenswrapper[4805]: I1203 15:07:37.028558 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cz7z6\" (UniqueName: \"kubernetes.io/projected/49d35952-8526-46a9-80f5-b0fe3a359abd-kube-api-access-cz7z6\") pod \"openstackclient\" (UID: \"49d35952-8526-46a9-80f5-b0fe3a359abd\") " pod="openstack/openstackclient" Dec 03 15:07:37 crc kubenswrapper[4805]: I1203 15:07:37.153634 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Dec 03 15:07:37 crc kubenswrapper[4805]: I1203 15:07:37.162145 4805 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/aodh-0"] Dec 03 15:07:37 crc kubenswrapper[4805]: I1203 15:07:37.162509 4805 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/aodh-0" podUID="ef0f7bdc-b48e-41a8-ae3d-9eac4016d70a" containerName="aodh-api" containerID="cri-o://ef96a03285bf5bb5faad845eb35c37c353fce53de018d916db07ac65f0cbc9e2" gracePeriod=30 Dec 03 15:07:37 crc kubenswrapper[4805]: I1203 15:07:37.162577 4805 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/aodh-0" podUID="ef0f7bdc-b48e-41a8-ae3d-9eac4016d70a" containerName="aodh-listener" containerID="cri-o://fb6610f613e5e1b6790967f098610eadabd6b001d42a97c2699e49aa08ac0397" gracePeriod=30 Dec 03 15:07:37 crc kubenswrapper[4805]: I1203 15:07:37.162565 4805 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/aodh-0" podUID="ef0f7bdc-b48e-41a8-ae3d-9eac4016d70a" containerName="aodh-notifier" containerID="cri-o://bf67c1c3e37fd095badea3841d170fb3b8345ca6a40d2106e16d2ccf4ed783e0" gracePeriod=30 Dec 03 15:07:37 crc kubenswrapper[4805]: I1203 15:07:37.162618 4805 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/aodh-0" podUID="ef0f7bdc-b48e-41a8-ae3d-9eac4016d70a" containerName="aodh-evaluator" containerID="cri-o://dfdd2ac28b14d459abf2c4b6a74db213eb06cb0368a0422d75becb5b25e09cb3" gracePeriod=30 Dec 03 15:07:37 crc kubenswrapper[4805]: I1203 15:07:37.604997 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstackclient"] Dec 03 15:07:37 crc kubenswrapper[4805]: I1203 15:07:37.664013 4805 generic.go:334] "Generic (PLEG): container finished" podID="ef0f7bdc-b48e-41a8-ae3d-9eac4016d70a" containerID="ef96a03285bf5bb5faad845eb35c37c353fce53de018d916db07ac65f0cbc9e2" exitCode=0 Dec 03 15:07:37 crc kubenswrapper[4805]: I1203 15:07:37.664068 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-0" event={"ID":"ef0f7bdc-b48e-41a8-ae3d-9eac4016d70a","Type":"ContainerDied","Data":"ef96a03285bf5bb5faad845eb35c37c353fce53de018d916db07ac65f0cbc9e2"} Dec 03 15:07:38 crc kubenswrapper[4805]: I1203 15:07:38.322640 4805 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/prometheus-metric-storage-0"] Dec 03 15:07:38 crc kubenswrapper[4805]: I1203 15:07:38.323279 4805 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/prometheus-metric-storage-0" podUID="2d49b230-da3b-4b61-bd0f-35408593cb08" containerName="prometheus" containerID="cri-o://35836fe083be2a43763fe982c9fb1d208fd22a505afe689684ab4ce61acf41c5" gracePeriod=600 Dec 03 15:07:38 crc kubenswrapper[4805]: I1203 15:07:38.323351 4805 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/prometheus-metric-storage-0" podUID="2d49b230-da3b-4b61-bd0f-35408593cb08" containerName="thanos-sidecar" containerID="cri-o://cf8d8dea50650b296384574cb5114c433e23304443326b7df4f5809103e103cb" gracePeriod=600 Dec 03 15:07:38 crc kubenswrapper[4805]: I1203 15:07:38.323413 4805 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/prometheus-metric-storage-0" podUID="2d49b230-da3b-4b61-bd0f-35408593cb08" containerName="config-reloader" containerID="cri-o://467fdf6d1cac465bdc39b9b7b49337586881b47a3b21d81736981bc1da42ae23" gracePeriod=600 Dec 03 15:07:38 crc kubenswrapper[4805]: I1203 15:07:38.673761 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstackclient" event={"ID":"49d35952-8526-46a9-80f5-b0fe3a359abd","Type":"ContainerStarted","Data":"239ce7b0ae1242eb57c978d1ff1b617dd6a15b1afd446ff3a1697ac015934863"} Dec 03 15:07:38 crc kubenswrapper[4805]: I1203 15:07:38.674336 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstackclient" event={"ID":"49d35952-8526-46a9-80f5-b0fe3a359abd","Type":"ContainerStarted","Data":"c56d8ccc1eee0a43f55f1ab112fdc19d120c0f1f1ebf215f7be8cd86b336b209"} Dec 03 15:07:38 crc kubenswrapper[4805]: I1203 15:07:38.676898 4805 generic.go:334] "Generic (PLEG): container finished" podID="ef0f7bdc-b48e-41a8-ae3d-9eac4016d70a" containerID="dfdd2ac28b14d459abf2c4b6a74db213eb06cb0368a0422d75becb5b25e09cb3" exitCode=0 Dec 03 15:07:38 crc kubenswrapper[4805]: I1203 15:07:38.676953 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-0" event={"ID":"ef0f7bdc-b48e-41a8-ae3d-9eac4016d70a","Type":"ContainerDied","Data":"dfdd2ac28b14d459abf2c4b6a74db213eb06cb0368a0422d75becb5b25e09cb3"} Dec 03 15:07:38 crc kubenswrapper[4805]: I1203 15:07:38.679614 4805 generic.go:334] "Generic (PLEG): container finished" podID="2d49b230-da3b-4b61-bd0f-35408593cb08" containerID="cf8d8dea50650b296384574cb5114c433e23304443326b7df4f5809103e103cb" exitCode=0 Dec 03 15:07:38 crc kubenswrapper[4805]: I1203 15:07:38.679635 4805 generic.go:334] "Generic (PLEG): container finished" podID="2d49b230-da3b-4b61-bd0f-35408593cb08" containerID="35836fe083be2a43763fe982c9fb1d208fd22a505afe689684ab4ce61acf41c5" exitCode=0 Dec 03 15:07:38 crc kubenswrapper[4805]: I1203 15:07:38.679653 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"2d49b230-da3b-4b61-bd0f-35408593cb08","Type":"ContainerDied","Data":"cf8d8dea50650b296384574cb5114c433e23304443326b7df4f5809103e103cb"} Dec 03 15:07:38 crc kubenswrapper[4805]: I1203 15:07:38.679669 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"2d49b230-da3b-4b61-bd0f-35408593cb08","Type":"ContainerDied","Data":"35836fe083be2a43763fe982c9fb1d208fd22a505afe689684ab4ce61acf41c5"} Dec 03 15:07:38 crc kubenswrapper[4805]: I1203 15:07:38.707257 4805 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/openstackclient" podStartSLOduration=2.707234681 podStartE2EDuration="2.707234681s" podCreationTimestamp="2025-12-03 15:07:36 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 15:07:38.702920457 +0000 UTC m=+3488.365837380" watchObservedRunningTime="2025-12-03 15:07:38.707234681 +0000 UTC m=+3488.370151594" Dec 03 15:07:39 crc kubenswrapper[4805]: I1203 15:07:39.094961 4805 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Dec 03 15:07:39 crc kubenswrapper[4805]: I1203 15:07:39.252329 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/8a387e8e-801f-4724-85a6-fbd066455197-openstack-config-secret\") pod \"8a387e8e-801f-4724-85a6-fbd066455197\" (UID: \"8a387e8e-801f-4724-85a6-fbd066455197\") " Dec 03 15:07:39 crc kubenswrapper[4805]: I1203 15:07:39.252375 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/8a387e8e-801f-4724-85a6-fbd066455197-openstack-config\") pod \"8a387e8e-801f-4724-85a6-fbd066455197\" (UID: \"8a387e8e-801f-4724-85a6-fbd066455197\") " Dec 03 15:07:39 crc kubenswrapper[4805]: I1203 15:07:39.252494 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8a387e8e-801f-4724-85a6-fbd066455197-combined-ca-bundle\") pod \"8a387e8e-801f-4724-85a6-fbd066455197\" (UID: \"8a387e8e-801f-4724-85a6-fbd066455197\") " Dec 03 15:07:39 crc kubenswrapper[4805]: I1203 15:07:39.252666 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ds9p7\" (UniqueName: \"kubernetes.io/projected/8a387e8e-801f-4724-85a6-fbd066455197-kube-api-access-ds9p7\") pod \"8a387e8e-801f-4724-85a6-fbd066455197\" (UID: \"8a387e8e-801f-4724-85a6-fbd066455197\") " Dec 03 15:07:39 crc kubenswrapper[4805]: I1203 15:07:39.258039 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8a387e8e-801f-4724-85a6-fbd066455197-kube-api-access-ds9p7" (OuterVolumeSpecName: "kube-api-access-ds9p7") pod "8a387e8e-801f-4724-85a6-fbd066455197" (UID: "8a387e8e-801f-4724-85a6-fbd066455197"). InnerVolumeSpecName "kube-api-access-ds9p7". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 15:07:39 crc kubenswrapper[4805]: I1203 15:07:39.279991 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8a387e8e-801f-4724-85a6-fbd066455197-openstack-config" (OuterVolumeSpecName: "openstack-config") pod "8a387e8e-801f-4724-85a6-fbd066455197" (UID: "8a387e8e-801f-4724-85a6-fbd066455197"). InnerVolumeSpecName "openstack-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 15:07:39 crc kubenswrapper[4805]: I1203 15:07:39.283383 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8a387e8e-801f-4724-85a6-fbd066455197-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "8a387e8e-801f-4724-85a6-fbd066455197" (UID: "8a387e8e-801f-4724-85a6-fbd066455197"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 15:07:39 crc kubenswrapper[4805]: I1203 15:07:39.302954 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8a387e8e-801f-4724-85a6-fbd066455197-openstack-config-secret" (OuterVolumeSpecName: "openstack-config-secret") pod "8a387e8e-801f-4724-85a6-fbd066455197" (UID: "8a387e8e-801f-4724-85a6-fbd066455197"). InnerVolumeSpecName "openstack-config-secret". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 15:07:39 crc kubenswrapper[4805]: I1203 15:07:39.354422 4805 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8a387e8e-801f-4724-85a6-fbd066455197-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 15:07:39 crc kubenswrapper[4805]: I1203 15:07:39.354455 4805 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ds9p7\" (UniqueName: \"kubernetes.io/projected/8a387e8e-801f-4724-85a6-fbd066455197-kube-api-access-ds9p7\") on node \"crc\" DevicePath \"\"" Dec 03 15:07:39 crc kubenswrapper[4805]: I1203 15:07:39.354466 4805 reconciler_common.go:293] "Volume detached for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/8a387e8e-801f-4724-85a6-fbd066455197-openstack-config-secret\") on node \"crc\" DevicePath \"\"" Dec 03 15:07:39 crc kubenswrapper[4805]: I1203 15:07:39.354476 4805 reconciler_common.go:293] "Volume detached for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/8a387e8e-801f-4724-85a6-fbd066455197-openstack-config\") on node \"crc\" DevicePath \"\"" Dec 03 15:07:39 crc kubenswrapper[4805]: I1203 15:07:39.715966 4805 generic.go:334] "Generic (PLEG): container finished" podID="8a387e8e-801f-4724-85a6-fbd066455197" containerID="d2436082da13f8316ca1b0be86c6c49eb13c4416ceecc8b65b3e8a3615ba5f90" exitCode=137 Dec 03 15:07:39 crc kubenswrapper[4805]: I1203 15:07:39.716052 4805 scope.go:117] "RemoveContainer" containerID="d2436082da13f8316ca1b0be86c6c49eb13c4416ceecc8b65b3e8a3615ba5f90" Dec 03 15:07:39 crc kubenswrapper[4805]: I1203 15:07:39.716214 4805 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Dec 03 15:07:39 crc kubenswrapper[4805]: I1203 15:07:39.722073 4805 generic.go:334] "Generic (PLEG): container finished" podID="ef0f7bdc-b48e-41a8-ae3d-9eac4016d70a" containerID="bf67c1c3e37fd095badea3841d170fb3b8345ca6a40d2106e16d2ccf4ed783e0" exitCode=0 Dec 03 15:07:39 crc kubenswrapper[4805]: I1203 15:07:39.722150 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-0" event={"ID":"ef0f7bdc-b48e-41a8-ae3d-9eac4016d70a","Type":"ContainerDied","Data":"bf67c1c3e37fd095badea3841d170fb3b8345ca6a40d2106e16d2ccf4ed783e0"} Dec 03 15:07:39 crc kubenswrapper[4805]: I1203 15:07:39.728140 4805 generic.go:334] "Generic (PLEG): container finished" podID="2d49b230-da3b-4b61-bd0f-35408593cb08" containerID="467fdf6d1cac465bdc39b9b7b49337586881b47a3b21d81736981bc1da42ae23" exitCode=0 Dec 03 15:07:39 crc kubenswrapper[4805]: I1203 15:07:39.729031 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"2d49b230-da3b-4b61-bd0f-35408593cb08","Type":"ContainerDied","Data":"467fdf6d1cac465bdc39b9b7b49337586881b47a3b21d81736981bc1da42ae23"} Dec 03 15:07:39 crc kubenswrapper[4805]: I1203 15:07:39.736242 4805 status_manager.go:861] "Pod was deleted and then recreated, skipping status update" pod="openstack/openstackclient" oldPodUID="8a387e8e-801f-4724-85a6-fbd066455197" podUID="49d35952-8526-46a9-80f5-b0fe3a359abd" Dec 03 15:07:39 crc kubenswrapper[4805]: I1203 15:07:39.744684 4805 scope.go:117] "RemoveContainer" containerID="d2436082da13f8316ca1b0be86c6c49eb13c4416ceecc8b65b3e8a3615ba5f90" Dec 03 15:07:39 crc kubenswrapper[4805]: E1203 15:07:39.745148 4805 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d2436082da13f8316ca1b0be86c6c49eb13c4416ceecc8b65b3e8a3615ba5f90\": container with ID starting with d2436082da13f8316ca1b0be86c6c49eb13c4416ceecc8b65b3e8a3615ba5f90 not found: ID does not exist" containerID="d2436082da13f8316ca1b0be86c6c49eb13c4416ceecc8b65b3e8a3615ba5f90" Dec 03 15:07:39 crc kubenswrapper[4805]: I1203 15:07:39.745196 4805 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d2436082da13f8316ca1b0be86c6c49eb13c4416ceecc8b65b3e8a3615ba5f90"} err="failed to get container status \"d2436082da13f8316ca1b0be86c6c49eb13c4416ceecc8b65b3e8a3615ba5f90\": rpc error: code = NotFound desc = could not find container \"d2436082da13f8316ca1b0be86c6c49eb13c4416ceecc8b65b3e8a3615ba5f90\": container with ID starting with d2436082da13f8316ca1b0be86c6c49eb13c4416ceecc8b65b3e8a3615ba5f90 not found: ID does not exist" Dec 03 15:07:39 crc kubenswrapper[4805]: I1203 15:07:39.830674 4805 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/prometheus-metric-storage-0" Dec 03 15:07:39 crc kubenswrapper[4805]: I1203 15:07:39.980303 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/2d49b230-da3b-4b61-bd0f-35408593cb08-tls-assets\") pod \"2d49b230-da3b-4b61-bd0f-35408593cb08\" (UID: \"2d49b230-da3b-4b61-bd0f-35408593cb08\") " Dec 03 15:07:39 crc kubenswrapper[4805]: I1203 15:07:39.980810 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"prometheus-metric-storage-rulefiles-0\" (UniqueName: \"kubernetes.io/configmap/2d49b230-da3b-4b61-bd0f-35408593cb08-prometheus-metric-storage-rulefiles-0\") pod \"2d49b230-da3b-4b61-bd0f-35408593cb08\" (UID: \"2d49b230-da3b-4b61-bd0f-35408593cb08\") " Dec 03 15:07:39 crc kubenswrapper[4805]: I1203 15:07:39.980836 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/2d49b230-da3b-4b61-bd0f-35408593cb08-config-out\") pod \"2d49b230-da3b-4b61-bd0f-35408593cb08\" (UID: \"2d49b230-da3b-4b61-bd0f-35408593cb08\") " Dec 03 15:07:39 crc kubenswrapper[4805]: I1203 15:07:39.980862 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"prometheus-metric-storage-db\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"2d49b230-da3b-4b61-bd0f-35408593cb08\" (UID: \"2d49b230-da3b-4b61-bd0f-35408593cb08\") " Dec 03 15:07:39 crc kubenswrapper[4805]: I1203 15:07:39.980946 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ppf4f\" (UniqueName: \"kubernetes.io/projected/2d49b230-da3b-4b61-bd0f-35408593cb08-kube-api-access-ppf4f\") pod \"2d49b230-da3b-4b61-bd0f-35408593cb08\" (UID: \"2d49b230-da3b-4b61-bd0f-35408593cb08\") " Dec 03 15:07:39 crc kubenswrapper[4805]: I1203 15:07:39.981039 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/2d49b230-da3b-4b61-bd0f-35408593cb08-web-config\") pod \"2d49b230-da3b-4b61-bd0f-35408593cb08\" (UID: \"2d49b230-da3b-4b61-bd0f-35408593cb08\") " Dec 03 15:07:39 crc kubenswrapper[4805]: I1203 15:07:39.981294 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/2d49b230-da3b-4b61-bd0f-35408593cb08-config\") pod \"2d49b230-da3b-4b61-bd0f-35408593cb08\" (UID: \"2d49b230-da3b-4b61-bd0f-35408593cb08\") " Dec 03 15:07:39 crc kubenswrapper[4805]: I1203 15:07:39.981364 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"thanos-prometheus-http-client-file\" (UniqueName: \"kubernetes.io/secret/2d49b230-da3b-4b61-bd0f-35408593cb08-thanos-prometheus-http-client-file\") pod \"2d49b230-da3b-4b61-bd0f-35408593cb08\" (UID: \"2d49b230-da3b-4b61-bd0f-35408593cb08\") " Dec 03 15:07:39 crc kubenswrapper[4805]: I1203 15:07:39.983316 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2d49b230-da3b-4b61-bd0f-35408593cb08-prometheus-metric-storage-rulefiles-0" (OuterVolumeSpecName: "prometheus-metric-storage-rulefiles-0") pod "2d49b230-da3b-4b61-bd0f-35408593cb08" (UID: "2d49b230-da3b-4b61-bd0f-35408593cb08"). InnerVolumeSpecName "prometheus-metric-storage-rulefiles-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 15:07:39 crc kubenswrapper[4805]: I1203 15:07:39.988338 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2d49b230-da3b-4b61-bd0f-35408593cb08-config" (OuterVolumeSpecName: "config") pod "2d49b230-da3b-4b61-bd0f-35408593cb08" (UID: "2d49b230-da3b-4b61-bd0f-35408593cb08"). InnerVolumeSpecName "config". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 15:07:39 crc kubenswrapper[4805]: I1203 15:07:39.988825 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2d49b230-da3b-4b61-bd0f-35408593cb08-kube-api-access-ppf4f" (OuterVolumeSpecName: "kube-api-access-ppf4f") pod "2d49b230-da3b-4b61-bd0f-35408593cb08" (UID: "2d49b230-da3b-4b61-bd0f-35408593cb08"). InnerVolumeSpecName "kube-api-access-ppf4f". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 15:07:39 crc kubenswrapper[4805]: I1203 15:07:39.989652 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2d49b230-da3b-4b61-bd0f-35408593cb08-tls-assets" (OuterVolumeSpecName: "tls-assets") pod "2d49b230-da3b-4b61-bd0f-35408593cb08" (UID: "2d49b230-da3b-4b61-bd0f-35408593cb08"). InnerVolumeSpecName "tls-assets". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 15:07:39 crc kubenswrapper[4805]: I1203 15:07:39.989670 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2d49b230-da3b-4b61-bd0f-35408593cb08-config-out" (OuterVolumeSpecName: "config-out") pod "2d49b230-da3b-4b61-bd0f-35408593cb08" (UID: "2d49b230-da3b-4b61-bd0f-35408593cb08"). InnerVolumeSpecName "config-out". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 15:07:39 crc kubenswrapper[4805]: I1203 15:07:39.991702 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2d49b230-da3b-4b61-bd0f-35408593cb08-thanos-prometheus-http-client-file" (OuterVolumeSpecName: "thanos-prometheus-http-client-file") pod "2d49b230-da3b-4b61-bd0f-35408593cb08" (UID: "2d49b230-da3b-4b61-bd0f-35408593cb08"). InnerVolumeSpecName "thanos-prometheus-http-client-file". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 15:07:39 crc kubenswrapper[4805]: I1203 15:07:39.999600 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage12-crc" (OuterVolumeSpecName: "prometheus-metric-storage-db") pod "2d49b230-da3b-4b61-bd0f-35408593cb08" (UID: "2d49b230-da3b-4b61-bd0f-35408593cb08"). InnerVolumeSpecName "local-storage12-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Dec 03 15:07:40 crc kubenswrapper[4805]: I1203 15:07:40.019428 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2d49b230-da3b-4b61-bd0f-35408593cb08-web-config" (OuterVolumeSpecName: "web-config") pod "2d49b230-da3b-4b61-bd0f-35408593cb08" (UID: "2d49b230-da3b-4b61-bd0f-35408593cb08"). InnerVolumeSpecName "web-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 15:07:40 crc kubenswrapper[4805]: I1203 15:07:40.084075 4805 reconciler_common.go:293] "Volume detached for volume \"thanos-prometheus-http-client-file\" (UniqueName: \"kubernetes.io/secret/2d49b230-da3b-4b61-bd0f-35408593cb08-thanos-prometheus-http-client-file\") on node \"crc\" DevicePath \"\"" Dec 03 15:07:40 crc kubenswrapper[4805]: I1203 15:07:40.084122 4805 reconciler_common.go:293] "Volume detached for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/2d49b230-da3b-4b61-bd0f-35408593cb08-tls-assets\") on node \"crc\" DevicePath \"\"" Dec 03 15:07:40 crc kubenswrapper[4805]: I1203 15:07:40.084137 4805 reconciler_common.go:293] "Volume detached for volume \"prometheus-metric-storage-rulefiles-0\" (UniqueName: \"kubernetes.io/configmap/2d49b230-da3b-4b61-bd0f-35408593cb08-prometheus-metric-storage-rulefiles-0\") on node \"crc\" DevicePath \"\"" Dec 03 15:07:40 crc kubenswrapper[4805]: I1203 15:07:40.084155 4805 reconciler_common.go:293] "Volume detached for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/2d49b230-da3b-4b61-bd0f-35408593cb08-config-out\") on node \"crc\" DevicePath \"\"" Dec 03 15:07:40 crc kubenswrapper[4805]: I1203 15:07:40.084197 4805 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") on node \"crc\" " Dec 03 15:07:40 crc kubenswrapper[4805]: I1203 15:07:40.084212 4805 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ppf4f\" (UniqueName: \"kubernetes.io/projected/2d49b230-da3b-4b61-bd0f-35408593cb08-kube-api-access-ppf4f\") on node \"crc\" DevicePath \"\"" Dec 03 15:07:40 crc kubenswrapper[4805]: I1203 15:07:40.084227 4805 reconciler_common.go:293] "Volume detached for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/2d49b230-da3b-4b61-bd0f-35408593cb08-web-config\") on node \"crc\" DevicePath \"\"" Dec 03 15:07:40 crc kubenswrapper[4805]: I1203 15:07:40.084239 4805 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/secret/2d49b230-da3b-4b61-bd0f-35408593cb08-config\") on node \"crc\" DevicePath \"\"" Dec 03 15:07:40 crc kubenswrapper[4805]: I1203 15:07:40.114212 4805 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage12-crc" (UniqueName: "kubernetes.io/local-volume/local-storage12-crc") on node "crc" Dec 03 15:07:40 crc kubenswrapper[4805]: I1203 15:07:40.185887 4805 reconciler_common.go:293] "Volume detached for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") on node \"crc\" DevicePath \"\"" Dec 03 15:07:40 crc kubenswrapper[4805]: I1203 15:07:40.711077 4805 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8a387e8e-801f-4724-85a6-fbd066455197" path="/var/lib/kubelet/pods/8a387e8e-801f-4724-85a6-fbd066455197/volumes" Dec 03 15:07:40 crc kubenswrapper[4805]: I1203 15:07:40.741155 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"2d49b230-da3b-4b61-bd0f-35408593cb08","Type":"ContainerDied","Data":"00cb4cbab348fb36d1a0a559de6a935b6e1b30e4966b7970b37c3ac1590fc5a8"} Dec 03 15:07:40 crc kubenswrapper[4805]: I1203 15:07:40.741217 4805 scope.go:117] "RemoveContainer" containerID="cf8d8dea50650b296384574cb5114c433e23304443326b7df4f5809103e103cb" Dec 03 15:07:40 crc kubenswrapper[4805]: I1203 15:07:40.741233 4805 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/prometheus-metric-storage-0" Dec 03 15:07:40 crc kubenswrapper[4805]: I1203 15:07:40.762140 4805 scope.go:117] "RemoveContainer" containerID="467fdf6d1cac465bdc39b9b7b49337586881b47a3b21d81736981bc1da42ae23" Dec 03 15:07:40 crc kubenswrapper[4805]: I1203 15:07:40.772160 4805 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/prometheus-metric-storage-0"] Dec 03 15:07:40 crc kubenswrapper[4805]: I1203 15:07:40.783284 4805 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/prometheus-metric-storage-0"] Dec 03 15:07:40 crc kubenswrapper[4805]: I1203 15:07:40.804335 4805 scope.go:117] "RemoveContainer" containerID="35836fe083be2a43763fe982c9fb1d208fd22a505afe689684ab4ce61acf41c5" Dec 03 15:07:40 crc kubenswrapper[4805]: I1203 15:07:40.809171 4805 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/prometheus-metric-storage-0"] Dec 03 15:07:40 crc kubenswrapper[4805]: E1203 15:07:40.809890 4805 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2d49b230-da3b-4b61-bd0f-35408593cb08" containerName="thanos-sidecar" Dec 03 15:07:40 crc kubenswrapper[4805]: I1203 15:07:40.809977 4805 state_mem.go:107] "Deleted CPUSet assignment" podUID="2d49b230-da3b-4b61-bd0f-35408593cb08" containerName="thanos-sidecar" Dec 03 15:07:40 crc kubenswrapper[4805]: E1203 15:07:40.810055 4805 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2d49b230-da3b-4b61-bd0f-35408593cb08" containerName="prometheus" Dec 03 15:07:40 crc kubenswrapper[4805]: I1203 15:07:40.810137 4805 state_mem.go:107] "Deleted CPUSet assignment" podUID="2d49b230-da3b-4b61-bd0f-35408593cb08" containerName="prometheus" Dec 03 15:07:40 crc kubenswrapper[4805]: E1203 15:07:40.810234 4805 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2d49b230-da3b-4b61-bd0f-35408593cb08" containerName="config-reloader" Dec 03 15:07:40 crc kubenswrapper[4805]: I1203 15:07:40.810297 4805 state_mem.go:107] "Deleted CPUSet assignment" podUID="2d49b230-da3b-4b61-bd0f-35408593cb08" containerName="config-reloader" Dec 03 15:07:40 crc kubenswrapper[4805]: E1203 15:07:40.810370 4805 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2d49b230-da3b-4b61-bd0f-35408593cb08" containerName="init-config-reloader" Dec 03 15:07:40 crc kubenswrapper[4805]: I1203 15:07:40.810437 4805 state_mem.go:107] "Deleted CPUSet assignment" podUID="2d49b230-da3b-4b61-bd0f-35408593cb08" containerName="init-config-reloader" Dec 03 15:07:40 crc kubenswrapper[4805]: I1203 15:07:40.810712 4805 memory_manager.go:354] "RemoveStaleState removing state" podUID="2d49b230-da3b-4b61-bd0f-35408593cb08" containerName="config-reloader" Dec 03 15:07:40 crc kubenswrapper[4805]: I1203 15:07:40.810795 4805 memory_manager.go:354] "RemoveStaleState removing state" podUID="2d49b230-da3b-4b61-bd0f-35408593cb08" containerName="prometheus" Dec 03 15:07:40 crc kubenswrapper[4805]: I1203 15:07:40.810899 4805 memory_manager.go:354] "RemoveStaleState removing state" podUID="2d49b230-da3b-4b61-bd0f-35408593cb08" containerName="thanos-sidecar" Dec 03 15:07:40 crc kubenswrapper[4805]: I1203 15:07:40.820049 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/prometheus-metric-storage-0" Dec 03 15:07:40 crc kubenswrapper[4805]: I1203 15:07:40.822441 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"prometheus-metric-storage-thanos-prometheus-http-client-file" Dec 03 15:07:40 crc kubenswrapper[4805]: I1203 15:07:40.822992 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"prometheus-metric-storage-web-config" Dec 03 15:07:40 crc kubenswrapper[4805]: I1203 15:07:40.823177 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"prometheus-metric-storage" Dec 03 15:07:40 crc kubenswrapper[4805]: I1203 15:07:40.823310 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-metric-storage-prometheus-svc" Dec 03 15:07:40 crc kubenswrapper[4805]: I1203 15:07:40.823586 4805 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"prometheus-metric-storage-rulefiles-0" Dec 03 15:07:40 crc kubenswrapper[4805]: I1203 15:07:40.823885 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"metric-storage-prometheus-dockercfg-kg4lr" Dec 03 15:07:40 crc kubenswrapper[4805]: I1203 15:07:40.833505 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/prometheus-metric-storage-0"] Dec 03 15:07:40 crc kubenswrapper[4805]: I1203 15:07:40.833649 4805 scope.go:117] "RemoveContainer" containerID="71ef45922e649b1ac11a86659ebc27a9c3917cc06c23e145cd0489cda5b944dc" Dec 03 15:07:40 crc kubenswrapper[4805]: I1203 15:07:40.851273 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"prometheus-metric-storage-tls-assets-0" Dec 03 15:07:41 crc kubenswrapper[4805]: I1203 15:07:41.002912 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"thanos-prometheus-http-client-file\" (UniqueName: \"kubernetes.io/secret/d7c44420-4173-45fe-bf74-5ed086364ea4-thanos-prometheus-http-client-file\") pod \"prometheus-metric-storage-0\" (UID: \"d7c44420-4173-45fe-bf74-5ed086364ea4\") " pod="openstack/prometheus-metric-storage-0" Dec 03 15:07:41 crc kubenswrapper[4805]: I1203 15:07:41.002991 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d7c44420-4173-45fe-bf74-5ed086364ea4-secret-combined-ca-bundle\") pod \"prometheus-metric-storage-0\" (UID: \"d7c44420-4173-45fe-bf74-5ed086364ea4\") " pod="openstack/prometheus-metric-storage-0" Dec 03 15:07:41 crc kubenswrapper[4805]: I1203 15:07:41.003041 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"prometheus-metric-storage-0\" (UID: \"d7c44420-4173-45fe-bf74-5ed086364ea4\") " pod="openstack/prometheus-metric-storage-0" Dec 03 15:07:41 crc kubenswrapper[4805]: I1203 15:07:41.003072 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"web-config-tls-secret-cert-cert-metric-storage-prometh-dc638c2d\" (UniqueName: \"kubernetes.io/secret/d7c44420-4173-45fe-bf74-5ed086364ea4-web-config-tls-secret-cert-cert-metric-storage-prometh-dc638c2d\") pod \"prometheus-metric-storage-0\" (UID: \"d7c44420-4173-45fe-bf74-5ed086364ea4\") " pod="openstack/prometheus-metric-storage-0" Dec 03 15:07:41 crc kubenswrapper[4805]: I1203 15:07:41.003109 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/d7c44420-4173-45fe-bf74-5ed086364ea4-config-out\") pod \"prometheus-metric-storage-0\" (UID: \"d7c44420-4173-45fe-bf74-5ed086364ea4\") " pod="openstack/prometheus-metric-storage-0" Dec 03 15:07:41 crc kubenswrapper[4805]: I1203 15:07:41.003148 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/d7c44420-4173-45fe-bf74-5ed086364ea4-tls-assets\") pod \"prometheus-metric-storage-0\" (UID: \"d7c44420-4173-45fe-bf74-5ed086364ea4\") " pod="openstack/prometheus-metric-storage-0" Dec 03 15:07:41 crc kubenswrapper[4805]: I1203 15:07:41.003184 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/d7c44420-4173-45fe-bf74-5ed086364ea4-web-config\") pod \"prometheus-metric-storage-0\" (UID: \"d7c44420-4173-45fe-bf74-5ed086364ea4\") " pod="openstack/prometheus-metric-storage-0" Dec 03 15:07:41 crc kubenswrapper[4805]: I1203 15:07:41.003216 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/d7c44420-4173-45fe-bf74-5ed086364ea4-config\") pod \"prometheus-metric-storage-0\" (UID: \"d7c44420-4173-45fe-bf74-5ed086364ea4\") " pod="openstack/prometheus-metric-storage-0" Dec 03 15:07:41 crc kubenswrapper[4805]: I1203 15:07:41.003242 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"prometheus-metric-storage-rulefiles-0\" (UniqueName: \"kubernetes.io/configmap/d7c44420-4173-45fe-bf74-5ed086364ea4-prometheus-metric-storage-rulefiles-0\") pod \"prometheus-metric-storage-0\" (UID: \"d7c44420-4173-45fe-bf74-5ed086364ea4\") " pod="openstack/prometheus-metric-storage-0" Dec 03 15:07:41 crc kubenswrapper[4805]: I1203 15:07:41.003274 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"web-config-tls-secret-key-cert-metric-storage-promethe-dc638c2d\" (UniqueName: \"kubernetes.io/secret/d7c44420-4173-45fe-bf74-5ed086364ea4-web-config-tls-secret-key-cert-metric-storage-promethe-dc638c2d\") pod \"prometheus-metric-storage-0\" (UID: \"d7c44420-4173-45fe-bf74-5ed086364ea4\") " pod="openstack/prometheus-metric-storage-0" Dec 03 15:07:41 crc kubenswrapper[4805]: I1203 15:07:41.003345 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4nq5c\" (UniqueName: \"kubernetes.io/projected/d7c44420-4173-45fe-bf74-5ed086364ea4-kube-api-access-4nq5c\") pod \"prometheus-metric-storage-0\" (UID: \"d7c44420-4173-45fe-bf74-5ed086364ea4\") " pod="openstack/prometheus-metric-storage-0" Dec 03 15:07:41 crc kubenswrapper[4805]: I1203 15:07:41.104654 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"thanos-prometheus-http-client-file\" (UniqueName: \"kubernetes.io/secret/d7c44420-4173-45fe-bf74-5ed086364ea4-thanos-prometheus-http-client-file\") pod \"prometheus-metric-storage-0\" (UID: \"d7c44420-4173-45fe-bf74-5ed086364ea4\") " pod="openstack/prometheus-metric-storage-0" Dec 03 15:07:41 crc kubenswrapper[4805]: I1203 15:07:41.104732 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d7c44420-4173-45fe-bf74-5ed086364ea4-secret-combined-ca-bundle\") pod \"prometheus-metric-storage-0\" (UID: \"d7c44420-4173-45fe-bf74-5ed086364ea4\") " pod="openstack/prometheus-metric-storage-0" Dec 03 15:07:41 crc kubenswrapper[4805]: I1203 15:07:41.107485 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"prometheus-metric-storage-0\" (UID: \"d7c44420-4173-45fe-bf74-5ed086364ea4\") " pod="openstack/prometheus-metric-storage-0" Dec 03 15:07:41 crc kubenswrapper[4805]: I1203 15:07:41.107562 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"web-config-tls-secret-cert-cert-metric-storage-prometh-dc638c2d\" (UniqueName: \"kubernetes.io/secret/d7c44420-4173-45fe-bf74-5ed086364ea4-web-config-tls-secret-cert-cert-metric-storage-prometh-dc638c2d\") pod \"prometheus-metric-storage-0\" (UID: \"d7c44420-4173-45fe-bf74-5ed086364ea4\") " pod="openstack/prometheus-metric-storage-0" Dec 03 15:07:41 crc kubenswrapper[4805]: I1203 15:07:41.107671 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/d7c44420-4173-45fe-bf74-5ed086364ea4-config-out\") pod \"prometheus-metric-storage-0\" (UID: \"d7c44420-4173-45fe-bf74-5ed086364ea4\") " pod="openstack/prometheus-metric-storage-0" Dec 03 15:07:41 crc kubenswrapper[4805]: I1203 15:07:41.107766 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/d7c44420-4173-45fe-bf74-5ed086364ea4-tls-assets\") pod \"prometheus-metric-storage-0\" (UID: \"d7c44420-4173-45fe-bf74-5ed086364ea4\") " pod="openstack/prometheus-metric-storage-0" Dec 03 15:07:41 crc kubenswrapper[4805]: I1203 15:07:41.107855 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/d7c44420-4173-45fe-bf74-5ed086364ea4-web-config\") pod \"prometheus-metric-storage-0\" (UID: \"d7c44420-4173-45fe-bf74-5ed086364ea4\") " pod="openstack/prometheus-metric-storage-0" Dec 03 15:07:41 crc kubenswrapper[4805]: I1203 15:07:41.107901 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/d7c44420-4173-45fe-bf74-5ed086364ea4-config\") pod \"prometheus-metric-storage-0\" (UID: \"d7c44420-4173-45fe-bf74-5ed086364ea4\") " pod="openstack/prometheus-metric-storage-0" Dec 03 15:07:41 crc kubenswrapper[4805]: I1203 15:07:41.107937 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"prometheus-metric-storage-rulefiles-0\" (UniqueName: \"kubernetes.io/configmap/d7c44420-4173-45fe-bf74-5ed086364ea4-prometheus-metric-storage-rulefiles-0\") pod \"prometheus-metric-storage-0\" (UID: \"d7c44420-4173-45fe-bf74-5ed086364ea4\") " pod="openstack/prometheus-metric-storage-0" Dec 03 15:07:41 crc kubenswrapper[4805]: I1203 15:07:41.107985 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"web-config-tls-secret-key-cert-metric-storage-promethe-dc638c2d\" (UniqueName: \"kubernetes.io/secret/d7c44420-4173-45fe-bf74-5ed086364ea4-web-config-tls-secret-key-cert-metric-storage-promethe-dc638c2d\") pod \"prometheus-metric-storage-0\" (UID: \"d7c44420-4173-45fe-bf74-5ed086364ea4\") " pod="openstack/prometheus-metric-storage-0" Dec 03 15:07:41 crc kubenswrapper[4805]: I1203 15:07:41.109722 4805 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"prometheus-metric-storage-0\" (UID: \"d7c44420-4173-45fe-bf74-5ed086364ea4\") device mount path \"/mnt/openstack/pv12\"" pod="openstack/prometheus-metric-storage-0" Dec 03 15:07:41 crc kubenswrapper[4805]: I1203 15:07:41.112326 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/d7c44420-4173-45fe-bf74-5ed086364ea4-config-out\") pod \"prometheus-metric-storage-0\" (UID: \"d7c44420-4173-45fe-bf74-5ed086364ea4\") " pod="openstack/prometheus-metric-storage-0" Dec 03 15:07:41 crc kubenswrapper[4805]: I1203 15:07:41.116028 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d7c44420-4173-45fe-bf74-5ed086364ea4-secret-combined-ca-bundle\") pod \"prometheus-metric-storage-0\" (UID: \"d7c44420-4173-45fe-bf74-5ed086364ea4\") " pod="openstack/prometheus-metric-storage-0" Dec 03 15:07:41 crc kubenswrapper[4805]: I1203 15:07:41.117343 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"thanos-prometheus-http-client-file\" (UniqueName: \"kubernetes.io/secret/d7c44420-4173-45fe-bf74-5ed086364ea4-thanos-prometheus-http-client-file\") pod \"prometheus-metric-storage-0\" (UID: \"d7c44420-4173-45fe-bf74-5ed086364ea4\") " pod="openstack/prometheus-metric-storage-0" Dec 03 15:07:41 crc kubenswrapper[4805]: I1203 15:07:41.117969 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"prometheus-metric-storage-rulefiles-0\" (UniqueName: \"kubernetes.io/configmap/d7c44420-4173-45fe-bf74-5ed086364ea4-prometheus-metric-storage-rulefiles-0\") pod \"prometheus-metric-storage-0\" (UID: \"d7c44420-4173-45fe-bf74-5ed086364ea4\") " pod="openstack/prometheus-metric-storage-0" Dec 03 15:07:41 crc kubenswrapper[4805]: I1203 15:07:41.117993 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"web-config-tls-secret-cert-cert-metric-storage-prometh-dc638c2d\" (UniqueName: \"kubernetes.io/secret/d7c44420-4173-45fe-bf74-5ed086364ea4-web-config-tls-secret-cert-cert-metric-storage-prometh-dc638c2d\") pod \"prometheus-metric-storage-0\" (UID: \"d7c44420-4173-45fe-bf74-5ed086364ea4\") " pod="openstack/prometheus-metric-storage-0" Dec 03 15:07:41 crc kubenswrapper[4805]: I1203 15:07:41.118524 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4nq5c\" (UniqueName: \"kubernetes.io/projected/d7c44420-4173-45fe-bf74-5ed086364ea4-kube-api-access-4nq5c\") pod \"prometheus-metric-storage-0\" (UID: \"d7c44420-4173-45fe-bf74-5ed086364ea4\") " pod="openstack/prometheus-metric-storage-0" Dec 03 15:07:41 crc kubenswrapper[4805]: I1203 15:07:41.123532 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/secret/d7c44420-4173-45fe-bf74-5ed086364ea4-config\") pod \"prometheus-metric-storage-0\" (UID: \"d7c44420-4173-45fe-bf74-5ed086364ea4\") " pod="openstack/prometheus-metric-storage-0" Dec 03 15:07:41 crc kubenswrapper[4805]: I1203 15:07:41.126545 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/d7c44420-4173-45fe-bf74-5ed086364ea4-web-config\") pod \"prometheus-metric-storage-0\" (UID: \"d7c44420-4173-45fe-bf74-5ed086364ea4\") " pod="openstack/prometheus-metric-storage-0" Dec 03 15:07:41 crc kubenswrapper[4805]: I1203 15:07:41.131540 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/d7c44420-4173-45fe-bf74-5ed086364ea4-tls-assets\") pod \"prometheus-metric-storage-0\" (UID: \"d7c44420-4173-45fe-bf74-5ed086364ea4\") " pod="openstack/prometheus-metric-storage-0" Dec 03 15:07:41 crc kubenswrapper[4805]: I1203 15:07:41.137289 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4nq5c\" (UniqueName: \"kubernetes.io/projected/d7c44420-4173-45fe-bf74-5ed086364ea4-kube-api-access-4nq5c\") pod \"prometheus-metric-storage-0\" (UID: \"d7c44420-4173-45fe-bf74-5ed086364ea4\") " pod="openstack/prometheus-metric-storage-0" Dec 03 15:07:41 crc kubenswrapper[4805]: I1203 15:07:41.138207 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"web-config-tls-secret-key-cert-metric-storage-promethe-dc638c2d\" (UniqueName: \"kubernetes.io/secret/d7c44420-4173-45fe-bf74-5ed086364ea4-web-config-tls-secret-key-cert-metric-storage-promethe-dc638c2d\") pod \"prometheus-metric-storage-0\" (UID: \"d7c44420-4173-45fe-bf74-5ed086364ea4\") " pod="openstack/prometheus-metric-storage-0" Dec 03 15:07:41 crc kubenswrapper[4805]: I1203 15:07:41.155522 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"prometheus-metric-storage-0\" (UID: \"d7c44420-4173-45fe-bf74-5ed086364ea4\") " pod="openstack/prometheus-metric-storage-0" Dec 03 15:07:41 crc kubenswrapper[4805]: I1203 15:07:41.174523 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/prometheus-metric-storage-0" Dec 03 15:07:41 crc kubenswrapper[4805]: I1203 15:07:41.446629 4805 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-0" Dec 03 15:07:41 crc kubenswrapper[4805]: I1203 15:07:41.630760 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/ef0f7bdc-b48e-41a8-ae3d-9eac4016d70a-public-tls-certs\") pod \"ef0f7bdc-b48e-41a8-ae3d-9eac4016d70a\" (UID: \"ef0f7bdc-b48e-41a8-ae3d-9eac4016d70a\") " Dec 03 15:07:41 crc kubenswrapper[4805]: I1203 15:07:41.631063 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/ef0f7bdc-b48e-41a8-ae3d-9eac4016d70a-internal-tls-certs\") pod \"ef0f7bdc-b48e-41a8-ae3d-9eac4016d70a\" (UID: \"ef0f7bdc-b48e-41a8-ae3d-9eac4016d70a\") " Dec 03 15:07:41 crc kubenswrapper[4805]: I1203 15:07:41.631128 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ef0f7bdc-b48e-41a8-ae3d-9eac4016d70a-combined-ca-bundle\") pod \"ef0f7bdc-b48e-41a8-ae3d-9eac4016d70a\" (UID: \"ef0f7bdc-b48e-41a8-ae3d-9eac4016d70a\") " Dec 03 15:07:41 crc kubenswrapper[4805]: I1203 15:07:41.631207 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gmtrk\" (UniqueName: \"kubernetes.io/projected/ef0f7bdc-b48e-41a8-ae3d-9eac4016d70a-kube-api-access-gmtrk\") pod \"ef0f7bdc-b48e-41a8-ae3d-9eac4016d70a\" (UID: \"ef0f7bdc-b48e-41a8-ae3d-9eac4016d70a\") " Dec 03 15:07:41 crc kubenswrapper[4805]: I1203 15:07:41.631252 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ef0f7bdc-b48e-41a8-ae3d-9eac4016d70a-scripts\") pod \"ef0f7bdc-b48e-41a8-ae3d-9eac4016d70a\" (UID: \"ef0f7bdc-b48e-41a8-ae3d-9eac4016d70a\") " Dec 03 15:07:41 crc kubenswrapper[4805]: I1203 15:07:41.631314 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ef0f7bdc-b48e-41a8-ae3d-9eac4016d70a-config-data\") pod \"ef0f7bdc-b48e-41a8-ae3d-9eac4016d70a\" (UID: \"ef0f7bdc-b48e-41a8-ae3d-9eac4016d70a\") " Dec 03 15:07:41 crc kubenswrapper[4805]: I1203 15:07:41.635673 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ef0f7bdc-b48e-41a8-ae3d-9eac4016d70a-kube-api-access-gmtrk" (OuterVolumeSpecName: "kube-api-access-gmtrk") pod "ef0f7bdc-b48e-41a8-ae3d-9eac4016d70a" (UID: "ef0f7bdc-b48e-41a8-ae3d-9eac4016d70a"). InnerVolumeSpecName "kube-api-access-gmtrk". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 15:07:41 crc kubenswrapper[4805]: I1203 15:07:41.636873 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ef0f7bdc-b48e-41a8-ae3d-9eac4016d70a-scripts" (OuterVolumeSpecName: "scripts") pod "ef0f7bdc-b48e-41a8-ae3d-9eac4016d70a" (UID: "ef0f7bdc-b48e-41a8-ae3d-9eac4016d70a"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 15:07:41 crc kubenswrapper[4805]: I1203 15:07:41.685071 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/prometheus-metric-storage-0"] Dec 03 15:07:41 crc kubenswrapper[4805]: I1203 15:07:41.694465 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ef0f7bdc-b48e-41a8-ae3d-9eac4016d70a-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "ef0f7bdc-b48e-41a8-ae3d-9eac4016d70a" (UID: "ef0f7bdc-b48e-41a8-ae3d-9eac4016d70a"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 15:07:41 crc kubenswrapper[4805]: I1203 15:07:41.698048 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ef0f7bdc-b48e-41a8-ae3d-9eac4016d70a-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "ef0f7bdc-b48e-41a8-ae3d-9eac4016d70a" (UID: "ef0f7bdc-b48e-41a8-ae3d-9eac4016d70a"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 15:07:41 crc kubenswrapper[4805]: I1203 15:07:41.734616 4805 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/ef0f7bdc-b48e-41a8-ae3d-9eac4016d70a-public-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 03 15:07:41 crc kubenswrapper[4805]: I1203 15:07:41.735765 4805 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/ef0f7bdc-b48e-41a8-ae3d-9eac4016d70a-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 03 15:07:41 crc kubenswrapper[4805]: I1203 15:07:41.735861 4805 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gmtrk\" (UniqueName: \"kubernetes.io/projected/ef0f7bdc-b48e-41a8-ae3d-9eac4016d70a-kube-api-access-gmtrk\") on node \"crc\" DevicePath \"\"" Dec 03 15:07:41 crc kubenswrapper[4805]: I1203 15:07:41.736102 4805 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ef0f7bdc-b48e-41a8-ae3d-9eac4016d70a-scripts\") on node \"crc\" DevicePath \"\"" Dec 03 15:07:41 crc kubenswrapper[4805]: I1203 15:07:41.747772 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ef0f7bdc-b48e-41a8-ae3d-9eac4016d70a-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "ef0f7bdc-b48e-41a8-ae3d-9eac4016d70a" (UID: "ef0f7bdc-b48e-41a8-ae3d-9eac4016d70a"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 15:07:41 crc kubenswrapper[4805]: I1203 15:07:41.748102 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ef0f7bdc-b48e-41a8-ae3d-9eac4016d70a-config-data" (OuterVolumeSpecName: "config-data") pod "ef0f7bdc-b48e-41a8-ae3d-9eac4016d70a" (UID: "ef0f7bdc-b48e-41a8-ae3d-9eac4016d70a"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 15:07:41 crc kubenswrapper[4805]: I1203 15:07:41.753927 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"d7c44420-4173-45fe-bf74-5ed086364ea4","Type":"ContainerStarted","Data":"a65f24232100da3ec26a09315e7aebfa9eb530da1408750d76417c07fc9f94c0"} Dec 03 15:07:41 crc kubenswrapper[4805]: I1203 15:07:41.757091 4805 generic.go:334] "Generic (PLEG): container finished" podID="ef0f7bdc-b48e-41a8-ae3d-9eac4016d70a" containerID="fb6610f613e5e1b6790967f098610eadabd6b001d42a97c2699e49aa08ac0397" exitCode=0 Dec 03 15:07:41 crc kubenswrapper[4805]: I1203 15:07:41.757152 4805 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-0" Dec 03 15:07:41 crc kubenswrapper[4805]: I1203 15:07:41.757174 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-0" event={"ID":"ef0f7bdc-b48e-41a8-ae3d-9eac4016d70a","Type":"ContainerDied","Data":"fb6610f613e5e1b6790967f098610eadabd6b001d42a97c2699e49aa08ac0397"} Dec 03 15:07:41 crc kubenswrapper[4805]: I1203 15:07:41.757236 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-0" event={"ID":"ef0f7bdc-b48e-41a8-ae3d-9eac4016d70a","Type":"ContainerDied","Data":"ffe45c7453df2f2a0b80ed7599a022e262f35174e810bde7bc2d76fad1256201"} Dec 03 15:07:41 crc kubenswrapper[4805]: I1203 15:07:41.757256 4805 scope.go:117] "RemoveContainer" containerID="fb6610f613e5e1b6790967f098610eadabd6b001d42a97c2699e49aa08ac0397" Dec 03 15:07:41 crc kubenswrapper[4805]: I1203 15:07:41.778529 4805 scope.go:117] "RemoveContainer" containerID="bf67c1c3e37fd095badea3841d170fb3b8345ca6a40d2106e16d2ccf4ed783e0" Dec 03 15:07:41 crc kubenswrapper[4805]: I1203 15:07:41.798833 4805 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/aodh-0"] Dec 03 15:07:41 crc kubenswrapper[4805]: I1203 15:07:41.811181 4805 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/aodh-0"] Dec 03 15:07:41 crc kubenswrapper[4805]: I1203 15:07:41.829215 4805 scope.go:117] "RemoveContainer" containerID="dfdd2ac28b14d459abf2c4b6a74db213eb06cb0368a0422d75becb5b25e09cb3" Dec 03 15:07:41 crc kubenswrapper[4805]: I1203 15:07:41.836225 4805 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/aodh-0"] Dec 03 15:07:41 crc kubenswrapper[4805]: E1203 15:07:41.836744 4805 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ef0f7bdc-b48e-41a8-ae3d-9eac4016d70a" containerName="aodh-api" Dec 03 15:07:41 crc kubenswrapper[4805]: I1203 15:07:41.836769 4805 state_mem.go:107] "Deleted CPUSet assignment" podUID="ef0f7bdc-b48e-41a8-ae3d-9eac4016d70a" containerName="aodh-api" Dec 03 15:07:41 crc kubenswrapper[4805]: E1203 15:07:41.836810 4805 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ef0f7bdc-b48e-41a8-ae3d-9eac4016d70a" containerName="aodh-listener" Dec 03 15:07:41 crc kubenswrapper[4805]: I1203 15:07:41.836818 4805 state_mem.go:107] "Deleted CPUSet assignment" podUID="ef0f7bdc-b48e-41a8-ae3d-9eac4016d70a" containerName="aodh-listener" Dec 03 15:07:41 crc kubenswrapper[4805]: E1203 15:07:41.836832 4805 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ef0f7bdc-b48e-41a8-ae3d-9eac4016d70a" containerName="aodh-evaluator" Dec 03 15:07:41 crc kubenswrapper[4805]: I1203 15:07:41.836842 4805 state_mem.go:107] "Deleted CPUSet assignment" podUID="ef0f7bdc-b48e-41a8-ae3d-9eac4016d70a" containerName="aodh-evaluator" Dec 03 15:07:41 crc kubenswrapper[4805]: E1203 15:07:41.836942 4805 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ef0f7bdc-b48e-41a8-ae3d-9eac4016d70a" containerName="aodh-notifier" Dec 03 15:07:41 crc kubenswrapper[4805]: I1203 15:07:41.836952 4805 state_mem.go:107] "Deleted CPUSet assignment" podUID="ef0f7bdc-b48e-41a8-ae3d-9eac4016d70a" containerName="aodh-notifier" Dec 03 15:07:41 crc kubenswrapper[4805]: I1203 15:07:41.837208 4805 memory_manager.go:354] "RemoveStaleState removing state" podUID="ef0f7bdc-b48e-41a8-ae3d-9eac4016d70a" containerName="aodh-listener" Dec 03 15:07:41 crc kubenswrapper[4805]: I1203 15:07:41.837236 4805 memory_manager.go:354] "RemoveStaleState removing state" podUID="ef0f7bdc-b48e-41a8-ae3d-9eac4016d70a" containerName="aodh-api" Dec 03 15:07:41 crc kubenswrapper[4805]: I1203 15:07:41.837252 4805 memory_manager.go:354] "RemoveStaleState removing state" podUID="ef0f7bdc-b48e-41a8-ae3d-9eac4016d70a" containerName="aodh-notifier" Dec 03 15:07:41 crc kubenswrapper[4805]: I1203 15:07:41.837265 4805 memory_manager.go:354] "RemoveStaleState removing state" podUID="ef0f7bdc-b48e-41a8-ae3d-9eac4016d70a" containerName="aodh-evaluator" Dec 03 15:07:41 crc kubenswrapper[4805]: I1203 15:07:41.838437 4805 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ef0f7bdc-b48e-41a8-ae3d-9eac4016d70a-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 15:07:41 crc kubenswrapper[4805]: I1203 15:07:41.838591 4805 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ef0f7bdc-b48e-41a8-ae3d-9eac4016d70a-config-data\") on node \"crc\" DevicePath \"\"" Dec 03 15:07:41 crc kubenswrapper[4805]: I1203 15:07:41.839704 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-0" Dec 03 15:07:41 crc kubenswrapper[4805]: I1203 15:07:41.842378 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-aodh-internal-svc" Dec 03 15:07:41 crc kubenswrapper[4805]: I1203 15:07:41.842446 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-aodh-public-svc" Dec 03 15:07:41 crc kubenswrapper[4805]: I1203 15:07:41.842452 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"telemetry-autoscaling-dockercfg-wc8kz" Dec 03 15:07:41 crc kubenswrapper[4805]: I1203 15:07:41.842390 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"aodh-config-data" Dec 03 15:07:41 crc kubenswrapper[4805]: I1203 15:07:41.842884 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"aodh-scripts" Dec 03 15:07:41 crc kubenswrapper[4805]: I1203 15:07:41.858199 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/aodh-0"] Dec 03 15:07:41 crc kubenswrapper[4805]: I1203 15:07:41.864180 4805 scope.go:117] "RemoveContainer" containerID="ef96a03285bf5bb5faad845eb35c37c353fce53de018d916db07ac65f0cbc9e2" Dec 03 15:07:41 crc kubenswrapper[4805]: I1203 15:07:41.889077 4805 scope.go:117] "RemoveContainer" containerID="fb6610f613e5e1b6790967f098610eadabd6b001d42a97c2699e49aa08ac0397" Dec 03 15:07:41 crc kubenswrapper[4805]: E1203 15:07:41.889491 4805 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"fb6610f613e5e1b6790967f098610eadabd6b001d42a97c2699e49aa08ac0397\": container with ID starting with fb6610f613e5e1b6790967f098610eadabd6b001d42a97c2699e49aa08ac0397 not found: ID does not exist" containerID="fb6610f613e5e1b6790967f098610eadabd6b001d42a97c2699e49aa08ac0397" Dec 03 15:07:41 crc kubenswrapper[4805]: I1203 15:07:41.889540 4805 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"fb6610f613e5e1b6790967f098610eadabd6b001d42a97c2699e49aa08ac0397"} err="failed to get container status \"fb6610f613e5e1b6790967f098610eadabd6b001d42a97c2699e49aa08ac0397\": rpc error: code = NotFound desc = could not find container \"fb6610f613e5e1b6790967f098610eadabd6b001d42a97c2699e49aa08ac0397\": container with ID starting with fb6610f613e5e1b6790967f098610eadabd6b001d42a97c2699e49aa08ac0397 not found: ID does not exist" Dec 03 15:07:41 crc kubenswrapper[4805]: I1203 15:07:41.889574 4805 scope.go:117] "RemoveContainer" containerID="bf67c1c3e37fd095badea3841d170fb3b8345ca6a40d2106e16d2ccf4ed783e0" Dec 03 15:07:41 crc kubenswrapper[4805]: E1203 15:07:41.889989 4805 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"bf67c1c3e37fd095badea3841d170fb3b8345ca6a40d2106e16d2ccf4ed783e0\": container with ID starting with bf67c1c3e37fd095badea3841d170fb3b8345ca6a40d2106e16d2ccf4ed783e0 not found: ID does not exist" containerID="bf67c1c3e37fd095badea3841d170fb3b8345ca6a40d2106e16d2ccf4ed783e0" Dec 03 15:07:41 crc kubenswrapper[4805]: I1203 15:07:41.890018 4805 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"bf67c1c3e37fd095badea3841d170fb3b8345ca6a40d2106e16d2ccf4ed783e0"} err="failed to get container status \"bf67c1c3e37fd095badea3841d170fb3b8345ca6a40d2106e16d2ccf4ed783e0\": rpc error: code = NotFound desc = could not find container \"bf67c1c3e37fd095badea3841d170fb3b8345ca6a40d2106e16d2ccf4ed783e0\": container with ID starting with bf67c1c3e37fd095badea3841d170fb3b8345ca6a40d2106e16d2ccf4ed783e0 not found: ID does not exist" Dec 03 15:07:41 crc kubenswrapper[4805]: I1203 15:07:41.890040 4805 scope.go:117] "RemoveContainer" containerID="dfdd2ac28b14d459abf2c4b6a74db213eb06cb0368a0422d75becb5b25e09cb3" Dec 03 15:07:41 crc kubenswrapper[4805]: E1203 15:07:41.890407 4805 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"dfdd2ac28b14d459abf2c4b6a74db213eb06cb0368a0422d75becb5b25e09cb3\": container with ID starting with dfdd2ac28b14d459abf2c4b6a74db213eb06cb0368a0422d75becb5b25e09cb3 not found: ID does not exist" containerID="dfdd2ac28b14d459abf2c4b6a74db213eb06cb0368a0422d75becb5b25e09cb3" Dec 03 15:07:41 crc kubenswrapper[4805]: I1203 15:07:41.890433 4805 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"dfdd2ac28b14d459abf2c4b6a74db213eb06cb0368a0422d75becb5b25e09cb3"} err="failed to get container status \"dfdd2ac28b14d459abf2c4b6a74db213eb06cb0368a0422d75becb5b25e09cb3\": rpc error: code = NotFound desc = could not find container \"dfdd2ac28b14d459abf2c4b6a74db213eb06cb0368a0422d75becb5b25e09cb3\": container with ID starting with dfdd2ac28b14d459abf2c4b6a74db213eb06cb0368a0422d75becb5b25e09cb3 not found: ID does not exist" Dec 03 15:07:41 crc kubenswrapper[4805]: I1203 15:07:41.890467 4805 scope.go:117] "RemoveContainer" containerID="ef96a03285bf5bb5faad845eb35c37c353fce53de018d916db07ac65f0cbc9e2" Dec 03 15:07:41 crc kubenswrapper[4805]: E1203 15:07:41.890920 4805 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ef96a03285bf5bb5faad845eb35c37c353fce53de018d916db07ac65f0cbc9e2\": container with ID starting with ef96a03285bf5bb5faad845eb35c37c353fce53de018d916db07ac65f0cbc9e2 not found: ID does not exist" containerID="ef96a03285bf5bb5faad845eb35c37c353fce53de018d916db07ac65f0cbc9e2" Dec 03 15:07:41 crc kubenswrapper[4805]: I1203 15:07:41.890943 4805 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ef96a03285bf5bb5faad845eb35c37c353fce53de018d916db07ac65f0cbc9e2"} err="failed to get container status \"ef96a03285bf5bb5faad845eb35c37c353fce53de018d916db07ac65f0cbc9e2\": rpc error: code = NotFound desc = could not find container \"ef96a03285bf5bb5faad845eb35c37c353fce53de018d916db07ac65f0cbc9e2\": container with ID starting with ef96a03285bf5bb5faad845eb35c37c353fce53de018d916db07ac65f0cbc9e2 not found: ID does not exist" Dec 03 15:07:41 crc kubenswrapper[4805]: I1203 15:07:41.939765 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d60bc69e-b77e-47ac-9a86-a50533f61711-scripts\") pod \"aodh-0\" (UID: \"d60bc69e-b77e-47ac-9a86-a50533f61711\") " pod="openstack/aodh-0" Dec 03 15:07:41 crc kubenswrapper[4805]: I1203 15:07:41.939818 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d60bc69e-b77e-47ac-9a86-a50533f61711-combined-ca-bundle\") pod \"aodh-0\" (UID: \"d60bc69e-b77e-47ac-9a86-a50533f61711\") " pod="openstack/aodh-0" Dec 03 15:07:41 crc kubenswrapper[4805]: I1203 15:07:41.939910 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/d60bc69e-b77e-47ac-9a86-a50533f61711-internal-tls-certs\") pod \"aodh-0\" (UID: \"d60bc69e-b77e-47ac-9a86-a50533f61711\") " pod="openstack/aodh-0" Dec 03 15:07:41 crc kubenswrapper[4805]: I1203 15:07:41.939977 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/d60bc69e-b77e-47ac-9a86-a50533f61711-public-tls-certs\") pod \"aodh-0\" (UID: \"d60bc69e-b77e-47ac-9a86-a50533f61711\") " pod="openstack/aodh-0" Dec 03 15:07:41 crc kubenswrapper[4805]: I1203 15:07:41.940060 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ppfpx\" (UniqueName: \"kubernetes.io/projected/d60bc69e-b77e-47ac-9a86-a50533f61711-kube-api-access-ppfpx\") pod \"aodh-0\" (UID: \"d60bc69e-b77e-47ac-9a86-a50533f61711\") " pod="openstack/aodh-0" Dec 03 15:07:41 crc kubenswrapper[4805]: I1203 15:07:41.940086 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d60bc69e-b77e-47ac-9a86-a50533f61711-config-data\") pod \"aodh-0\" (UID: \"d60bc69e-b77e-47ac-9a86-a50533f61711\") " pod="openstack/aodh-0" Dec 03 15:07:42 crc kubenswrapper[4805]: I1203 15:07:42.041990 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d60bc69e-b77e-47ac-9a86-a50533f61711-scripts\") pod \"aodh-0\" (UID: \"d60bc69e-b77e-47ac-9a86-a50533f61711\") " pod="openstack/aodh-0" Dec 03 15:07:42 crc kubenswrapper[4805]: I1203 15:07:42.042043 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d60bc69e-b77e-47ac-9a86-a50533f61711-combined-ca-bundle\") pod \"aodh-0\" (UID: \"d60bc69e-b77e-47ac-9a86-a50533f61711\") " pod="openstack/aodh-0" Dec 03 15:07:42 crc kubenswrapper[4805]: I1203 15:07:42.042101 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/d60bc69e-b77e-47ac-9a86-a50533f61711-internal-tls-certs\") pod \"aodh-0\" (UID: \"d60bc69e-b77e-47ac-9a86-a50533f61711\") " pod="openstack/aodh-0" Dec 03 15:07:42 crc kubenswrapper[4805]: I1203 15:07:42.042133 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/d60bc69e-b77e-47ac-9a86-a50533f61711-public-tls-certs\") pod \"aodh-0\" (UID: \"d60bc69e-b77e-47ac-9a86-a50533f61711\") " pod="openstack/aodh-0" Dec 03 15:07:42 crc kubenswrapper[4805]: I1203 15:07:42.042185 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ppfpx\" (UniqueName: \"kubernetes.io/projected/d60bc69e-b77e-47ac-9a86-a50533f61711-kube-api-access-ppfpx\") pod \"aodh-0\" (UID: \"d60bc69e-b77e-47ac-9a86-a50533f61711\") " pod="openstack/aodh-0" Dec 03 15:07:42 crc kubenswrapper[4805]: I1203 15:07:42.042213 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d60bc69e-b77e-47ac-9a86-a50533f61711-config-data\") pod \"aodh-0\" (UID: \"d60bc69e-b77e-47ac-9a86-a50533f61711\") " pod="openstack/aodh-0" Dec 03 15:07:42 crc kubenswrapper[4805]: I1203 15:07:42.046546 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d60bc69e-b77e-47ac-9a86-a50533f61711-combined-ca-bundle\") pod \"aodh-0\" (UID: \"d60bc69e-b77e-47ac-9a86-a50533f61711\") " pod="openstack/aodh-0" Dec 03 15:07:42 crc kubenswrapper[4805]: I1203 15:07:42.046885 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d60bc69e-b77e-47ac-9a86-a50533f61711-scripts\") pod \"aodh-0\" (UID: \"d60bc69e-b77e-47ac-9a86-a50533f61711\") " pod="openstack/aodh-0" Dec 03 15:07:42 crc kubenswrapper[4805]: I1203 15:07:42.047019 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/d60bc69e-b77e-47ac-9a86-a50533f61711-internal-tls-certs\") pod \"aodh-0\" (UID: \"d60bc69e-b77e-47ac-9a86-a50533f61711\") " pod="openstack/aodh-0" Dec 03 15:07:42 crc kubenswrapper[4805]: I1203 15:07:42.047597 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d60bc69e-b77e-47ac-9a86-a50533f61711-config-data\") pod \"aodh-0\" (UID: \"d60bc69e-b77e-47ac-9a86-a50533f61711\") " pod="openstack/aodh-0" Dec 03 15:07:42 crc kubenswrapper[4805]: I1203 15:07:42.047957 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/d60bc69e-b77e-47ac-9a86-a50533f61711-public-tls-certs\") pod \"aodh-0\" (UID: \"d60bc69e-b77e-47ac-9a86-a50533f61711\") " pod="openstack/aodh-0" Dec 03 15:07:42 crc kubenswrapper[4805]: I1203 15:07:42.059214 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ppfpx\" (UniqueName: \"kubernetes.io/projected/d60bc69e-b77e-47ac-9a86-a50533f61711-kube-api-access-ppfpx\") pod \"aodh-0\" (UID: \"d60bc69e-b77e-47ac-9a86-a50533f61711\") " pod="openstack/aodh-0" Dec 03 15:07:42 crc kubenswrapper[4805]: I1203 15:07:42.165069 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-0" Dec 03 15:07:43 crc kubenswrapper[4805]: I1203 15:07:42.621196 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/aodh-0"] Dec 03 15:07:43 crc kubenswrapper[4805]: I1203 15:07:42.710172 4805 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2d49b230-da3b-4b61-bd0f-35408593cb08" path="/var/lib/kubelet/pods/2d49b230-da3b-4b61-bd0f-35408593cb08/volumes" Dec 03 15:07:43 crc kubenswrapper[4805]: I1203 15:07:42.711649 4805 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ef0f7bdc-b48e-41a8-ae3d-9eac4016d70a" path="/var/lib/kubelet/pods/ef0f7bdc-b48e-41a8-ae3d-9eac4016d70a/volumes" Dec 03 15:07:43 crc kubenswrapper[4805]: I1203 15:07:42.768802 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-0" event={"ID":"d60bc69e-b77e-47ac-9a86-a50533f61711","Type":"ContainerStarted","Data":"348b523ea3abdf9a6864ac8ba27b81fb90f5d07c9c6b91133ca8ca100cc1d401"} Dec 03 15:07:43 crc kubenswrapper[4805]: I1203 15:07:42.802281 4805 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/prometheus-metric-storage-0" podUID="2d49b230-da3b-4b61-bd0f-35408593cb08" containerName="prometheus" probeResult="failure" output="Get \"http://10.217.1.10:9090/-/ready\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 03 15:07:43 crc kubenswrapper[4805]: I1203 15:07:43.779894 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-0" event={"ID":"d60bc69e-b77e-47ac-9a86-a50533f61711","Type":"ContainerStarted","Data":"b6244e91da7b185eb4899fd3b52a1a8356420a93544d4555e64c6dd85f412d57"} Dec 03 15:07:44 crc kubenswrapper[4805]: I1203 15:07:44.791596 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-0" event={"ID":"d60bc69e-b77e-47ac-9a86-a50533f61711","Type":"ContainerStarted","Data":"de129d6dcf6dbf4a79fec418b3e50087c1668058c15adcda0c51eae694f465ce"} Dec 03 15:07:45 crc kubenswrapper[4805]: I1203 15:07:45.804128 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-0" event={"ID":"d60bc69e-b77e-47ac-9a86-a50533f61711","Type":"ContainerStarted","Data":"374e25e9dfc63ccf7ee10520c0f4af5e17c37e5b2a010ab02f4c9f45e9a599da"} Dec 03 15:07:45 crc kubenswrapper[4805]: I1203 15:07:45.806333 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"d7c44420-4173-45fe-bf74-5ed086364ea4","Type":"ContainerStarted","Data":"894956bdcafeba1a2eae20d8f53ec468f00b495dc09e3affa45032f7834cdd29"} Dec 03 15:07:47 crc kubenswrapper[4805]: I1203 15:07:47.830467 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-0" event={"ID":"d60bc69e-b77e-47ac-9a86-a50533f61711","Type":"ContainerStarted","Data":"eabbde92660ffc9b7960552812807370f6c0f71b2e1e3967895c09dbcdec0ee4"} Dec 03 15:07:47 crc kubenswrapper[4805]: I1203 15:07:47.860259 4805 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/aodh-0" podStartSLOduration=2.884596758 podStartE2EDuration="6.860235395s" podCreationTimestamp="2025-12-03 15:07:41 +0000 UTC" firstStartedPulling="2025-12-03 15:07:42.622574307 +0000 UTC m=+3492.285491230" lastFinishedPulling="2025-12-03 15:07:46.598212944 +0000 UTC m=+3496.261129867" observedRunningTime="2025-12-03 15:07:47.858796667 +0000 UTC m=+3497.521713600" watchObservedRunningTime="2025-12-03 15:07:47.860235395 +0000 UTC m=+3497.523152368" Dec 03 15:07:52 crc kubenswrapper[4805]: I1203 15:07:52.883737 4805 generic.go:334] "Generic (PLEG): container finished" podID="d7c44420-4173-45fe-bf74-5ed086364ea4" containerID="894956bdcafeba1a2eae20d8f53ec468f00b495dc09e3affa45032f7834cdd29" exitCode=0 Dec 03 15:07:52 crc kubenswrapper[4805]: I1203 15:07:52.883877 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"d7c44420-4173-45fe-bf74-5ed086364ea4","Type":"ContainerDied","Data":"894956bdcafeba1a2eae20d8f53ec468f00b495dc09e3affa45032f7834cdd29"} Dec 03 15:07:53 crc kubenswrapper[4805]: I1203 15:07:53.907114 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"d7c44420-4173-45fe-bf74-5ed086364ea4","Type":"ContainerStarted","Data":"048c55494b984dd592e98c8291e407fc1ebaaebdb6f45a57dccf0eed800ea2f2"} Dec 03 15:07:56 crc kubenswrapper[4805]: I1203 15:07:56.958427 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"d7c44420-4173-45fe-bf74-5ed086364ea4","Type":"ContainerStarted","Data":"e0fecc8750cd2830f57ad4347ef1f659b9c5365badd25705fea3633133b1c604"} Dec 03 15:07:56 crc kubenswrapper[4805]: I1203 15:07:56.958895 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"d7c44420-4173-45fe-bf74-5ed086364ea4","Type":"ContainerStarted","Data":"38d5882460472e27d4ce5803cfbdb0352f15ab198d24ac83c6062b2c7580f76a"} Dec 03 15:07:56 crc kubenswrapper[4805]: I1203 15:07:56.994947 4805 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/prometheus-metric-storage-0" podStartSLOduration=16.994920113 podStartE2EDuration="16.994920113s" podCreationTimestamp="2025-12-03 15:07:40 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 15:07:56.989198881 +0000 UTC m=+3506.652115804" watchObservedRunningTime="2025-12-03 15:07:56.994920113 +0000 UTC m=+3506.657837056" Dec 03 15:08:01 crc kubenswrapper[4805]: I1203 15:08:01.175527 4805 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/prometheus-metric-storage-0" Dec 03 15:08:11 crc kubenswrapper[4805]: I1203 15:08:11.175759 4805 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/prometheus-metric-storage-0" Dec 03 15:08:11 crc kubenswrapper[4805]: I1203 15:08:11.181728 4805 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/prometheus-metric-storage-0" Dec 03 15:08:12 crc kubenswrapper[4805]: I1203 15:08:12.124816 4805 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/prometheus-metric-storage-0" Dec 03 15:08:45 crc kubenswrapper[4805]: I1203 15:08:45.731804 4805 scope.go:117] "RemoveContainer" containerID="7324ec81fc734ace11a78f898680039c33e91f55796fb2e1b43dbb0722390a4e" Dec 03 15:08:45 crc kubenswrapper[4805]: I1203 15:08:45.753711 4805 scope.go:117] "RemoveContainer" containerID="3797ba396ee119b5e8cb994e9b2aeef70af1f115ba9a92078848e669283054e6" Dec 03 15:08:45 crc kubenswrapper[4805]: I1203 15:08:45.779512 4805 scope.go:117] "RemoveContainer" containerID="962645f025d997f376be2e9db91f639ab21b690752350e7f3ad6d738ea6e7456" Dec 03 15:08:45 crc kubenswrapper[4805]: I1203 15:08:45.799996 4805 scope.go:117] "RemoveContainer" containerID="e05fe671c29954321872dfb71956fb33f0b62d4d8bc62c1bd80b77224e98a33a" Dec 03 15:09:13 crc kubenswrapper[4805]: I1203 15:09:13.917073 4805 patch_prober.go:28] interesting pod/machine-config-daemon-gskh4 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 03 15:09:13 crc kubenswrapper[4805]: I1203 15:09:13.917657 4805 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-gskh4" podUID="6098937f-e3f6-45e8-a647-4994a79cd711" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 03 15:09:43 crc kubenswrapper[4805]: I1203 15:09:43.917621 4805 patch_prober.go:28] interesting pod/machine-config-daemon-gskh4 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 03 15:09:43 crc kubenswrapper[4805]: I1203 15:09:43.918349 4805 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-gskh4" podUID="6098937f-e3f6-45e8-a647-4994a79cd711" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 03 15:10:13 crc kubenswrapper[4805]: I1203 15:10:13.917934 4805 patch_prober.go:28] interesting pod/machine-config-daemon-gskh4 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 03 15:10:13 crc kubenswrapper[4805]: I1203 15:10:13.918643 4805 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-gskh4" podUID="6098937f-e3f6-45e8-a647-4994a79cd711" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 03 15:10:13 crc kubenswrapper[4805]: I1203 15:10:13.918712 4805 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-gskh4" Dec 03 15:10:13 crc kubenswrapper[4805]: I1203 15:10:13.919934 4805 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"e54004d4fb8c135b56b16241c1140f37c36df51d3d40b05d6724caf5e9950f98"} pod="openshift-machine-config-operator/machine-config-daemon-gskh4" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 03 15:10:13 crc kubenswrapper[4805]: I1203 15:10:13.920056 4805 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-gskh4" podUID="6098937f-e3f6-45e8-a647-4994a79cd711" containerName="machine-config-daemon" containerID="cri-o://e54004d4fb8c135b56b16241c1140f37c36df51d3d40b05d6724caf5e9950f98" gracePeriod=600 Dec 03 15:10:14 crc kubenswrapper[4805]: I1203 15:10:14.373885 4805 generic.go:334] "Generic (PLEG): container finished" podID="6098937f-e3f6-45e8-a647-4994a79cd711" containerID="e54004d4fb8c135b56b16241c1140f37c36df51d3d40b05d6724caf5e9950f98" exitCode=0 Dec 03 15:10:14 crc kubenswrapper[4805]: I1203 15:10:14.373935 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-gskh4" event={"ID":"6098937f-e3f6-45e8-a647-4994a79cd711","Type":"ContainerDied","Data":"e54004d4fb8c135b56b16241c1140f37c36df51d3d40b05d6724caf5e9950f98"} Dec 03 15:10:14 crc kubenswrapper[4805]: I1203 15:10:14.373971 4805 scope.go:117] "RemoveContainer" containerID="13e0da4d7fc8f4085be601d845d688938c16ea7941e52ca511e39a37d6e0ba4e" Dec 03 15:10:14 crc kubenswrapper[4805]: E1203 15:10:14.471170 4805 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod6098937f_e3f6_45e8_a647_4994a79cd711.slice/crio-conmon-e54004d4fb8c135b56b16241c1140f37c36df51d3d40b05d6724caf5e9950f98.scope\": RecentStats: unable to find data in memory cache]" Dec 03 15:10:15 crc kubenswrapper[4805]: I1203 15:10:15.384761 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-gskh4" event={"ID":"6098937f-e3f6-45e8-a647-4994a79cd711","Type":"ContainerStarted","Data":"cbf73c6c9ba740c0a523c84ca23c0851292d6eba7a0fb47d38599acca964e9cc"} Dec 03 15:10:23 crc kubenswrapper[4805]: I1203 15:10:23.520094 4805 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_telemetry-operator-controller-manager-cc9d8f87b-qfmmx_8cbfcd07-e136-4bfd-b3c6-6615c911d440/manager/0.log" Dec 03 15:10:25 crc kubenswrapper[4805]: I1203 15:10:25.468616 4805 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/prometheus-metric-storage-0"] Dec 03 15:10:25 crc kubenswrapper[4805]: I1203 15:10:25.469516 4805 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/prometheus-metric-storage-0" podUID="d7c44420-4173-45fe-bf74-5ed086364ea4" containerName="prometheus" containerID="cri-o://048c55494b984dd592e98c8291e407fc1ebaaebdb6f45a57dccf0eed800ea2f2" gracePeriod=600 Dec 03 15:10:25 crc kubenswrapper[4805]: I1203 15:10:25.469669 4805 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/prometheus-metric-storage-0" podUID="d7c44420-4173-45fe-bf74-5ed086364ea4" containerName="thanos-sidecar" containerID="cri-o://e0fecc8750cd2830f57ad4347ef1f659b9c5365badd25705fea3633133b1c604" gracePeriod=600 Dec 03 15:10:25 crc kubenswrapper[4805]: I1203 15:10:25.469717 4805 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/prometheus-metric-storage-0" podUID="d7c44420-4173-45fe-bf74-5ed086364ea4" containerName="config-reloader" containerID="cri-o://38d5882460472e27d4ce5803cfbdb0352f15ab198d24ac83c6062b2c7580f76a" gracePeriod=600 Dec 03 15:10:26 crc kubenswrapper[4805]: I1203 15:10:26.175729 4805 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/prometheus-metric-storage-0" podUID="d7c44420-4173-45fe-bf74-5ed086364ea4" containerName="prometheus" probeResult="failure" output="Get \"https://10.217.1.12:9090/-/ready\": dial tcp 10.217.1.12:9090: connect: connection refused" Dec 03 15:10:26 crc kubenswrapper[4805]: I1203 15:10:26.505505 4805 generic.go:334] "Generic (PLEG): container finished" podID="d7c44420-4173-45fe-bf74-5ed086364ea4" containerID="e0fecc8750cd2830f57ad4347ef1f659b9c5365badd25705fea3633133b1c604" exitCode=0 Dec 03 15:10:26 crc kubenswrapper[4805]: I1203 15:10:26.505548 4805 generic.go:334] "Generic (PLEG): container finished" podID="d7c44420-4173-45fe-bf74-5ed086364ea4" containerID="38d5882460472e27d4ce5803cfbdb0352f15ab198d24ac83c6062b2c7580f76a" exitCode=0 Dec 03 15:10:26 crc kubenswrapper[4805]: I1203 15:10:26.505558 4805 generic.go:334] "Generic (PLEG): container finished" podID="d7c44420-4173-45fe-bf74-5ed086364ea4" containerID="048c55494b984dd592e98c8291e407fc1ebaaebdb6f45a57dccf0eed800ea2f2" exitCode=0 Dec 03 15:10:26 crc kubenswrapper[4805]: I1203 15:10:26.505576 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"d7c44420-4173-45fe-bf74-5ed086364ea4","Type":"ContainerDied","Data":"e0fecc8750cd2830f57ad4347ef1f659b9c5365badd25705fea3633133b1c604"} Dec 03 15:10:26 crc kubenswrapper[4805]: I1203 15:10:26.505622 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"d7c44420-4173-45fe-bf74-5ed086364ea4","Type":"ContainerDied","Data":"38d5882460472e27d4ce5803cfbdb0352f15ab198d24ac83c6062b2c7580f76a"} Dec 03 15:10:26 crc kubenswrapper[4805]: I1203 15:10:26.505632 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"d7c44420-4173-45fe-bf74-5ed086364ea4","Type":"ContainerDied","Data":"048c55494b984dd592e98c8291e407fc1ebaaebdb6f45a57dccf0eed800ea2f2"} Dec 03 15:10:26 crc kubenswrapper[4805]: I1203 15:10:26.505643 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"d7c44420-4173-45fe-bf74-5ed086364ea4","Type":"ContainerDied","Data":"a65f24232100da3ec26a09315e7aebfa9eb530da1408750d76417c07fc9f94c0"} Dec 03 15:10:26 crc kubenswrapper[4805]: I1203 15:10:26.505652 4805 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="a65f24232100da3ec26a09315e7aebfa9eb530da1408750d76417c07fc9f94c0" Dec 03 15:10:26 crc kubenswrapper[4805]: I1203 15:10:26.575156 4805 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/prometheus-metric-storage-0" Dec 03 15:10:26 crc kubenswrapper[4805]: I1203 15:10:26.721032 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"web-config-tls-secret-cert-cert-metric-storage-prometh-dc638c2d\" (UniqueName: \"kubernetes.io/secret/d7c44420-4173-45fe-bf74-5ed086364ea4-web-config-tls-secret-cert-cert-metric-storage-prometh-dc638c2d\") pod \"d7c44420-4173-45fe-bf74-5ed086364ea4\" (UID: \"d7c44420-4173-45fe-bf74-5ed086364ea4\") " Dec 03 15:10:26 crc kubenswrapper[4805]: I1203 15:10:26.721140 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/d7c44420-4173-45fe-bf74-5ed086364ea4-config-out\") pod \"d7c44420-4173-45fe-bf74-5ed086364ea4\" (UID: \"d7c44420-4173-45fe-bf74-5ed086364ea4\") " Dec 03 15:10:26 crc kubenswrapper[4805]: I1203 15:10:26.721189 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/d7c44420-4173-45fe-bf74-5ed086364ea4-web-config\") pod \"d7c44420-4173-45fe-bf74-5ed086364ea4\" (UID: \"d7c44420-4173-45fe-bf74-5ed086364ea4\") " Dec 03 15:10:26 crc kubenswrapper[4805]: I1203 15:10:26.721232 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/d7c44420-4173-45fe-bf74-5ed086364ea4-config\") pod \"d7c44420-4173-45fe-bf74-5ed086364ea4\" (UID: \"d7c44420-4173-45fe-bf74-5ed086364ea4\") " Dec 03 15:10:26 crc kubenswrapper[4805]: I1203 15:10:26.721259 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4nq5c\" (UniqueName: \"kubernetes.io/projected/d7c44420-4173-45fe-bf74-5ed086364ea4-kube-api-access-4nq5c\") pod \"d7c44420-4173-45fe-bf74-5ed086364ea4\" (UID: \"d7c44420-4173-45fe-bf74-5ed086364ea4\") " Dec 03 15:10:26 crc kubenswrapper[4805]: I1203 15:10:26.721293 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"prometheus-metric-storage-rulefiles-0\" (UniqueName: \"kubernetes.io/configmap/d7c44420-4173-45fe-bf74-5ed086364ea4-prometheus-metric-storage-rulefiles-0\") pod \"d7c44420-4173-45fe-bf74-5ed086364ea4\" (UID: \"d7c44420-4173-45fe-bf74-5ed086364ea4\") " Dec 03 15:10:26 crc kubenswrapper[4805]: I1203 15:10:26.721339 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/d7c44420-4173-45fe-bf74-5ed086364ea4-tls-assets\") pod \"d7c44420-4173-45fe-bf74-5ed086364ea4\" (UID: \"d7c44420-4173-45fe-bf74-5ed086364ea4\") " Dec 03 15:10:26 crc kubenswrapper[4805]: I1203 15:10:26.721416 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"thanos-prometheus-http-client-file\" (UniqueName: \"kubernetes.io/secret/d7c44420-4173-45fe-bf74-5ed086364ea4-thanos-prometheus-http-client-file\") pod \"d7c44420-4173-45fe-bf74-5ed086364ea4\" (UID: \"d7c44420-4173-45fe-bf74-5ed086364ea4\") " Dec 03 15:10:26 crc kubenswrapper[4805]: I1203 15:10:26.721438 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d7c44420-4173-45fe-bf74-5ed086364ea4-secret-combined-ca-bundle\") pod \"d7c44420-4173-45fe-bf74-5ed086364ea4\" (UID: \"d7c44420-4173-45fe-bf74-5ed086364ea4\") " Dec 03 15:10:26 crc kubenswrapper[4805]: I1203 15:10:26.721582 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"prometheus-metric-storage-db\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"d7c44420-4173-45fe-bf74-5ed086364ea4\" (UID: \"d7c44420-4173-45fe-bf74-5ed086364ea4\") " Dec 03 15:10:26 crc kubenswrapper[4805]: I1203 15:10:26.721622 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"web-config-tls-secret-key-cert-metric-storage-promethe-dc638c2d\" (UniqueName: \"kubernetes.io/secret/d7c44420-4173-45fe-bf74-5ed086364ea4-web-config-tls-secret-key-cert-metric-storage-promethe-dc638c2d\") pod \"d7c44420-4173-45fe-bf74-5ed086364ea4\" (UID: \"d7c44420-4173-45fe-bf74-5ed086364ea4\") " Dec 03 15:10:26 crc kubenswrapper[4805]: I1203 15:10:26.722308 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d7c44420-4173-45fe-bf74-5ed086364ea4-prometheus-metric-storage-rulefiles-0" (OuterVolumeSpecName: "prometheus-metric-storage-rulefiles-0") pod "d7c44420-4173-45fe-bf74-5ed086364ea4" (UID: "d7c44420-4173-45fe-bf74-5ed086364ea4"). InnerVolumeSpecName "prometheus-metric-storage-rulefiles-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 15:10:26 crc kubenswrapper[4805]: I1203 15:10:26.722746 4805 reconciler_common.go:293] "Volume detached for volume \"prometheus-metric-storage-rulefiles-0\" (UniqueName: \"kubernetes.io/configmap/d7c44420-4173-45fe-bf74-5ed086364ea4-prometheus-metric-storage-rulefiles-0\") on node \"crc\" DevicePath \"\"" Dec 03 15:10:26 crc kubenswrapper[4805]: I1203 15:10:26.730345 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d7c44420-4173-45fe-bf74-5ed086364ea4-secret-combined-ca-bundle" (OuterVolumeSpecName: "secret-combined-ca-bundle") pod "d7c44420-4173-45fe-bf74-5ed086364ea4" (UID: "d7c44420-4173-45fe-bf74-5ed086364ea4"). InnerVolumeSpecName "secret-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 15:10:26 crc kubenswrapper[4805]: I1203 15:10:26.730388 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d7c44420-4173-45fe-bf74-5ed086364ea4-config" (OuterVolumeSpecName: "config") pod "d7c44420-4173-45fe-bf74-5ed086364ea4" (UID: "d7c44420-4173-45fe-bf74-5ed086364ea4"). InnerVolumeSpecName "config". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 15:10:26 crc kubenswrapper[4805]: I1203 15:10:26.732050 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d7c44420-4173-45fe-bf74-5ed086364ea4-tls-assets" (OuterVolumeSpecName: "tls-assets") pod "d7c44420-4173-45fe-bf74-5ed086364ea4" (UID: "d7c44420-4173-45fe-bf74-5ed086364ea4"). InnerVolumeSpecName "tls-assets". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 15:10:26 crc kubenswrapper[4805]: I1203 15:10:26.732553 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d7c44420-4173-45fe-bf74-5ed086364ea4-web-config-tls-secret-key-cert-metric-storage-promethe-dc638c2d" (OuterVolumeSpecName: "web-config-tls-secret-key-cert-metric-storage-promethe-dc638c2d") pod "d7c44420-4173-45fe-bf74-5ed086364ea4" (UID: "d7c44420-4173-45fe-bf74-5ed086364ea4"). InnerVolumeSpecName "web-config-tls-secret-key-cert-metric-storage-promethe-dc638c2d". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 15:10:26 crc kubenswrapper[4805]: I1203 15:10:26.733030 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d7c44420-4173-45fe-bf74-5ed086364ea4-kube-api-access-4nq5c" (OuterVolumeSpecName: "kube-api-access-4nq5c") pod "d7c44420-4173-45fe-bf74-5ed086364ea4" (UID: "d7c44420-4173-45fe-bf74-5ed086364ea4"). InnerVolumeSpecName "kube-api-access-4nq5c". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 15:10:26 crc kubenswrapper[4805]: I1203 15:10:26.733172 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d7c44420-4173-45fe-bf74-5ed086364ea4-web-config-tls-secret-cert-cert-metric-storage-prometh-dc638c2d" (OuterVolumeSpecName: "web-config-tls-secret-cert-cert-metric-storage-prometh-dc638c2d") pod "d7c44420-4173-45fe-bf74-5ed086364ea4" (UID: "d7c44420-4173-45fe-bf74-5ed086364ea4"). InnerVolumeSpecName "web-config-tls-secret-cert-cert-metric-storage-prometh-dc638c2d". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 15:10:26 crc kubenswrapper[4805]: I1203 15:10:26.735889 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage12-crc" (OuterVolumeSpecName: "prometheus-metric-storage-db") pod "d7c44420-4173-45fe-bf74-5ed086364ea4" (UID: "d7c44420-4173-45fe-bf74-5ed086364ea4"). InnerVolumeSpecName "local-storage12-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Dec 03 15:10:26 crc kubenswrapper[4805]: I1203 15:10:26.744231 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d7c44420-4173-45fe-bf74-5ed086364ea4-thanos-prometheus-http-client-file" (OuterVolumeSpecName: "thanos-prometheus-http-client-file") pod "d7c44420-4173-45fe-bf74-5ed086364ea4" (UID: "d7c44420-4173-45fe-bf74-5ed086364ea4"). InnerVolumeSpecName "thanos-prometheus-http-client-file". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 15:10:26 crc kubenswrapper[4805]: I1203 15:10:26.751651 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d7c44420-4173-45fe-bf74-5ed086364ea4-config-out" (OuterVolumeSpecName: "config-out") pod "d7c44420-4173-45fe-bf74-5ed086364ea4" (UID: "d7c44420-4173-45fe-bf74-5ed086364ea4"). InnerVolumeSpecName "config-out". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 15:10:26 crc kubenswrapper[4805]: I1203 15:10:26.825078 4805 reconciler_common.go:293] "Volume detached for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/d7c44420-4173-45fe-bf74-5ed086364ea4-config-out\") on node \"crc\" DevicePath \"\"" Dec 03 15:10:26 crc kubenswrapper[4805]: I1203 15:10:26.825120 4805 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/secret/d7c44420-4173-45fe-bf74-5ed086364ea4-config\") on node \"crc\" DevicePath \"\"" Dec 03 15:10:26 crc kubenswrapper[4805]: I1203 15:10:26.825134 4805 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4nq5c\" (UniqueName: \"kubernetes.io/projected/d7c44420-4173-45fe-bf74-5ed086364ea4-kube-api-access-4nq5c\") on node \"crc\" DevicePath \"\"" Dec 03 15:10:26 crc kubenswrapper[4805]: I1203 15:10:26.825149 4805 reconciler_common.go:293] "Volume detached for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/d7c44420-4173-45fe-bf74-5ed086364ea4-tls-assets\") on node \"crc\" DevicePath \"\"" Dec 03 15:10:26 crc kubenswrapper[4805]: I1203 15:10:26.825163 4805 reconciler_common.go:293] "Volume detached for volume \"thanos-prometheus-http-client-file\" (UniqueName: \"kubernetes.io/secret/d7c44420-4173-45fe-bf74-5ed086364ea4-thanos-prometheus-http-client-file\") on node \"crc\" DevicePath \"\"" Dec 03 15:10:26 crc kubenswrapper[4805]: I1203 15:10:26.825176 4805 reconciler_common.go:293] "Volume detached for volume \"secret-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d7c44420-4173-45fe-bf74-5ed086364ea4-secret-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 15:10:26 crc kubenswrapper[4805]: I1203 15:10:26.825203 4805 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") on node \"crc\" " Dec 03 15:10:26 crc kubenswrapper[4805]: I1203 15:10:26.825221 4805 reconciler_common.go:293] "Volume detached for volume \"web-config-tls-secret-key-cert-metric-storage-promethe-dc638c2d\" (UniqueName: \"kubernetes.io/secret/d7c44420-4173-45fe-bf74-5ed086364ea4-web-config-tls-secret-key-cert-metric-storage-promethe-dc638c2d\") on node \"crc\" DevicePath \"\"" Dec 03 15:10:26 crc kubenswrapper[4805]: I1203 15:10:26.825236 4805 reconciler_common.go:293] "Volume detached for volume \"web-config-tls-secret-cert-cert-metric-storage-prometh-dc638c2d\" (UniqueName: \"kubernetes.io/secret/d7c44420-4173-45fe-bf74-5ed086364ea4-web-config-tls-secret-cert-cert-metric-storage-prometh-dc638c2d\") on node \"crc\" DevicePath \"\"" Dec 03 15:10:26 crc kubenswrapper[4805]: I1203 15:10:26.848376 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d7c44420-4173-45fe-bf74-5ed086364ea4-web-config" (OuterVolumeSpecName: "web-config") pod "d7c44420-4173-45fe-bf74-5ed086364ea4" (UID: "d7c44420-4173-45fe-bf74-5ed086364ea4"). InnerVolumeSpecName "web-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 15:10:26 crc kubenswrapper[4805]: I1203 15:10:26.855223 4805 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage12-crc" (UniqueName: "kubernetes.io/local-volume/local-storage12-crc") on node "crc" Dec 03 15:10:26 crc kubenswrapper[4805]: I1203 15:10:26.926213 4805 reconciler_common.go:293] "Volume detached for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/d7c44420-4173-45fe-bf74-5ed086364ea4-web-config\") on node \"crc\" DevicePath \"\"" Dec 03 15:10:26 crc kubenswrapper[4805]: I1203 15:10:26.926249 4805 reconciler_common.go:293] "Volume detached for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") on node \"crc\" DevicePath \"\"" Dec 03 15:10:27 crc kubenswrapper[4805]: I1203 15:10:27.515067 4805 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/prometheus-metric-storage-0" Dec 03 15:10:27 crc kubenswrapper[4805]: I1203 15:10:27.552342 4805 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/prometheus-metric-storage-0"] Dec 03 15:10:27 crc kubenswrapper[4805]: I1203 15:10:27.561564 4805 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/prometheus-metric-storage-0"] Dec 03 15:10:28 crc kubenswrapper[4805]: I1203 15:10:28.373621 4805 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/prometheus-metric-storage-0"] Dec 03 15:10:28 crc kubenswrapper[4805]: E1203 15:10:28.374426 4805 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d7c44420-4173-45fe-bf74-5ed086364ea4" containerName="prometheus" Dec 03 15:10:28 crc kubenswrapper[4805]: I1203 15:10:28.374452 4805 state_mem.go:107] "Deleted CPUSet assignment" podUID="d7c44420-4173-45fe-bf74-5ed086364ea4" containerName="prometheus" Dec 03 15:10:28 crc kubenswrapper[4805]: E1203 15:10:28.374495 4805 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d7c44420-4173-45fe-bf74-5ed086364ea4" containerName="config-reloader" Dec 03 15:10:28 crc kubenswrapper[4805]: I1203 15:10:28.374504 4805 state_mem.go:107] "Deleted CPUSet assignment" podUID="d7c44420-4173-45fe-bf74-5ed086364ea4" containerName="config-reloader" Dec 03 15:10:28 crc kubenswrapper[4805]: E1203 15:10:28.374518 4805 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d7c44420-4173-45fe-bf74-5ed086364ea4" containerName="init-config-reloader" Dec 03 15:10:28 crc kubenswrapper[4805]: I1203 15:10:28.374527 4805 state_mem.go:107] "Deleted CPUSet assignment" podUID="d7c44420-4173-45fe-bf74-5ed086364ea4" containerName="init-config-reloader" Dec 03 15:10:28 crc kubenswrapper[4805]: E1203 15:10:28.374546 4805 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d7c44420-4173-45fe-bf74-5ed086364ea4" containerName="thanos-sidecar" Dec 03 15:10:28 crc kubenswrapper[4805]: I1203 15:10:28.374554 4805 state_mem.go:107] "Deleted CPUSet assignment" podUID="d7c44420-4173-45fe-bf74-5ed086364ea4" containerName="thanos-sidecar" Dec 03 15:10:28 crc kubenswrapper[4805]: I1203 15:10:28.374795 4805 memory_manager.go:354] "RemoveStaleState removing state" podUID="d7c44420-4173-45fe-bf74-5ed086364ea4" containerName="prometheus" Dec 03 15:10:28 crc kubenswrapper[4805]: I1203 15:10:28.374822 4805 memory_manager.go:354] "RemoveStaleState removing state" podUID="d7c44420-4173-45fe-bf74-5ed086364ea4" containerName="config-reloader" Dec 03 15:10:28 crc kubenswrapper[4805]: I1203 15:10:28.375810 4805 memory_manager.go:354] "RemoveStaleState removing state" podUID="d7c44420-4173-45fe-bf74-5ed086364ea4" containerName="thanos-sidecar" Dec 03 15:10:28 crc kubenswrapper[4805]: I1203 15:10:28.378425 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/prometheus-metric-storage-0" Dec 03 15:10:28 crc kubenswrapper[4805]: I1203 15:10:28.380686 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"prometheus-metric-storage" Dec 03 15:10:28 crc kubenswrapper[4805]: I1203 15:10:28.380927 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"prometheus-metric-storage-web-config" Dec 03 15:10:28 crc kubenswrapper[4805]: I1203 15:10:28.381881 4805 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"prometheus-metric-storage-rulefiles-0" Dec 03 15:10:28 crc kubenswrapper[4805]: I1203 15:10:28.382999 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"metric-storage-prometheus-dockercfg-kg4lr" Dec 03 15:10:28 crc kubenswrapper[4805]: I1203 15:10:28.383007 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-metric-storage-prometheus-svc" Dec 03 15:10:28 crc kubenswrapper[4805]: I1203 15:10:28.383535 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"prometheus-metric-storage-thanos-prometheus-http-client-file" Dec 03 15:10:28 crc kubenswrapper[4805]: I1203 15:10:28.390536 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"prometheus-metric-storage-tls-assets-0" Dec 03 15:10:28 crc kubenswrapper[4805]: I1203 15:10:28.400138 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/prometheus-metric-storage-0"] Dec 03 15:10:28 crc kubenswrapper[4805]: I1203 15:10:28.453416 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"web-config-tls-secret-key-cert-metric-storage-promethe-dc638c2d\" (UniqueName: \"kubernetes.io/secret/9e35b852-1838-49da-a140-f048788f6635-web-config-tls-secret-key-cert-metric-storage-promethe-dc638c2d\") pod \"prometheus-metric-storage-0\" (UID: \"9e35b852-1838-49da-a140-f048788f6635\") " pod="openstack/prometheus-metric-storage-0" Dec 03 15:10:28 crc kubenswrapper[4805]: I1203 15:10:28.453632 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"thanos-prometheus-http-client-file\" (UniqueName: \"kubernetes.io/secret/9e35b852-1838-49da-a140-f048788f6635-thanos-prometheus-http-client-file\") pod \"prometheus-metric-storage-0\" (UID: \"9e35b852-1838-49da-a140-f048788f6635\") " pod="openstack/prometheus-metric-storage-0" Dec 03 15:10:28 crc kubenswrapper[4805]: I1203 15:10:28.453702 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/9e35b852-1838-49da-a140-f048788f6635-config-out\") pod \"prometheus-metric-storage-0\" (UID: \"9e35b852-1838-49da-a140-f048788f6635\") " pod="openstack/prometheus-metric-storage-0" Dec 03 15:10:28 crc kubenswrapper[4805]: I1203 15:10:28.453763 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"prometheus-metric-storage-db\" (UniqueName: \"kubernetes.io/empty-dir/9e35b852-1838-49da-a140-f048788f6635-prometheus-metric-storage-db\") pod \"prometheus-metric-storage-0\" (UID: \"9e35b852-1838-49da-a140-f048788f6635\") " pod="openstack/prometheus-metric-storage-0" Dec 03 15:10:28 crc kubenswrapper[4805]: I1203 15:10:28.453957 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/9e35b852-1838-49da-a140-f048788f6635-config\") pod \"prometheus-metric-storage-0\" (UID: \"9e35b852-1838-49da-a140-f048788f6635\") " pod="openstack/prometheus-metric-storage-0" Dec 03 15:10:28 crc kubenswrapper[4805]: I1203 15:10:28.454075 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-z2qxg\" (UniqueName: \"kubernetes.io/projected/9e35b852-1838-49da-a140-f048788f6635-kube-api-access-z2qxg\") pod \"prometheus-metric-storage-0\" (UID: \"9e35b852-1838-49da-a140-f048788f6635\") " pod="openstack/prometheus-metric-storage-0" Dec 03 15:10:28 crc kubenswrapper[4805]: I1203 15:10:28.454135 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/9e35b852-1838-49da-a140-f048788f6635-web-config\") pod \"prometheus-metric-storage-0\" (UID: \"9e35b852-1838-49da-a140-f048788f6635\") " pod="openstack/prometheus-metric-storage-0" Dec 03 15:10:28 crc kubenswrapper[4805]: I1203 15:10:28.454159 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"prometheus-metric-storage-rulefiles-0\" (UniqueName: \"kubernetes.io/configmap/9e35b852-1838-49da-a140-f048788f6635-prometheus-metric-storage-rulefiles-0\") pod \"prometheus-metric-storage-0\" (UID: \"9e35b852-1838-49da-a140-f048788f6635\") " pod="openstack/prometheus-metric-storage-0" Dec 03 15:10:28 crc kubenswrapper[4805]: I1203 15:10:28.454232 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9e35b852-1838-49da-a140-f048788f6635-secret-combined-ca-bundle\") pod \"prometheus-metric-storage-0\" (UID: \"9e35b852-1838-49da-a140-f048788f6635\") " pod="openstack/prometheus-metric-storage-0" Dec 03 15:10:28 crc kubenswrapper[4805]: I1203 15:10:28.454508 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/9e35b852-1838-49da-a140-f048788f6635-tls-assets\") pod \"prometheus-metric-storage-0\" (UID: \"9e35b852-1838-49da-a140-f048788f6635\") " pod="openstack/prometheus-metric-storage-0" Dec 03 15:10:28 crc kubenswrapper[4805]: I1203 15:10:28.454631 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"web-config-tls-secret-cert-cert-metric-storage-prometh-dc638c2d\" (UniqueName: \"kubernetes.io/secret/9e35b852-1838-49da-a140-f048788f6635-web-config-tls-secret-cert-cert-metric-storage-prometh-dc638c2d\") pod \"prometheus-metric-storage-0\" (UID: \"9e35b852-1838-49da-a140-f048788f6635\") " pod="openstack/prometheus-metric-storage-0" Dec 03 15:10:28 crc kubenswrapper[4805]: I1203 15:10:28.556245 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-z2qxg\" (UniqueName: \"kubernetes.io/projected/9e35b852-1838-49da-a140-f048788f6635-kube-api-access-z2qxg\") pod \"prometheus-metric-storage-0\" (UID: \"9e35b852-1838-49da-a140-f048788f6635\") " pod="openstack/prometheus-metric-storage-0" Dec 03 15:10:28 crc kubenswrapper[4805]: I1203 15:10:28.556653 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/9e35b852-1838-49da-a140-f048788f6635-web-config\") pod \"prometheus-metric-storage-0\" (UID: \"9e35b852-1838-49da-a140-f048788f6635\") " pod="openstack/prometheus-metric-storage-0" Dec 03 15:10:28 crc kubenswrapper[4805]: I1203 15:10:28.556679 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"prometheus-metric-storage-rulefiles-0\" (UniqueName: \"kubernetes.io/configmap/9e35b852-1838-49da-a140-f048788f6635-prometheus-metric-storage-rulefiles-0\") pod \"prometheus-metric-storage-0\" (UID: \"9e35b852-1838-49da-a140-f048788f6635\") " pod="openstack/prometheus-metric-storage-0" Dec 03 15:10:28 crc kubenswrapper[4805]: I1203 15:10:28.557469 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9e35b852-1838-49da-a140-f048788f6635-secret-combined-ca-bundle\") pod \"prometheus-metric-storage-0\" (UID: \"9e35b852-1838-49da-a140-f048788f6635\") " pod="openstack/prometheus-metric-storage-0" Dec 03 15:10:28 crc kubenswrapper[4805]: I1203 15:10:28.557533 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/9e35b852-1838-49da-a140-f048788f6635-tls-assets\") pod \"prometheus-metric-storage-0\" (UID: \"9e35b852-1838-49da-a140-f048788f6635\") " pod="openstack/prometheus-metric-storage-0" Dec 03 15:10:28 crc kubenswrapper[4805]: I1203 15:10:28.557581 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"web-config-tls-secret-cert-cert-metric-storage-prometh-dc638c2d\" (UniqueName: \"kubernetes.io/secret/9e35b852-1838-49da-a140-f048788f6635-web-config-tls-secret-cert-cert-metric-storage-prometh-dc638c2d\") pod \"prometheus-metric-storage-0\" (UID: \"9e35b852-1838-49da-a140-f048788f6635\") " pod="openstack/prometheus-metric-storage-0" Dec 03 15:10:28 crc kubenswrapper[4805]: I1203 15:10:28.557532 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"prometheus-metric-storage-rulefiles-0\" (UniqueName: \"kubernetes.io/configmap/9e35b852-1838-49da-a140-f048788f6635-prometheus-metric-storage-rulefiles-0\") pod \"prometheus-metric-storage-0\" (UID: \"9e35b852-1838-49da-a140-f048788f6635\") " pod="openstack/prometheus-metric-storage-0" Dec 03 15:10:28 crc kubenswrapper[4805]: I1203 15:10:28.557672 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"web-config-tls-secret-key-cert-metric-storage-promethe-dc638c2d\" (UniqueName: \"kubernetes.io/secret/9e35b852-1838-49da-a140-f048788f6635-web-config-tls-secret-key-cert-metric-storage-promethe-dc638c2d\") pod \"prometheus-metric-storage-0\" (UID: \"9e35b852-1838-49da-a140-f048788f6635\") " pod="openstack/prometheus-metric-storage-0" Dec 03 15:10:28 crc kubenswrapper[4805]: I1203 15:10:28.557796 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"thanos-prometheus-http-client-file\" (UniqueName: \"kubernetes.io/secret/9e35b852-1838-49da-a140-f048788f6635-thanos-prometheus-http-client-file\") pod \"prometheus-metric-storage-0\" (UID: \"9e35b852-1838-49da-a140-f048788f6635\") " pod="openstack/prometheus-metric-storage-0" Dec 03 15:10:28 crc kubenswrapper[4805]: I1203 15:10:28.557836 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/9e35b852-1838-49da-a140-f048788f6635-config-out\") pod \"prometheus-metric-storage-0\" (UID: \"9e35b852-1838-49da-a140-f048788f6635\") " pod="openstack/prometheus-metric-storage-0" Dec 03 15:10:28 crc kubenswrapper[4805]: I1203 15:10:28.557972 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"prometheus-metric-storage-db\" (UniqueName: \"kubernetes.io/empty-dir/9e35b852-1838-49da-a140-f048788f6635-prometheus-metric-storage-db\") pod \"prometheus-metric-storage-0\" (UID: \"9e35b852-1838-49da-a140-f048788f6635\") " pod="openstack/prometheus-metric-storage-0" Dec 03 15:10:28 crc kubenswrapper[4805]: I1203 15:10:28.558059 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/9e35b852-1838-49da-a140-f048788f6635-config\") pod \"prometheus-metric-storage-0\" (UID: \"9e35b852-1838-49da-a140-f048788f6635\") " pod="openstack/prometheus-metric-storage-0" Dec 03 15:10:28 crc kubenswrapper[4805]: I1203 15:10:28.561001 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"prometheus-metric-storage-db\" (UniqueName: \"kubernetes.io/empty-dir/9e35b852-1838-49da-a140-f048788f6635-prometheus-metric-storage-db\") pod \"prometheus-metric-storage-0\" (UID: \"9e35b852-1838-49da-a140-f048788f6635\") " pod="openstack/prometheus-metric-storage-0" Dec 03 15:10:28 crc kubenswrapper[4805]: I1203 15:10:28.561994 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"thanos-prometheus-http-client-file\" (UniqueName: \"kubernetes.io/secret/9e35b852-1838-49da-a140-f048788f6635-thanos-prometheus-http-client-file\") pod \"prometheus-metric-storage-0\" (UID: \"9e35b852-1838-49da-a140-f048788f6635\") " pod="openstack/prometheus-metric-storage-0" Dec 03 15:10:28 crc kubenswrapper[4805]: I1203 15:10:28.563609 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"web-config-tls-secret-key-cert-metric-storage-promethe-dc638c2d\" (UniqueName: \"kubernetes.io/secret/9e35b852-1838-49da-a140-f048788f6635-web-config-tls-secret-key-cert-metric-storage-promethe-dc638c2d\") pod \"prometheus-metric-storage-0\" (UID: \"9e35b852-1838-49da-a140-f048788f6635\") " pod="openstack/prometheus-metric-storage-0" Dec 03 15:10:28 crc kubenswrapper[4805]: I1203 15:10:28.564398 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"web-config-tls-secret-cert-cert-metric-storage-prometh-dc638c2d\" (UniqueName: \"kubernetes.io/secret/9e35b852-1838-49da-a140-f048788f6635-web-config-tls-secret-cert-cert-metric-storage-prometh-dc638c2d\") pod \"prometheus-metric-storage-0\" (UID: \"9e35b852-1838-49da-a140-f048788f6635\") " pod="openstack/prometheus-metric-storage-0" Dec 03 15:10:28 crc kubenswrapper[4805]: I1203 15:10:28.564779 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/9e35b852-1838-49da-a140-f048788f6635-tls-assets\") pod \"prometheus-metric-storage-0\" (UID: \"9e35b852-1838-49da-a140-f048788f6635\") " pod="openstack/prometheus-metric-storage-0" Dec 03 15:10:28 crc kubenswrapper[4805]: I1203 15:10:28.565468 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/9e35b852-1838-49da-a140-f048788f6635-config-out\") pod \"prometheus-metric-storage-0\" (UID: \"9e35b852-1838-49da-a140-f048788f6635\") " pod="openstack/prometheus-metric-storage-0" Dec 03 15:10:28 crc kubenswrapper[4805]: I1203 15:10:28.565812 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/secret/9e35b852-1838-49da-a140-f048788f6635-config\") pod \"prometheus-metric-storage-0\" (UID: \"9e35b852-1838-49da-a140-f048788f6635\") " pod="openstack/prometheus-metric-storage-0" Dec 03 15:10:28 crc kubenswrapper[4805]: I1203 15:10:28.565897 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/9e35b852-1838-49da-a140-f048788f6635-web-config\") pod \"prometheus-metric-storage-0\" (UID: \"9e35b852-1838-49da-a140-f048788f6635\") " pod="openstack/prometheus-metric-storage-0" Dec 03 15:10:28 crc kubenswrapper[4805]: I1203 15:10:28.566630 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9e35b852-1838-49da-a140-f048788f6635-secret-combined-ca-bundle\") pod \"prometheus-metric-storage-0\" (UID: \"9e35b852-1838-49da-a140-f048788f6635\") " pod="openstack/prometheus-metric-storage-0" Dec 03 15:10:28 crc kubenswrapper[4805]: I1203 15:10:28.575388 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-z2qxg\" (UniqueName: \"kubernetes.io/projected/9e35b852-1838-49da-a140-f048788f6635-kube-api-access-z2qxg\") pod \"prometheus-metric-storage-0\" (UID: \"9e35b852-1838-49da-a140-f048788f6635\") " pod="openstack/prometheus-metric-storage-0" Dec 03 15:10:28 crc kubenswrapper[4805]: I1203 15:10:28.702087 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/prometheus-metric-storage-0" Dec 03 15:10:28 crc kubenswrapper[4805]: I1203 15:10:28.729682 4805 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d7c44420-4173-45fe-bf74-5ed086364ea4" path="/var/lib/kubelet/pods/d7c44420-4173-45fe-bf74-5ed086364ea4/volumes" Dec 03 15:10:29 crc kubenswrapper[4805]: I1203 15:10:29.238990 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/prometheus-metric-storage-0"] Dec 03 15:10:29 crc kubenswrapper[4805]: W1203 15:10:29.246380 4805 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod9e35b852_1838_49da_a140_f048788f6635.slice/crio-111dbdd1292248118175356f2124f21600ff2eb246fe046e9d8fa3e52f21587c WatchSource:0}: Error finding container 111dbdd1292248118175356f2124f21600ff2eb246fe046e9d8fa3e52f21587c: Status 404 returned error can't find the container with id 111dbdd1292248118175356f2124f21600ff2eb246fe046e9d8fa3e52f21587c Dec 03 15:10:29 crc kubenswrapper[4805]: I1203 15:10:29.534639 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"9e35b852-1838-49da-a140-f048788f6635","Type":"ContainerStarted","Data":"111dbdd1292248118175356f2124f21600ff2eb246fe046e9d8fa3e52f21587c"} Dec 03 15:10:33 crc kubenswrapper[4805]: I1203 15:10:33.580131 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"9e35b852-1838-49da-a140-f048788f6635","Type":"ContainerStarted","Data":"e4497f355686372593cee84b0a311f2e56fa24ad7f3aed9b51b7bbfbb220ce7c"} Dec 03 15:10:39 crc kubenswrapper[4805]: I1203 15:10:39.653168 4805 generic.go:334] "Generic (PLEG): container finished" podID="9e35b852-1838-49da-a140-f048788f6635" containerID="e4497f355686372593cee84b0a311f2e56fa24ad7f3aed9b51b7bbfbb220ce7c" exitCode=0 Dec 03 15:10:39 crc kubenswrapper[4805]: I1203 15:10:39.653365 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"9e35b852-1838-49da-a140-f048788f6635","Type":"ContainerDied","Data":"e4497f355686372593cee84b0a311f2e56fa24ad7f3aed9b51b7bbfbb220ce7c"} Dec 03 15:10:40 crc kubenswrapper[4805]: I1203 15:10:40.667106 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"9e35b852-1838-49da-a140-f048788f6635","Type":"ContainerStarted","Data":"3f32476151f16cefda9576a8077a18738591448ef92afdba7d6ba1cfe707d756"} Dec 03 15:10:44 crc kubenswrapper[4805]: I1203 15:10:44.712362 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"9e35b852-1838-49da-a140-f048788f6635","Type":"ContainerStarted","Data":"c6738ea6a72122506d9eaaacc460bb54d669b6aa4a8ffdb2c2a9d8829326e697"} Dec 03 15:10:45 crc kubenswrapper[4805]: I1203 15:10:45.728758 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"9e35b852-1838-49da-a140-f048788f6635","Type":"ContainerStarted","Data":"a023a91b7c3b39787116d2c8c1a0da144673bf6d87052338e976a9af4bcc7df5"} Dec 03 15:10:45 crc kubenswrapper[4805]: I1203 15:10:45.762814 4805 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/prometheus-metric-storage-0" podStartSLOduration=17.762789177 podStartE2EDuration="17.762789177s" podCreationTimestamp="2025-12-03 15:10:28 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 15:10:45.757495627 +0000 UTC m=+3675.420412570" watchObservedRunningTime="2025-12-03 15:10:45.762789177 +0000 UTC m=+3675.425706110" Dec 03 15:10:48 crc kubenswrapper[4805]: I1203 15:10:48.706575 4805 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/prometheus-metric-storage-0" Dec 03 15:10:58 crc kubenswrapper[4805]: I1203 15:10:58.709572 4805 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/prometheus-metric-storage-0" Dec 03 15:10:58 crc kubenswrapper[4805]: I1203 15:10:58.712533 4805 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/prometheus-metric-storage-0" Dec 03 15:10:58 crc kubenswrapper[4805]: I1203 15:10:58.862154 4805 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/prometheus-metric-storage-0" Dec 03 15:11:24 crc kubenswrapper[4805]: I1203 15:11:24.853908 4805 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-vmvlx"] Dec 03 15:11:24 crc kubenswrapper[4805]: I1203 15:11:24.857072 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-vmvlx" Dec 03 15:11:24 crc kubenswrapper[4805]: I1203 15:11:24.869410 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-vmvlx"] Dec 03 15:11:25 crc kubenswrapper[4805]: I1203 15:11:25.049130 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/46870522-a59e-4525-b5fd-452bec4d362b-catalog-content\") pod \"redhat-operators-vmvlx\" (UID: \"46870522-a59e-4525-b5fd-452bec4d362b\") " pod="openshift-marketplace/redhat-operators-vmvlx" Dec 03 15:11:25 crc kubenswrapper[4805]: I1203 15:11:25.049191 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/46870522-a59e-4525-b5fd-452bec4d362b-utilities\") pod \"redhat-operators-vmvlx\" (UID: \"46870522-a59e-4525-b5fd-452bec4d362b\") " pod="openshift-marketplace/redhat-operators-vmvlx" Dec 03 15:11:25 crc kubenswrapper[4805]: I1203 15:11:25.049397 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7n79z\" (UniqueName: \"kubernetes.io/projected/46870522-a59e-4525-b5fd-452bec4d362b-kube-api-access-7n79z\") pod \"redhat-operators-vmvlx\" (UID: \"46870522-a59e-4525-b5fd-452bec4d362b\") " pod="openshift-marketplace/redhat-operators-vmvlx" Dec 03 15:11:25 crc kubenswrapper[4805]: I1203 15:11:25.151485 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/46870522-a59e-4525-b5fd-452bec4d362b-catalog-content\") pod \"redhat-operators-vmvlx\" (UID: \"46870522-a59e-4525-b5fd-452bec4d362b\") " pod="openshift-marketplace/redhat-operators-vmvlx" Dec 03 15:11:25 crc kubenswrapper[4805]: I1203 15:11:25.152113 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/46870522-a59e-4525-b5fd-452bec4d362b-utilities\") pod \"redhat-operators-vmvlx\" (UID: \"46870522-a59e-4525-b5fd-452bec4d362b\") " pod="openshift-marketplace/redhat-operators-vmvlx" Dec 03 15:11:25 crc kubenswrapper[4805]: I1203 15:11:25.152125 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/46870522-a59e-4525-b5fd-452bec4d362b-catalog-content\") pod \"redhat-operators-vmvlx\" (UID: \"46870522-a59e-4525-b5fd-452bec4d362b\") " pod="openshift-marketplace/redhat-operators-vmvlx" Dec 03 15:11:25 crc kubenswrapper[4805]: I1203 15:11:25.152275 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7n79z\" (UniqueName: \"kubernetes.io/projected/46870522-a59e-4525-b5fd-452bec4d362b-kube-api-access-7n79z\") pod \"redhat-operators-vmvlx\" (UID: \"46870522-a59e-4525-b5fd-452bec4d362b\") " pod="openshift-marketplace/redhat-operators-vmvlx" Dec 03 15:11:25 crc kubenswrapper[4805]: I1203 15:11:25.152377 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/46870522-a59e-4525-b5fd-452bec4d362b-utilities\") pod \"redhat-operators-vmvlx\" (UID: \"46870522-a59e-4525-b5fd-452bec4d362b\") " pod="openshift-marketplace/redhat-operators-vmvlx" Dec 03 15:11:25 crc kubenswrapper[4805]: I1203 15:11:25.178751 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7n79z\" (UniqueName: \"kubernetes.io/projected/46870522-a59e-4525-b5fd-452bec4d362b-kube-api-access-7n79z\") pod \"redhat-operators-vmvlx\" (UID: \"46870522-a59e-4525-b5fd-452bec4d362b\") " pod="openshift-marketplace/redhat-operators-vmvlx" Dec 03 15:11:25 crc kubenswrapper[4805]: I1203 15:11:25.187672 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-vmvlx" Dec 03 15:11:25 crc kubenswrapper[4805]: I1203 15:11:25.761491 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-vmvlx"] Dec 03 15:11:26 crc kubenswrapper[4805]: I1203 15:11:26.136576 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-vmvlx" event={"ID":"46870522-a59e-4525-b5fd-452bec4d362b","Type":"ContainerStarted","Data":"1135817e28cc496283cd78507506d99e61d52c17f113524d13e4819dc7b81b6c"} Dec 03 15:11:26 crc kubenswrapper[4805]: I1203 15:11:26.137942 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-vmvlx" event={"ID":"46870522-a59e-4525-b5fd-452bec4d362b","Type":"ContainerStarted","Data":"fb4da7ea66f8b58412baafe3ea6ac1723ccd899a3fc335ec20faad39b8eb2b37"} Dec 03 15:11:26 crc kubenswrapper[4805]: I1203 15:11:26.138808 4805 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 03 15:11:26 crc kubenswrapper[4805]: E1203 15:11:26.363364 4805 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod46870522_a59e_4525_b5fd_452bec4d362b.slice/crio-conmon-1135817e28cc496283cd78507506d99e61d52c17f113524d13e4819dc7b81b6c.scope\": RecentStats: unable to find data in memory cache]" Dec 03 15:11:27 crc kubenswrapper[4805]: I1203 15:11:27.149403 4805 generic.go:334] "Generic (PLEG): container finished" podID="46870522-a59e-4525-b5fd-452bec4d362b" containerID="1135817e28cc496283cd78507506d99e61d52c17f113524d13e4819dc7b81b6c" exitCode=0 Dec 03 15:11:27 crc kubenswrapper[4805]: I1203 15:11:27.149508 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-vmvlx" event={"ID":"46870522-a59e-4525-b5fd-452bec4d362b","Type":"ContainerDied","Data":"1135817e28cc496283cd78507506d99e61d52c17f113524d13e4819dc7b81b6c"} Dec 03 15:11:28 crc kubenswrapper[4805]: I1203 15:11:28.161479 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-vmvlx" event={"ID":"46870522-a59e-4525-b5fd-452bec4d362b","Type":"ContainerStarted","Data":"77c28a5291f1dd9620171541a00ba97f126d9846fbd141605e590afdadc110b3"} Dec 03 15:11:33 crc kubenswrapper[4805]: I1203 15:11:33.061927 4805 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/aodh-624d-account-create-update-xd9kw"] Dec 03 15:11:33 crc kubenswrapper[4805]: I1203 15:11:33.074386 4805 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/aodh-db-create-9xnk5"] Dec 03 15:11:33 crc kubenswrapper[4805]: I1203 15:11:33.083037 4805 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/aodh-624d-account-create-update-xd9kw"] Dec 03 15:11:33 crc kubenswrapper[4805]: I1203 15:11:33.095756 4805 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/aodh-db-create-9xnk5"] Dec 03 15:11:34 crc kubenswrapper[4805]: I1203 15:11:34.709305 4805 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c3131917-28b3-4dd5-a575-7517f266b36c" path="/var/lib/kubelet/pods/c3131917-28b3-4dd5-a575-7517f266b36c/volumes" Dec 03 15:11:34 crc kubenswrapper[4805]: I1203 15:11:34.713031 4805 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ff3965d5-1e07-45d0-acb0-f1f25716821f" path="/var/lib/kubelet/pods/ff3965d5-1e07-45d0-acb0-f1f25716821f/volumes" Dec 03 15:11:37 crc kubenswrapper[4805]: I1203 15:11:37.250393 4805 generic.go:334] "Generic (PLEG): container finished" podID="46870522-a59e-4525-b5fd-452bec4d362b" containerID="77c28a5291f1dd9620171541a00ba97f126d9846fbd141605e590afdadc110b3" exitCode=0 Dec 03 15:11:37 crc kubenswrapper[4805]: I1203 15:11:37.250442 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-vmvlx" event={"ID":"46870522-a59e-4525-b5fd-452bec4d362b","Type":"ContainerDied","Data":"77c28a5291f1dd9620171541a00ba97f126d9846fbd141605e590afdadc110b3"} Dec 03 15:11:39 crc kubenswrapper[4805]: I1203 15:11:39.269300 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-vmvlx" event={"ID":"46870522-a59e-4525-b5fd-452bec4d362b","Type":"ContainerStarted","Data":"bc24751412043860962da494df8125f35012c09ca0121fb537567c9a2fa73eb2"} Dec 03 15:11:39 crc kubenswrapper[4805]: I1203 15:11:39.288380 4805 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-vmvlx" podStartSLOduration=3.182635788 podStartE2EDuration="15.288360461s" podCreationTimestamp="2025-12-03 15:11:24 +0000 UTC" firstStartedPulling="2025-12-03 15:11:26.1385483 +0000 UTC m=+3715.801465223" lastFinishedPulling="2025-12-03 15:11:38.244272973 +0000 UTC m=+3727.907189896" observedRunningTime="2025-12-03 15:11:39.284399906 +0000 UTC m=+3728.947316829" watchObservedRunningTime="2025-12-03 15:11:39.288360461 +0000 UTC m=+3728.951277384" Dec 03 15:11:45 crc kubenswrapper[4805]: I1203 15:11:45.037256 4805 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/aodh-db-sync-znm4q"] Dec 03 15:11:45 crc kubenswrapper[4805]: I1203 15:11:45.049855 4805 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/aodh-db-sync-znm4q"] Dec 03 15:11:45 crc kubenswrapper[4805]: I1203 15:11:45.188815 4805 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-vmvlx" Dec 03 15:11:45 crc kubenswrapper[4805]: I1203 15:11:45.189233 4805 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-vmvlx" Dec 03 15:11:45 crc kubenswrapper[4805]: I1203 15:11:45.892864 4805 scope.go:117] "RemoveContainer" containerID="8265f7bd78602bdcbae0a347221e8ffbb0b0c660a3740f8759fbec420e1b17eb" Dec 03 15:11:45 crc kubenswrapper[4805]: I1203 15:11:45.935565 4805 scope.go:117] "RemoveContainer" containerID="802e74b1d3b9b5ea1ce841167f986d4201dcce31e8b8090f7bee58a18f2709b9" Dec 03 15:11:46 crc kubenswrapper[4805]: I1203 15:11:46.239634 4805 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-vmvlx" podUID="46870522-a59e-4525-b5fd-452bec4d362b" containerName="registry-server" probeResult="failure" output=< Dec 03 15:11:46 crc kubenswrapper[4805]: timeout: failed to connect service ":50051" within 1s Dec 03 15:11:46 crc kubenswrapper[4805]: > Dec 03 15:11:46 crc kubenswrapper[4805]: I1203 15:11:46.707633 4805 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="07c5b863-2051-46da-b63b-87210297c6cb" path="/var/lib/kubelet/pods/07c5b863-2051-46da-b63b-87210297c6cb/volumes" Dec 03 15:11:55 crc kubenswrapper[4805]: I1203 15:11:55.477276 4805 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-vmvlx" Dec 03 15:11:55 crc kubenswrapper[4805]: I1203 15:11:55.549342 4805 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-vmvlx" Dec 03 15:11:56 crc kubenswrapper[4805]: I1203 15:11:56.052727 4805 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-vmvlx"] Dec 03 15:11:57 crc kubenswrapper[4805]: I1203 15:11:57.449306 4805 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-vmvlx" podUID="46870522-a59e-4525-b5fd-452bec4d362b" containerName="registry-server" containerID="cri-o://bc24751412043860962da494df8125f35012c09ca0121fb537567c9a2fa73eb2" gracePeriod=2 Dec 03 15:11:58 crc kubenswrapper[4805]: I1203 15:11:58.475001 4805 generic.go:334] "Generic (PLEG): container finished" podID="46870522-a59e-4525-b5fd-452bec4d362b" containerID="bc24751412043860962da494df8125f35012c09ca0121fb537567c9a2fa73eb2" exitCode=0 Dec 03 15:11:58 crc kubenswrapper[4805]: I1203 15:11:58.475104 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-vmvlx" event={"ID":"46870522-a59e-4525-b5fd-452bec4d362b","Type":"ContainerDied","Data":"bc24751412043860962da494df8125f35012c09ca0121fb537567c9a2fa73eb2"} Dec 03 15:11:59 crc kubenswrapper[4805]: I1203 15:11:59.486931 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-vmvlx" event={"ID":"46870522-a59e-4525-b5fd-452bec4d362b","Type":"ContainerDied","Data":"fb4da7ea66f8b58412baafe3ea6ac1723ccd899a3fc335ec20faad39b8eb2b37"} Dec 03 15:11:59 crc kubenswrapper[4805]: I1203 15:11:59.487199 4805 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="fb4da7ea66f8b58412baafe3ea6ac1723ccd899a3fc335ec20faad39b8eb2b37" Dec 03 15:11:59 crc kubenswrapper[4805]: I1203 15:11:59.569016 4805 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-vmvlx" Dec 03 15:11:59 crc kubenswrapper[4805]: I1203 15:11:59.616793 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/46870522-a59e-4525-b5fd-452bec4d362b-utilities\") pod \"46870522-a59e-4525-b5fd-452bec4d362b\" (UID: \"46870522-a59e-4525-b5fd-452bec4d362b\") " Dec 03 15:11:59 crc kubenswrapper[4805]: I1203 15:11:59.616983 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/46870522-a59e-4525-b5fd-452bec4d362b-catalog-content\") pod \"46870522-a59e-4525-b5fd-452bec4d362b\" (UID: \"46870522-a59e-4525-b5fd-452bec4d362b\") " Dec 03 15:11:59 crc kubenswrapper[4805]: I1203 15:11:59.617072 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7n79z\" (UniqueName: \"kubernetes.io/projected/46870522-a59e-4525-b5fd-452bec4d362b-kube-api-access-7n79z\") pod \"46870522-a59e-4525-b5fd-452bec4d362b\" (UID: \"46870522-a59e-4525-b5fd-452bec4d362b\") " Dec 03 15:11:59 crc kubenswrapper[4805]: I1203 15:11:59.617793 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/46870522-a59e-4525-b5fd-452bec4d362b-utilities" (OuterVolumeSpecName: "utilities") pod "46870522-a59e-4525-b5fd-452bec4d362b" (UID: "46870522-a59e-4525-b5fd-452bec4d362b"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 15:11:59 crc kubenswrapper[4805]: I1203 15:11:59.623099 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/46870522-a59e-4525-b5fd-452bec4d362b-kube-api-access-7n79z" (OuterVolumeSpecName: "kube-api-access-7n79z") pod "46870522-a59e-4525-b5fd-452bec4d362b" (UID: "46870522-a59e-4525-b5fd-452bec4d362b"). InnerVolumeSpecName "kube-api-access-7n79z". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 15:11:59 crc kubenswrapper[4805]: I1203 15:11:59.719302 4805 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/46870522-a59e-4525-b5fd-452bec4d362b-utilities\") on node \"crc\" DevicePath \"\"" Dec 03 15:11:59 crc kubenswrapper[4805]: I1203 15:11:59.719339 4805 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7n79z\" (UniqueName: \"kubernetes.io/projected/46870522-a59e-4525-b5fd-452bec4d362b-kube-api-access-7n79z\") on node \"crc\" DevicePath \"\"" Dec 03 15:11:59 crc kubenswrapper[4805]: I1203 15:11:59.726650 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/46870522-a59e-4525-b5fd-452bec4d362b-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "46870522-a59e-4525-b5fd-452bec4d362b" (UID: "46870522-a59e-4525-b5fd-452bec4d362b"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 15:11:59 crc kubenswrapper[4805]: I1203 15:11:59.821631 4805 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/46870522-a59e-4525-b5fd-452bec4d362b-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 03 15:12:00 crc kubenswrapper[4805]: I1203 15:12:00.497416 4805 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-vmvlx" Dec 03 15:12:00 crc kubenswrapper[4805]: I1203 15:12:00.547520 4805 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-vmvlx"] Dec 03 15:12:00 crc kubenswrapper[4805]: I1203 15:12:00.567468 4805 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-vmvlx"] Dec 03 15:12:00 crc kubenswrapper[4805]: I1203 15:12:00.706772 4805 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="46870522-a59e-4525-b5fd-452bec4d362b" path="/var/lib/kubelet/pods/46870522-a59e-4525-b5fd-452bec4d362b/volumes" Dec 03 15:12:16 crc kubenswrapper[4805]: I1203 15:12:16.060998 4805 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-5w2gr"] Dec 03 15:12:16 crc kubenswrapper[4805]: E1203 15:12:16.061917 4805 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="46870522-a59e-4525-b5fd-452bec4d362b" containerName="extract-utilities" Dec 03 15:12:16 crc kubenswrapper[4805]: I1203 15:12:16.061933 4805 state_mem.go:107] "Deleted CPUSet assignment" podUID="46870522-a59e-4525-b5fd-452bec4d362b" containerName="extract-utilities" Dec 03 15:12:16 crc kubenswrapper[4805]: E1203 15:12:16.061946 4805 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="46870522-a59e-4525-b5fd-452bec4d362b" containerName="extract-content" Dec 03 15:12:16 crc kubenswrapper[4805]: I1203 15:12:16.061952 4805 state_mem.go:107] "Deleted CPUSet assignment" podUID="46870522-a59e-4525-b5fd-452bec4d362b" containerName="extract-content" Dec 03 15:12:16 crc kubenswrapper[4805]: E1203 15:12:16.061994 4805 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="46870522-a59e-4525-b5fd-452bec4d362b" containerName="registry-server" Dec 03 15:12:16 crc kubenswrapper[4805]: I1203 15:12:16.062001 4805 state_mem.go:107] "Deleted CPUSet assignment" podUID="46870522-a59e-4525-b5fd-452bec4d362b" containerName="registry-server" Dec 03 15:12:16 crc kubenswrapper[4805]: I1203 15:12:16.062183 4805 memory_manager.go:354] "RemoveStaleState removing state" podUID="46870522-a59e-4525-b5fd-452bec4d362b" containerName="registry-server" Dec 03 15:12:16 crc kubenswrapper[4805]: I1203 15:12:16.063872 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-5w2gr" Dec 03 15:12:16 crc kubenswrapper[4805]: I1203 15:12:16.122227 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-5w2gr"] Dec 03 15:12:16 crc kubenswrapper[4805]: I1203 15:12:16.150955 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bsvz2\" (UniqueName: \"kubernetes.io/projected/fcfe071f-c650-4f17-9130-b2564be1d44f-kube-api-access-bsvz2\") pod \"certified-operators-5w2gr\" (UID: \"fcfe071f-c650-4f17-9130-b2564be1d44f\") " pod="openshift-marketplace/certified-operators-5w2gr" Dec 03 15:12:16 crc kubenswrapper[4805]: I1203 15:12:16.151129 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/fcfe071f-c650-4f17-9130-b2564be1d44f-utilities\") pod \"certified-operators-5w2gr\" (UID: \"fcfe071f-c650-4f17-9130-b2564be1d44f\") " pod="openshift-marketplace/certified-operators-5w2gr" Dec 03 15:12:16 crc kubenswrapper[4805]: I1203 15:12:16.151473 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/fcfe071f-c650-4f17-9130-b2564be1d44f-catalog-content\") pod \"certified-operators-5w2gr\" (UID: \"fcfe071f-c650-4f17-9130-b2564be1d44f\") " pod="openshift-marketplace/certified-operators-5w2gr" Dec 03 15:12:16 crc kubenswrapper[4805]: I1203 15:12:16.253633 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/fcfe071f-c650-4f17-9130-b2564be1d44f-catalog-content\") pod \"certified-operators-5w2gr\" (UID: \"fcfe071f-c650-4f17-9130-b2564be1d44f\") " pod="openshift-marketplace/certified-operators-5w2gr" Dec 03 15:12:16 crc kubenswrapper[4805]: I1203 15:12:16.253772 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bsvz2\" (UniqueName: \"kubernetes.io/projected/fcfe071f-c650-4f17-9130-b2564be1d44f-kube-api-access-bsvz2\") pod \"certified-operators-5w2gr\" (UID: \"fcfe071f-c650-4f17-9130-b2564be1d44f\") " pod="openshift-marketplace/certified-operators-5w2gr" Dec 03 15:12:16 crc kubenswrapper[4805]: I1203 15:12:16.253852 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/fcfe071f-c650-4f17-9130-b2564be1d44f-utilities\") pod \"certified-operators-5w2gr\" (UID: \"fcfe071f-c650-4f17-9130-b2564be1d44f\") " pod="openshift-marketplace/certified-operators-5w2gr" Dec 03 15:12:16 crc kubenswrapper[4805]: I1203 15:12:16.254279 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/fcfe071f-c650-4f17-9130-b2564be1d44f-catalog-content\") pod \"certified-operators-5w2gr\" (UID: \"fcfe071f-c650-4f17-9130-b2564be1d44f\") " pod="openshift-marketplace/certified-operators-5w2gr" Dec 03 15:12:16 crc kubenswrapper[4805]: I1203 15:12:16.254371 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/fcfe071f-c650-4f17-9130-b2564be1d44f-utilities\") pod \"certified-operators-5w2gr\" (UID: \"fcfe071f-c650-4f17-9130-b2564be1d44f\") " pod="openshift-marketplace/certified-operators-5w2gr" Dec 03 15:12:16 crc kubenswrapper[4805]: I1203 15:12:16.273707 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bsvz2\" (UniqueName: \"kubernetes.io/projected/fcfe071f-c650-4f17-9130-b2564be1d44f-kube-api-access-bsvz2\") pod \"certified-operators-5w2gr\" (UID: \"fcfe071f-c650-4f17-9130-b2564be1d44f\") " pod="openshift-marketplace/certified-operators-5w2gr" Dec 03 15:12:16 crc kubenswrapper[4805]: I1203 15:12:16.392021 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-5w2gr" Dec 03 15:12:16 crc kubenswrapper[4805]: I1203 15:12:16.901446 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-5w2gr"] Dec 03 15:12:17 crc kubenswrapper[4805]: I1203 15:12:17.660475 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-5w2gr" event={"ID":"fcfe071f-c650-4f17-9130-b2564be1d44f","Type":"ContainerStarted","Data":"54664a10c3eca170a670e79e6ded9489e1f06f7ea5b8580b08118d3c382f935e"} Dec 03 15:12:18 crc kubenswrapper[4805]: I1203 15:12:18.671343 4805 generic.go:334] "Generic (PLEG): container finished" podID="fcfe071f-c650-4f17-9130-b2564be1d44f" containerID="e9d343bc6e2131c8837e0e3729e89035187959e3721f348fed3e119e9d4afa38" exitCode=0 Dec 03 15:12:18 crc kubenswrapper[4805]: I1203 15:12:18.671457 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-5w2gr" event={"ID":"fcfe071f-c650-4f17-9130-b2564be1d44f","Type":"ContainerDied","Data":"e9d343bc6e2131c8837e0e3729e89035187959e3721f348fed3e119e9d4afa38"} Dec 03 15:12:21 crc kubenswrapper[4805]: I1203 15:12:21.706179 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-5w2gr" event={"ID":"fcfe071f-c650-4f17-9130-b2564be1d44f","Type":"ContainerStarted","Data":"3281abdcb636f86f160518adfd65f97eaeea0840cf1f23c4bc0f547621c951e0"} Dec 03 15:12:22 crc kubenswrapper[4805]: I1203 15:12:22.717184 4805 generic.go:334] "Generic (PLEG): container finished" podID="fcfe071f-c650-4f17-9130-b2564be1d44f" containerID="3281abdcb636f86f160518adfd65f97eaeea0840cf1f23c4bc0f547621c951e0" exitCode=0 Dec 03 15:12:22 crc kubenswrapper[4805]: I1203 15:12:22.717260 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-5w2gr" event={"ID":"fcfe071f-c650-4f17-9130-b2564be1d44f","Type":"ContainerDied","Data":"3281abdcb636f86f160518adfd65f97eaeea0840cf1f23c4bc0f547621c951e0"} Dec 03 15:12:25 crc kubenswrapper[4805]: I1203 15:12:25.028591 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-5w2gr" event={"ID":"fcfe071f-c650-4f17-9130-b2564be1d44f","Type":"ContainerStarted","Data":"7d3a89aa81794403f2659ff62d3c987df7b23abd0412666b214f7cf5a222197f"} Dec 03 15:12:25 crc kubenswrapper[4805]: I1203 15:12:25.052322 4805 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-5w2gr" podStartSLOduration=3.471147208 podStartE2EDuration="9.052304723s" podCreationTimestamp="2025-12-03 15:12:16 +0000 UTC" firstStartedPulling="2025-12-03 15:12:18.673343326 +0000 UTC m=+3768.336260249" lastFinishedPulling="2025-12-03 15:12:24.254500841 +0000 UTC m=+3773.917417764" observedRunningTime="2025-12-03 15:12:25.046968442 +0000 UTC m=+3774.709885375" watchObservedRunningTime="2025-12-03 15:12:25.052304723 +0000 UTC m=+3774.715221646" Dec 03 15:12:25 crc kubenswrapper[4805]: I1203 15:12:25.300441 4805 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_telemetry-operator-controller-manager-cc9d8f87b-qfmmx_8cbfcd07-e136-4bfd-b3c6-6615c911d440/manager/0.log" Dec 03 15:12:26 crc kubenswrapper[4805]: I1203 15:12:26.393158 4805 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-5w2gr" Dec 03 15:12:26 crc kubenswrapper[4805]: I1203 15:12:26.394120 4805 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-5w2gr" Dec 03 15:12:26 crc kubenswrapper[4805]: I1203 15:12:26.464488 4805 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-5w2gr" Dec 03 15:12:26 crc kubenswrapper[4805]: I1203 15:12:26.672360 4805 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/aodh-0"] Dec 03 15:12:26 crc kubenswrapper[4805]: I1203 15:12:26.672695 4805 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/aodh-0" podUID="d60bc69e-b77e-47ac-9a86-a50533f61711" containerName="aodh-api" containerID="cri-o://b6244e91da7b185eb4899fd3b52a1a8356420a93544d4555e64c6dd85f412d57" gracePeriod=30 Dec 03 15:12:26 crc kubenswrapper[4805]: I1203 15:12:26.672797 4805 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/aodh-0" podUID="d60bc69e-b77e-47ac-9a86-a50533f61711" containerName="aodh-notifier" containerID="cri-o://374e25e9dfc63ccf7ee10520c0f4af5e17c37e5b2a010ab02f4c9f45e9a599da" gracePeriod=30 Dec 03 15:12:26 crc kubenswrapper[4805]: I1203 15:12:26.672849 4805 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/aodh-0" podUID="d60bc69e-b77e-47ac-9a86-a50533f61711" containerName="aodh-evaluator" containerID="cri-o://de129d6dcf6dbf4a79fec418b3e50087c1668058c15adcda0c51eae694f465ce" gracePeriod=30 Dec 03 15:12:26 crc kubenswrapper[4805]: I1203 15:12:26.673049 4805 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/aodh-0" podUID="d60bc69e-b77e-47ac-9a86-a50533f61711" containerName="aodh-listener" containerID="cri-o://eabbde92660ffc9b7960552812807370f6c0f71b2e1e3967895c09dbcdec0ee4" gracePeriod=30 Dec 03 15:12:30 crc kubenswrapper[4805]: I1203 15:12:30.076972 4805 generic.go:334] "Generic (PLEG): container finished" podID="d60bc69e-b77e-47ac-9a86-a50533f61711" containerID="eabbde92660ffc9b7960552812807370f6c0f71b2e1e3967895c09dbcdec0ee4" exitCode=0 Dec 03 15:12:30 crc kubenswrapper[4805]: I1203 15:12:30.077250 4805 generic.go:334] "Generic (PLEG): container finished" podID="d60bc69e-b77e-47ac-9a86-a50533f61711" containerID="374e25e9dfc63ccf7ee10520c0f4af5e17c37e5b2a010ab02f4c9f45e9a599da" exitCode=0 Dec 03 15:12:30 crc kubenswrapper[4805]: I1203 15:12:30.077259 4805 generic.go:334] "Generic (PLEG): container finished" podID="d60bc69e-b77e-47ac-9a86-a50533f61711" containerID="de129d6dcf6dbf4a79fec418b3e50087c1668058c15adcda0c51eae694f465ce" exitCode=0 Dec 03 15:12:30 crc kubenswrapper[4805]: I1203 15:12:30.077269 4805 generic.go:334] "Generic (PLEG): container finished" podID="d60bc69e-b77e-47ac-9a86-a50533f61711" containerID="b6244e91da7b185eb4899fd3b52a1a8356420a93544d4555e64c6dd85f412d57" exitCode=0 Dec 03 15:12:30 crc kubenswrapper[4805]: I1203 15:12:30.077075 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-0" event={"ID":"d60bc69e-b77e-47ac-9a86-a50533f61711","Type":"ContainerDied","Data":"eabbde92660ffc9b7960552812807370f6c0f71b2e1e3967895c09dbcdec0ee4"} Dec 03 15:12:30 crc kubenswrapper[4805]: I1203 15:12:30.077308 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-0" event={"ID":"d60bc69e-b77e-47ac-9a86-a50533f61711","Type":"ContainerDied","Data":"374e25e9dfc63ccf7ee10520c0f4af5e17c37e5b2a010ab02f4c9f45e9a599da"} Dec 03 15:12:30 crc kubenswrapper[4805]: I1203 15:12:30.077324 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-0" event={"ID":"d60bc69e-b77e-47ac-9a86-a50533f61711","Type":"ContainerDied","Data":"de129d6dcf6dbf4a79fec418b3e50087c1668058c15adcda0c51eae694f465ce"} Dec 03 15:12:30 crc kubenswrapper[4805]: I1203 15:12:30.077333 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-0" event={"ID":"d60bc69e-b77e-47ac-9a86-a50533f61711","Type":"ContainerDied","Data":"b6244e91da7b185eb4899fd3b52a1a8356420a93544d4555e64c6dd85f412d57"} Dec 03 15:12:31 crc kubenswrapper[4805]: I1203 15:12:31.094201 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-0" event={"ID":"d60bc69e-b77e-47ac-9a86-a50533f61711","Type":"ContainerDied","Data":"348b523ea3abdf9a6864ac8ba27b81fb90f5d07c9c6b91133ca8ca100cc1d401"} Dec 03 15:12:31 crc kubenswrapper[4805]: I1203 15:12:31.094526 4805 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="348b523ea3abdf9a6864ac8ba27b81fb90f5d07c9c6b91133ca8ca100cc1d401" Dec 03 15:12:31 crc kubenswrapper[4805]: I1203 15:12:31.135049 4805 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-0" Dec 03 15:12:31 crc kubenswrapper[4805]: I1203 15:12:31.233689 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/d60bc69e-b77e-47ac-9a86-a50533f61711-internal-tls-certs\") pod \"d60bc69e-b77e-47ac-9a86-a50533f61711\" (UID: \"d60bc69e-b77e-47ac-9a86-a50533f61711\") " Dec 03 15:12:31 crc kubenswrapper[4805]: I1203 15:12:31.233761 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ppfpx\" (UniqueName: \"kubernetes.io/projected/d60bc69e-b77e-47ac-9a86-a50533f61711-kube-api-access-ppfpx\") pod \"d60bc69e-b77e-47ac-9a86-a50533f61711\" (UID: \"d60bc69e-b77e-47ac-9a86-a50533f61711\") " Dec 03 15:12:31 crc kubenswrapper[4805]: I1203 15:12:31.233886 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d60bc69e-b77e-47ac-9a86-a50533f61711-scripts\") pod \"d60bc69e-b77e-47ac-9a86-a50533f61711\" (UID: \"d60bc69e-b77e-47ac-9a86-a50533f61711\") " Dec 03 15:12:31 crc kubenswrapper[4805]: I1203 15:12:31.233929 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/d60bc69e-b77e-47ac-9a86-a50533f61711-public-tls-certs\") pod \"d60bc69e-b77e-47ac-9a86-a50533f61711\" (UID: \"d60bc69e-b77e-47ac-9a86-a50533f61711\") " Dec 03 15:12:31 crc kubenswrapper[4805]: I1203 15:12:31.234051 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d60bc69e-b77e-47ac-9a86-a50533f61711-config-data\") pod \"d60bc69e-b77e-47ac-9a86-a50533f61711\" (UID: \"d60bc69e-b77e-47ac-9a86-a50533f61711\") " Dec 03 15:12:31 crc kubenswrapper[4805]: I1203 15:12:31.234101 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d60bc69e-b77e-47ac-9a86-a50533f61711-combined-ca-bundle\") pod \"d60bc69e-b77e-47ac-9a86-a50533f61711\" (UID: \"d60bc69e-b77e-47ac-9a86-a50533f61711\") " Dec 03 15:12:31 crc kubenswrapper[4805]: I1203 15:12:31.239533 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d60bc69e-b77e-47ac-9a86-a50533f61711-scripts" (OuterVolumeSpecName: "scripts") pod "d60bc69e-b77e-47ac-9a86-a50533f61711" (UID: "d60bc69e-b77e-47ac-9a86-a50533f61711"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 15:12:31 crc kubenswrapper[4805]: I1203 15:12:31.261901 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d60bc69e-b77e-47ac-9a86-a50533f61711-kube-api-access-ppfpx" (OuterVolumeSpecName: "kube-api-access-ppfpx") pod "d60bc69e-b77e-47ac-9a86-a50533f61711" (UID: "d60bc69e-b77e-47ac-9a86-a50533f61711"). InnerVolumeSpecName "kube-api-access-ppfpx". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 15:12:31 crc kubenswrapper[4805]: I1203 15:12:31.303957 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d60bc69e-b77e-47ac-9a86-a50533f61711-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "d60bc69e-b77e-47ac-9a86-a50533f61711" (UID: "d60bc69e-b77e-47ac-9a86-a50533f61711"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 15:12:31 crc kubenswrapper[4805]: I1203 15:12:31.317061 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d60bc69e-b77e-47ac-9a86-a50533f61711-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "d60bc69e-b77e-47ac-9a86-a50533f61711" (UID: "d60bc69e-b77e-47ac-9a86-a50533f61711"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 15:12:31 crc kubenswrapper[4805]: I1203 15:12:31.336386 4805 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ppfpx\" (UniqueName: \"kubernetes.io/projected/d60bc69e-b77e-47ac-9a86-a50533f61711-kube-api-access-ppfpx\") on node \"crc\" DevicePath \"\"" Dec 03 15:12:31 crc kubenswrapper[4805]: I1203 15:12:31.336417 4805 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d60bc69e-b77e-47ac-9a86-a50533f61711-scripts\") on node \"crc\" DevicePath \"\"" Dec 03 15:12:31 crc kubenswrapper[4805]: I1203 15:12:31.336425 4805 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/d60bc69e-b77e-47ac-9a86-a50533f61711-public-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 03 15:12:31 crc kubenswrapper[4805]: I1203 15:12:31.336433 4805 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/d60bc69e-b77e-47ac-9a86-a50533f61711-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 03 15:12:31 crc kubenswrapper[4805]: I1203 15:12:31.365072 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d60bc69e-b77e-47ac-9a86-a50533f61711-config-data" (OuterVolumeSpecName: "config-data") pod "d60bc69e-b77e-47ac-9a86-a50533f61711" (UID: "d60bc69e-b77e-47ac-9a86-a50533f61711"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 15:12:31 crc kubenswrapper[4805]: I1203 15:12:31.372415 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d60bc69e-b77e-47ac-9a86-a50533f61711-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "d60bc69e-b77e-47ac-9a86-a50533f61711" (UID: "d60bc69e-b77e-47ac-9a86-a50533f61711"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 15:12:31 crc kubenswrapper[4805]: I1203 15:12:31.438191 4805 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d60bc69e-b77e-47ac-9a86-a50533f61711-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 15:12:31 crc kubenswrapper[4805]: I1203 15:12:31.438228 4805 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d60bc69e-b77e-47ac-9a86-a50533f61711-config-data\") on node \"crc\" DevicePath \"\"" Dec 03 15:12:32 crc kubenswrapper[4805]: I1203 15:12:32.102652 4805 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-0" Dec 03 15:12:32 crc kubenswrapper[4805]: I1203 15:12:32.140514 4805 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/aodh-0"] Dec 03 15:12:32 crc kubenswrapper[4805]: I1203 15:12:32.152799 4805 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/aodh-0"] Dec 03 15:12:32 crc kubenswrapper[4805]: I1203 15:12:32.183498 4805 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/aodh-0"] Dec 03 15:12:32 crc kubenswrapper[4805]: E1203 15:12:32.184228 4805 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d60bc69e-b77e-47ac-9a86-a50533f61711" containerName="aodh-notifier" Dec 03 15:12:32 crc kubenswrapper[4805]: I1203 15:12:32.184256 4805 state_mem.go:107] "Deleted CPUSet assignment" podUID="d60bc69e-b77e-47ac-9a86-a50533f61711" containerName="aodh-notifier" Dec 03 15:12:32 crc kubenswrapper[4805]: E1203 15:12:32.184271 4805 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d60bc69e-b77e-47ac-9a86-a50533f61711" containerName="aodh-api" Dec 03 15:12:32 crc kubenswrapper[4805]: I1203 15:12:32.184281 4805 state_mem.go:107] "Deleted CPUSet assignment" podUID="d60bc69e-b77e-47ac-9a86-a50533f61711" containerName="aodh-api" Dec 03 15:12:32 crc kubenswrapper[4805]: E1203 15:12:32.184310 4805 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d60bc69e-b77e-47ac-9a86-a50533f61711" containerName="aodh-evaluator" Dec 03 15:12:32 crc kubenswrapper[4805]: I1203 15:12:32.184319 4805 state_mem.go:107] "Deleted CPUSet assignment" podUID="d60bc69e-b77e-47ac-9a86-a50533f61711" containerName="aodh-evaluator" Dec 03 15:12:32 crc kubenswrapper[4805]: E1203 15:12:32.184335 4805 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d60bc69e-b77e-47ac-9a86-a50533f61711" containerName="aodh-listener" Dec 03 15:12:32 crc kubenswrapper[4805]: I1203 15:12:32.184345 4805 state_mem.go:107] "Deleted CPUSet assignment" podUID="d60bc69e-b77e-47ac-9a86-a50533f61711" containerName="aodh-listener" Dec 03 15:12:32 crc kubenswrapper[4805]: I1203 15:12:32.184594 4805 memory_manager.go:354] "RemoveStaleState removing state" podUID="d60bc69e-b77e-47ac-9a86-a50533f61711" containerName="aodh-notifier" Dec 03 15:12:32 crc kubenswrapper[4805]: I1203 15:12:32.184629 4805 memory_manager.go:354] "RemoveStaleState removing state" podUID="d60bc69e-b77e-47ac-9a86-a50533f61711" containerName="aodh-api" Dec 03 15:12:32 crc kubenswrapper[4805]: I1203 15:12:32.184655 4805 memory_manager.go:354] "RemoveStaleState removing state" podUID="d60bc69e-b77e-47ac-9a86-a50533f61711" containerName="aodh-evaluator" Dec 03 15:12:32 crc kubenswrapper[4805]: I1203 15:12:32.184670 4805 memory_manager.go:354] "RemoveStaleState removing state" podUID="d60bc69e-b77e-47ac-9a86-a50533f61711" containerName="aodh-listener" Dec 03 15:12:32 crc kubenswrapper[4805]: I1203 15:12:32.189178 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-0" Dec 03 15:12:32 crc kubenswrapper[4805]: I1203 15:12:32.192183 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"telemetry-autoscaling-dockercfg-wc8kz" Dec 03 15:12:32 crc kubenswrapper[4805]: I1203 15:12:32.192528 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"aodh-scripts" Dec 03 15:12:32 crc kubenswrapper[4805]: I1203 15:12:32.192824 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-aodh-internal-svc" Dec 03 15:12:32 crc kubenswrapper[4805]: I1203 15:12:32.197353 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"aodh-config-data" Dec 03 15:12:32 crc kubenswrapper[4805]: I1203 15:12:32.197728 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-aodh-public-svc" Dec 03 15:12:32 crc kubenswrapper[4805]: I1203 15:12:32.217581 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/aodh-0"] Dec 03 15:12:32 crc kubenswrapper[4805]: I1203 15:12:32.357722 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-njz58\" (UniqueName: \"kubernetes.io/projected/4297360e-d110-459b-b8b9-75f4a201921e-kube-api-access-njz58\") pod \"aodh-0\" (UID: \"4297360e-d110-459b-b8b9-75f4a201921e\") " pod="openstack/aodh-0" Dec 03 15:12:32 crc kubenswrapper[4805]: I1203 15:12:32.357781 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4297360e-d110-459b-b8b9-75f4a201921e-scripts\") pod \"aodh-0\" (UID: \"4297360e-d110-459b-b8b9-75f4a201921e\") " pod="openstack/aodh-0" Dec 03 15:12:32 crc kubenswrapper[4805]: I1203 15:12:32.357827 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/4297360e-d110-459b-b8b9-75f4a201921e-public-tls-certs\") pod \"aodh-0\" (UID: \"4297360e-d110-459b-b8b9-75f4a201921e\") " pod="openstack/aodh-0" Dec 03 15:12:32 crc kubenswrapper[4805]: I1203 15:12:32.357867 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4297360e-d110-459b-b8b9-75f4a201921e-combined-ca-bundle\") pod \"aodh-0\" (UID: \"4297360e-d110-459b-b8b9-75f4a201921e\") " pod="openstack/aodh-0" Dec 03 15:12:32 crc kubenswrapper[4805]: I1203 15:12:32.357985 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4297360e-d110-459b-b8b9-75f4a201921e-config-data\") pod \"aodh-0\" (UID: \"4297360e-d110-459b-b8b9-75f4a201921e\") " pod="openstack/aodh-0" Dec 03 15:12:32 crc kubenswrapper[4805]: I1203 15:12:32.358031 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/4297360e-d110-459b-b8b9-75f4a201921e-internal-tls-certs\") pod \"aodh-0\" (UID: \"4297360e-d110-459b-b8b9-75f4a201921e\") " pod="openstack/aodh-0" Dec 03 15:12:32 crc kubenswrapper[4805]: I1203 15:12:32.459206 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/4297360e-d110-459b-b8b9-75f4a201921e-public-tls-certs\") pod \"aodh-0\" (UID: \"4297360e-d110-459b-b8b9-75f4a201921e\") " pod="openstack/aodh-0" Dec 03 15:12:32 crc kubenswrapper[4805]: I1203 15:12:32.459254 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4297360e-d110-459b-b8b9-75f4a201921e-combined-ca-bundle\") pod \"aodh-0\" (UID: \"4297360e-d110-459b-b8b9-75f4a201921e\") " pod="openstack/aodh-0" Dec 03 15:12:32 crc kubenswrapper[4805]: I1203 15:12:32.459359 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4297360e-d110-459b-b8b9-75f4a201921e-config-data\") pod \"aodh-0\" (UID: \"4297360e-d110-459b-b8b9-75f4a201921e\") " pod="openstack/aodh-0" Dec 03 15:12:32 crc kubenswrapper[4805]: I1203 15:12:32.459386 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/4297360e-d110-459b-b8b9-75f4a201921e-internal-tls-certs\") pod \"aodh-0\" (UID: \"4297360e-d110-459b-b8b9-75f4a201921e\") " pod="openstack/aodh-0" Dec 03 15:12:32 crc kubenswrapper[4805]: I1203 15:12:32.459427 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-njz58\" (UniqueName: \"kubernetes.io/projected/4297360e-d110-459b-b8b9-75f4a201921e-kube-api-access-njz58\") pod \"aodh-0\" (UID: \"4297360e-d110-459b-b8b9-75f4a201921e\") " pod="openstack/aodh-0" Dec 03 15:12:32 crc kubenswrapper[4805]: I1203 15:12:32.459460 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4297360e-d110-459b-b8b9-75f4a201921e-scripts\") pod \"aodh-0\" (UID: \"4297360e-d110-459b-b8b9-75f4a201921e\") " pod="openstack/aodh-0" Dec 03 15:12:32 crc kubenswrapper[4805]: I1203 15:12:32.463956 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4297360e-d110-459b-b8b9-75f4a201921e-scripts\") pod \"aodh-0\" (UID: \"4297360e-d110-459b-b8b9-75f4a201921e\") " pod="openstack/aodh-0" Dec 03 15:12:32 crc kubenswrapper[4805]: I1203 15:12:32.464250 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/4297360e-d110-459b-b8b9-75f4a201921e-public-tls-certs\") pod \"aodh-0\" (UID: \"4297360e-d110-459b-b8b9-75f4a201921e\") " pod="openstack/aodh-0" Dec 03 15:12:32 crc kubenswrapper[4805]: I1203 15:12:32.464511 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/4297360e-d110-459b-b8b9-75f4a201921e-internal-tls-certs\") pod \"aodh-0\" (UID: \"4297360e-d110-459b-b8b9-75f4a201921e\") " pod="openstack/aodh-0" Dec 03 15:12:32 crc kubenswrapper[4805]: I1203 15:12:32.472580 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4297360e-d110-459b-b8b9-75f4a201921e-combined-ca-bundle\") pod \"aodh-0\" (UID: \"4297360e-d110-459b-b8b9-75f4a201921e\") " pod="openstack/aodh-0" Dec 03 15:12:32 crc kubenswrapper[4805]: I1203 15:12:32.472731 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4297360e-d110-459b-b8b9-75f4a201921e-config-data\") pod \"aodh-0\" (UID: \"4297360e-d110-459b-b8b9-75f4a201921e\") " pod="openstack/aodh-0" Dec 03 15:12:32 crc kubenswrapper[4805]: I1203 15:12:32.475608 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-njz58\" (UniqueName: \"kubernetes.io/projected/4297360e-d110-459b-b8b9-75f4a201921e-kube-api-access-njz58\") pod \"aodh-0\" (UID: \"4297360e-d110-459b-b8b9-75f4a201921e\") " pod="openstack/aodh-0" Dec 03 15:12:32 crc kubenswrapper[4805]: I1203 15:12:32.512628 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-0" Dec 03 15:12:32 crc kubenswrapper[4805]: I1203 15:12:32.710637 4805 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d60bc69e-b77e-47ac-9a86-a50533f61711" path="/var/lib/kubelet/pods/d60bc69e-b77e-47ac-9a86-a50533f61711/volumes" Dec 03 15:12:33 crc kubenswrapper[4805]: W1203 15:12:33.349597 4805 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod4297360e_d110_459b_b8b9_75f4a201921e.slice/crio-767b11f88aa70b3a6db186a1319c50de854e5f877414c4bd75570e23e6780d9b WatchSource:0}: Error finding container 767b11f88aa70b3a6db186a1319c50de854e5f877414c4bd75570e23e6780d9b: Status 404 returned error can't find the container with id 767b11f88aa70b3a6db186a1319c50de854e5f877414c4bd75570e23e6780d9b Dec 03 15:12:33 crc kubenswrapper[4805]: I1203 15:12:33.356832 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/aodh-0"] Dec 03 15:12:34 crc kubenswrapper[4805]: I1203 15:12:34.127654 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-0" event={"ID":"4297360e-d110-459b-b8b9-75f4a201921e","Type":"ContainerStarted","Data":"767b11f88aa70b3a6db186a1319c50de854e5f877414c4bd75570e23e6780d9b"} Dec 03 15:12:36 crc kubenswrapper[4805]: I1203 15:12:36.452035 4805 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-5w2gr" Dec 03 15:12:36 crc kubenswrapper[4805]: I1203 15:12:36.518549 4805 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-5w2gr"] Dec 03 15:12:37 crc kubenswrapper[4805]: I1203 15:12:37.156474 4805 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-5w2gr" podUID="fcfe071f-c650-4f17-9130-b2564be1d44f" containerName="registry-server" containerID="cri-o://7d3a89aa81794403f2659ff62d3c987df7b23abd0412666b214f7cf5a222197f" gracePeriod=2 Dec 03 15:12:37 crc kubenswrapper[4805]: I1203 15:12:37.156829 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-0" event={"ID":"4297360e-d110-459b-b8b9-75f4a201921e","Type":"ContainerStarted","Data":"33d1266a1f04895391794f1aa3fd4b94db2cd4ebf42673fadf0db16c3b367ff2"} Dec 03 15:12:38 crc kubenswrapper[4805]: I1203 15:12:38.174028 4805 generic.go:334] "Generic (PLEG): container finished" podID="fcfe071f-c650-4f17-9130-b2564be1d44f" containerID="7d3a89aa81794403f2659ff62d3c987df7b23abd0412666b214f7cf5a222197f" exitCode=0 Dec 03 15:12:38 crc kubenswrapper[4805]: I1203 15:12:38.174197 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-5w2gr" event={"ID":"fcfe071f-c650-4f17-9130-b2564be1d44f","Type":"ContainerDied","Data":"7d3a89aa81794403f2659ff62d3c987df7b23abd0412666b214f7cf5a222197f"} Dec 03 15:12:38 crc kubenswrapper[4805]: I1203 15:12:38.174606 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-5w2gr" event={"ID":"fcfe071f-c650-4f17-9130-b2564be1d44f","Type":"ContainerDied","Data":"54664a10c3eca170a670e79e6ded9489e1f06f7ea5b8580b08118d3c382f935e"} Dec 03 15:12:38 crc kubenswrapper[4805]: I1203 15:12:38.174628 4805 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="54664a10c3eca170a670e79e6ded9489e1f06f7ea5b8580b08118d3c382f935e" Dec 03 15:12:38 crc kubenswrapper[4805]: I1203 15:12:38.230440 4805 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-5w2gr" Dec 03 15:12:38 crc kubenswrapper[4805]: I1203 15:12:38.409386 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/fcfe071f-c650-4f17-9130-b2564be1d44f-utilities\") pod \"fcfe071f-c650-4f17-9130-b2564be1d44f\" (UID: \"fcfe071f-c650-4f17-9130-b2564be1d44f\") " Dec 03 15:12:38 crc kubenswrapper[4805]: I1203 15:12:38.409447 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bsvz2\" (UniqueName: \"kubernetes.io/projected/fcfe071f-c650-4f17-9130-b2564be1d44f-kube-api-access-bsvz2\") pod \"fcfe071f-c650-4f17-9130-b2564be1d44f\" (UID: \"fcfe071f-c650-4f17-9130-b2564be1d44f\") " Dec 03 15:12:38 crc kubenswrapper[4805]: I1203 15:12:38.409695 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/fcfe071f-c650-4f17-9130-b2564be1d44f-catalog-content\") pod \"fcfe071f-c650-4f17-9130-b2564be1d44f\" (UID: \"fcfe071f-c650-4f17-9130-b2564be1d44f\") " Dec 03 15:12:38 crc kubenswrapper[4805]: I1203 15:12:38.410293 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/fcfe071f-c650-4f17-9130-b2564be1d44f-utilities" (OuterVolumeSpecName: "utilities") pod "fcfe071f-c650-4f17-9130-b2564be1d44f" (UID: "fcfe071f-c650-4f17-9130-b2564be1d44f"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 15:12:38 crc kubenswrapper[4805]: I1203 15:12:38.416300 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fcfe071f-c650-4f17-9130-b2564be1d44f-kube-api-access-bsvz2" (OuterVolumeSpecName: "kube-api-access-bsvz2") pod "fcfe071f-c650-4f17-9130-b2564be1d44f" (UID: "fcfe071f-c650-4f17-9130-b2564be1d44f"). InnerVolumeSpecName "kube-api-access-bsvz2". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 15:12:38 crc kubenswrapper[4805]: I1203 15:12:38.462199 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/fcfe071f-c650-4f17-9130-b2564be1d44f-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "fcfe071f-c650-4f17-9130-b2564be1d44f" (UID: "fcfe071f-c650-4f17-9130-b2564be1d44f"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 15:12:38 crc kubenswrapper[4805]: I1203 15:12:38.511684 4805 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/fcfe071f-c650-4f17-9130-b2564be1d44f-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 03 15:12:38 crc kubenswrapper[4805]: I1203 15:12:38.511723 4805 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/fcfe071f-c650-4f17-9130-b2564be1d44f-utilities\") on node \"crc\" DevicePath \"\"" Dec 03 15:12:38 crc kubenswrapper[4805]: I1203 15:12:38.511732 4805 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bsvz2\" (UniqueName: \"kubernetes.io/projected/fcfe071f-c650-4f17-9130-b2564be1d44f-kube-api-access-bsvz2\") on node \"crc\" DevicePath \"\"" Dec 03 15:12:39 crc kubenswrapper[4805]: I1203 15:12:39.182929 4805 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-5w2gr" Dec 03 15:12:39 crc kubenswrapper[4805]: I1203 15:12:39.214944 4805 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-5w2gr"] Dec 03 15:12:39 crc kubenswrapper[4805]: I1203 15:12:39.225019 4805 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-5w2gr"] Dec 03 15:12:40 crc kubenswrapper[4805]: I1203 15:12:40.194787 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-0" event={"ID":"4297360e-d110-459b-b8b9-75f4a201921e","Type":"ContainerStarted","Data":"d36903336767d165f2f4e47ab10a29db6ae4c42bd990e9f774aa336f8b820619"} Dec 03 15:12:40 crc kubenswrapper[4805]: I1203 15:12:40.718377 4805 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="fcfe071f-c650-4f17-9130-b2564be1d44f" path="/var/lib/kubelet/pods/fcfe071f-c650-4f17-9130-b2564be1d44f/volumes" Dec 03 15:12:42 crc kubenswrapper[4805]: I1203 15:12:42.233809 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-0" event={"ID":"4297360e-d110-459b-b8b9-75f4a201921e","Type":"ContainerStarted","Data":"72575d8ebf082d477322e64027f44bfc34edb83961203d482e520d87431f7f08"} Dec 03 15:12:43 crc kubenswrapper[4805]: I1203 15:12:43.246826 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-0" event={"ID":"4297360e-d110-459b-b8b9-75f4a201921e","Type":"ContainerStarted","Data":"4d9e45ef8847c425c9e087a7f6d1cfbfd33eb45d53896d79349267a55e64f90c"} Dec 03 15:12:43 crc kubenswrapper[4805]: I1203 15:12:43.271569 4805 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/aodh-0" podStartSLOduration=2.327442541 podStartE2EDuration="11.271550959s" podCreationTimestamp="2025-12-03 15:12:32 +0000 UTC" firstStartedPulling="2025-12-03 15:12:33.352627708 +0000 UTC m=+3783.015544641" lastFinishedPulling="2025-12-03 15:12:42.296736136 +0000 UTC m=+3791.959653059" observedRunningTime="2025-12-03 15:12:43.265063787 +0000 UTC m=+3792.927980720" watchObservedRunningTime="2025-12-03 15:12:43.271550959 +0000 UTC m=+3792.934467882" Dec 03 15:12:43 crc kubenswrapper[4805]: I1203 15:12:43.919550 4805 patch_prober.go:28] interesting pod/machine-config-daemon-gskh4 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 03 15:12:43 crc kubenswrapper[4805]: I1203 15:12:43.919612 4805 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-gskh4" podUID="6098937f-e3f6-45e8-a647-4994a79cd711" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 03 15:12:46 crc kubenswrapper[4805]: I1203 15:12:46.058327 4805 scope.go:117] "RemoveContainer" containerID="0746524baeba3bd134675117e4c1d81496a866fffa6dd8f2fc7b3675e9635340" Dec 03 15:13:13 crc kubenswrapper[4805]: I1203 15:13:13.917153 4805 patch_prober.go:28] interesting pod/machine-config-daemon-gskh4 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 03 15:13:13 crc kubenswrapper[4805]: I1203 15:13:13.918631 4805 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-gskh4" podUID="6098937f-e3f6-45e8-a647-4994a79cd711" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 03 15:13:43 crc kubenswrapper[4805]: I1203 15:13:43.917323 4805 patch_prober.go:28] interesting pod/machine-config-daemon-gskh4 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 03 15:13:43 crc kubenswrapper[4805]: I1203 15:13:43.917768 4805 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-gskh4" podUID="6098937f-e3f6-45e8-a647-4994a79cd711" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 03 15:13:43 crc kubenswrapper[4805]: I1203 15:13:43.917815 4805 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-gskh4" Dec 03 15:13:43 crc kubenswrapper[4805]: I1203 15:13:43.918537 4805 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"cbf73c6c9ba740c0a523c84ca23c0851292d6eba7a0fb47d38599acca964e9cc"} pod="openshift-machine-config-operator/machine-config-daemon-gskh4" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 03 15:13:43 crc kubenswrapper[4805]: I1203 15:13:43.918591 4805 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-gskh4" podUID="6098937f-e3f6-45e8-a647-4994a79cd711" containerName="machine-config-daemon" containerID="cri-o://cbf73c6c9ba740c0a523c84ca23c0851292d6eba7a0fb47d38599acca964e9cc" gracePeriod=600 Dec 03 15:13:45 crc kubenswrapper[4805]: E1203 15:13:45.911747 4805 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-gskh4_openshift-machine-config-operator(6098937f-e3f6-45e8-a647-4994a79cd711)\"" pod="openshift-machine-config-operator/machine-config-daemon-gskh4" podUID="6098937f-e3f6-45e8-a647-4994a79cd711" Dec 03 15:13:46 crc kubenswrapper[4805]: I1203 15:13:46.007789 4805 generic.go:334] "Generic (PLEG): container finished" podID="6098937f-e3f6-45e8-a647-4994a79cd711" containerID="cbf73c6c9ba740c0a523c84ca23c0851292d6eba7a0fb47d38599acca964e9cc" exitCode=0 Dec 03 15:13:46 crc kubenswrapper[4805]: I1203 15:13:46.007883 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-gskh4" event={"ID":"6098937f-e3f6-45e8-a647-4994a79cd711","Type":"ContainerDied","Data":"cbf73c6c9ba740c0a523c84ca23c0851292d6eba7a0fb47d38599acca964e9cc"} Dec 03 15:13:46 crc kubenswrapper[4805]: I1203 15:13:46.008018 4805 scope.go:117] "RemoveContainer" containerID="e54004d4fb8c135b56b16241c1140f37c36df51d3d40b05d6724caf5e9950f98" Dec 03 15:13:46 crc kubenswrapper[4805]: I1203 15:13:46.008770 4805 scope.go:117] "RemoveContainer" containerID="cbf73c6c9ba740c0a523c84ca23c0851292d6eba7a0fb47d38599acca964e9cc" Dec 03 15:13:46 crc kubenswrapper[4805]: E1203 15:13:46.009120 4805 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-gskh4_openshift-machine-config-operator(6098937f-e3f6-45e8-a647-4994a79cd711)\"" pod="openshift-machine-config-operator/machine-config-daemon-gskh4" podUID="6098937f-e3f6-45e8-a647-4994a79cd711" Dec 03 15:13:46 crc kubenswrapper[4805]: I1203 15:13:46.447005 4805 scope.go:117] "RemoveContainer" containerID="de129d6dcf6dbf4a79fec418b3e50087c1668058c15adcda0c51eae694f465ce" Dec 03 15:13:46 crc kubenswrapper[4805]: I1203 15:13:46.467694 4805 scope.go:117] "RemoveContainer" containerID="374e25e9dfc63ccf7ee10520c0f4af5e17c37e5b2a010ab02f4c9f45e9a599da" Dec 03 15:13:46 crc kubenswrapper[4805]: I1203 15:13:46.489936 4805 scope.go:117] "RemoveContainer" containerID="b6244e91da7b185eb4899fd3b52a1a8356420a93544d4555e64c6dd85f412d57" Dec 03 15:13:46 crc kubenswrapper[4805]: I1203 15:13:46.507378 4805 scope.go:117] "RemoveContainer" containerID="894956bdcafeba1a2eae20d8f53ec468f00b495dc09e3affa45032f7834cdd29" Dec 03 15:13:47 crc kubenswrapper[4805]: I1203 15:13:47.025252 4805 scope.go:117] "RemoveContainer" containerID="cbf73c6c9ba740c0a523c84ca23c0851292d6eba7a0fb47d38599acca964e9cc" Dec 03 15:13:47 crc kubenswrapper[4805]: E1203 15:13:47.025569 4805 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-gskh4_openshift-machine-config-operator(6098937f-e3f6-45e8-a647-4994a79cd711)\"" pod="openshift-machine-config-operator/machine-config-daemon-gskh4" podUID="6098937f-e3f6-45e8-a647-4994a79cd711" Dec 03 15:13:57 crc kubenswrapper[4805]: I1203 15:13:57.694867 4805 scope.go:117] "RemoveContainer" containerID="cbf73c6c9ba740c0a523c84ca23c0851292d6eba7a0fb47d38599acca964e9cc" Dec 03 15:13:57 crc kubenswrapper[4805]: E1203 15:13:57.695635 4805 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-gskh4_openshift-machine-config-operator(6098937f-e3f6-45e8-a647-4994a79cd711)\"" pod="openshift-machine-config-operator/machine-config-daemon-gskh4" podUID="6098937f-e3f6-45e8-a647-4994a79cd711" Dec 03 15:14:11 crc kubenswrapper[4805]: I1203 15:14:11.695432 4805 scope.go:117] "RemoveContainer" containerID="cbf73c6c9ba740c0a523c84ca23c0851292d6eba7a0fb47d38599acca964e9cc" Dec 03 15:14:11 crc kubenswrapper[4805]: E1203 15:14:11.696238 4805 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-gskh4_openshift-machine-config-operator(6098937f-e3f6-45e8-a647-4994a79cd711)\"" pod="openshift-machine-config-operator/machine-config-daemon-gskh4" podUID="6098937f-e3f6-45e8-a647-4994a79cd711" Dec 03 15:14:26 crc kubenswrapper[4805]: I1203 15:14:26.694801 4805 scope.go:117] "RemoveContainer" containerID="cbf73c6c9ba740c0a523c84ca23c0851292d6eba7a0fb47d38599acca964e9cc" Dec 03 15:14:26 crc kubenswrapper[4805]: E1203 15:14:26.695617 4805 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-gskh4_openshift-machine-config-operator(6098937f-e3f6-45e8-a647-4994a79cd711)\"" pod="openshift-machine-config-operator/machine-config-daemon-gskh4" podUID="6098937f-e3f6-45e8-a647-4994a79cd711" Dec 03 15:14:27 crc kubenswrapper[4805]: I1203 15:14:27.116875 4805 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_telemetry-operator-controller-manager-cc9d8f87b-qfmmx_8cbfcd07-e136-4bfd-b3c6-6615c911d440/manager/0.log" Dec 03 15:14:30 crc kubenswrapper[4805]: I1203 15:14:30.663562 4805 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/prometheus-metric-storage-0"] Dec 03 15:14:30 crc kubenswrapper[4805]: I1203 15:14:30.665641 4805 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/prometheus-metric-storage-0" podUID="9e35b852-1838-49da-a140-f048788f6635" containerName="prometheus" containerID="cri-o://3f32476151f16cefda9576a8077a18738591448ef92afdba7d6ba1cfe707d756" gracePeriod=600 Dec 03 15:14:30 crc kubenswrapper[4805]: I1203 15:14:30.665680 4805 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/prometheus-metric-storage-0" podUID="9e35b852-1838-49da-a140-f048788f6635" containerName="thanos-sidecar" containerID="cri-o://a023a91b7c3b39787116d2c8c1a0da144673bf6d87052338e976a9af4bcc7df5" gracePeriod=600 Dec 03 15:14:30 crc kubenswrapper[4805]: I1203 15:14:30.665686 4805 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/prometheus-metric-storage-0" podUID="9e35b852-1838-49da-a140-f048788f6635" containerName="config-reloader" containerID="cri-o://c6738ea6a72122506d9eaaacc460bb54d669b6aa4a8ffdb2c2a9d8829326e697" gracePeriod=600 Dec 03 15:14:31 crc kubenswrapper[4805]: E1203 15:14:31.372205 4805 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod9e35b852_1838_49da_a140_f048788f6635.slice/crio-3f32476151f16cefda9576a8077a18738591448ef92afdba7d6ba1cfe707d756.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod9e35b852_1838_49da_a140_f048788f6635.slice/crio-conmon-3f32476151f16cefda9576a8077a18738591448ef92afdba7d6ba1cfe707d756.scope\": RecentStats: unable to find data in memory cache]" Dec 03 15:14:31 crc kubenswrapper[4805]: I1203 15:14:31.509041 4805 generic.go:334] "Generic (PLEG): container finished" podID="9e35b852-1838-49da-a140-f048788f6635" containerID="a023a91b7c3b39787116d2c8c1a0da144673bf6d87052338e976a9af4bcc7df5" exitCode=0 Dec 03 15:14:31 crc kubenswrapper[4805]: I1203 15:14:31.509280 4805 generic.go:334] "Generic (PLEG): container finished" podID="9e35b852-1838-49da-a140-f048788f6635" containerID="c6738ea6a72122506d9eaaacc460bb54d669b6aa4a8ffdb2c2a9d8829326e697" exitCode=0 Dec 03 15:14:31 crc kubenswrapper[4805]: I1203 15:14:31.509289 4805 generic.go:334] "Generic (PLEG): container finished" podID="9e35b852-1838-49da-a140-f048788f6635" containerID="3f32476151f16cefda9576a8077a18738591448ef92afdba7d6ba1cfe707d756" exitCode=0 Dec 03 15:14:31 crc kubenswrapper[4805]: I1203 15:14:31.509129 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"9e35b852-1838-49da-a140-f048788f6635","Type":"ContainerDied","Data":"a023a91b7c3b39787116d2c8c1a0da144673bf6d87052338e976a9af4bcc7df5"} Dec 03 15:14:31 crc kubenswrapper[4805]: I1203 15:14:31.509324 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"9e35b852-1838-49da-a140-f048788f6635","Type":"ContainerDied","Data":"c6738ea6a72122506d9eaaacc460bb54d669b6aa4a8ffdb2c2a9d8829326e697"} Dec 03 15:14:31 crc kubenswrapper[4805]: I1203 15:14:31.509335 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"9e35b852-1838-49da-a140-f048788f6635","Type":"ContainerDied","Data":"3f32476151f16cefda9576a8077a18738591448ef92afdba7d6ba1cfe707d756"} Dec 03 15:14:31 crc kubenswrapper[4805]: I1203 15:14:31.670499 4805 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/prometheus-metric-storage-0" Dec 03 15:14:31 crc kubenswrapper[4805]: I1203 15:14:31.780626 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"web-config-tls-secret-key-cert-metric-storage-promethe-dc638c2d\" (UniqueName: \"kubernetes.io/secret/9e35b852-1838-49da-a140-f048788f6635-web-config-tls-secret-key-cert-metric-storage-promethe-dc638c2d\") pod \"9e35b852-1838-49da-a140-f048788f6635\" (UID: \"9e35b852-1838-49da-a140-f048788f6635\") " Dec 03 15:14:31 crc kubenswrapper[4805]: I1203 15:14:31.780704 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/9e35b852-1838-49da-a140-f048788f6635-config-out\") pod \"9e35b852-1838-49da-a140-f048788f6635\" (UID: \"9e35b852-1838-49da-a140-f048788f6635\") " Dec 03 15:14:31 crc kubenswrapper[4805]: I1203 15:14:31.780762 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"thanos-prometheus-http-client-file\" (UniqueName: \"kubernetes.io/secret/9e35b852-1838-49da-a140-f048788f6635-thanos-prometheus-http-client-file\") pod \"9e35b852-1838-49da-a140-f048788f6635\" (UID: \"9e35b852-1838-49da-a140-f048788f6635\") " Dec 03 15:14:31 crc kubenswrapper[4805]: I1203 15:14:31.780807 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9e35b852-1838-49da-a140-f048788f6635-secret-combined-ca-bundle\") pod \"9e35b852-1838-49da-a140-f048788f6635\" (UID: \"9e35b852-1838-49da-a140-f048788f6635\") " Dec 03 15:14:31 crc kubenswrapper[4805]: I1203 15:14:31.780893 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"prometheus-metric-storage-rulefiles-0\" (UniqueName: \"kubernetes.io/configmap/9e35b852-1838-49da-a140-f048788f6635-prometheus-metric-storage-rulefiles-0\") pod \"9e35b852-1838-49da-a140-f048788f6635\" (UID: \"9e35b852-1838-49da-a140-f048788f6635\") " Dec 03 15:14:31 crc kubenswrapper[4805]: I1203 15:14:31.780928 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"prometheus-metric-storage-db\" (UniqueName: \"kubernetes.io/empty-dir/9e35b852-1838-49da-a140-f048788f6635-prometheus-metric-storage-db\") pod \"9e35b852-1838-49da-a140-f048788f6635\" (UID: \"9e35b852-1838-49da-a140-f048788f6635\") " Dec 03 15:14:31 crc kubenswrapper[4805]: I1203 15:14:31.780989 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/9e35b852-1838-49da-a140-f048788f6635-config\") pod \"9e35b852-1838-49da-a140-f048788f6635\" (UID: \"9e35b852-1838-49da-a140-f048788f6635\") " Dec 03 15:14:31 crc kubenswrapper[4805]: I1203 15:14:31.781027 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-z2qxg\" (UniqueName: \"kubernetes.io/projected/9e35b852-1838-49da-a140-f048788f6635-kube-api-access-z2qxg\") pod \"9e35b852-1838-49da-a140-f048788f6635\" (UID: \"9e35b852-1838-49da-a140-f048788f6635\") " Dec 03 15:14:31 crc kubenswrapper[4805]: I1203 15:14:31.781068 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/9e35b852-1838-49da-a140-f048788f6635-web-config\") pod \"9e35b852-1838-49da-a140-f048788f6635\" (UID: \"9e35b852-1838-49da-a140-f048788f6635\") " Dec 03 15:14:31 crc kubenswrapper[4805]: I1203 15:14:31.781172 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/9e35b852-1838-49da-a140-f048788f6635-tls-assets\") pod \"9e35b852-1838-49da-a140-f048788f6635\" (UID: \"9e35b852-1838-49da-a140-f048788f6635\") " Dec 03 15:14:31 crc kubenswrapper[4805]: I1203 15:14:31.781238 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"web-config-tls-secret-cert-cert-metric-storage-prometh-dc638c2d\" (UniqueName: \"kubernetes.io/secret/9e35b852-1838-49da-a140-f048788f6635-web-config-tls-secret-cert-cert-metric-storage-prometh-dc638c2d\") pod \"9e35b852-1838-49da-a140-f048788f6635\" (UID: \"9e35b852-1838-49da-a140-f048788f6635\") " Dec 03 15:14:31 crc kubenswrapper[4805]: I1203 15:14:31.782558 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9e35b852-1838-49da-a140-f048788f6635-prometheus-metric-storage-rulefiles-0" (OuterVolumeSpecName: "prometheus-metric-storage-rulefiles-0") pod "9e35b852-1838-49da-a140-f048788f6635" (UID: "9e35b852-1838-49da-a140-f048788f6635"). InnerVolumeSpecName "prometheus-metric-storage-rulefiles-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 15:14:31 crc kubenswrapper[4805]: I1203 15:14:31.782603 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9e35b852-1838-49da-a140-f048788f6635-prometheus-metric-storage-db" (OuterVolumeSpecName: "prometheus-metric-storage-db") pod "9e35b852-1838-49da-a140-f048788f6635" (UID: "9e35b852-1838-49da-a140-f048788f6635"). InnerVolumeSpecName "prometheus-metric-storage-db". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 15:14:31 crc kubenswrapper[4805]: I1203 15:14:31.789388 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9e35b852-1838-49da-a140-f048788f6635-web-config-tls-secret-key-cert-metric-storage-promethe-dc638c2d" (OuterVolumeSpecName: "web-config-tls-secret-key-cert-metric-storage-promethe-dc638c2d") pod "9e35b852-1838-49da-a140-f048788f6635" (UID: "9e35b852-1838-49da-a140-f048788f6635"). InnerVolumeSpecName "web-config-tls-secret-key-cert-metric-storage-promethe-dc638c2d". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 15:14:31 crc kubenswrapper[4805]: I1203 15:14:31.789877 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9e35b852-1838-49da-a140-f048788f6635-web-config-tls-secret-cert-cert-metric-storage-prometh-dc638c2d" (OuterVolumeSpecName: "web-config-tls-secret-cert-cert-metric-storage-prometh-dc638c2d") pod "9e35b852-1838-49da-a140-f048788f6635" (UID: "9e35b852-1838-49da-a140-f048788f6635"). InnerVolumeSpecName "web-config-tls-secret-cert-cert-metric-storage-prometh-dc638c2d". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 15:14:31 crc kubenswrapper[4805]: I1203 15:14:31.790093 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9e35b852-1838-49da-a140-f048788f6635-tls-assets" (OuterVolumeSpecName: "tls-assets") pod "9e35b852-1838-49da-a140-f048788f6635" (UID: "9e35b852-1838-49da-a140-f048788f6635"). InnerVolumeSpecName "tls-assets". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 15:14:31 crc kubenswrapper[4805]: I1203 15:14:31.791165 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9e35b852-1838-49da-a140-f048788f6635-kube-api-access-z2qxg" (OuterVolumeSpecName: "kube-api-access-z2qxg") pod "9e35b852-1838-49da-a140-f048788f6635" (UID: "9e35b852-1838-49da-a140-f048788f6635"). InnerVolumeSpecName "kube-api-access-z2qxg". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 15:14:31 crc kubenswrapper[4805]: I1203 15:14:31.791882 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9e35b852-1838-49da-a140-f048788f6635-thanos-prometheus-http-client-file" (OuterVolumeSpecName: "thanos-prometheus-http-client-file") pod "9e35b852-1838-49da-a140-f048788f6635" (UID: "9e35b852-1838-49da-a140-f048788f6635"). InnerVolumeSpecName "thanos-prometheus-http-client-file". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 15:14:31 crc kubenswrapper[4805]: I1203 15:14:31.794158 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9e35b852-1838-49da-a140-f048788f6635-config" (OuterVolumeSpecName: "config") pod "9e35b852-1838-49da-a140-f048788f6635" (UID: "9e35b852-1838-49da-a140-f048788f6635"). InnerVolumeSpecName "config". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 15:14:31 crc kubenswrapper[4805]: I1203 15:14:31.794648 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9e35b852-1838-49da-a140-f048788f6635-config-out" (OuterVolumeSpecName: "config-out") pod "9e35b852-1838-49da-a140-f048788f6635" (UID: "9e35b852-1838-49da-a140-f048788f6635"). InnerVolumeSpecName "config-out". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 15:14:31 crc kubenswrapper[4805]: I1203 15:14:31.801390 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9e35b852-1838-49da-a140-f048788f6635-secret-combined-ca-bundle" (OuterVolumeSpecName: "secret-combined-ca-bundle") pod "9e35b852-1838-49da-a140-f048788f6635" (UID: "9e35b852-1838-49da-a140-f048788f6635"). InnerVolumeSpecName "secret-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 15:14:31 crc kubenswrapper[4805]: I1203 15:14:31.870717 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9e35b852-1838-49da-a140-f048788f6635-web-config" (OuterVolumeSpecName: "web-config") pod "9e35b852-1838-49da-a140-f048788f6635" (UID: "9e35b852-1838-49da-a140-f048788f6635"). InnerVolumeSpecName "web-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 15:14:31 crc kubenswrapper[4805]: I1203 15:14:31.886615 4805 reconciler_common.go:293] "Volume detached for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/9e35b852-1838-49da-a140-f048788f6635-web-config\") on node \"crc\" DevicePath \"\"" Dec 03 15:14:31 crc kubenswrapper[4805]: I1203 15:14:31.886643 4805 reconciler_common.go:293] "Volume detached for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/9e35b852-1838-49da-a140-f048788f6635-tls-assets\") on node \"crc\" DevicePath \"\"" Dec 03 15:14:31 crc kubenswrapper[4805]: I1203 15:14:31.886652 4805 reconciler_common.go:293] "Volume detached for volume \"web-config-tls-secret-cert-cert-metric-storage-prometh-dc638c2d\" (UniqueName: \"kubernetes.io/secret/9e35b852-1838-49da-a140-f048788f6635-web-config-tls-secret-cert-cert-metric-storage-prometh-dc638c2d\") on node \"crc\" DevicePath \"\"" Dec 03 15:14:31 crc kubenswrapper[4805]: I1203 15:14:31.886666 4805 reconciler_common.go:293] "Volume detached for volume \"web-config-tls-secret-key-cert-metric-storage-promethe-dc638c2d\" (UniqueName: \"kubernetes.io/secret/9e35b852-1838-49da-a140-f048788f6635-web-config-tls-secret-key-cert-metric-storage-promethe-dc638c2d\") on node \"crc\" DevicePath \"\"" Dec 03 15:14:31 crc kubenswrapper[4805]: I1203 15:14:31.886676 4805 reconciler_common.go:293] "Volume detached for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/9e35b852-1838-49da-a140-f048788f6635-config-out\") on node \"crc\" DevicePath \"\"" Dec 03 15:14:31 crc kubenswrapper[4805]: I1203 15:14:31.886685 4805 reconciler_common.go:293] "Volume detached for volume \"thanos-prometheus-http-client-file\" (UniqueName: \"kubernetes.io/secret/9e35b852-1838-49da-a140-f048788f6635-thanos-prometheus-http-client-file\") on node \"crc\" DevicePath \"\"" Dec 03 15:14:31 crc kubenswrapper[4805]: I1203 15:14:31.886694 4805 reconciler_common.go:293] "Volume detached for volume \"secret-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9e35b852-1838-49da-a140-f048788f6635-secret-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 15:14:31 crc kubenswrapper[4805]: I1203 15:14:31.886704 4805 reconciler_common.go:293] "Volume detached for volume \"prometheus-metric-storage-rulefiles-0\" (UniqueName: \"kubernetes.io/configmap/9e35b852-1838-49da-a140-f048788f6635-prometheus-metric-storage-rulefiles-0\") on node \"crc\" DevicePath \"\"" Dec 03 15:14:31 crc kubenswrapper[4805]: I1203 15:14:31.886714 4805 reconciler_common.go:293] "Volume detached for volume \"prometheus-metric-storage-db\" (UniqueName: \"kubernetes.io/empty-dir/9e35b852-1838-49da-a140-f048788f6635-prometheus-metric-storage-db\") on node \"crc\" DevicePath \"\"" Dec 03 15:14:31 crc kubenswrapper[4805]: I1203 15:14:31.886722 4805 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/secret/9e35b852-1838-49da-a140-f048788f6635-config\") on node \"crc\" DevicePath \"\"" Dec 03 15:14:31 crc kubenswrapper[4805]: I1203 15:14:31.886730 4805 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-z2qxg\" (UniqueName: \"kubernetes.io/projected/9e35b852-1838-49da-a140-f048788f6635-kube-api-access-z2qxg\") on node \"crc\" DevicePath \"\"" Dec 03 15:14:32 crc kubenswrapper[4805]: I1203 15:14:32.526866 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"9e35b852-1838-49da-a140-f048788f6635","Type":"ContainerDied","Data":"111dbdd1292248118175356f2124f21600ff2eb246fe046e9d8fa3e52f21587c"} Dec 03 15:14:32 crc kubenswrapper[4805]: I1203 15:14:32.526955 4805 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/prometheus-metric-storage-0" Dec 03 15:14:32 crc kubenswrapper[4805]: I1203 15:14:32.527261 4805 scope.go:117] "RemoveContainer" containerID="a023a91b7c3b39787116d2c8c1a0da144673bf6d87052338e976a9af4bcc7df5" Dec 03 15:14:32 crc kubenswrapper[4805]: I1203 15:14:32.565329 4805 scope.go:117] "RemoveContainer" containerID="c6738ea6a72122506d9eaaacc460bb54d669b6aa4a8ffdb2c2a9d8829326e697" Dec 03 15:14:32 crc kubenswrapper[4805]: I1203 15:14:32.576871 4805 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/prometheus-metric-storage-0"] Dec 03 15:14:32 crc kubenswrapper[4805]: I1203 15:14:32.586912 4805 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/prometheus-metric-storage-0"] Dec 03 15:14:32 crc kubenswrapper[4805]: I1203 15:14:32.596887 4805 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/prometheus-metric-storage-0"] Dec 03 15:14:32 crc kubenswrapper[4805]: I1203 15:14:32.597023 4805 scope.go:117] "RemoveContainer" containerID="3f32476151f16cefda9576a8077a18738591448ef92afdba7d6ba1cfe707d756" Dec 03 15:14:32 crc kubenswrapper[4805]: E1203 15:14:32.597297 4805 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9e35b852-1838-49da-a140-f048788f6635" containerName="thanos-sidecar" Dec 03 15:14:32 crc kubenswrapper[4805]: I1203 15:14:32.597315 4805 state_mem.go:107] "Deleted CPUSet assignment" podUID="9e35b852-1838-49da-a140-f048788f6635" containerName="thanos-sidecar" Dec 03 15:14:32 crc kubenswrapper[4805]: E1203 15:14:32.597377 4805 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9e35b852-1838-49da-a140-f048788f6635" containerName="config-reloader" Dec 03 15:14:32 crc kubenswrapper[4805]: I1203 15:14:32.597387 4805 state_mem.go:107] "Deleted CPUSet assignment" podUID="9e35b852-1838-49da-a140-f048788f6635" containerName="config-reloader" Dec 03 15:14:32 crc kubenswrapper[4805]: E1203 15:14:32.597402 4805 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fcfe071f-c650-4f17-9130-b2564be1d44f" containerName="extract-utilities" Dec 03 15:14:32 crc kubenswrapper[4805]: I1203 15:14:32.597420 4805 state_mem.go:107] "Deleted CPUSet assignment" podUID="fcfe071f-c650-4f17-9130-b2564be1d44f" containerName="extract-utilities" Dec 03 15:14:32 crc kubenswrapper[4805]: E1203 15:14:32.597428 4805 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9e35b852-1838-49da-a140-f048788f6635" containerName="init-config-reloader" Dec 03 15:14:32 crc kubenswrapper[4805]: I1203 15:14:32.597434 4805 state_mem.go:107] "Deleted CPUSet assignment" podUID="9e35b852-1838-49da-a140-f048788f6635" containerName="init-config-reloader" Dec 03 15:14:32 crc kubenswrapper[4805]: E1203 15:14:32.597449 4805 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fcfe071f-c650-4f17-9130-b2564be1d44f" containerName="registry-server" Dec 03 15:14:32 crc kubenswrapper[4805]: I1203 15:14:32.597455 4805 state_mem.go:107] "Deleted CPUSet assignment" podUID="fcfe071f-c650-4f17-9130-b2564be1d44f" containerName="registry-server" Dec 03 15:14:32 crc kubenswrapper[4805]: E1203 15:14:32.597478 4805 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9e35b852-1838-49da-a140-f048788f6635" containerName="prometheus" Dec 03 15:14:32 crc kubenswrapper[4805]: I1203 15:14:32.597484 4805 state_mem.go:107] "Deleted CPUSet assignment" podUID="9e35b852-1838-49da-a140-f048788f6635" containerName="prometheus" Dec 03 15:14:32 crc kubenswrapper[4805]: E1203 15:14:32.597501 4805 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fcfe071f-c650-4f17-9130-b2564be1d44f" containerName="extract-content" Dec 03 15:14:32 crc kubenswrapper[4805]: I1203 15:14:32.597508 4805 state_mem.go:107] "Deleted CPUSet assignment" podUID="fcfe071f-c650-4f17-9130-b2564be1d44f" containerName="extract-content" Dec 03 15:14:32 crc kubenswrapper[4805]: I1203 15:14:32.597672 4805 memory_manager.go:354] "RemoveStaleState removing state" podUID="9e35b852-1838-49da-a140-f048788f6635" containerName="thanos-sidecar" Dec 03 15:14:32 crc kubenswrapper[4805]: I1203 15:14:32.597691 4805 memory_manager.go:354] "RemoveStaleState removing state" podUID="fcfe071f-c650-4f17-9130-b2564be1d44f" containerName="registry-server" Dec 03 15:14:32 crc kubenswrapper[4805]: I1203 15:14:32.597709 4805 memory_manager.go:354] "RemoveStaleState removing state" podUID="9e35b852-1838-49da-a140-f048788f6635" containerName="config-reloader" Dec 03 15:14:32 crc kubenswrapper[4805]: I1203 15:14:32.597721 4805 memory_manager.go:354] "RemoveStaleState removing state" podUID="9e35b852-1838-49da-a140-f048788f6635" containerName="prometheus" Dec 03 15:14:32 crc kubenswrapper[4805]: I1203 15:14:32.599523 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/prometheus-metric-storage-0" Dec 03 15:14:32 crc kubenswrapper[4805]: I1203 15:14:32.601044 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2cf69241-d762-41ec-9512-84643c4d9dc1-secret-combined-ca-bundle\") pod \"prometheus-metric-storage-0\" (UID: \"2cf69241-d762-41ec-9512-84643c4d9dc1\") " pod="openstack/prometheus-metric-storage-0" Dec 03 15:14:32 crc kubenswrapper[4805]: I1203 15:14:32.601102 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/2cf69241-d762-41ec-9512-84643c4d9dc1-config-out\") pod \"prometheus-metric-storage-0\" (UID: \"2cf69241-d762-41ec-9512-84643c4d9dc1\") " pod="openstack/prometheus-metric-storage-0" Dec 03 15:14:32 crc kubenswrapper[4805]: I1203 15:14:32.601133 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/2cf69241-d762-41ec-9512-84643c4d9dc1-config\") pod \"prometheus-metric-storage-0\" (UID: \"2cf69241-d762-41ec-9512-84643c4d9dc1\") " pod="openstack/prometheus-metric-storage-0" Dec 03 15:14:32 crc kubenswrapper[4805]: I1203 15:14:32.601197 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/2cf69241-d762-41ec-9512-84643c4d9dc1-web-config\") pod \"prometheus-metric-storage-0\" (UID: \"2cf69241-d762-41ec-9512-84643c4d9dc1\") " pod="openstack/prometheus-metric-storage-0" Dec 03 15:14:32 crc kubenswrapper[4805]: I1203 15:14:32.601244 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"prometheus-metric-storage-db\" (UniqueName: \"kubernetes.io/empty-dir/2cf69241-d762-41ec-9512-84643c4d9dc1-prometheus-metric-storage-db\") pod \"prometheus-metric-storage-0\" (UID: \"2cf69241-d762-41ec-9512-84643c4d9dc1\") " pod="openstack/prometheus-metric-storage-0" Dec 03 15:14:32 crc kubenswrapper[4805]: I1203 15:14:32.601283 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"web-config-tls-secret-cert-cert-metric-storage-prometh-dc638c2d\" (UniqueName: \"kubernetes.io/secret/2cf69241-d762-41ec-9512-84643c4d9dc1-web-config-tls-secret-cert-cert-metric-storage-prometh-dc638c2d\") pod \"prometheus-metric-storage-0\" (UID: \"2cf69241-d762-41ec-9512-84643c4d9dc1\") " pod="openstack/prometheus-metric-storage-0" Dec 03 15:14:32 crc kubenswrapper[4805]: I1203 15:14:32.601314 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"thanos-prometheus-http-client-file\" (UniqueName: \"kubernetes.io/secret/2cf69241-d762-41ec-9512-84643c4d9dc1-thanos-prometheus-http-client-file\") pod \"prometheus-metric-storage-0\" (UID: \"2cf69241-d762-41ec-9512-84643c4d9dc1\") " pod="openstack/prometheus-metric-storage-0" Dec 03 15:14:32 crc kubenswrapper[4805]: I1203 15:14:32.601398 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"prometheus-metric-storage-rulefiles-0\" (UniqueName: \"kubernetes.io/configmap/2cf69241-d762-41ec-9512-84643c4d9dc1-prometheus-metric-storage-rulefiles-0\") pod \"prometheus-metric-storage-0\" (UID: \"2cf69241-d762-41ec-9512-84643c4d9dc1\") " pod="openstack/prometheus-metric-storage-0" Dec 03 15:14:32 crc kubenswrapper[4805]: I1203 15:14:32.601430 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"web-config-tls-secret-key-cert-metric-storage-promethe-dc638c2d\" (UniqueName: \"kubernetes.io/secret/2cf69241-d762-41ec-9512-84643c4d9dc1-web-config-tls-secret-key-cert-metric-storage-promethe-dc638c2d\") pod \"prometheus-metric-storage-0\" (UID: \"2cf69241-d762-41ec-9512-84643c4d9dc1\") " pod="openstack/prometheus-metric-storage-0" Dec 03 15:14:32 crc kubenswrapper[4805]: I1203 15:14:32.601454 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hxfz2\" (UniqueName: \"kubernetes.io/projected/2cf69241-d762-41ec-9512-84643c4d9dc1-kube-api-access-hxfz2\") pod \"prometheus-metric-storage-0\" (UID: \"2cf69241-d762-41ec-9512-84643c4d9dc1\") " pod="openstack/prometheus-metric-storage-0" Dec 03 15:14:32 crc kubenswrapper[4805]: I1203 15:14:32.601457 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"prometheus-metric-storage" Dec 03 15:14:32 crc kubenswrapper[4805]: I1203 15:14:32.601547 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"prometheus-metric-storage-thanos-prometheus-http-client-file" Dec 03 15:14:32 crc kubenswrapper[4805]: I1203 15:14:32.601779 4805 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"prometheus-metric-storage-rulefiles-0" Dec 03 15:14:32 crc kubenswrapper[4805]: I1203 15:14:32.601810 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/2cf69241-d762-41ec-9512-84643c4d9dc1-tls-assets\") pod \"prometheus-metric-storage-0\" (UID: \"2cf69241-d762-41ec-9512-84643c4d9dc1\") " pod="openstack/prometheus-metric-storage-0" Dec 03 15:14:32 crc kubenswrapper[4805]: I1203 15:14:32.604071 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-metric-storage-prometheus-svc" Dec 03 15:14:32 crc kubenswrapper[4805]: I1203 15:14:32.604347 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"metric-storage-prometheus-dockercfg-kg4lr" Dec 03 15:14:32 crc kubenswrapper[4805]: I1203 15:14:32.604548 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"prometheus-metric-storage-web-config" Dec 03 15:14:32 crc kubenswrapper[4805]: I1203 15:14:32.612169 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"prometheus-metric-storage-tls-assets-0" Dec 03 15:14:32 crc kubenswrapper[4805]: I1203 15:14:32.628695 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/prometheus-metric-storage-0"] Dec 03 15:14:32 crc kubenswrapper[4805]: I1203 15:14:32.636045 4805 scope.go:117] "RemoveContainer" containerID="e4497f355686372593cee84b0a311f2e56fa24ad7f3aed9b51b7bbfbb220ce7c" Dec 03 15:14:32 crc kubenswrapper[4805]: I1203 15:14:32.702507 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/2cf69241-d762-41ec-9512-84643c4d9dc1-web-config\") pod \"prometheus-metric-storage-0\" (UID: \"2cf69241-d762-41ec-9512-84643c4d9dc1\") " pod="openstack/prometheus-metric-storage-0" Dec 03 15:14:32 crc kubenswrapper[4805]: I1203 15:14:32.702890 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"prometheus-metric-storage-db\" (UniqueName: \"kubernetes.io/empty-dir/2cf69241-d762-41ec-9512-84643c4d9dc1-prometheus-metric-storage-db\") pod \"prometheus-metric-storage-0\" (UID: \"2cf69241-d762-41ec-9512-84643c4d9dc1\") " pod="openstack/prometheus-metric-storage-0" Dec 03 15:14:32 crc kubenswrapper[4805]: I1203 15:14:32.702918 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"web-config-tls-secret-cert-cert-metric-storage-prometh-dc638c2d\" (UniqueName: \"kubernetes.io/secret/2cf69241-d762-41ec-9512-84643c4d9dc1-web-config-tls-secret-cert-cert-metric-storage-prometh-dc638c2d\") pod \"prometheus-metric-storage-0\" (UID: \"2cf69241-d762-41ec-9512-84643c4d9dc1\") " pod="openstack/prometheus-metric-storage-0" Dec 03 15:14:32 crc kubenswrapper[4805]: I1203 15:14:32.702953 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"thanos-prometheus-http-client-file\" (UniqueName: \"kubernetes.io/secret/2cf69241-d762-41ec-9512-84643c4d9dc1-thanos-prometheus-http-client-file\") pod \"prometheus-metric-storage-0\" (UID: \"2cf69241-d762-41ec-9512-84643c4d9dc1\") " pod="openstack/prometheus-metric-storage-0" Dec 03 15:14:32 crc kubenswrapper[4805]: I1203 15:14:32.702991 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"prometheus-metric-storage-rulefiles-0\" (UniqueName: \"kubernetes.io/configmap/2cf69241-d762-41ec-9512-84643c4d9dc1-prometheus-metric-storage-rulefiles-0\") pod \"prometheus-metric-storage-0\" (UID: \"2cf69241-d762-41ec-9512-84643c4d9dc1\") " pod="openstack/prometheus-metric-storage-0" Dec 03 15:14:32 crc kubenswrapper[4805]: I1203 15:14:32.703675 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"web-config-tls-secret-key-cert-metric-storage-promethe-dc638c2d\" (UniqueName: \"kubernetes.io/secret/2cf69241-d762-41ec-9512-84643c4d9dc1-web-config-tls-secret-key-cert-metric-storage-promethe-dc638c2d\") pod \"prometheus-metric-storage-0\" (UID: \"2cf69241-d762-41ec-9512-84643c4d9dc1\") " pod="openstack/prometheus-metric-storage-0" Dec 03 15:14:32 crc kubenswrapper[4805]: I1203 15:14:32.703703 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hxfz2\" (UniqueName: \"kubernetes.io/projected/2cf69241-d762-41ec-9512-84643c4d9dc1-kube-api-access-hxfz2\") pod \"prometheus-metric-storage-0\" (UID: \"2cf69241-d762-41ec-9512-84643c4d9dc1\") " pod="openstack/prometheus-metric-storage-0" Dec 03 15:14:32 crc kubenswrapper[4805]: I1203 15:14:32.703818 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"prometheus-metric-storage-db\" (UniqueName: \"kubernetes.io/empty-dir/2cf69241-d762-41ec-9512-84643c4d9dc1-prometheus-metric-storage-db\") pod \"prometheus-metric-storage-0\" (UID: \"2cf69241-d762-41ec-9512-84643c4d9dc1\") " pod="openstack/prometheus-metric-storage-0" Dec 03 15:14:32 crc kubenswrapper[4805]: I1203 15:14:32.703827 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/2cf69241-d762-41ec-9512-84643c4d9dc1-tls-assets\") pod \"prometheus-metric-storage-0\" (UID: \"2cf69241-d762-41ec-9512-84643c4d9dc1\") " pod="openstack/prometheus-metric-storage-0" Dec 03 15:14:32 crc kubenswrapper[4805]: I1203 15:14:32.703984 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"prometheus-metric-storage-rulefiles-0\" (UniqueName: \"kubernetes.io/configmap/2cf69241-d762-41ec-9512-84643c4d9dc1-prometheus-metric-storage-rulefiles-0\") pod \"prometheus-metric-storage-0\" (UID: \"2cf69241-d762-41ec-9512-84643c4d9dc1\") " pod="openstack/prometheus-metric-storage-0" Dec 03 15:14:32 crc kubenswrapper[4805]: I1203 15:14:32.704178 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2cf69241-d762-41ec-9512-84643c4d9dc1-secret-combined-ca-bundle\") pod \"prometheus-metric-storage-0\" (UID: \"2cf69241-d762-41ec-9512-84643c4d9dc1\") " pod="openstack/prometheus-metric-storage-0" Dec 03 15:14:32 crc kubenswrapper[4805]: I1203 15:14:32.704241 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/2cf69241-d762-41ec-9512-84643c4d9dc1-config-out\") pod \"prometheus-metric-storage-0\" (UID: \"2cf69241-d762-41ec-9512-84643c4d9dc1\") " pod="openstack/prometheus-metric-storage-0" Dec 03 15:14:32 crc kubenswrapper[4805]: I1203 15:14:32.704269 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/2cf69241-d762-41ec-9512-84643c4d9dc1-config\") pod \"prometheus-metric-storage-0\" (UID: \"2cf69241-d762-41ec-9512-84643c4d9dc1\") " pod="openstack/prometheus-metric-storage-0" Dec 03 15:14:32 crc kubenswrapper[4805]: I1203 15:14:32.708872 4805 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9e35b852-1838-49da-a140-f048788f6635" path="/var/lib/kubelet/pods/9e35b852-1838-49da-a140-f048788f6635/volumes" Dec 03 15:14:32 crc kubenswrapper[4805]: I1203 15:14:32.710636 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/secret/2cf69241-d762-41ec-9512-84643c4d9dc1-config\") pod \"prometheus-metric-storage-0\" (UID: \"2cf69241-d762-41ec-9512-84643c4d9dc1\") " pod="openstack/prometheus-metric-storage-0" Dec 03 15:14:32 crc kubenswrapper[4805]: I1203 15:14:32.713370 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/2cf69241-d762-41ec-9512-84643c4d9dc1-web-config\") pod \"prometheus-metric-storage-0\" (UID: \"2cf69241-d762-41ec-9512-84643c4d9dc1\") " pod="openstack/prometheus-metric-storage-0" Dec 03 15:14:32 crc kubenswrapper[4805]: I1203 15:14:32.715654 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/2cf69241-d762-41ec-9512-84643c4d9dc1-config-out\") pod \"prometheus-metric-storage-0\" (UID: \"2cf69241-d762-41ec-9512-84643c4d9dc1\") " pod="openstack/prometheus-metric-storage-0" Dec 03 15:14:32 crc kubenswrapper[4805]: I1203 15:14:32.716391 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"thanos-prometheus-http-client-file\" (UniqueName: \"kubernetes.io/secret/2cf69241-d762-41ec-9512-84643c4d9dc1-thanos-prometheus-http-client-file\") pod \"prometheus-metric-storage-0\" (UID: \"2cf69241-d762-41ec-9512-84643c4d9dc1\") " pod="openstack/prometheus-metric-storage-0" Dec 03 15:14:32 crc kubenswrapper[4805]: I1203 15:14:32.723075 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"web-config-tls-secret-key-cert-metric-storage-promethe-dc638c2d\" (UniqueName: \"kubernetes.io/secret/2cf69241-d762-41ec-9512-84643c4d9dc1-web-config-tls-secret-key-cert-metric-storage-promethe-dc638c2d\") pod \"prometheus-metric-storage-0\" (UID: \"2cf69241-d762-41ec-9512-84643c4d9dc1\") " pod="openstack/prometheus-metric-storage-0" Dec 03 15:14:32 crc kubenswrapper[4805]: I1203 15:14:32.724115 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"web-config-tls-secret-cert-cert-metric-storage-prometh-dc638c2d\" (UniqueName: \"kubernetes.io/secret/2cf69241-d762-41ec-9512-84643c4d9dc1-web-config-tls-secret-cert-cert-metric-storage-prometh-dc638c2d\") pod \"prometheus-metric-storage-0\" (UID: \"2cf69241-d762-41ec-9512-84643c4d9dc1\") " pod="openstack/prometheus-metric-storage-0" Dec 03 15:14:32 crc kubenswrapper[4805]: I1203 15:14:32.725371 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2cf69241-d762-41ec-9512-84643c4d9dc1-secret-combined-ca-bundle\") pod \"prometheus-metric-storage-0\" (UID: \"2cf69241-d762-41ec-9512-84643c4d9dc1\") " pod="openstack/prometheus-metric-storage-0" Dec 03 15:14:32 crc kubenswrapper[4805]: I1203 15:14:32.727733 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/2cf69241-d762-41ec-9512-84643c4d9dc1-tls-assets\") pod \"prometheus-metric-storage-0\" (UID: \"2cf69241-d762-41ec-9512-84643c4d9dc1\") " pod="openstack/prometheus-metric-storage-0" Dec 03 15:14:32 crc kubenswrapper[4805]: I1203 15:14:32.729883 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hxfz2\" (UniqueName: \"kubernetes.io/projected/2cf69241-d762-41ec-9512-84643c4d9dc1-kube-api-access-hxfz2\") pod \"prometheus-metric-storage-0\" (UID: \"2cf69241-d762-41ec-9512-84643c4d9dc1\") " pod="openstack/prometheus-metric-storage-0" Dec 03 15:14:32 crc kubenswrapper[4805]: I1203 15:14:32.947670 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/prometheus-metric-storage-0" Dec 03 15:14:33 crc kubenswrapper[4805]: I1203 15:14:33.944526 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/prometheus-metric-storage-0"] Dec 03 15:14:34 crc kubenswrapper[4805]: I1203 15:14:34.551409 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"2cf69241-d762-41ec-9512-84643c4d9dc1","Type":"ContainerStarted","Data":"b8edd55a8faf76202b236e29635a42fa178c171fdeddd0ee85f4146b40011ed1"} Dec 03 15:14:37 crc kubenswrapper[4805]: I1203 15:14:37.602695 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"2cf69241-d762-41ec-9512-84643c4d9dc1","Type":"ContainerStarted","Data":"3843c6aed9e7fa07d9ad18f5639dd45d3bc93af1a6ca7ac33d9098365cfd48cf"} Dec 03 15:14:41 crc kubenswrapper[4805]: I1203 15:14:41.700216 4805 scope.go:117] "RemoveContainer" containerID="cbf73c6c9ba740c0a523c84ca23c0851292d6eba7a0fb47d38599acca964e9cc" Dec 03 15:14:41 crc kubenswrapper[4805]: E1203 15:14:41.701810 4805 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-gskh4_openshift-machine-config-operator(6098937f-e3f6-45e8-a647-4994a79cd711)\"" pod="openshift-machine-config-operator/machine-config-daemon-gskh4" podUID="6098937f-e3f6-45e8-a647-4994a79cd711" Dec 03 15:14:41 crc kubenswrapper[4805]: E1203 15:14:41.724488 4805 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod9e35b852_1838_49da_a140_f048788f6635.slice/crio-3f32476151f16cefda9576a8077a18738591448ef92afdba7d6ba1cfe707d756.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod9e35b852_1838_49da_a140_f048788f6635.slice/crio-conmon-3f32476151f16cefda9576a8077a18738591448ef92afdba7d6ba1cfe707d756.scope\": RecentStats: unable to find data in memory cache]" Dec 03 15:14:45 crc kubenswrapper[4805]: I1203 15:14:45.675278 4805 generic.go:334] "Generic (PLEG): container finished" podID="2cf69241-d762-41ec-9512-84643c4d9dc1" containerID="3843c6aed9e7fa07d9ad18f5639dd45d3bc93af1a6ca7ac33d9098365cfd48cf" exitCode=0 Dec 03 15:14:45 crc kubenswrapper[4805]: I1203 15:14:45.675378 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"2cf69241-d762-41ec-9512-84643c4d9dc1","Type":"ContainerDied","Data":"3843c6aed9e7fa07d9ad18f5639dd45d3bc93af1a6ca7ac33d9098365cfd48cf"} Dec 03 15:14:46 crc kubenswrapper[4805]: I1203 15:14:46.552616 4805 scope.go:117] "RemoveContainer" containerID="048c55494b984dd592e98c8291e407fc1ebaaebdb6f45a57dccf0eed800ea2f2" Dec 03 15:14:46 crc kubenswrapper[4805]: I1203 15:14:46.579014 4805 scope.go:117] "RemoveContainer" containerID="e0fecc8750cd2830f57ad4347ef1f659b9c5365badd25705fea3633133b1c604" Dec 03 15:14:46 crc kubenswrapper[4805]: I1203 15:14:46.599540 4805 scope.go:117] "RemoveContainer" containerID="38d5882460472e27d4ce5803cfbdb0352f15ab198d24ac83c6062b2c7580f76a" Dec 03 15:14:46 crc kubenswrapper[4805]: I1203 15:14:46.616423 4805 scope.go:117] "RemoveContainer" containerID="eabbde92660ffc9b7960552812807370f6c0f71b2e1e3967895c09dbcdec0ee4" Dec 03 15:14:47 crc kubenswrapper[4805]: I1203 15:14:47.698509 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"2cf69241-d762-41ec-9512-84643c4d9dc1","Type":"ContainerStarted","Data":"ca3d35c268b7f75b189ac1c1a68e98e65cbf92c246a149125f0021b8cabd752e"} Dec 03 15:14:48 crc kubenswrapper[4805]: I1203 15:14:48.138047 4805 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-7hxsd"] Dec 03 15:14:48 crc kubenswrapper[4805]: I1203 15:14:48.141103 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-7hxsd" Dec 03 15:14:48 crc kubenswrapper[4805]: I1203 15:14:48.174886 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-7hxsd"] Dec 03 15:14:48 crc kubenswrapper[4805]: I1203 15:14:48.230642 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2slcx\" (UniqueName: \"kubernetes.io/projected/3ee20797-d380-4b2b-b604-ca4d159b3c07-kube-api-access-2slcx\") pod \"community-operators-7hxsd\" (UID: \"3ee20797-d380-4b2b-b604-ca4d159b3c07\") " pod="openshift-marketplace/community-operators-7hxsd" Dec 03 15:14:48 crc kubenswrapper[4805]: I1203 15:14:48.230690 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3ee20797-d380-4b2b-b604-ca4d159b3c07-utilities\") pod \"community-operators-7hxsd\" (UID: \"3ee20797-d380-4b2b-b604-ca4d159b3c07\") " pod="openshift-marketplace/community-operators-7hxsd" Dec 03 15:14:48 crc kubenswrapper[4805]: I1203 15:14:48.230941 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3ee20797-d380-4b2b-b604-ca4d159b3c07-catalog-content\") pod \"community-operators-7hxsd\" (UID: \"3ee20797-d380-4b2b-b604-ca4d159b3c07\") " pod="openshift-marketplace/community-operators-7hxsd" Dec 03 15:14:48 crc kubenswrapper[4805]: I1203 15:14:48.333387 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2slcx\" (UniqueName: \"kubernetes.io/projected/3ee20797-d380-4b2b-b604-ca4d159b3c07-kube-api-access-2slcx\") pod \"community-operators-7hxsd\" (UID: \"3ee20797-d380-4b2b-b604-ca4d159b3c07\") " pod="openshift-marketplace/community-operators-7hxsd" Dec 03 15:14:48 crc kubenswrapper[4805]: I1203 15:14:48.333470 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3ee20797-d380-4b2b-b604-ca4d159b3c07-utilities\") pod \"community-operators-7hxsd\" (UID: \"3ee20797-d380-4b2b-b604-ca4d159b3c07\") " pod="openshift-marketplace/community-operators-7hxsd" Dec 03 15:14:48 crc kubenswrapper[4805]: I1203 15:14:48.333645 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3ee20797-d380-4b2b-b604-ca4d159b3c07-catalog-content\") pod \"community-operators-7hxsd\" (UID: \"3ee20797-d380-4b2b-b604-ca4d159b3c07\") " pod="openshift-marketplace/community-operators-7hxsd" Dec 03 15:14:48 crc kubenswrapper[4805]: I1203 15:14:48.334253 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3ee20797-d380-4b2b-b604-ca4d159b3c07-utilities\") pod \"community-operators-7hxsd\" (UID: \"3ee20797-d380-4b2b-b604-ca4d159b3c07\") " pod="openshift-marketplace/community-operators-7hxsd" Dec 03 15:14:48 crc kubenswrapper[4805]: I1203 15:14:48.334312 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3ee20797-d380-4b2b-b604-ca4d159b3c07-catalog-content\") pod \"community-operators-7hxsd\" (UID: \"3ee20797-d380-4b2b-b604-ca4d159b3c07\") " pod="openshift-marketplace/community-operators-7hxsd" Dec 03 15:14:48 crc kubenswrapper[4805]: I1203 15:14:48.362706 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2slcx\" (UniqueName: \"kubernetes.io/projected/3ee20797-d380-4b2b-b604-ca4d159b3c07-kube-api-access-2slcx\") pod \"community-operators-7hxsd\" (UID: \"3ee20797-d380-4b2b-b604-ca4d159b3c07\") " pod="openshift-marketplace/community-operators-7hxsd" Dec 03 15:14:48 crc kubenswrapper[4805]: I1203 15:14:48.482240 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-7hxsd" Dec 03 15:14:49 crc kubenswrapper[4805]: I1203 15:14:49.023375 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-7hxsd"] Dec 03 15:14:49 crc kubenswrapper[4805]: I1203 15:14:49.722178 4805 generic.go:334] "Generic (PLEG): container finished" podID="3ee20797-d380-4b2b-b604-ca4d159b3c07" containerID="a542bcdef97581d5bca30ebf1d9dada3ea84509e2ecbefd9e2169e752bf0e672" exitCode=0 Dec 03 15:14:49 crc kubenswrapper[4805]: I1203 15:14:49.722392 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-7hxsd" event={"ID":"3ee20797-d380-4b2b-b604-ca4d159b3c07","Type":"ContainerDied","Data":"a542bcdef97581d5bca30ebf1d9dada3ea84509e2ecbefd9e2169e752bf0e672"} Dec 03 15:14:49 crc kubenswrapper[4805]: I1203 15:14:49.722516 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-7hxsd" event={"ID":"3ee20797-d380-4b2b-b604-ca4d159b3c07","Type":"ContainerStarted","Data":"5214c0a85068b4589efd58e3cab66d2eb2188f11f367ae2c3616f8b78adbf6e1"} Dec 03 15:14:51 crc kubenswrapper[4805]: I1203 15:14:51.745115 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"2cf69241-d762-41ec-9512-84643c4d9dc1","Type":"ContainerStarted","Data":"ed39b4fe7b04e307e83b798baedf330746a895727fe4e48d546f8f0d2ec68a75"} Dec 03 15:14:51 crc kubenswrapper[4805]: I1203 15:14:51.748127 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-7hxsd" event={"ID":"3ee20797-d380-4b2b-b604-ca4d159b3c07","Type":"ContainerStarted","Data":"e78d8c40bff7cfd30b394b8b8bb1c090cf0448f912c63389093ca3769616303e"} Dec 03 15:14:51 crc kubenswrapper[4805]: E1203 15:14:51.997560 4805 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod9e35b852_1838_49da_a140_f048788f6635.slice/crio-3f32476151f16cefda9576a8077a18738591448ef92afdba7d6ba1cfe707d756.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod9e35b852_1838_49da_a140_f048788f6635.slice/crio-conmon-3f32476151f16cefda9576a8077a18738591448ef92afdba7d6ba1cfe707d756.scope\": RecentStats: unable to find data in memory cache]" Dec 03 15:14:52 crc kubenswrapper[4805]: I1203 15:14:52.762050 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"2cf69241-d762-41ec-9512-84643c4d9dc1","Type":"ContainerStarted","Data":"6447db5e808e7441091d6706ba346008f8a6dfb69dfcbbd1a3701c7c59f3728e"} Dec 03 15:14:52 crc kubenswrapper[4805]: I1203 15:14:52.801163 4805 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/prometheus-metric-storage-0" podStartSLOduration=20.801143558 podStartE2EDuration="20.801143558s" podCreationTimestamp="2025-12-03 15:14:32 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 15:14:52.793742928 +0000 UTC m=+3922.456659851" watchObservedRunningTime="2025-12-03 15:14:52.801143558 +0000 UTC m=+3922.464060481" Dec 03 15:14:52 crc kubenswrapper[4805]: I1203 15:14:52.948645 4805 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/prometheus-metric-storage-0" Dec 03 15:14:54 crc kubenswrapper[4805]: I1203 15:14:54.784222 4805 generic.go:334] "Generic (PLEG): container finished" podID="3ee20797-d380-4b2b-b604-ca4d159b3c07" containerID="e78d8c40bff7cfd30b394b8b8bb1c090cf0448f912c63389093ca3769616303e" exitCode=0 Dec 03 15:14:54 crc kubenswrapper[4805]: I1203 15:14:54.784321 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-7hxsd" event={"ID":"3ee20797-d380-4b2b-b604-ca4d159b3c07","Type":"ContainerDied","Data":"e78d8c40bff7cfd30b394b8b8bb1c090cf0448f912c63389093ca3769616303e"} Dec 03 15:14:55 crc kubenswrapper[4805]: I1203 15:14:55.694753 4805 scope.go:117] "RemoveContainer" containerID="cbf73c6c9ba740c0a523c84ca23c0851292d6eba7a0fb47d38599acca964e9cc" Dec 03 15:14:55 crc kubenswrapper[4805]: E1203 15:14:55.695495 4805 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-gskh4_openshift-machine-config-operator(6098937f-e3f6-45e8-a647-4994a79cd711)\"" pod="openshift-machine-config-operator/machine-config-daemon-gskh4" podUID="6098937f-e3f6-45e8-a647-4994a79cd711" Dec 03 15:14:55 crc kubenswrapper[4805]: I1203 15:14:55.798554 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-7hxsd" event={"ID":"3ee20797-d380-4b2b-b604-ca4d159b3c07","Type":"ContainerStarted","Data":"70e2a0e349297cac98c24271fca3809a1b64a17320db20cbb8bd5eda80a20364"} Dec 03 15:14:55 crc kubenswrapper[4805]: I1203 15:14:55.824178 4805 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-7hxsd" podStartSLOduration=2.2322615949999998 podStartE2EDuration="7.824157501s" podCreationTimestamp="2025-12-03 15:14:48 +0000 UTC" firstStartedPulling="2025-12-03 15:14:49.723913553 +0000 UTC m=+3919.386830486" lastFinishedPulling="2025-12-03 15:14:55.315809469 +0000 UTC m=+3924.978726392" observedRunningTime="2025-12-03 15:14:55.816085114 +0000 UTC m=+3925.479002047" watchObservedRunningTime="2025-12-03 15:14:55.824157501 +0000 UTC m=+3925.487074424" Dec 03 15:14:58 crc kubenswrapper[4805]: I1203 15:14:58.483219 4805 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-7hxsd" Dec 03 15:14:58 crc kubenswrapper[4805]: I1203 15:14:58.483639 4805 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-7hxsd" Dec 03 15:14:58 crc kubenswrapper[4805]: I1203 15:14:58.539445 4805 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-7hxsd" Dec 03 15:15:00 crc kubenswrapper[4805]: I1203 15:15:00.164769 4805 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29412915-wzdgx"] Dec 03 15:15:00 crc kubenswrapper[4805]: I1203 15:15:00.166722 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29412915-wzdgx" Dec 03 15:15:00 crc kubenswrapper[4805]: I1203 15:15:00.170517 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Dec 03 15:15:00 crc kubenswrapper[4805]: I1203 15:15:00.170636 4805 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Dec 03 15:15:00 crc kubenswrapper[4805]: I1203 15:15:00.203446 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29412915-wzdgx"] Dec 03 15:15:00 crc kubenswrapper[4805]: I1203 15:15:00.270237 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pjsx8\" (UniqueName: \"kubernetes.io/projected/4d6fdcb7-4fb1-4739-aa7e-71c91a241733-kube-api-access-pjsx8\") pod \"collect-profiles-29412915-wzdgx\" (UID: \"4d6fdcb7-4fb1-4739-aa7e-71c91a241733\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29412915-wzdgx" Dec 03 15:15:00 crc kubenswrapper[4805]: I1203 15:15:00.270332 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/4d6fdcb7-4fb1-4739-aa7e-71c91a241733-config-volume\") pod \"collect-profiles-29412915-wzdgx\" (UID: \"4d6fdcb7-4fb1-4739-aa7e-71c91a241733\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29412915-wzdgx" Dec 03 15:15:00 crc kubenswrapper[4805]: I1203 15:15:00.270788 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/4d6fdcb7-4fb1-4739-aa7e-71c91a241733-secret-volume\") pod \"collect-profiles-29412915-wzdgx\" (UID: \"4d6fdcb7-4fb1-4739-aa7e-71c91a241733\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29412915-wzdgx" Dec 03 15:15:00 crc kubenswrapper[4805]: I1203 15:15:00.373143 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pjsx8\" (UniqueName: \"kubernetes.io/projected/4d6fdcb7-4fb1-4739-aa7e-71c91a241733-kube-api-access-pjsx8\") pod \"collect-profiles-29412915-wzdgx\" (UID: \"4d6fdcb7-4fb1-4739-aa7e-71c91a241733\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29412915-wzdgx" Dec 03 15:15:00 crc kubenswrapper[4805]: I1203 15:15:00.373228 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/4d6fdcb7-4fb1-4739-aa7e-71c91a241733-config-volume\") pod \"collect-profiles-29412915-wzdgx\" (UID: \"4d6fdcb7-4fb1-4739-aa7e-71c91a241733\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29412915-wzdgx" Dec 03 15:15:00 crc kubenswrapper[4805]: I1203 15:15:00.373350 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/4d6fdcb7-4fb1-4739-aa7e-71c91a241733-secret-volume\") pod \"collect-profiles-29412915-wzdgx\" (UID: \"4d6fdcb7-4fb1-4739-aa7e-71c91a241733\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29412915-wzdgx" Dec 03 15:15:00 crc kubenswrapper[4805]: I1203 15:15:00.374443 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/4d6fdcb7-4fb1-4739-aa7e-71c91a241733-config-volume\") pod \"collect-profiles-29412915-wzdgx\" (UID: \"4d6fdcb7-4fb1-4739-aa7e-71c91a241733\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29412915-wzdgx" Dec 03 15:15:00 crc kubenswrapper[4805]: I1203 15:15:00.380821 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/4d6fdcb7-4fb1-4739-aa7e-71c91a241733-secret-volume\") pod \"collect-profiles-29412915-wzdgx\" (UID: \"4d6fdcb7-4fb1-4739-aa7e-71c91a241733\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29412915-wzdgx" Dec 03 15:15:00 crc kubenswrapper[4805]: I1203 15:15:00.395145 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pjsx8\" (UniqueName: \"kubernetes.io/projected/4d6fdcb7-4fb1-4739-aa7e-71c91a241733-kube-api-access-pjsx8\") pod \"collect-profiles-29412915-wzdgx\" (UID: \"4d6fdcb7-4fb1-4739-aa7e-71c91a241733\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29412915-wzdgx" Dec 03 15:15:00 crc kubenswrapper[4805]: I1203 15:15:00.496293 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29412915-wzdgx" Dec 03 15:15:00 crc kubenswrapper[4805]: I1203 15:15:00.976254 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29412915-wzdgx"] Dec 03 15:15:00 crc kubenswrapper[4805]: W1203 15:15:00.979859 4805 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod4d6fdcb7_4fb1_4739_aa7e_71c91a241733.slice/crio-2f1a9fefb8e232ece7a3f271abeceda9e565bf2157d2d27dcb26d8c3ee833920 WatchSource:0}: Error finding container 2f1a9fefb8e232ece7a3f271abeceda9e565bf2157d2d27dcb26d8c3ee833920: Status 404 returned error can't find the container with id 2f1a9fefb8e232ece7a3f271abeceda9e565bf2157d2d27dcb26d8c3ee833920 Dec 03 15:15:01 crc kubenswrapper[4805]: I1203 15:15:01.884433 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29412915-wzdgx" event={"ID":"4d6fdcb7-4fb1-4739-aa7e-71c91a241733","Type":"ContainerStarted","Data":"933fe1a6ff56a31d99abd6cf5171c7d0fd2a7f77542a514925312251634050dd"} Dec 03 15:15:01 crc kubenswrapper[4805]: I1203 15:15:01.885127 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29412915-wzdgx" event={"ID":"4d6fdcb7-4fb1-4739-aa7e-71c91a241733","Type":"ContainerStarted","Data":"2f1a9fefb8e232ece7a3f271abeceda9e565bf2157d2d27dcb26d8c3ee833920"} Dec 03 15:15:01 crc kubenswrapper[4805]: I1203 15:15:01.908202 4805 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/collect-profiles-29412915-wzdgx" podStartSLOduration=1.9081784019999999 podStartE2EDuration="1.908178402s" podCreationTimestamp="2025-12-03 15:15:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 15:15:01.902748036 +0000 UTC m=+3931.565664959" watchObservedRunningTime="2025-12-03 15:15:01.908178402 +0000 UTC m=+3931.571095335" Dec 03 15:15:02 crc kubenswrapper[4805]: E1203 15:15:02.300158 4805 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod9e35b852_1838_49da_a140_f048788f6635.slice/crio-conmon-3f32476151f16cefda9576a8077a18738591448ef92afdba7d6ba1cfe707d756.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod9e35b852_1838_49da_a140_f048788f6635.slice/crio-3f32476151f16cefda9576a8077a18738591448ef92afdba7d6ba1cfe707d756.scope\": RecentStats: unable to find data in memory cache]" Dec 03 15:15:02 crc kubenswrapper[4805]: I1203 15:15:02.895424 4805 generic.go:334] "Generic (PLEG): container finished" podID="4d6fdcb7-4fb1-4739-aa7e-71c91a241733" containerID="933fe1a6ff56a31d99abd6cf5171c7d0fd2a7f77542a514925312251634050dd" exitCode=0 Dec 03 15:15:02 crc kubenswrapper[4805]: I1203 15:15:02.895643 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29412915-wzdgx" event={"ID":"4d6fdcb7-4fb1-4739-aa7e-71c91a241733","Type":"ContainerDied","Data":"933fe1a6ff56a31d99abd6cf5171c7d0fd2a7f77542a514925312251634050dd"} Dec 03 15:15:02 crc kubenswrapper[4805]: I1203 15:15:02.948252 4805 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/prometheus-metric-storage-0" Dec 03 15:15:02 crc kubenswrapper[4805]: I1203 15:15:02.953824 4805 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/prometheus-metric-storage-0" Dec 03 15:15:03 crc kubenswrapper[4805]: I1203 15:15:03.916999 4805 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/prometheus-metric-storage-0" Dec 03 15:15:04 crc kubenswrapper[4805]: I1203 15:15:04.373982 4805 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29412915-wzdgx" Dec 03 15:15:04 crc kubenswrapper[4805]: I1203 15:15:04.468593 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/4d6fdcb7-4fb1-4739-aa7e-71c91a241733-config-volume\") pod \"4d6fdcb7-4fb1-4739-aa7e-71c91a241733\" (UID: \"4d6fdcb7-4fb1-4739-aa7e-71c91a241733\") " Dec 03 15:15:04 crc kubenswrapper[4805]: I1203 15:15:04.468715 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/4d6fdcb7-4fb1-4739-aa7e-71c91a241733-secret-volume\") pod \"4d6fdcb7-4fb1-4739-aa7e-71c91a241733\" (UID: \"4d6fdcb7-4fb1-4739-aa7e-71c91a241733\") " Dec 03 15:15:04 crc kubenswrapper[4805]: I1203 15:15:04.468825 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pjsx8\" (UniqueName: \"kubernetes.io/projected/4d6fdcb7-4fb1-4739-aa7e-71c91a241733-kube-api-access-pjsx8\") pod \"4d6fdcb7-4fb1-4739-aa7e-71c91a241733\" (UID: \"4d6fdcb7-4fb1-4739-aa7e-71c91a241733\") " Dec 03 15:15:04 crc kubenswrapper[4805]: I1203 15:15:04.471159 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4d6fdcb7-4fb1-4739-aa7e-71c91a241733-config-volume" (OuterVolumeSpecName: "config-volume") pod "4d6fdcb7-4fb1-4739-aa7e-71c91a241733" (UID: "4d6fdcb7-4fb1-4739-aa7e-71c91a241733"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 15:15:04 crc kubenswrapper[4805]: I1203 15:15:04.475442 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4d6fdcb7-4fb1-4739-aa7e-71c91a241733-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "4d6fdcb7-4fb1-4739-aa7e-71c91a241733" (UID: "4d6fdcb7-4fb1-4739-aa7e-71c91a241733"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 15:15:04 crc kubenswrapper[4805]: I1203 15:15:04.475624 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4d6fdcb7-4fb1-4739-aa7e-71c91a241733-kube-api-access-pjsx8" (OuterVolumeSpecName: "kube-api-access-pjsx8") pod "4d6fdcb7-4fb1-4739-aa7e-71c91a241733" (UID: "4d6fdcb7-4fb1-4739-aa7e-71c91a241733"). InnerVolumeSpecName "kube-api-access-pjsx8". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 15:15:04 crc kubenswrapper[4805]: I1203 15:15:04.571574 4805 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/4d6fdcb7-4fb1-4739-aa7e-71c91a241733-config-volume\") on node \"crc\" DevicePath \"\"" Dec 03 15:15:04 crc kubenswrapper[4805]: I1203 15:15:04.571605 4805 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/4d6fdcb7-4fb1-4739-aa7e-71c91a241733-secret-volume\") on node \"crc\" DevicePath \"\"" Dec 03 15:15:04 crc kubenswrapper[4805]: I1203 15:15:04.571615 4805 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pjsx8\" (UniqueName: \"kubernetes.io/projected/4d6fdcb7-4fb1-4739-aa7e-71c91a241733-kube-api-access-pjsx8\") on node \"crc\" DevicePath \"\"" Dec 03 15:15:04 crc kubenswrapper[4805]: I1203 15:15:04.923852 4805 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29412915-wzdgx" Dec 03 15:15:04 crc kubenswrapper[4805]: I1203 15:15:04.923882 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29412915-wzdgx" event={"ID":"4d6fdcb7-4fb1-4739-aa7e-71c91a241733","Type":"ContainerDied","Data":"2f1a9fefb8e232ece7a3f271abeceda9e565bf2157d2d27dcb26d8c3ee833920"} Dec 03 15:15:04 crc kubenswrapper[4805]: I1203 15:15:04.924253 4805 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="2f1a9fefb8e232ece7a3f271abeceda9e565bf2157d2d27dcb26d8c3ee833920" Dec 03 15:15:05 crc kubenswrapper[4805]: I1203 15:15:05.161894 4805 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29412870-46tbm"] Dec 03 15:15:05 crc kubenswrapper[4805]: I1203 15:15:05.171817 4805 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29412870-46tbm"] Dec 03 15:15:06 crc kubenswrapper[4805]: I1203 15:15:06.708993 4805 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5aae8010-7566-4f0b-ba0f-8d905bc9e455" path="/var/lib/kubelet/pods/5aae8010-7566-4f0b-ba0f-8d905bc9e455/volumes" Dec 03 15:15:08 crc kubenswrapper[4805]: I1203 15:15:08.538074 4805 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-7hxsd" Dec 03 15:15:08 crc kubenswrapper[4805]: I1203 15:15:08.588147 4805 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-7hxsd"] Dec 03 15:15:08 crc kubenswrapper[4805]: I1203 15:15:08.962450 4805 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-7hxsd" podUID="3ee20797-d380-4b2b-b604-ca4d159b3c07" containerName="registry-server" containerID="cri-o://70e2a0e349297cac98c24271fca3809a1b64a17320db20cbb8bd5eda80a20364" gracePeriod=2 Dec 03 15:15:09 crc kubenswrapper[4805]: I1203 15:15:09.508280 4805 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-7hxsd" Dec 03 15:15:09 crc kubenswrapper[4805]: I1203 15:15:09.584399 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3ee20797-d380-4b2b-b604-ca4d159b3c07-utilities\") pod \"3ee20797-d380-4b2b-b604-ca4d159b3c07\" (UID: \"3ee20797-d380-4b2b-b604-ca4d159b3c07\") " Dec 03 15:15:09 crc kubenswrapper[4805]: I1203 15:15:09.584749 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3ee20797-d380-4b2b-b604-ca4d159b3c07-catalog-content\") pod \"3ee20797-d380-4b2b-b604-ca4d159b3c07\" (UID: \"3ee20797-d380-4b2b-b604-ca4d159b3c07\") " Dec 03 15:15:09 crc kubenswrapper[4805]: I1203 15:15:09.584867 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2slcx\" (UniqueName: \"kubernetes.io/projected/3ee20797-d380-4b2b-b604-ca4d159b3c07-kube-api-access-2slcx\") pod \"3ee20797-d380-4b2b-b604-ca4d159b3c07\" (UID: \"3ee20797-d380-4b2b-b604-ca4d159b3c07\") " Dec 03 15:15:09 crc kubenswrapper[4805]: I1203 15:15:09.585846 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3ee20797-d380-4b2b-b604-ca4d159b3c07-utilities" (OuterVolumeSpecName: "utilities") pod "3ee20797-d380-4b2b-b604-ca4d159b3c07" (UID: "3ee20797-d380-4b2b-b604-ca4d159b3c07"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 15:15:09 crc kubenswrapper[4805]: I1203 15:15:09.592135 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3ee20797-d380-4b2b-b604-ca4d159b3c07-kube-api-access-2slcx" (OuterVolumeSpecName: "kube-api-access-2slcx") pod "3ee20797-d380-4b2b-b604-ca4d159b3c07" (UID: "3ee20797-d380-4b2b-b604-ca4d159b3c07"). InnerVolumeSpecName "kube-api-access-2slcx". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 15:15:09 crc kubenswrapper[4805]: I1203 15:15:09.635917 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3ee20797-d380-4b2b-b604-ca4d159b3c07-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "3ee20797-d380-4b2b-b604-ca4d159b3c07" (UID: "3ee20797-d380-4b2b-b604-ca4d159b3c07"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 15:15:09 crc kubenswrapper[4805]: I1203 15:15:09.687874 4805 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3ee20797-d380-4b2b-b604-ca4d159b3c07-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 03 15:15:09 crc kubenswrapper[4805]: I1203 15:15:09.687930 4805 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2slcx\" (UniqueName: \"kubernetes.io/projected/3ee20797-d380-4b2b-b604-ca4d159b3c07-kube-api-access-2slcx\") on node \"crc\" DevicePath \"\"" Dec 03 15:15:09 crc kubenswrapper[4805]: I1203 15:15:09.687951 4805 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3ee20797-d380-4b2b-b604-ca4d159b3c07-utilities\") on node \"crc\" DevicePath \"\"" Dec 03 15:15:09 crc kubenswrapper[4805]: I1203 15:15:09.694236 4805 scope.go:117] "RemoveContainer" containerID="cbf73c6c9ba740c0a523c84ca23c0851292d6eba7a0fb47d38599acca964e9cc" Dec 03 15:15:09 crc kubenswrapper[4805]: E1203 15:15:09.694811 4805 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-gskh4_openshift-machine-config-operator(6098937f-e3f6-45e8-a647-4994a79cd711)\"" pod="openshift-machine-config-operator/machine-config-daemon-gskh4" podUID="6098937f-e3f6-45e8-a647-4994a79cd711" Dec 03 15:15:09 crc kubenswrapper[4805]: I1203 15:15:09.975542 4805 generic.go:334] "Generic (PLEG): container finished" podID="3ee20797-d380-4b2b-b604-ca4d159b3c07" containerID="70e2a0e349297cac98c24271fca3809a1b64a17320db20cbb8bd5eda80a20364" exitCode=0 Dec 03 15:15:09 crc kubenswrapper[4805]: I1203 15:15:09.975601 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-7hxsd" event={"ID":"3ee20797-d380-4b2b-b604-ca4d159b3c07","Type":"ContainerDied","Data":"70e2a0e349297cac98c24271fca3809a1b64a17320db20cbb8bd5eda80a20364"} Dec 03 15:15:09 crc kubenswrapper[4805]: I1203 15:15:09.975969 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-7hxsd" event={"ID":"3ee20797-d380-4b2b-b604-ca4d159b3c07","Type":"ContainerDied","Data":"5214c0a85068b4589efd58e3cab66d2eb2188f11f367ae2c3616f8b78adbf6e1"} Dec 03 15:15:09 crc kubenswrapper[4805]: I1203 15:15:09.975649 4805 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-7hxsd" Dec 03 15:15:09 crc kubenswrapper[4805]: I1203 15:15:09.975996 4805 scope.go:117] "RemoveContainer" containerID="70e2a0e349297cac98c24271fca3809a1b64a17320db20cbb8bd5eda80a20364" Dec 03 15:15:10 crc kubenswrapper[4805]: I1203 15:15:10.006481 4805 scope.go:117] "RemoveContainer" containerID="e78d8c40bff7cfd30b394b8b8bb1c090cf0448f912c63389093ca3769616303e" Dec 03 15:15:10 crc kubenswrapper[4805]: I1203 15:15:10.022000 4805 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-7hxsd"] Dec 03 15:15:10 crc kubenswrapper[4805]: I1203 15:15:10.035751 4805 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-7hxsd"] Dec 03 15:15:10 crc kubenswrapper[4805]: I1203 15:15:10.061795 4805 scope.go:117] "RemoveContainer" containerID="a542bcdef97581d5bca30ebf1d9dada3ea84509e2ecbefd9e2169e752bf0e672" Dec 03 15:15:10 crc kubenswrapper[4805]: I1203 15:15:10.098784 4805 scope.go:117] "RemoveContainer" containerID="70e2a0e349297cac98c24271fca3809a1b64a17320db20cbb8bd5eda80a20364" Dec 03 15:15:10 crc kubenswrapper[4805]: E1203 15:15:10.099354 4805 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"70e2a0e349297cac98c24271fca3809a1b64a17320db20cbb8bd5eda80a20364\": container with ID starting with 70e2a0e349297cac98c24271fca3809a1b64a17320db20cbb8bd5eda80a20364 not found: ID does not exist" containerID="70e2a0e349297cac98c24271fca3809a1b64a17320db20cbb8bd5eda80a20364" Dec 03 15:15:10 crc kubenswrapper[4805]: I1203 15:15:10.099395 4805 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"70e2a0e349297cac98c24271fca3809a1b64a17320db20cbb8bd5eda80a20364"} err="failed to get container status \"70e2a0e349297cac98c24271fca3809a1b64a17320db20cbb8bd5eda80a20364\": rpc error: code = NotFound desc = could not find container \"70e2a0e349297cac98c24271fca3809a1b64a17320db20cbb8bd5eda80a20364\": container with ID starting with 70e2a0e349297cac98c24271fca3809a1b64a17320db20cbb8bd5eda80a20364 not found: ID does not exist" Dec 03 15:15:10 crc kubenswrapper[4805]: I1203 15:15:10.099419 4805 scope.go:117] "RemoveContainer" containerID="e78d8c40bff7cfd30b394b8b8bb1c090cf0448f912c63389093ca3769616303e" Dec 03 15:15:10 crc kubenswrapper[4805]: E1203 15:15:10.099853 4805 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e78d8c40bff7cfd30b394b8b8bb1c090cf0448f912c63389093ca3769616303e\": container with ID starting with e78d8c40bff7cfd30b394b8b8bb1c090cf0448f912c63389093ca3769616303e not found: ID does not exist" containerID="e78d8c40bff7cfd30b394b8b8bb1c090cf0448f912c63389093ca3769616303e" Dec 03 15:15:10 crc kubenswrapper[4805]: I1203 15:15:10.099884 4805 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e78d8c40bff7cfd30b394b8b8bb1c090cf0448f912c63389093ca3769616303e"} err="failed to get container status \"e78d8c40bff7cfd30b394b8b8bb1c090cf0448f912c63389093ca3769616303e\": rpc error: code = NotFound desc = could not find container \"e78d8c40bff7cfd30b394b8b8bb1c090cf0448f912c63389093ca3769616303e\": container with ID starting with e78d8c40bff7cfd30b394b8b8bb1c090cf0448f912c63389093ca3769616303e not found: ID does not exist" Dec 03 15:15:10 crc kubenswrapper[4805]: I1203 15:15:10.099902 4805 scope.go:117] "RemoveContainer" containerID="a542bcdef97581d5bca30ebf1d9dada3ea84509e2ecbefd9e2169e752bf0e672" Dec 03 15:15:10 crc kubenswrapper[4805]: E1203 15:15:10.100327 4805 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a542bcdef97581d5bca30ebf1d9dada3ea84509e2ecbefd9e2169e752bf0e672\": container with ID starting with a542bcdef97581d5bca30ebf1d9dada3ea84509e2ecbefd9e2169e752bf0e672 not found: ID does not exist" containerID="a542bcdef97581d5bca30ebf1d9dada3ea84509e2ecbefd9e2169e752bf0e672" Dec 03 15:15:10 crc kubenswrapper[4805]: I1203 15:15:10.100356 4805 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a542bcdef97581d5bca30ebf1d9dada3ea84509e2ecbefd9e2169e752bf0e672"} err="failed to get container status \"a542bcdef97581d5bca30ebf1d9dada3ea84509e2ecbefd9e2169e752bf0e672\": rpc error: code = NotFound desc = could not find container \"a542bcdef97581d5bca30ebf1d9dada3ea84509e2ecbefd9e2169e752bf0e672\": container with ID starting with a542bcdef97581d5bca30ebf1d9dada3ea84509e2ecbefd9e2169e752bf0e672 not found: ID does not exist" Dec 03 15:15:10 crc kubenswrapper[4805]: I1203 15:15:10.709811 4805 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3ee20797-d380-4b2b-b604-ca4d159b3c07" path="/var/lib/kubelet/pods/3ee20797-d380-4b2b-b604-ca4d159b3c07/volumes" Dec 03 15:15:12 crc kubenswrapper[4805]: E1203 15:15:12.545111 4805 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod9e35b852_1838_49da_a140_f048788f6635.slice/crio-conmon-3f32476151f16cefda9576a8077a18738591448ef92afdba7d6ba1cfe707d756.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod9e35b852_1838_49da_a140_f048788f6635.slice/crio-3f32476151f16cefda9576a8077a18738591448ef92afdba7d6ba1cfe707d756.scope\": RecentStats: unable to find data in memory cache]" Dec 03 15:15:22 crc kubenswrapper[4805]: I1203 15:15:22.694006 4805 scope.go:117] "RemoveContainer" containerID="cbf73c6c9ba740c0a523c84ca23c0851292d6eba7a0fb47d38599acca964e9cc" Dec 03 15:15:22 crc kubenswrapper[4805]: E1203 15:15:22.694795 4805 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-gskh4_openshift-machine-config-operator(6098937f-e3f6-45e8-a647-4994a79cd711)\"" pod="openshift-machine-config-operator/machine-config-daemon-gskh4" podUID="6098937f-e3f6-45e8-a647-4994a79cd711" Dec 03 15:15:22 crc kubenswrapper[4805]: E1203 15:15:22.815568 4805 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod9e35b852_1838_49da_a140_f048788f6635.slice/crio-3f32476151f16cefda9576a8077a18738591448ef92afdba7d6ba1cfe707d756.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod9e35b852_1838_49da_a140_f048788f6635.slice/crio-conmon-3f32476151f16cefda9576a8077a18738591448ef92afdba7d6ba1cfe707d756.scope\": RecentStats: unable to find data in memory cache]" Dec 03 15:15:34 crc kubenswrapper[4805]: I1203 15:15:34.695065 4805 scope.go:117] "RemoveContainer" containerID="cbf73c6c9ba740c0a523c84ca23c0851292d6eba7a0fb47d38599acca964e9cc" Dec 03 15:15:34 crc kubenswrapper[4805]: E1203 15:15:34.695885 4805 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-gskh4_openshift-machine-config-operator(6098937f-e3f6-45e8-a647-4994a79cd711)\"" pod="openshift-machine-config-operator/machine-config-daemon-gskh4" podUID="6098937f-e3f6-45e8-a647-4994a79cd711" Dec 03 15:15:46 crc kubenswrapper[4805]: I1203 15:15:46.687983 4805 scope.go:117] "RemoveContainer" containerID="36e5b2355468d0b356654a149f8c3bcfd0ca25fcf904e7d3354aa72bd9a641ec" Dec 03 15:15:49 crc kubenswrapper[4805]: I1203 15:15:49.695477 4805 scope.go:117] "RemoveContainer" containerID="cbf73c6c9ba740c0a523c84ca23c0851292d6eba7a0fb47d38599acca964e9cc" Dec 03 15:15:49 crc kubenswrapper[4805]: E1203 15:15:49.696256 4805 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-gskh4_openshift-machine-config-operator(6098937f-e3f6-45e8-a647-4994a79cd711)\"" pod="openshift-machine-config-operator/machine-config-daemon-gskh4" podUID="6098937f-e3f6-45e8-a647-4994a79cd711" Dec 03 15:16:02 crc kubenswrapper[4805]: I1203 15:16:02.696786 4805 scope.go:117] "RemoveContainer" containerID="cbf73c6c9ba740c0a523c84ca23c0851292d6eba7a0fb47d38599acca964e9cc" Dec 03 15:16:02 crc kubenswrapper[4805]: E1203 15:16:02.698190 4805 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-gskh4_openshift-machine-config-operator(6098937f-e3f6-45e8-a647-4994a79cd711)\"" pod="openshift-machine-config-operator/machine-config-daemon-gskh4" podUID="6098937f-e3f6-45e8-a647-4994a79cd711" Dec 03 15:16:15 crc kubenswrapper[4805]: I1203 15:16:15.694793 4805 scope.go:117] "RemoveContainer" containerID="cbf73c6c9ba740c0a523c84ca23c0851292d6eba7a0fb47d38599acca964e9cc" Dec 03 15:16:15 crc kubenswrapper[4805]: E1203 15:16:15.695699 4805 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-gskh4_openshift-machine-config-operator(6098937f-e3f6-45e8-a647-4994a79cd711)\"" pod="openshift-machine-config-operator/machine-config-daemon-gskh4" podUID="6098937f-e3f6-45e8-a647-4994a79cd711" Dec 03 15:16:29 crc kubenswrapper[4805]: I1203 15:16:29.694907 4805 scope.go:117] "RemoveContainer" containerID="cbf73c6c9ba740c0a523c84ca23c0851292d6eba7a0fb47d38599acca964e9cc" Dec 03 15:16:29 crc kubenswrapper[4805]: E1203 15:16:29.695779 4805 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-gskh4_openshift-machine-config-operator(6098937f-e3f6-45e8-a647-4994a79cd711)\"" pod="openshift-machine-config-operator/machine-config-daemon-gskh4" podUID="6098937f-e3f6-45e8-a647-4994a79cd711" Dec 03 15:16:30 crc kubenswrapper[4805]: I1203 15:16:30.554265 4805 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_telemetry-operator-controller-manager-cc9d8f87b-qfmmx_8cbfcd07-e136-4bfd-b3c6-6615c911d440/manager/0.log" Dec 03 15:16:40 crc kubenswrapper[4805]: I1203 15:16:40.700664 4805 scope.go:117] "RemoveContainer" containerID="cbf73c6c9ba740c0a523c84ca23c0851292d6eba7a0fb47d38599acca964e9cc" Dec 03 15:16:40 crc kubenswrapper[4805]: E1203 15:16:40.701495 4805 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-gskh4_openshift-machine-config-operator(6098937f-e3f6-45e8-a647-4994a79cd711)\"" pod="openshift-machine-config-operator/machine-config-daemon-gskh4" podUID="6098937f-e3f6-45e8-a647-4994a79cd711" Dec 03 15:16:52 crc kubenswrapper[4805]: I1203 15:16:52.161032 4805 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-n5xns/must-gather-5mnzj"] Dec 03 15:16:52 crc kubenswrapper[4805]: E1203 15:16:52.162355 4805 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3ee20797-d380-4b2b-b604-ca4d159b3c07" containerName="registry-server" Dec 03 15:16:52 crc kubenswrapper[4805]: I1203 15:16:52.162375 4805 state_mem.go:107] "Deleted CPUSet assignment" podUID="3ee20797-d380-4b2b-b604-ca4d159b3c07" containerName="registry-server" Dec 03 15:16:52 crc kubenswrapper[4805]: E1203 15:16:52.162402 4805 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3ee20797-d380-4b2b-b604-ca4d159b3c07" containerName="extract-utilities" Dec 03 15:16:52 crc kubenswrapper[4805]: I1203 15:16:52.162409 4805 state_mem.go:107] "Deleted CPUSet assignment" podUID="3ee20797-d380-4b2b-b604-ca4d159b3c07" containerName="extract-utilities" Dec 03 15:16:52 crc kubenswrapper[4805]: E1203 15:16:52.162523 4805 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4d6fdcb7-4fb1-4739-aa7e-71c91a241733" containerName="collect-profiles" Dec 03 15:16:52 crc kubenswrapper[4805]: I1203 15:16:52.162532 4805 state_mem.go:107] "Deleted CPUSet assignment" podUID="4d6fdcb7-4fb1-4739-aa7e-71c91a241733" containerName="collect-profiles" Dec 03 15:16:52 crc kubenswrapper[4805]: E1203 15:16:52.162563 4805 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3ee20797-d380-4b2b-b604-ca4d159b3c07" containerName="extract-content" Dec 03 15:16:52 crc kubenswrapper[4805]: I1203 15:16:52.162569 4805 state_mem.go:107] "Deleted CPUSet assignment" podUID="3ee20797-d380-4b2b-b604-ca4d159b3c07" containerName="extract-content" Dec 03 15:16:52 crc kubenswrapper[4805]: I1203 15:16:52.162865 4805 memory_manager.go:354] "RemoveStaleState removing state" podUID="4d6fdcb7-4fb1-4739-aa7e-71c91a241733" containerName="collect-profiles" Dec 03 15:16:52 crc kubenswrapper[4805]: I1203 15:16:52.162880 4805 memory_manager.go:354] "RemoveStaleState removing state" podUID="3ee20797-d380-4b2b-b604-ca4d159b3c07" containerName="registry-server" Dec 03 15:16:52 crc kubenswrapper[4805]: I1203 15:16:52.164126 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-n5xns/must-gather-5mnzj" Dec 03 15:16:52 crc kubenswrapper[4805]: I1203 15:16:52.165556 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-must-gather-n5xns"/"default-dockercfg-d7rkg" Dec 03 15:16:52 crc kubenswrapper[4805]: I1203 15:16:52.166048 4805 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-must-gather-n5xns"/"kube-root-ca.crt" Dec 03 15:16:52 crc kubenswrapper[4805]: I1203 15:16:52.169155 4805 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-must-gather-n5xns"/"openshift-service-ca.crt" Dec 03 15:16:52 crc kubenswrapper[4805]: I1203 15:16:52.179704 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-must-gather-n5xns/must-gather-5mnzj"] Dec 03 15:16:52 crc kubenswrapper[4805]: I1203 15:16:52.283213 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vhp6q\" (UniqueName: \"kubernetes.io/projected/08ff2c92-4d49-4ac0-84c5-3403610205da-kube-api-access-vhp6q\") pod \"must-gather-5mnzj\" (UID: \"08ff2c92-4d49-4ac0-84c5-3403610205da\") " pod="openshift-must-gather-n5xns/must-gather-5mnzj" Dec 03 15:16:52 crc kubenswrapper[4805]: I1203 15:16:52.283404 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/08ff2c92-4d49-4ac0-84c5-3403610205da-must-gather-output\") pod \"must-gather-5mnzj\" (UID: \"08ff2c92-4d49-4ac0-84c5-3403610205da\") " pod="openshift-must-gather-n5xns/must-gather-5mnzj" Dec 03 15:16:52 crc kubenswrapper[4805]: I1203 15:16:52.385162 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/08ff2c92-4d49-4ac0-84c5-3403610205da-must-gather-output\") pod \"must-gather-5mnzj\" (UID: \"08ff2c92-4d49-4ac0-84c5-3403610205da\") " pod="openshift-must-gather-n5xns/must-gather-5mnzj" Dec 03 15:16:52 crc kubenswrapper[4805]: I1203 15:16:52.385281 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vhp6q\" (UniqueName: \"kubernetes.io/projected/08ff2c92-4d49-4ac0-84c5-3403610205da-kube-api-access-vhp6q\") pod \"must-gather-5mnzj\" (UID: \"08ff2c92-4d49-4ac0-84c5-3403610205da\") " pod="openshift-must-gather-n5xns/must-gather-5mnzj" Dec 03 15:16:52 crc kubenswrapper[4805]: I1203 15:16:52.386713 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/08ff2c92-4d49-4ac0-84c5-3403610205da-must-gather-output\") pod \"must-gather-5mnzj\" (UID: \"08ff2c92-4d49-4ac0-84c5-3403610205da\") " pod="openshift-must-gather-n5xns/must-gather-5mnzj" Dec 03 15:16:52 crc kubenswrapper[4805]: I1203 15:16:52.409975 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vhp6q\" (UniqueName: \"kubernetes.io/projected/08ff2c92-4d49-4ac0-84c5-3403610205da-kube-api-access-vhp6q\") pod \"must-gather-5mnzj\" (UID: \"08ff2c92-4d49-4ac0-84c5-3403610205da\") " pod="openshift-must-gather-n5xns/must-gather-5mnzj" Dec 03 15:16:52 crc kubenswrapper[4805]: I1203 15:16:52.490788 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-n5xns/must-gather-5mnzj" Dec 03 15:16:53 crc kubenswrapper[4805]: I1203 15:16:53.078322 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-must-gather-n5xns/must-gather-5mnzj"] Dec 03 15:16:53 crc kubenswrapper[4805]: W1203 15:16:53.083014 4805 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod08ff2c92_4d49_4ac0_84c5_3403610205da.slice/crio-2d2eb33a971d1f0548a5d3f2f9415ecc21721f07ad05ad9c1d45875758f5ed75 WatchSource:0}: Error finding container 2d2eb33a971d1f0548a5d3f2f9415ecc21721f07ad05ad9c1d45875758f5ed75: Status 404 returned error can't find the container with id 2d2eb33a971d1f0548a5d3f2f9415ecc21721f07ad05ad9c1d45875758f5ed75 Dec 03 15:16:53 crc kubenswrapper[4805]: I1203 15:16:53.085377 4805 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 03 15:16:54 crc kubenswrapper[4805]: I1203 15:16:54.060924 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-n5xns/must-gather-5mnzj" event={"ID":"08ff2c92-4d49-4ac0-84c5-3403610205da","Type":"ContainerStarted","Data":"2d2eb33a971d1f0548a5d3f2f9415ecc21721f07ad05ad9c1d45875758f5ed75"} Dec 03 15:16:55 crc kubenswrapper[4805]: I1203 15:16:55.694638 4805 scope.go:117] "RemoveContainer" containerID="cbf73c6c9ba740c0a523c84ca23c0851292d6eba7a0fb47d38599acca964e9cc" Dec 03 15:16:55 crc kubenswrapper[4805]: E1203 15:16:55.695475 4805 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-gskh4_openshift-machine-config-operator(6098937f-e3f6-45e8-a647-4994a79cd711)\"" pod="openshift-machine-config-operator/machine-config-daemon-gskh4" podUID="6098937f-e3f6-45e8-a647-4994a79cd711" Dec 03 15:16:58 crc kubenswrapper[4805]: I1203 15:16:58.120089 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-n5xns/must-gather-5mnzj" event={"ID":"08ff2c92-4d49-4ac0-84c5-3403610205da","Type":"ContainerStarted","Data":"7cfbc3fe2001fe4c7e6cf09443ce6a6e860a40da1f2c016d071c05fbe7033424"} Dec 03 15:16:58 crc kubenswrapper[4805]: I1203 15:16:58.811681 4805 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-b9vtg"] Dec 03 15:16:58 crc kubenswrapper[4805]: I1203 15:16:58.814444 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-b9vtg" Dec 03 15:16:58 crc kubenswrapper[4805]: I1203 15:16:58.824874 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-b9vtg"] Dec 03 15:16:58 crc kubenswrapper[4805]: I1203 15:16:58.929885 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/629e29e5-26f0-4ad2-93cb-b7101572c7ec-catalog-content\") pod \"redhat-marketplace-b9vtg\" (UID: \"629e29e5-26f0-4ad2-93cb-b7101572c7ec\") " pod="openshift-marketplace/redhat-marketplace-b9vtg" Dec 03 15:16:58 crc kubenswrapper[4805]: I1203 15:16:58.930097 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/629e29e5-26f0-4ad2-93cb-b7101572c7ec-utilities\") pod \"redhat-marketplace-b9vtg\" (UID: \"629e29e5-26f0-4ad2-93cb-b7101572c7ec\") " pod="openshift-marketplace/redhat-marketplace-b9vtg" Dec 03 15:16:58 crc kubenswrapper[4805]: I1203 15:16:58.930457 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7fwq7\" (UniqueName: \"kubernetes.io/projected/629e29e5-26f0-4ad2-93cb-b7101572c7ec-kube-api-access-7fwq7\") pod \"redhat-marketplace-b9vtg\" (UID: \"629e29e5-26f0-4ad2-93cb-b7101572c7ec\") " pod="openshift-marketplace/redhat-marketplace-b9vtg" Dec 03 15:16:59 crc kubenswrapper[4805]: I1203 15:16:59.032935 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/629e29e5-26f0-4ad2-93cb-b7101572c7ec-utilities\") pod \"redhat-marketplace-b9vtg\" (UID: \"629e29e5-26f0-4ad2-93cb-b7101572c7ec\") " pod="openshift-marketplace/redhat-marketplace-b9vtg" Dec 03 15:16:59 crc kubenswrapper[4805]: I1203 15:16:59.033157 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7fwq7\" (UniqueName: \"kubernetes.io/projected/629e29e5-26f0-4ad2-93cb-b7101572c7ec-kube-api-access-7fwq7\") pod \"redhat-marketplace-b9vtg\" (UID: \"629e29e5-26f0-4ad2-93cb-b7101572c7ec\") " pod="openshift-marketplace/redhat-marketplace-b9vtg" Dec 03 15:16:59 crc kubenswrapper[4805]: I1203 15:16:59.033227 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/629e29e5-26f0-4ad2-93cb-b7101572c7ec-catalog-content\") pod \"redhat-marketplace-b9vtg\" (UID: \"629e29e5-26f0-4ad2-93cb-b7101572c7ec\") " pod="openshift-marketplace/redhat-marketplace-b9vtg" Dec 03 15:16:59 crc kubenswrapper[4805]: I1203 15:16:59.033641 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/629e29e5-26f0-4ad2-93cb-b7101572c7ec-utilities\") pod \"redhat-marketplace-b9vtg\" (UID: \"629e29e5-26f0-4ad2-93cb-b7101572c7ec\") " pod="openshift-marketplace/redhat-marketplace-b9vtg" Dec 03 15:16:59 crc kubenswrapper[4805]: I1203 15:16:59.033782 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/629e29e5-26f0-4ad2-93cb-b7101572c7ec-catalog-content\") pod \"redhat-marketplace-b9vtg\" (UID: \"629e29e5-26f0-4ad2-93cb-b7101572c7ec\") " pod="openshift-marketplace/redhat-marketplace-b9vtg" Dec 03 15:16:59 crc kubenswrapper[4805]: I1203 15:16:59.056193 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7fwq7\" (UniqueName: \"kubernetes.io/projected/629e29e5-26f0-4ad2-93cb-b7101572c7ec-kube-api-access-7fwq7\") pod \"redhat-marketplace-b9vtg\" (UID: \"629e29e5-26f0-4ad2-93cb-b7101572c7ec\") " pod="openshift-marketplace/redhat-marketplace-b9vtg" Dec 03 15:16:59 crc kubenswrapper[4805]: I1203 15:16:59.133884 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-n5xns/must-gather-5mnzj" event={"ID":"08ff2c92-4d49-4ac0-84c5-3403610205da","Type":"ContainerStarted","Data":"313a0a66fa07afa0a361dc21ea0df9772ee5f6169321650079c0add6695204d7"} Dec 03 15:16:59 crc kubenswrapper[4805]: I1203 15:16:59.139250 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-b9vtg" Dec 03 15:16:59 crc kubenswrapper[4805]: I1203 15:16:59.158357 4805 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-must-gather-n5xns/must-gather-5mnzj" podStartSLOduration=2.459070559 podStartE2EDuration="7.158334705s" podCreationTimestamp="2025-12-03 15:16:52 +0000 UTC" firstStartedPulling="2025-12-03 15:16:53.085113614 +0000 UTC m=+4042.748030537" lastFinishedPulling="2025-12-03 15:16:57.78437776 +0000 UTC m=+4047.447294683" observedRunningTime="2025-12-03 15:16:59.153218027 +0000 UTC m=+4048.816134950" watchObservedRunningTime="2025-12-03 15:16:59.158334705 +0000 UTC m=+4048.821251628" Dec 03 15:16:59 crc kubenswrapper[4805]: I1203 15:16:59.687866 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-b9vtg"] Dec 03 15:16:59 crc kubenswrapper[4805]: W1203 15:16:59.697224 4805 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod629e29e5_26f0_4ad2_93cb_b7101572c7ec.slice/crio-2ad22f9c48dee5edaf6d1cb75fff07f37c64fcdd60c37cd36fca27b050b4b86f WatchSource:0}: Error finding container 2ad22f9c48dee5edaf6d1cb75fff07f37c64fcdd60c37cd36fca27b050b4b86f: Status 404 returned error can't find the container with id 2ad22f9c48dee5edaf6d1cb75fff07f37c64fcdd60c37cd36fca27b050b4b86f Dec 03 15:17:00 crc kubenswrapper[4805]: I1203 15:17:00.148356 4805 generic.go:334] "Generic (PLEG): container finished" podID="629e29e5-26f0-4ad2-93cb-b7101572c7ec" containerID="c93c7390d71c68e650f9079c4bcde56605294da20bf42c762602c14f3e17c2d3" exitCode=0 Dec 03 15:17:00 crc kubenswrapper[4805]: I1203 15:17:00.148489 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-b9vtg" event={"ID":"629e29e5-26f0-4ad2-93cb-b7101572c7ec","Type":"ContainerDied","Data":"c93c7390d71c68e650f9079c4bcde56605294da20bf42c762602c14f3e17c2d3"} Dec 03 15:17:00 crc kubenswrapper[4805]: I1203 15:17:00.148785 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-b9vtg" event={"ID":"629e29e5-26f0-4ad2-93cb-b7101572c7ec","Type":"ContainerStarted","Data":"2ad22f9c48dee5edaf6d1cb75fff07f37c64fcdd60c37cd36fca27b050b4b86f"} Dec 03 15:17:01 crc kubenswrapper[4805]: I1203 15:17:01.160139 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-b9vtg" event={"ID":"629e29e5-26f0-4ad2-93cb-b7101572c7ec","Type":"ContainerStarted","Data":"006a66803d3aff3883e5a387c7b33a5744a8a843dbc7f2df540e3216ac83a320"} Dec 03 15:17:02 crc kubenswrapper[4805]: I1203 15:17:02.171533 4805 generic.go:334] "Generic (PLEG): container finished" podID="629e29e5-26f0-4ad2-93cb-b7101572c7ec" containerID="006a66803d3aff3883e5a387c7b33a5744a8a843dbc7f2df540e3216ac83a320" exitCode=0 Dec 03 15:17:02 crc kubenswrapper[4805]: I1203 15:17:02.171716 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-b9vtg" event={"ID":"629e29e5-26f0-4ad2-93cb-b7101572c7ec","Type":"ContainerDied","Data":"006a66803d3aff3883e5a387c7b33a5744a8a843dbc7f2df540e3216ac83a320"} Dec 03 15:17:03 crc kubenswrapper[4805]: I1203 15:17:03.191761 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-b9vtg" event={"ID":"629e29e5-26f0-4ad2-93cb-b7101572c7ec","Type":"ContainerStarted","Data":"3bd591b644e4bba107c0cfe41fc5d38a6f2782a079c0240e730a0542cc3eaa73"} Dec 03 15:17:03 crc kubenswrapper[4805]: E1203 15:17:03.231777 4805 upgradeaware.go:427] Error proxying data from client to backend: readfrom tcp 38.102.83.2:47378->38.102.83.2:35135: write tcp 38.102.83.2:47378->38.102.83.2:35135: write: broken pipe Dec 03 15:17:04 crc kubenswrapper[4805]: I1203 15:17:04.562854 4805 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-b9vtg" podStartSLOduration=3.770623296 podStartE2EDuration="6.562822719s" podCreationTimestamp="2025-12-03 15:16:58 +0000 UTC" firstStartedPulling="2025-12-03 15:17:00.151371301 +0000 UTC m=+4049.814288224" lastFinishedPulling="2025-12-03 15:17:02.943570724 +0000 UTC m=+4052.606487647" observedRunningTime="2025-12-03 15:17:03.224354122 +0000 UTC m=+4052.887271045" watchObservedRunningTime="2025-12-03 15:17:04.562822719 +0000 UTC m=+4054.225739642" Dec 03 15:17:04 crc kubenswrapper[4805]: I1203 15:17:04.573466 4805 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-n5xns/crc-debug-s9275"] Dec 03 15:17:04 crc kubenswrapper[4805]: I1203 15:17:04.574831 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-n5xns/crc-debug-s9275" Dec 03 15:17:04 crc kubenswrapper[4805]: I1203 15:17:04.759091 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/085d787e-1813-460e-8c3d-c55d60231820-host\") pod \"crc-debug-s9275\" (UID: \"085d787e-1813-460e-8c3d-c55d60231820\") " pod="openshift-must-gather-n5xns/crc-debug-s9275" Dec 03 15:17:04 crc kubenswrapper[4805]: I1203 15:17:04.759150 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jd9x6\" (UniqueName: \"kubernetes.io/projected/085d787e-1813-460e-8c3d-c55d60231820-kube-api-access-jd9x6\") pod \"crc-debug-s9275\" (UID: \"085d787e-1813-460e-8c3d-c55d60231820\") " pod="openshift-must-gather-n5xns/crc-debug-s9275" Dec 03 15:17:04 crc kubenswrapper[4805]: I1203 15:17:04.860924 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/085d787e-1813-460e-8c3d-c55d60231820-host\") pod \"crc-debug-s9275\" (UID: \"085d787e-1813-460e-8c3d-c55d60231820\") " pod="openshift-must-gather-n5xns/crc-debug-s9275" Dec 03 15:17:04 crc kubenswrapper[4805]: I1203 15:17:04.860979 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jd9x6\" (UniqueName: \"kubernetes.io/projected/085d787e-1813-460e-8c3d-c55d60231820-kube-api-access-jd9x6\") pod \"crc-debug-s9275\" (UID: \"085d787e-1813-460e-8c3d-c55d60231820\") " pod="openshift-must-gather-n5xns/crc-debug-s9275" Dec 03 15:17:04 crc kubenswrapper[4805]: I1203 15:17:04.861309 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/085d787e-1813-460e-8c3d-c55d60231820-host\") pod \"crc-debug-s9275\" (UID: \"085d787e-1813-460e-8c3d-c55d60231820\") " pod="openshift-must-gather-n5xns/crc-debug-s9275" Dec 03 15:17:04 crc kubenswrapper[4805]: I1203 15:17:04.891575 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jd9x6\" (UniqueName: \"kubernetes.io/projected/085d787e-1813-460e-8c3d-c55d60231820-kube-api-access-jd9x6\") pod \"crc-debug-s9275\" (UID: \"085d787e-1813-460e-8c3d-c55d60231820\") " pod="openshift-must-gather-n5xns/crc-debug-s9275" Dec 03 15:17:04 crc kubenswrapper[4805]: I1203 15:17:04.907208 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-n5xns/crc-debug-s9275" Dec 03 15:17:05 crc kubenswrapper[4805]: I1203 15:17:05.212416 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-n5xns/crc-debug-s9275" event={"ID":"085d787e-1813-460e-8c3d-c55d60231820","Type":"ContainerStarted","Data":"e03f2e1f7438fbafc04ea4d4b671ed82fade7bcfc1d34f2c551f65ee49edf571"} Dec 03 15:17:09 crc kubenswrapper[4805]: I1203 15:17:09.140039 4805 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-b9vtg" Dec 03 15:17:09 crc kubenswrapper[4805]: I1203 15:17:09.140668 4805 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-b9vtg" Dec 03 15:17:09 crc kubenswrapper[4805]: I1203 15:17:09.209563 4805 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-b9vtg" Dec 03 15:17:09 crc kubenswrapper[4805]: I1203 15:17:09.352132 4805 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-b9vtg" Dec 03 15:17:09 crc kubenswrapper[4805]: I1203 15:17:09.460756 4805 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-b9vtg"] Dec 03 15:17:09 crc kubenswrapper[4805]: I1203 15:17:09.695288 4805 scope.go:117] "RemoveContainer" containerID="cbf73c6c9ba740c0a523c84ca23c0851292d6eba7a0fb47d38599acca964e9cc" Dec 03 15:17:09 crc kubenswrapper[4805]: E1203 15:17:09.695614 4805 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-gskh4_openshift-machine-config-operator(6098937f-e3f6-45e8-a647-4994a79cd711)\"" pod="openshift-machine-config-operator/machine-config-daemon-gskh4" podUID="6098937f-e3f6-45e8-a647-4994a79cd711" Dec 03 15:17:11 crc kubenswrapper[4805]: I1203 15:17:11.302905 4805 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-b9vtg" podUID="629e29e5-26f0-4ad2-93cb-b7101572c7ec" containerName="registry-server" containerID="cri-o://3bd591b644e4bba107c0cfe41fc5d38a6f2782a079c0240e730a0542cc3eaa73" gracePeriod=2 Dec 03 15:17:12 crc kubenswrapper[4805]: I1203 15:17:12.345595 4805 generic.go:334] "Generic (PLEG): container finished" podID="629e29e5-26f0-4ad2-93cb-b7101572c7ec" containerID="3bd591b644e4bba107c0cfe41fc5d38a6f2782a079c0240e730a0542cc3eaa73" exitCode=0 Dec 03 15:17:12 crc kubenswrapper[4805]: I1203 15:17:12.345982 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-b9vtg" event={"ID":"629e29e5-26f0-4ad2-93cb-b7101572c7ec","Type":"ContainerDied","Data":"3bd591b644e4bba107c0cfe41fc5d38a6f2782a079c0240e730a0542cc3eaa73"} Dec 03 15:17:19 crc kubenswrapper[4805]: E1203 15:17:19.140590 4805 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 3bd591b644e4bba107c0cfe41fc5d38a6f2782a079c0240e730a0542cc3eaa73 is running failed: container process not found" containerID="3bd591b644e4bba107c0cfe41fc5d38a6f2782a079c0240e730a0542cc3eaa73" cmd=["grpc_health_probe","-addr=:50051"] Dec 03 15:17:19 crc kubenswrapper[4805]: E1203 15:17:19.141594 4805 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 3bd591b644e4bba107c0cfe41fc5d38a6f2782a079c0240e730a0542cc3eaa73 is running failed: container process not found" containerID="3bd591b644e4bba107c0cfe41fc5d38a6f2782a079c0240e730a0542cc3eaa73" cmd=["grpc_health_probe","-addr=:50051"] Dec 03 15:17:19 crc kubenswrapper[4805]: E1203 15:17:19.142192 4805 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 3bd591b644e4bba107c0cfe41fc5d38a6f2782a079c0240e730a0542cc3eaa73 is running failed: container process not found" containerID="3bd591b644e4bba107c0cfe41fc5d38a6f2782a079c0240e730a0542cc3eaa73" cmd=["grpc_health_probe","-addr=:50051"] Dec 03 15:17:19 crc kubenswrapper[4805]: E1203 15:17:19.142239 4805 prober.go:104] "Probe errored" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 3bd591b644e4bba107c0cfe41fc5d38a6f2782a079c0240e730a0542cc3eaa73 is running failed: container process not found" probeType="Readiness" pod="openshift-marketplace/redhat-marketplace-b9vtg" podUID="629e29e5-26f0-4ad2-93cb-b7101572c7ec" containerName="registry-server" Dec 03 15:17:19 crc kubenswrapper[4805]: E1203 15:17:19.615858 4805 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:6ab858aed98e4fe57e6b144da8e90ad5d6698bb4cc5521206f5c05809f0f9296" Dec 03 15:17:19 crc kubenswrapper[4805]: E1203 15:17:19.616039 4805 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:container-00,Image:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:6ab858aed98e4fe57e6b144da8e90ad5d6698bb4cc5521206f5c05809f0f9296,Command:[chroot /host bash -c echo 'TOOLBOX_NAME=toolbox-osp' > /root/.toolboxrc ; rm -rf \"/var/tmp/sos-osp\" && mkdir -p \"/var/tmp/sos-osp\" && sudo podman rm --force toolbox-osp; sudo --preserve-env podman pull --authfile /var/lib/kubelet/config.json registry.redhat.io/rhel9/support-tools && toolbox sos report --batch --all-logs --only-plugins block,cifs,crio,devicemapper,devices,firewall_tables,firewalld,iscsi,lvm2,memory,multipath,nfs,nis,nvme,podman,process,processor,selinux,scsi,udev,logs,crypto --tmp-dir=\"/var/tmp/sos-osp\" && if [[ \"$(ls /var/log/pods/*/{*.log.*,*/*.log.*} 2>/dev/null)\" != '' ]]; then tar --ignore-failed-read --warning=no-file-changed -cJf \"/var/tmp/sos-osp/podlogs.tar.xz\" --transform 's,^,podlogs/,' /var/log/pods/*/{*.log.*,*/*.log.*} || true; fi],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:TMOUT,Value:900,ValueFrom:nil,},EnvVar{Name:HOST,Value:/host,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:host,ReadOnly:false,MountPath:/host,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-jd9x6,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:nil,Privileged:*true,SELinuxOptions:nil,RunAsUser:*0,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod crc-debug-s9275_openshift-must-gather-n5xns(085d787e-1813-460e-8c3d-c55d60231820): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 03 15:17:19 crc kubenswrapper[4805]: E1203 15:17:19.617756 4805 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"container-00\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openshift-must-gather-n5xns/crc-debug-s9275" podUID="085d787e-1813-460e-8c3d-c55d60231820" Dec 03 15:17:19 crc kubenswrapper[4805]: I1203 15:17:19.936263 4805 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-b9vtg" Dec 03 15:17:20 crc kubenswrapper[4805]: I1203 15:17:20.025241 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7fwq7\" (UniqueName: \"kubernetes.io/projected/629e29e5-26f0-4ad2-93cb-b7101572c7ec-kube-api-access-7fwq7\") pod \"629e29e5-26f0-4ad2-93cb-b7101572c7ec\" (UID: \"629e29e5-26f0-4ad2-93cb-b7101572c7ec\") " Dec 03 15:17:20 crc kubenswrapper[4805]: I1203 15:17:20.025371 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/629e29e5-26f0-4ad2-93cb-b7101572c7ec-utilities\") pod \"629e29e5-26f0-4ad2-93cb-b7101572c7ec\" (UID: \"629e29e5-26f0-4ad2-93cb-b7101572c7ec\") " Dec 03 15:17:20 crc kubenswrapper[4805]: I1203 15:17:20.025415 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/629e29e5-26f0-4ad2-93cb-b7101572c7ec-catalog-content\") pod \"629e29e5-26f0-4ad2-93cb-b7101572c7ec\" (UID: \"629e29e5-26f0-4ad2-93cb-b7101572c7ec\") " Dec 03 15:17:20 crc kubenswrapper[4805]: I1203 15:17:20.026448 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/629e29e5-26f0-4ad2-93cb-b7101572c7ec-utilities" (OuterVolumeSpecName: "utilities") pod "629e29e5-26f0-4ad2-93cb-b7101572c7ec" (UID: "629e29e5-26f0-4ad2-93cb-b7101572c7ec"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 15:17:20 crc kubenswrapper[4805]: I1203 15:17:20.032349 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/629e29e5-26f0-4ad2-93cb-b7101572c7ec-kube-api-access-7fwq7" (OuterVolumeSpecName: "kube-api-access-7fwq7") pod "629e29e5-26f0-4ad2-93cb-b7101572c7ec" (UID: "629e29e5-26f0-4ad2-93cb-b7101572c7ec"). InnerVolumeSpecName "kube-api-access-7fwq7". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 15:17:20 crc kubenswrapper[4805]: I1203 15:17:20.045777 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/629e29e5-26f0-4ad2-93cb-b7101572c7ec-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "629e29e5-26f0-4ad2-93cb-b7101572c7ec" (UID: "629e29e5-26f0-4ad2-93cb-b7101572c7ec"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 15:17:20 crc kubenswrapper[4805]: I1203 15:17:20.127851 4805 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7fwq7\" (UniqueName: \"kubernetes.io/projected/629e29e5-26f0-4ad2-93cb-b7101572c7ec-kube-api-access-7fwq7\") on node \"crc\" DevicePath \"\"" Dec 03 15:17:20 crc kubenswrapper[4805]: I1203 15:17:20.127891 4805 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/629e29e5-26f0-4ad2-93cb-b7101572c7ec-utilities\") on node \"crc\" DevicePath \"\"" Dec 03 15:17:20 crc kubenswrapper[4805]: I1203 15:17:20.127902 4805 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/629e29e5-26f0-4ad2-93cb-b7101572c7ec-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 03 15:17:20 crc kubenswrapper[4805]: I1203 15:17:20.444523 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-b9vtg" event={"ID":"629e29e5-26f0-4ad2-93cb-b7101572c7ec","Type":"ContainerDied","Data":"2ad22f9c48dee5edaf6d1cb75fff07f37c64fcdd60c37cd36fca27b050b4b86f"} Dec 03 15:17:20 crc kubenswrapper[4805]: I1203 15:17:20.444897 4805 scope.go:117] "RemoveContainer" containerID="3bd591b644e4bba107c0cfe41fc5d38a6f2782a079c0240e730a0542cc3eaa73" Dec 03 15:17:20 crc kubenswrapper[4805]: I1203 15:17:20.444671 4805 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-b9vtg" Dec 03 15:17:20 crc kubenswrapper[4805]: E1203 15:17:20.446573 4805 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"container-00\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:6ab858aed98e4fe57e6b144da8e90ad5d6698bb4cc5521206f5c05809f0f9296\\\"\"" pod="openshift-must-gather-n5xns/crc-debug-s9275" podUID="085d787e-1813-460e-8c3d-c55d60231820" Dec 03 15:17:20 crc kubenswrapper[4805]: I1203 15:17:20.495219 4805 scope.go:117] "RemoveContainer" containerID="006a66803d3aff3883e5a387c7b33a5744a8a843dbc7f2df540e3216ac83a320" Dec 03 15:17:20 crc kubenswrapper[4805]: I1203 15:17:20.529087 4805 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-b9vtg"] Dec 03 15:17:20 crc kubenswrapper[4805]: I1203 15:17:20.546763 4805 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-b9vtg"] Dec 03 15:17:20 crc kubenswrapper[4805]: I1203 15:17:20.561216 4805 scope.go:117] "RemoveContainer" containerID="c93c7390d71c68e650f9079c4bcde56605294da20bf42c762602c14f3e17c2d3" Dec 03 15:17:20 crc kubenswrapper[4805]: I1203 15:17:20.710944 4805 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="629e29e5-26f0-4ad2-93cb-b7101572c7ec" path="/var/lib/kubelet/pods/629e29e5-26f0-4ad2-93cb-b7101572c7ec/volumes" Dec 03 15:17:24 crc kubenswrapper[4805]: I1203 15:17:24.670608 4805 scope.go:117] "RemoveContainer" containerID="cbf73c6c9ba740c0a523c84ca23c0851292d6eba7a0fb47d38599acca964e9cc" Dec 03 15:17:24 crc kubenswrapper[4805]: E1203 15:17:24.671770 4805 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-gskh4_openshift-machine-config-operator(6098937f-e3f6-45e8-a647-4994a79cd711)\"" pod="openshift-machine-config-operator/machine-config-daemon-gskh4" podUID="6098937f-e3f6-45e8-a647-4994a79cd711" Dec 03 15:17:34 crc kubenswrapper[4805]: I1203 15:17:34.845921 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-n5xns/crc-debug-s9275" event={"ID":"085d787e-1813-460e-8c3d-c55d60231820","Type":"ContainerStarted","Data":"dc7bbaeac341c8029cb0d6353b619fb0fe48d797e008b06df3b2585a93983909"} Dec 03 15:17:34 crc kubenswrapper[4805]: I1203 15:17:34.867511 4805 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-must-gather-n5xns/crc-debug-s9275" podStartSLOduration=1.587972012 podStartE2EDuration="30.867481663s" podCreationTimestamp="2025-12-03 15:17:04 +0000 UTC" firstStartedPulling="2025-12-03 15:17:04.942589685 +0000 UTC m=+4054.605506608" lastFinishedPulling="2025-12-03 15:17:34.222099326 +0000 UTC m=+4083.885016259" observedRunningTime="2025-12-03 15:17:34.859826946 +0000 UTC m=+4084.522743869" watchObservedRunningTime="2025-12-03 15:17:34.867481663 +0000 UTC m=+4084.530398586" Dec 03 15:17:38 crc kubenswrapper[4805]: I1203 15:17:38.695076 4805 scope.go:117] "RemoveContainer" containerID="cbf73c6c9ba740c0a523c84ca23c0851292d6eba7a0fb47d38599acca964e9cc" Dec 03 15:17:38 crc kubenswrapper[4805]: E1203 15:17:38.696025 4805 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-gskh4_openshift-machine-config-operator(6098937f-e3f6-45e8-a647-4994a79cd711)\"" pod="openshift-machine-config-operator/machine-config-daemon-gskh4" podUID="6098937f-e3f6-45e8-a647-4994a79cd711" Dec 03 15:17:46 crc kubenswrapper[4805]: I1203 15:17:46.844567 4805 scope.go:117] "RemoveContainer" containerID="1135817e28cc496283cd78507506d99e61d52c17f113524d13e4819dc7b81b6c" Dec 03 15:17:47 crc kubenswrapper[4805]: I1203 15:17:47.587376 4805 scope.go:117] "RemoveContainer" containerID="bc24751412043860962da494df8125f35012c09ca0121fb537567c9a2fa73eb2" Dec 03 15:17:47 crc kubenswrapper[4805]: I1203 15:17:47.630960 4805 scope.go:117] "RemoveContainer" containerID="77c28a5291f1dd9620171541a00ba97f126d9846fbd141605e590afdadc110b3" Dec 03 15:17:50 crc kubenswrapper[4805]: I1203 15:17:50.703516 4805 scope.go:117] "RemoveContainer" containerID="cbf73c6c9ba740c0a523c84ca23c0851292d6eba7a0fb47d38599acca964e9cc" Dec 03 15:17:50 crc kubenswrapper[4805]: E1203 15:17:50.704098 4805 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-gskh4_openshift-machine-config-operator(6098937f-e3f6-45e8-a647-4994a79cd711)\"" pod="openshift-machine-config-operator/machine-config-daemon-gskh4" podUID="6098937f-e3f6-45e8-a647-4994a79cd711" Dec 03 15:17:53 crc kubenswrapper[4805]: I1203 15:17:53.054201 4805 generic.go:334] "Generic (PLEG): container finished" podID="085d787e-1813-460e-8c3d-c55d60231820" containerID="dc7bbaeac341c8029cb0d6353b619fb0fe48d797e008b06df3b2585a93983909" exitCode=0 Dec 03 15:17:53 crc kubenswrapper[4805]: I1203 15:17:53.054242 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-n5xns/crc-debug-s9275" event={"ID":"085d787e-1813-460e-8c3d-c55d60231820","Type":"ContainerDied","Data":"dc7bbaeac341c8029cb0d6353b619fb0fe48d797e008b06df3b2585a93983909"} Dec 03 15:17:54 crc kubenswrapper[4805]: I1203 15:17:54.191056 4805 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-n5xns/crc-debug-s9275" Dec 03 15:17:54 crc kubenswrapper[4805]: I1203 15:17:54.225267 4805 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-n5xns/crc-debug-s9275"] Dec 03 15:17:54 crc kubenswrapper[4805]: I1203 15:17:54.235724 4805 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-n5xns/crc-debug-s9275"] Dec 03 15:17:54 crc kubenswrapper[4805]: I1203 15:17:54.268801 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jd9x6\" (UniqueName: \"kubernetes.io/projected/085d787e-1813-460e-8c3d-c55d60231820-kube-api-access-jd9x6\") pod \"085d787e-1813-460e-8c3d-c55d60231820\" (UID: \"085d787e-1813-460e-8c3d-c55d60231820\") " Dec 03 15:17:54 crc kubenswrapper[4805]: I1203 15:17:54.269102 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/085d787e-1813-460e-8c3d-c55d60231820-host\") pod \"085d787e-1813-460e-8c3d-c55d60231820\" (UID: \"085d787e-1813-460e-8c3d-c55d60231820\") " Dec 03 15:17:54 crc kubenswrapper[4805]: I1203 15:17:54.269343 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/085d787e-1813-460e-8c3d-c55d60231820-host" (OuterVolumeSpecName: "host") pod "085d787e-1813-460e-8c3d-c55d60231820" (UID: "085d787e-1813-460e-8c3d-c55d60231820"). InnerVolumeSpecName "host". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 03 15:17:54 crc kubenswrapper[4805]: I1203 15:17:54.269945 4805 reconciler_common.go:293] "Volume detached for volume \"host\" (UniqueName: \"kubernetes.io/host-path/085d787e-1813-460e-8c3d-c55d60231820-host\") on node \"crc\" DevicePath \"\"" Dec 03 15:17:54 crc kubenswrapper[4805]: I1203 15:17:54.276247 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/085d787e-1813-460e-8c3d-c55d60231820-kube-api-access-jd9x6" (OuterVolumeSpecName: "kube-api-access-jd9x6") pod "085d787e-1813-460e-8c3d-c55d60231820" (UID: "085d787e-1813-460e-8c3d-c55d60231820"). InnerVolumeSpecName "kube-api-access-jd9x6". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 15:17:54 crc kubenswrapper[4805]: I1203 15:17:54.371937 4805 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jd9x6\" (UniqueName: \"kubernetes.io/projected/085d787e-1813-460e-8c3d-c55d60231820-kube-api-access-jd9x6\") on node \"crc\" DevicePath \"\"" Dec 03 15:17:54 crc kubenswrapper[4805]: I1203 15:17:54.715063 4805 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="085d787e-1813-460e-8c3d-c55d60231820" path="/var/lib/kubelet/pods/085d787e-1813-460e-8c3d-c55d60231820/volumes" Dec 03 15:17:55 crc kubenswrapper[4805]: I1203 15:17:55.078681 4805 scope.go:117] "RemoveContainer" containerID="dc7bbaeac341c8029cb0d6353b619fb0fe48d797e008b06df3b2585a93983909" Dec 03 15:17:55 crc kubenswrapper[4805]: I1203 15:17:55.078746 4805 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-n5xns/crc-debug-s9275" Dec 03 15:17:55 crc kubenswrapper[4805]: I1203 15:17:55.500199 4805 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-n5xns/crc-debug-vptzf"] Dec 03 15:17:55 crc kubenswrapper[4805]: E1203 15:17:55.500727 4805 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="085d787e-1813-460e-8c3d-c55d60231820" containerName="container-00" Dec 03 15:17:55 crc kubenswrapper[4805]: I1203 15:17:55.500749 4805 state_mem.go:107] "Deleted CPUSet assignment" podUID="085d787e-1813-460e-8c3d-c55d60231820" containerName="container-00" Dec 03 15:17:55 crc kubenswrapper[4805]: E1203 15:17:55.500809 4805 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="629e29e5-26f0-4ad2-93cb-b7101572c7ec" containerName="registry-server" Dec 03 15:17:55 crc kubenswrapper[4805]: I1203 15:17:55.500818 4805 state_mem.go:107] "Deleted CPUSet assignment" podUID="629e29e5-26f0-4ad2-93cb-b7101572c7ec" containerName="registry-server" Dec 03 15:17:55 crc kubenswrapper[4805]: E1203 15:17:55.500827 4805 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="629e29e5-26f0-4ad2-93cb-b7101572c7ec" containerName="extract-content" Dec 03 15:17:55 crc kubenswrapper[4805]: I1203 15:17:55.500835 4805 state_mem.go:107] "Deleted CPUSet assignment" podUID="629e29e5-26f0-4ad2-93cb-b7101572c7ec" containerName="extract-content" Dec 03 15:17:55 crc kubenswrapper[4805]: E1203 15:17:55.500874 4805 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="629e29e5-26f0-4ad2-93cb-b7101572c7ec" containerName="extract-utilities" Dec 03 15:17:55 crc kubenswrapper[4805]: I1203 15:17:55.500905 4805 state_mem.go:107] "Deleted CPUSet assignment" podUID="629e29e5-26f0-4ad2-93cb-b7101572c7ec" containerName="extract-utilities" Dec 03 15:17:55 crc kubenswrapper[4805]: I1203 15:17:55.501160 4805 memory_manager.go:354] "RemoveStaleState removing state" podUID="085d787e-1813-460e-8c3d-c55d60231820" containerName="container-00" Dec 03 15:17:55 crc kubenswrapper[4805]: I1203 15:17:55.501182 4805 memory_manager.go:354] "RemoveStaleState removing state" podUID="629e29e5-26f0-4ad2-93cb-b7101572c7ec" containerName="registry-server" Dec 03 15:17:55 crc kubenswrapper[4805]: I1203 15:17:55.502170 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-n5xns/crc-debug-vptzf" Dec 03 15:17:55 crc kubenswrapper[4805]: I1203 15:17:55.592944 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-m9sz4\" (UniqueName: \"kubernetes.io/projected/e77ba0b2-883a-410f-ab6b-c9fb086a31d6-kube-api-access-m9sz4\") pod \"crc-debug-vptzf\" (UID: \"e77ba0b2-883a-410f-ab6b-c9fb086a31d6\") " pod="openshift-must-gather-n5xns/crc-debug-vptzf" Dec 03 15:17:55 crc kubenswrapper[4805]: I1203 15:17:55.593164 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/e77ba0b2-883a-410f-ab6b-c9fb086a31d6-host\") pod \"crc-debug-vptzf\" (UID: \"e77ba0b2-883a-410f-ab6b-c9fb086a31d6\") " pod="openshift-must-gather-n5xns/crc-debug-vptzf" Dec 03 15:17:55 crc kubenswrapper[4805]: I1203 15:17:55.695744 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/e77ba0b2-883a-410f-ab6b-c9fb086a31d6-host\") pod \"crc-debug-vptzf\" (UID: \"e77ba0b2-883a-410f-ab6b-c9fb086a31d6\") " pod="openshift-must-gather-n5xns/crc-debug-vptzf" Dec 03 15:17:55 crc kubenswrapper[4805]: I1203 15:17:55.695873 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-m9sz4\" (UniqueName: \"kubernetes.io/projected/e77ba0b2-883a-410f-ab6b-c9fb086a31d6-kube-api-access-m9sz4\") pod \"crc-debug-vptzf\" (UID: \"e77ba0b2-883a-410f-ab6b-c9fb086a31d6\") " pod="openshift-must-gather-n5xns/crc-debug-vptzf" Dec 03 15:17:55 crc kubenswrapper[4805]: I1203 15:17:55.696348 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/e77ba0b2-883a-410f-ab6b-c9fb086a31d6-host\") pod \"crc-debug-vptzf\" (UID: \"e77ba0b2-883a-410f-ab6b-c9fb086a31d6\") " pod="openshift-must-gather-n5xns/crc-debug-vptzf" Dec 03 15:17:55 crc kubenswrapper[4805]: I1203 15:17:55.721819 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-m9sz4\" (UniqueName: \"kubernetes.io/projected/e77ba0b2-883a-410f-ab6b-c9fb086a31d6-kube-api-access-m9sz4\") pod \"crc-debug-vptzf\" (UID: \"e77ba0b2-883a-410f-ab6b-c9fb086a31d6\") " pod="openshift-must-gather-n5xns/crc-debug-vptzf" Dec 03 15:17:55 crc kubenswrapper[4805]: I1203 15:17:55.824140 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-n5xns/crc-debug-vptzf" Dec 03 15:17:56 crc kubenswrapper[4805]: I1203 15:17:56.091867 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-n5xns/crc-debug-vptzf" event={"ID":"e77ba0b2-883a-410f-ab6b-c9fb086a31d6","Type":"ContainerStarted","Data":"f8f5f0a407c86a36669320f716e889dddae0734ce32537dd89af343fac0e28f0"} Dec 03 15:17:57 crc kubenswrapper[4805]: I1203 15:17:57.104684 4805 generic.go:334] "Generic (PLEG): container finished" podID="e77ba0b2-883a-410f-ab6b-c9fb086a31d6" containerID="341fb393b9c282af25232d6075c9d2c64085bbd686b935116a3d347c78779b9a" exitCode=1 Dec 03 15:17:57 crc kubenswrapper[4805]: I1203 15:17:57.104730 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-n5xns/crc-debug-vptzf" event={"ID":"e77ba0b2-883a-410f-ab6b-c9fb086a31d6","Type":"ContainerDied","Data":"341fb393b9c282af25232d6075c9d2c64085bbd686b935116a3d347c78779b9a"} Dec 03 15:17:57 crc kubenswrapper[4805]: I1203 15:17:57.140241 4805 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-n5xns/crc-debug-vptzf"] Dec 03 15:17:57 crc kubenswrapper[4805]: I1203 15:17:57.150543 4805 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-n5xns/crc-debug-vptzf"] Dec 03 15:17:58 crc kubenswrapper[4805]: I1203 15:17:58.272679 4805 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-n5xns/crc-debug-vptzf" Dec 03 15:17:58 crc kubenswrapper[4805]: I1203 15:17:58.367392 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-m9sz4\" (UniqueName: \"kubernetes.io/projected/e77ba0b2-883a-410f-ab6b-c9fb086a31d6-kube-api-access-m9sz4\") pod \"e77ba0b2-883a-410f-ab6b-c9fb086a31d6\" (UID: \"e77ba0b2-883a-410f-ab6b-c9fb086a31d6\") " Dec 03 15:17:58 crc kubenswrapper[4805]: I1203 15:17:58.367620 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/e77ba0b2-883a-410f-ab6b-c9fb086a31d6-host\") pod \"e77ba0b2-883a-410f-ab6b-c9fb086a31d6\" (UID: \"e77ba0b2-883a-410f-ab6b-c9fb086a31d6\") " Dec 03 15:17:58 crc kubenswrapper[4805]: I1203 15:17:58.368278 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/e77ba0b2-883a-410f-ab6b-c9fb086a31d6-host" (OuterVolumeSpecName: "host") pod "e77ba0b2-883a-410f-ab6b-c9fb086a31d6" (UID: "e77ba0b2-883a-410f-ab6b-c9fb086a31d6"). InnerVolumeSpecName "host". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 03 15:17:58 crc kubenswrapper[4805]: I1203 15:17:58.377101 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e77ba0b2-883a-410f-ab6b-c9fb086a31d6-kube-api-access-m9sz4" (OuterVolumeSpecName: "kube-api-access-m9sz4") pod "e77ba0b2-883a-410f-ab6b-c9fb086a31d6" (UID: "e77ba0b2-883a-410f-ab6b-c9fb086a31d6"). InnerVolumeSpecName "kube-api-access-m9sz4". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 15:17:58 crc kubenswrapper[4805]: I1203 15:17:58.469944 4805 reconciler_common.go:293] "Volume detached for volume \"host\" (UniqueName: \"kubernetes.io/host-path/e77ba0b2-883a-410f-ab6b-c9fb086a31d6-host\") on node \"crc\" DevicePath \"\"" Dec 03 15:17:58 crc kubenswrapper[4805]: I1203 15:17:58.469994 4805 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-m9sz4\" (UniqueName: \"kubernetes.io/projected/e77ba0b2-883a-410f-ab6b-c9fb086a31d6-kube-api-access-m9sz4\") on node \"crc\" DevicePath \"\"" Dec 03 15:17:58 crc kubenswrapper[4805]: I1203 15:17:58.711535 4805 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e77ba0b2-883a-410f-ab6b-c9fb086a31d6" path="/var/lib/kubelet/pods/e77ba0b2-883a-410f-ab6b-c9fb086a31d6/volumes" Dec 03 15:17:59 crc kubenswrapper[4805]: I1203 15:17:59.129553 4805 scope.go:117] "RemoveContainer" containerID="341fb393b9c282af25232d6075c9d2c64085bbd686b935116a3d347c78779b9a" Dec 03 15:17:59 crc kubenswrapper[4805]: I1203 15:17:59.130152 4805 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-n5xns/crc-debug-vptzf" Dec 03 15:18:03 crc kubenswrapper[4805]: I1203 15:18:03.694384 4805 scope.go:117] "RemoveContainer" containerID="cbf73c6c9ba740c0a523c84ca23c0851292d6eba7a0fb47d38599acca964e9cc" Dec 03 15:18:03 crc kubenswrapper[4805]: E1203 15:18:03.695185 4805 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-gskh4_openshift-machine-config-operator(6098937f-e3f6-45e8-a647-4994a79cd711)\"" pod="openshift-machine-config-operator/machine-config-daemon-gskh4" podUID="6098937f-e3f6-45e8-a647-4994a79cd711" Dec 03 15:18:14 crc kubenswrapper[4805]: I1203 15:18:14.694384 4805 scope.go:117] "RemoveContainer" containerID="cbf73c6c9ba740c0a523c84ca23c0851292d6eba7a0fb47d38599acca964e9cc" Dec 03 15:18:14 crc kubenswrapper[4805]: E1203 15:18:14.696315 4805 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-gskh4_openshift-machine-config-operator(6098937f-e3f6-45e8-a647-4994a79cd711)\"" pod="openshift-machine-config-operator/machine-config-daemon-gskh4" podUID="6098937f-e3f6-45e8-a647-4994a79cd711" Dec 03 15:18:28 crc kubenswrapper[4805]: I1203 15:18:28.694169 4805 scope.go:117] "RemoveContainer" containerID="cbf73c6c9ba740c0a523c84ca23c0851292d6eba7a0fb47d38599acca964e9cc" Dec 03 15:18:28 crc kubenswrapper[4805]: E1203 15:18:28.695006 4805 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-gskh4_openshift-machine-config-operator(6098937f-e3f6-45e8-a647-4994a79cd711)\"" pod="openshift-machine-config-operator/machine-config-daemon-gskh4" podUID="6098937f-e3f6-45e8-a647-4994a79cd711" Dec 03 15:18:35 crc kubenswrapper[4805]: I1203 15:18:35.840825 4805 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_alertmanager-metric-storage-0_a0412bf9-ae66-4d99-9bb3-ea2fe4637678/init-config-reloader/0.log" Dec 03 15:18:36 crc kubenswrapper[4805]: I1203 15:18:36.046548 4805 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_alertmanager-metric-storage-0_a0412bf9-ae66-4d99-9bb3-ea2fe4637678/init-config-reloader/0.log" Dec 03 15:18:36 crc kubenswrapper[4805]: I1203 15:18:36.061133 4805 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_alertmanager-metric-storage-0_a0412bf9-ae66-4d99-9bb3-ea2fe4637678/alertmanager/0.log" Dec 03 15:18:36 crc kubenswrapper[4805]: I1203 15:18:36.085229 4805 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_alertmanager-metric-storage-0_a0412bf9-ae66-4d99-9bb3-ea2fe4637678/config-reloader/0.log" Dec 03 15:18:36 crc kubenswrapper[4805]: I1203 15:18:36.248190 4805 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_aodh-0_4297360e-d110-459b-b8b9-75f4a201921e/aodh-evaluator/0.log" Dec 03 15:18:36 crc kubenswrapper[4805]: I1203 15:18:36.270905 4805 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_aodh-0_4297360e-d110-459b-b8b9-75f4a201921e/aodh-api/0.log" Dec 03 15:18:36 crc kubenswrapper[4805]: I1203 15:18:36.315388 4805 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_aodh-0_4297360e-d110-459b-b8b9-75f4a201921e/aodh-listener/0.log" Dec 03 15:18:36 crc kubenswrapper[4805]: I1203 15:18:36.397637 4805 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_aodh-0_4297360e-d110-459b-b8b9-75f4a201921e/aodh-notifier/0.log" Dec 03 15:18:36 crc kubenswrapper[4805]: I1203 15:18:36.547488 4805 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-api-58794647f4-rtjlq_5a55a37b-c678-481e-bf36-a8961f5b26ed/barbican-api-log/0.log" Dec 03 15:18:36 crc kubenswrapper[4805]: I1203 15:18:36.560907 4805 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-api-58794647f4-rtjlq_5a55a37b-c678-481e-bf36-a8961f5b26ed/barbican-api/0.log" Dec 03 15:18:36 crc kubenswrapper[4805]: I1203 15:18:36.738853 4805 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-keystone-listener-8484c694-6frjp_18349f43-5c89-4c11-8ff2-fee30d6cceb7/barbican-keystone-listener/0.log" Dec 03 15:18:36 crc kubenswrapper[4805]: I1203 15:18:36.847643 4805 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-keystone-listener-8484c694-6frjp_18349f43-5c89-4c11-8ff2-fee30d6cceb7/barbican-keystone-listener-log/0.log" Dec 03 15:18:36 crc kubenswrapper[4805]: I1203 15:18:36.905981 4805 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-worker-8559568497-vwxwh_93ad1063-a321-44a3-bd91-c00c7ba1ef96/barbican-worker/0.log" Dec 03 15:18:36 crc kubenswrapper[4805]: I1203 15:18:36.970150 4805 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-worker-8559568497-vwxwh_93ad1063-a321-44a3-bd91-c00c7ba1ef96/barbican-worker-log/0.log" Dec 03 15:18:37 crc kubenswrapper[4805]: I1203 15:18:37.061978 4805 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_bootstrap-edpm-deployment-openstack-edpm-ipam-brwqg_7dde2b60-d63b-4989-b76d-ec802af8f951/bootstrap-edpm-deployment-openstack-edpm-ipam/0.log" Dec 03 15:18:37 crc kubenswrapper[4805]: I1203 15:18:37.217036 4805 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_df32824d-ca3b-4ee5-88a2-1e4591fafc04/ceilometer-central-agent/0.log" Dec 03 15:18:37 crc kubenswrapper[4805]: I1203 15:18:37.281690 4805 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_df32824d-ca3b-4ee5-88a2-1e4591fafc04/ceilometer-notification-agent/0.log" Dec 03 15:18:37 crc kubenswrapper[4805]: I1203 15:18:37.325818 4805 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_df32824d-ca3b-4ee5-88a2-1e4591fafc04/proxy-httpd/0.log" Dec 03 15:18:37 crc kubenswrapper[4805]: I1203 15:18:37.408526 4805 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_df32824d-ca3b-4ee5-88a2-1e4591fafc04/sg-core/0.log" Dec 03 15:18:37 crc kubenswrapper[4805]: I1203 15:18:37.519153 4805 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-api-0_7b5b9f5a-5af6-41d7-be26-3f4c091098b7/cinder-api/0.log" Dec 03 15:18:37 crc kubenswrapper[4805]: I1203 15:18:37.596303 4805 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-api-0_7b5b9f5a-5af6-41d7-be26-3f4c091098b7/cinder-api-log/0.log" Dec 03 15:18:38 crc kubenswrapper[4805]: I1203 15:18:38.090803 4805 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-scheduler-0_6d34a555-7003-4796-a15a-85a71c489bc9/cinder-scheduler/0.log" Dec 03 15:18:38 crc kubenswrapper[4805]: I1203 15:18:38.163583 4805 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-scheduler-0_6d34a555-7003-4796-a15a-85a71c489bc9/probe/0.log" Dec 03 15:18:38 crc kubenswrapper[4805]: I1203 15:18:38.184219 4805 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_configure-network-edpm-deployment-openstack-edpm-ipam-6wncz_546a9824-86ca-46ed-b6fa-3384de230115/configure-network-edpm-deployment-openstack-edpm-ipam/0.log" Dec 03 15:18:38 crc kubenswrapper[4805]: I1203 15:18:38.416763 4805 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_configure-os-edpm-deployment-openstack-edpm-ipam-5bpz7_2aa97cdd-3fbb-4e80-b82a-fb688c8c92d0/configure-os-edpm-deployment-openstack-edpm-ipam/0.log" Dec 03 15:18:38 crc kubenswrapper[4805]: I1203 15:18:38.453887 4805 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_dnsmasq-dns-bb85b8995-trmpv_7d390cdb-18d5-49b3-8b24-1c2203442042/init/0.log" Dec 03 15:18:38 crc kubenswrapper[4805]: I1203 15:18:38.672691 4805 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_dnsmasq-dns-bb85b8995-trmpv_7d390cdb-18d5-49b3-8b24-1c2203442042/init/0.log" Dec 03 15:18:38 crc kubenswrapper[4805]: I1203 15:18:38.724021 4805 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_dnsmasq-dns-bb85b8995-trmpv_7d390cdb-18d5-49b3-8b24-1c2203442042/dnsmasq-dns/0.log" Dec 03 15:18:38 crc kubenswrapper[4805]: I1203 15:18:38.755577 4805 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_download-cache-edpm-deployment-openstack-edpm-ipam-nvbww_a6f87a69-c7bb-44d8-9f65-82b1b4f8885a/download-cache-edpm-deployment-openstack-edpm-ipam/0.log" Dec 03 15:18:39 crc kubenswrapper[4805]: I1203 15:18:39.136762 4805 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-external-api-0_6a4ca0f4-cff3-4c57-ab1b-3a01d7d6576c/glance-log/0.log" Dec 03 15:18:39 crc kubenswrapper[4805]: I1203 15:18:39.138636 4805 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-external-api-0_6a4ca0f4-cff3-4c57-ab1b-3a01d7d6576c/glance-httpd/0.log" Dec 03 15:18:39 crc kubenswrapper[4805]: I1203 15:18:39.368218 4805 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-internal-api-0_cbb9e604-033b-4586-b0fd-7300a5e30897/glance-log/0.log" Dec 03 15:18:39 crc kubenswrapper[4805]: I1203 15:18:39.383105 4805 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-internal-api-0_cbb9e604-033b-4586-b0fd-7300a5e30897/glance-httpd/0.log" Dec 03 15:18:39 crc kubenswrapper[4805]: I1203 15:18:39.695342 4805 scope.go:117] "RemoveContainer" containerID="cbf73c6c9ba740c0a523c84ca23c0851292d6eba7a0fb47d38599acca964e9cc" Dec 03 15:18:39 crc kubenswrapper[4805]: E1203 15:18:39.695661 4805 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-gskh4_openshift-machine-config-operator(6098937f-e3f6-45e8-a647-4994a79cd711)\"" pod="openshift-machine-config-operator/machine-config-daemon-gskh4" podUID="6098937f-e3f6-45e8-a647-4994a79cd711" Dec 03 15:18:40 crc kubenswrapper[4805]: I1203 15:18:40.385584 4805 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_heat-engine-76dcfc959c-swfs4_854cde2d-c692-4ab4-9172-02ad6a2b2f6a/heat-engine/0.log" Dec 03 15:18:40 crc kubenswrapper[4805]: I1203 15:18:40.610287 4805 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_install-certs-edpm-deployment-openstack-edpm-ipam-tw4dn_9cd62272-1077-4e13-8e99-c158fbb8ba1a/install-certs-edpm-deployment-openstack-edpm-ipam/0.log" Dec 03 15:18:40 crc kubenswrapper[4805]: I1203 15:18:40.684655 4805 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_heat-cfnapi-68589fbcdd-snszz_cde129c0-8757-4996-8974-c553a5a48f17/heat-cfnapi/0.log" Dec 03 15:18:40 crc kubenswrapper[4805]: I1203 15:18:40.758172 4805 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_heat-api-84746cbddc-kc7hm_078bbd61-70d1-4e2c-ab8c-eeba4127f662/heat-api/0.log" Dec 03 15:18:40 crc kubenswrapper[4805]: I1203 15:18:40.855998 4805 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_install-os-edpm-deployment-openstack-edpm-ipam-v79c2_b96a6fd3-b6a0-4580-9504-9ac4237ad37e/install-os-edpm-deployment-openstack-edpm-ipam/0.log" Dec 03 15:18:41 crc kubenswrapper[4805]: I1203 15:18:41.025173 4805 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_keystone-585fb5cffc-n26gq_3e723db7-3ea3-4360-b5b8-49f1e0524f8b/keystone-api/0.log" Dec 03 15:18:41 crc kubenswrapper[4805]: I1203 15:18:41.062863 4805 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_keystone-cron-29412901-wgktp_ae49c111-f5d4-4884-857c-d03296d16312/keystone-cron/0.log" Dec 03 15:18:41 crc kubenswrapper[4805]: I1203 15:18:41.090349 4805 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_kube-state-metrics-0_91f65152-9858-4a6f-ba45-9e0556490322/kube-state-metrics/0.log" Dec 03 15:18:41 crc kubenswrapper[4805]: I1203 15:18:41.272047 4805 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_libvirt-edpm-deployment-openstack-edpm-ipam-lmlms_1776968b-6613-4b93-bb53-a91c3b6c3d2c/libvirt-edpm-deployment-openstack-edpm-ipam/0.log" Dec 03 15:18:41 crc kubenswrapper[4805]: I1203 15:18:41.519611 4805 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-9764b468f-7sr7n_31537e49-7f76-4c8e-8c14-3fe55f9615b9/neutron-api/0.log" Dec 03 15:18:41 crc kubenswrapper[4805]: I1203 15:18:41.531202 4805 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-9764b468f-7sr7n_31537e49-7f76-4c8e-8c14-3fe55f9615b9/neutron-httpd/0.log" Dec 03 15:18:41 crc kubenswrapper[4805]: I1203 15:18:41.676614 4805 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-metadata-edpm-deployment-openstack-edpm-ipam-s4fzz_779c86b3-bd17-4893-b976-e668607675f2/neutron-metadata-edpm-deployment-openstack-edpm-ipam/0.log" Dec 03 15:18:41 crc kubenswrapper[4805]: I1203 15:18:41.929407 4805 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-api-0_65386594-9cac-40ef-b3a8-2f3ee225bfb6/nova-api-log/0.log" Dec 03 15:18:42 crc kubenswrapper[4805]: I1203 15:18:42.169998 4805 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell0-conductor-0_bc143d9b-fc34-4d12-a500-67de79d3c71a/nova-cell0-conductor-conductor/0.log" Dec 03 15:18:42 crc kubenswrapper[4805]: I1203 15:18:42.179240 4805 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-api-0_65386594-9cac-40ef-b3a8-2f3ee225bfb6/nova-api-api/0.log" Dec 03 15:18:42 crc kubenswrapper[4805]: I1203 15:18:42.299891 4805 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell1-conductor-0_b3635de5-a6a4-465c-8793-bce58d357da6/nova-cell1-conductor-conductor/0.log" Dec 03 15:18:42 crc kubenswrapper[4805]: I1203 15:18:42.503787 4805 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell1-novncproxy-0_d386b865-1a97-468b-ba0e-7733abb94034/nova-cell1-novncproxy-novncproxy/0.log" Dec 03 15:18:42 crc kubenswrapper[4805]: I1203 15:18:42.849396 4805 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-edpm-deployment-openstack-edpm-ipam-75h75_7e18e1db-7529-412b-bf01-17cb78e689b9/nova-edpm-deployment-openstack-edpm-ipam/0.log" Dec 03 15:18:43 crc kubenswrapper[4805]: I1203 15:18:43.096087 4805 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-metadata-0_9acb704f-0e77-458e-b31e-7b32cf8619e5/nova-metadata-log/0.log" Dec 03 15:18:43 crc kubenswrapper[4805]: I1203 15:18:43.498864 4805 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-scheduler-0_e30e8472-484d-47a9-bb8f-839d41ec04be/nova-scheduler-scheduler/0.log" Dec 03 15:18:43 crc kubenswrapper[4805]: I1203 15:18:43.559064 4805 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-cell1-galera-0_c13cbc51-432a-419a-9469-37fd31a82dd9/mysql-bootstrap/0.log" Dec 03 15:18:43 crc kubenswrapper[4805]: I1203 15:18:43.749358 4805 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-cell1-galera-0_c13cbc51-432a-419a-9469-37fd31a82dd9/galera/0.log" Dec 03 15:18:43 crc kubenswrapper[4805]: I1203 15:18:43.756742 4805 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-cell1-galera-0_c13cbc51-432a-419a-9469-37fd31a82dd9/mysql-bootstrap/0.log" Dec 03 15:18:44 crc kubenswrapper[4805]: I1203 15:18:44.213116 4805 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-galera-0_5b4bb902-dba6-4ec6-97bf-532d996e918d/mysql-bootstrap/0.log" Dec 03 15:18:44 crc kubenswrapper[4805]: I1203 15:18:44.427521 4805 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-metadata-0_9acb704f-0e77-458e-b31e-7b32cf8619e5/nova-metadata-metadata/0.log" Dec 03 15:18:44 crc kubenswrapper[4805]: I1203 15:18:44.443802 4805 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-galera-0_5b4bb902-dba6-4ec6-97bf-532d996e918d/mysql-bootstrap/0.log" Dec 03 15:18:44 crc kubenswrapper[4805]: I1203 15:18:44.471801 4805 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-galera-0_5b4bb902-dba6-4ec6-97bf-532d996e918d/galera/0.log" Dec 03 15:18:44 crc kubenswrapper[4805]: I1203 15:18:44.636453 4805 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstackclient_49d35952-8526-46a9-80f5-b0fe3a359abd/openstackclient/0.log" Dec 03 15:18:44 crc kubenswrapper[4805]: I1203 15:18:44.780518 4805 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-metrics-7kmnp_76002c62-88cb-4f74-816a-3b2be4f8ba4c/openstack-network-exporter/0.log" Dec 03 15:18:44 crc kubenswrapper[4805]: I1203 15:18:44.933058 4805 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-dqwfk_4335461c-d070-4c88-9d40-980109c57b4c/ovsdb-server-init/0.log" Dec 03 15:18:45 crc kubenswrapper[4805]: I1203 15:18:45.040727 4805 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-dqwfk_4335461c-d070-4c88-9d40-980109c57b4c/ovs-vswitchd/0.log" Dec 03 15:18:45 crc kubenswrapper[4805]: I1203 15:18:45.079912 4805 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-dqwfk_4335461c-d070-4c88-9d40-980109c57b4c/ovsdb-server-init/0.log" Dec 03 15:18:45 crc kubenswrapper[4805]: I1203 15:18:45.165624 4805 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-dqwfk_4335461c-d070-4c88-9d40-980109c57b4c/ovsdb-server/0.log" Dec 03 15:18:45 crc kubenswrapper[4805]: I1203 15:18:45.348502 4805 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-zhnwt_139abefe-12d4-4d8f-bdc6-43e71c818a65/ovn-controller/0.log" Dec 03 15:18:45 crc kubenswrapper[4805]: I1203 15:18:45.602522 4805 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-edpm-deployment-openstack-edpm-ipam-hbx5q_af960cbb-259d-409e-9140-2d66af349200/ovn-edpm-deployment-openstack-edpm-ipam/0.log" Dec 03 15:18:45 crc kubenswrapper[4805]: I1203 15:18:45.623292 4805 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-northd-0_8e6ca40c-f7f2-445e-8e22-71f8d4d4693b/openstack-network-exporter/0.log" Dec 03 15:18:45 crc kubenswrapper[4805]: I1203 15:18:45.702321 4805 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-northd-0_8e6ca40c-f7f2-445e-8e22-71f8d4d4693b/ovn-northd/0.log" Dec 03 15:18:45 crc kubenswrapper[4805]: I1203 15:18:45.857495 4805 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-nb-0_00ac9536-1677-4a3a-be82-7c8d8df91656/ovsdbserver-nb/0.log" Dec 03 15:18:45 crc kubenswrapper[4805]: I1203 15:18:45.877126 4805 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-nb-0_00ac9536-1677-4a3a-be82-7c8d8df91656/openstack-network-exporter/0.log" Dec 03 15:18:46 crc kubenswrapper[4805]: I1203 15:18:46.076621 4805 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-sb-0_17f337d8-be50-4c6c-8364-208b394be6d3/ovsdbserver-sb/0.log" Dec 03 15:18:46 crc kubenswrapper[4805]: I1203 15:18:46.093484 4805 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-sb-0_17f337d8-be50-4c6c-8364-208b394be6d3/openstack-network-exporter/0.log" Dec 03 15:18:46 crc kubenswrapper[4805]: I1203 15:18:46.387318 4805 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_placement-98cb4d7c6-cbjgt_d099a9d3-b941-499a-829f-7590e1fe7b0a/placement-log/0.log" Dec 03 15:18:46 crc kubenswrapper[4805]: I1203 15:18:46.430853 4805 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_placement-98cb4d7c6-cbjgt_d099a9d3-b941-499a-829f-7590e1fe7b0a/placement-api/0.log" Dec 03 15:18:46 crc kubenswrapper[4805]: I1203 15:18:46.726546 4805 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_prometheus-metric-storage-0_2cf69241-d762-41ec-9512-84643c4d9dc1/init-config-reloader/0.log" Dec 03 15:18:46 crc kubenswrapper[4805]: I1203 15:18:46.940812 4805 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_prometheus-metric-storage-0_2cf69241-d762-41ec-9512-84643c4d9dc1/config-reloader/0.log" Dec 03 15:18:46 crc kubenswrapper[4805]: I1203 15:18:46.952758 4805 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_prometheus-metric-storage-0_2cf69241-d762-41ec-9512-84643c4d9dc1/init-config-reloader/0.log" Dec 03 15:18:46 crc kubenswrapper[4805]: I1203 15:18:46.964557 4805 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_prometheus-metric-storage-0_2cf69241-d762-41ec-9512-84643c4d9dc1/prometheus/0.log" Dec 03 15:18:47 crc kubenswrapper[4805]: I1203 15:18:47.005908 4805 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_prometheus-metric-storage-0_2cf69241-d762-41ec-9512-84643c4d9dc1/thanos-sidecar/0.log" Dec 03 15:18:47 crc kubenswrapper[4805]: I1203 15:18:47.155963 4805 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-cell1-server-0_fff4a752-b61c-4c51-bc0b-428c9affbb00/setup-container/0.log" Dec 03 15:18:47 crc kubenswrapper[4805]: I1203 15:18:47.444314 4805 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-cell1-server-0_fff4a752-b61c-4c51-bc0b-428c9affbb00/setup-container/0.log" Dec 03 15:18:47 crc kubenswrapper[4805]: I1203 15:18:47.469027 4805 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-cell1-server-0_fff4a752-b61c-4c51-bc0b-428c9affbb00/rabbitmq/0.log" Dec 03 15:18:47 crc kubenswrapper[4805]: I1203 15:18:47.582013 4805 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-server-0_0664c16b-c578-4685-b1ba-98c5d5f3fe07/setup-container/0.log" Dec 03 15:18:47 crc kubenswrapper[4805]: I1203 15:18:47.713402 4805 scope.go:117] "RemoveContainer" containerID="3281abdcb636f86f160518adfd65f97eaeea0840cf1f23c4bc0f547621c951e0" Dec 03 15:18:47 crc kubenswrapper[4805]: I1203 15:18:47.731782 4805 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-server-0_0664c16b-c578-4685-b1ba-98c5d5f3fe07/setup-container/0.log" Dec 03 15:18:47 crc kubenswrapper[4805]: I1203 15:18:47.744544 4805 scope.go:117] "RemoveContainer" containerID="7d3a89aa81794403f2659ff62d3c987df7b23abd0412666b214f7cf5a222197f" Dec 03 15:18:47 crc kubenswrapper[4805]: I1203 15:18:47.784739 4805 scope.go:117] "RemoveContainer" containerID="e9d343bc6e2131c8837e0e3729e89035187959e3721f348fed3e119e9d4afa38" Dec 03 15:18:47 crc kubenswrapper[4805]: I1203 15:18:47.815423 4805 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-server-0_0664c16b-c578-4685-b1ba-98c5d5f3fe07/rabbitmq/0.log" Dec 03 15:18:47 crc kubenswrapper[4805]: I1203 15:18:47.966291 4805 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_reboot-os-edpm-deployment-openstack-edpm-ipam-k5scz_ba862dc9-b0f4-4ba3-888a-5eae1601dfdb/reboot-os-edpm-deployment-openstack-edpm-ipam/0.log" Dec 03 15:18:48 crc kubenswrapper[4805]: I1203 15:18:48.031692 4805 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_redhat-edpm-deployment-openstack-edpm-ipam-ltzhs_30551b1d-dce7-4c0f-85c9-300f32bf0fc5/redhat-edpm-deployment-openstack-edpm-ipam/0.log" Dec 03 15:18:48 crc kubenswrapper[4805]: I1203 15:18:48.211484 4805 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_repo-setup-edpm-deployment-openstack-edpm-ipam-lghw6_a63daecb-effd-412b-9aeb-9400cd81d763/repo-setup-edpm-deployment-openstack-edpm-ipam/0.log" Dec 03 15:18:48 crc kubenswrapper[4805]: I1203 15:18:48.703023 4805 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_run-os-edpm-deployment-openstack-edpm-ipam-lp9h9_3183d9eb-f2b5-4ac6-8cb1-917cff8a8ded/run-os-edpm-deployment-openstack-edpm-ipam/0.log" Dec 03 15:18:48 crc kubenswrapper[4805]: I1203 15:18:48.786094 4805 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ssh-known-hosts-edpm-deployment-hzqxj_5221f418-44bd-4db1-96c8-788e0df137c7/ssh-known-hosts-edpm-deployment/0.log" Dec 03 15:18:49 crc kubenswrapper[4805]: I1203 15:18:49.049096 4805 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-proxy-78bdd85b87-5tf8n_0fdb9aee-e509-4e04-88b7-61fc216cfc34/proxy-server/0.log" Dec 03 15:18:49 crc kubenswrapper[4805]: I1203 15:18:49.194454 4805 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-ring-rebalance-h55gd_acfd02d5-fb0a-4427-95e0-4c9f956ea137/swift-ring-rebalance/0.log" Dec 03 15:18:49 crc kubenswrapper[4805]: I1203 15:18:49.196009 4805 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-proxy-78bdd85b87-5tf8n_0fdb9aee-e509-4e04-88b7-61fc216cfc34/proxy-httpd/0.log" Dec 03 15:18:49 crc kubenswrapper[4805]: I1203 15:18:49.385157 4805 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_4fbc8d7f-d47f-4d91-915b-504820aac08a/account-auditor/0.log" Dec 03 15:18:49 crc kubenswrapper[4805]: I1203 15:18:49.464821 4805 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_4fbc8d7f-d47f-4d91-915b-504820aac08a/account-reaper/0.log" Dec 03 15:18:49 crc kubenswrapper[4805]: I1203 15:18:49.495890 4805 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_4fbc8d7f-d47f-4d91-915b-504820aac08a/account-replicator/0.log" Dec 03 15:18:49 crc kubenswrapper[4805]: I1203 15:18:49.538189 4805 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_4fbc8d7f-d47f-4d91-915b-504820aac08a/account-server/0.log" Dec 03 15:18:49 crc kubenswrapper[4805]: I1203 15:18:49.683270 4805 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_4fbc8d7f-d47f-4d91-915b-504820aac08a/container-auditor/0.log" Dec 03 15:18:49 crc kubenswrapper[4805]: I1203 15:18:49.709273 4805 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_4fbc8d7f-d47f-4d91-915b-504820aac08a/container-server/0.log" Dec 03 15:18:49 crc kubenswrapper[4805]: I1203 15:18:49.738573 4805 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_4fbc8d7f-d47f-4d91-915b-504820aac08a/container-replicator/0.log" Dec 03 15:18:49 crc kubenswrapper[4805]: I1203 15:18:49.842737 4805 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_4fbc8d7f-d47f-4d91-915b-504820aac08a/container-updater/0.log" Dec 03 15:18:49 crc kubenswrapper[4805]: I1203 15:18:49.956979 4805 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_4fbc8d7f-d47f-4d91-915b-504820aac08a/object-auditor/0.log" Dec 03 15:18:49 crc kubenswrapper[4805]: I1203 15:18:49.967905 4805 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_4fbc8d7f-d47f-4d91-915b-504820aac08a/object-replicator/0.log" Dec 03 15:18:49 crc kubenswrapper[4805]: I1203 15:18:49.986781 4805 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_4fbc8d7f-d47f-4d91-915b-504820aac08a/object-expirer/0.log" Dec 03 15:18:50 crc kubenswrapper[4805]: I1203 15:18:50.301137 4805 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_4fbc8d7f-d47f-4d91-915b-504820aac08a/object-server/0.log" Dec 03 15:18:50 crc kubenswrapper[4805]: I1203 15:18:50.359709 4805 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_4fbc8d7f-d47f-4d91-915b-504820aac08a/object-updater/0.log" Dec 03 15:18:50 crc kubenswrapper[4805]: I1203 15:18:50.785676 4805 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_4fbc8d7f-d47f-4d91-915b-504820aac08a/rsync/0.log" Dec 03 15:18:50 crc kubenswrapper[4805]: I1203 15:18:50.787183 4805 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_4fbc8d7f-d47f-4d91-915b-504820aac08a/swift-recon-cron/0.log" Dec 03 15:18:50 crc kubenswrapper[4805]: I1203 15:18:50.854206 4805 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_telemetry-edpm-deployment-openstack-edpm-ipam-ctxx4_4c03480a-7dce-48dd-9e5d-1aea5c540a72/telemetry-edpm-deployment-openstack-edpm-ipam/0.log" Dec 03 15:18:51 crc kubenswrapper[4805]: I1203 15:18:51.035650 4805 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_validate-network-edpm-deployment-openstack-edpm-ipam-7px67_dd64758f-9fe5-4f00-8f60-4bba03a0f98a/validate-network-edpm-deployment-openstack-edpm-ipam/0.log" Dec 03 15:18:51 crc kubenswrapper[4805]: I1203 15:18:51.694965 4805 scope.go:117] "RemoveContainer" containerID="cbf73c6c9ba740c0a523c84ca23c0851292d6eba7a0fb47d38599acca964e9cc" Dec 03 15:18:52 crc kubenswrapper[4805]: I1203 15:18:52.763075 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-gskh4" event={"ID":"6098937f-e3f6-45e8-a647-4994a79cd711","Type":"ContainerStarted","Data":"c1a31f4f6e3e9fa851f0e438617fecf09bad8f6021091f02eae4b71f6b3e374c"} Dec 03 15:18:59 crc kubenswrapper[4805]: I1203 15:18:59.007441 4805 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_memcached-0_3c84e84f-43fa-4163-81b2-489ec6a0d834/memcached/0.log" Dec 03 15:19:22 crc kubenswrapper[4805]: I1203 15:19:22.416477 4805 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_01b9b0248b373c1370a002b5a413a9a23eeeadf6c66c17ba9e1f19741esbt54_7811f073-78a5-49a2-8b6e-336505b43756/util/0.log" Dec 03 15:19:22 crc kubenswrapper[4805]: I1203 15:19:22.614546 4805 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_01b9b0248b373c1370a002b5a413a9a23eeeadf6c66c17ba9e1f19741esbt54_7811f073-78a5-49a2-8b6e-336505b43756/util/0.log" Dec 03 15:19:22 crc kubenswrapper[4805]: I1203 15:19:22.635208 4805 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_01b9b0248b373c1370a002b5a413a9a23eeeadf6c66c17ba9e1f19741esbt54_7811f073-78a5-49a2-8b6e-336505b43756/pull/0.log" Dec 03 15:19:22 crc kubenswrapper[4805]: I1203 15:19:22.649739 4805 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_01b9b0248b373c1370a002b5a413a9a23eeeadf6c66c17ba9e1f19741esbt54_7811f073-78a5-49a2-8b6e-336505b43756/pull/0.log" Dec 03 15:19:22 crc kubenswrapper[4805]: I1203 15:19:22.839954 4805 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_01b9b0248b373c1370a002b5a413a9a23eeeadf6c66c17ba9e1f19741esbt54_7811f073-78a5-49a2-8b6e-336505b43756/extract/0.log" Dec 03 15:19:22 crc kubenswrapper[4805]: I1203 15:19:22.855123 4805 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_01b9b0248b373c1370a002b5a413a9a23eeeadf6c66c17ba9e1f19741esbt54_7811f073-78a5-49a2-8b6e-336505b43756/util/0.log" Dec 03 15:19:22 crc kubenswrapper[4805]: I1203 15:19:22.872646 4805 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_01b9b0248b373c1370a002b5a413a9a23eeeadf6c66c17ba9e1f19741esbt54_7811f073-78a5-49a2-8b6e-336505b43756/pull/0.log" Dec 03 15:19:23 crc kubenswrapper[4805]: I1203 15:19:23.160374 4805 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_barbican-operator-controller-manager-7d9dfd778-jc4dw_cbf3813c-fb86-49d2-8fac-25487ff08601/kube-rbac-proxy/0.log" Dec 03 15:19:23 crc kubenswrapper[4805]: I1203 15:19:23.254271 4805 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_cinder-operator-controller-manager-859b6ccc6-2nzgx_f9b6b508-1cf6-40aa-829e-b9cf264921e5/kube-rbac-proxy/0.log" Dec 03 15:19:23 crc kubenswrapper[4805]: I1203 15:19:23.375472 4805 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_barbican-operator-controller-manager-7d9dfd778-jc4dw_cbf3813c-fb86-49d2-8fac-25487ff08601/manager/0.log" Dec 03 15:19:23 crc kubenswrapper[4805]: I1203 15:19:23.426444 4805 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_cinder-operator-controller-manager-859b6ccc6-2nzgx_f9b6b508-1cf6-40aa-829e-b9cf264921e5/manager/0.log" Dec 03 15:19:23 crc kubenswrapper[4805]: I1203 15:19:23.534359 4805 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_designate-operator-controller-manager-78b4bc895b-w56sr_326305da-3798-498c-9b47-0aeb71bd2205/kube-rbac-proxy/0.log" Dec 03 15:19:23 crc kubenswrapper[4805]: I1203 15:19:23.580471 4805 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_designate-operator-controller-manager-78b4bc895b-w56sr_326305da-3798-498c-9b47-0aeb71bd2205/manager/0.log" Dec 03 15:19:23 crc kubenswrapper[4805]: I1203 15:19:23.709461 4805 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_glance-operator-controller-manager-77987cd8cd-pt9fw_5615ce63-1ec1-4a77-8e44-49e6d114eec6/kube-rbac-proxy/0.log" Dec 03 15:19:23 crc kubenswrapper[4805]: I1203 15:19:23.851801 4805 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_glance-operator-controller-manager-77987cd8cd-pt9fw_5615ce63-1ec1-4a77-8e44-49e6d114eec6/manager/0.log" Dec 03 15:19:23 crc kubenswrapper[4805]: I1203 15:19:23.931708 4805 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_heat-operator-controller-manager-5f64f6f8bb-86sbc_a748cfe4-2e54-4a7b-a353-e1dfea6c14b3/kube-rbac-proxy/0.log" Dec 03 15:19:24 crc kubenswrapper[4805]: I1203 15:19:24.037325 4805 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_heat-operator-controller-manager-5f64f6f8bb-86sbc_a748cfe4-2e54-4a7b-a353-e1dfea6c14b3/manager/0.log" Dec 03 15:19:24 crc kubenswrapper[4805]: I1203 15:19:24.089829 4805 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_horizon-operator-controller-manager-68c6d99b8f-kxbr8_b29e4120-0cbc-4de5-a4b6-97b97761f49a/kube-rbac-proxy/0.log" Dec 03 15:19:24 crc kubenswrapper[4805]: I1203 15:19:24.408258 4805 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_horizon-operator-controller-manager-68c6d99b8f-kxbr8_b29e4120-0cbc-4de5-a4b6-97b97761f49a/manager/0.log" Dec 03 15:19:24 crc kubenswrapper[4805]: I1203 15:19:24.510584 4805 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_infra-operator-controller-manager-57548d458d-77t9s_f21790d8-af47-4be7-8758-f038bda76908/kube-rbac-proxy/0.log" Dec 03 15:19:24 crc kubenswrapper[4805]: I1203 15:19:24.663683 4805 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ironic-operator-controller-manager-6c548fd776-mz4pw_1ad4fa11-2f11-4a06-9944-bba8c7445e84/kube-rbac-proxy/0.log" Dec 03 15:19:24 crc kubenswrapper[4805]: I1203 15:19:24.772833 4805 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_infra-operator-controller-manager-57548d458d-77t9s_f21790d8-af47-4be7-8758-f038bda76908/manager/0.log" Dec 03 15:19:24 crc kubenswrapper[4805]: I1203 15:19:24.803323 4805 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ironic-operator-controller-manager-6c548fd776-mz4pw_1ad4fa11-2f11-4a06-9944-bba8c7445e84/manager/0.log" Dec 03 15:19:24 crc kubenswrapper[4805]: I1203 15:19:24.920038 4805 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_keystone-operator-controller-manager-7765d96ddf-qzzz9_ffa72db0-588a-4a78-a1cd-8710f5ce4f41/kube-rbac-proxy/0.log" Dec 03 15:19:25 crc kubenswrapper[4805]: I1203 15:19:25.110029 4805 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_keystone-operator-controller-manager-7765d96ddf-qzzz9_ffa72db0-588a-4a78-a1cd-8710f5ce4f41/manager/0.log" Dec 03 15:19:25 crc kubenswrapper[4805]: I1203 15:19:25.145746 4805 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_manila-operator-controller-manager-7c79b5df47-d2thc_967f26e7-3027-4e23-9f02-d111d714a396/manager/0.log" Dec 03 15:19:25 crc kubenswrapper[4805]: I1203 15:19:25.227240 4805 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_manila-operator-controller-manager-7c79b5df47-d2thc_967f26e7-3027-4e23-9f02-d111d714a396/kube-rbac-proxy/0.log" Dec 03 15:19:25 crc kubenswrapper[4805]: I1203 15:19:25.330048 4805 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_mariadb-operator-controller-manager-56bbcc9d85-n6m2d_2d72009f-b9ef-4c9d-b224-5b133a3cd93a/kube-rbac-proxy/0.log" Dec 03 15:19:25 crc kubenswrapper[4805]: I1203 15:19:25.415178 4805 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_mariadb-operator-controller-manager-56bbcc9d85-n6m2d_2d72009f-b9ef-4c9d-b224-5b133a3cd93a/manager/0.log" Dec 03 15:19:25 crc kubenswrapper[4805]: I1203 15:19:25.523322 4805 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_neutron-operator-controller-manager-5fdfd5b6b5-4mlhk_58aa635e-157c-4ae2-ad0c-568c144638d1/kube-rbac-proxy/0.log" Dec 03 15:19:25 crc kubenswrapper[4805]: I1203 15:19:25.601958 4805 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_neutron-operator-controller-manager-5fdfd5b6b5-4mlhk_58aa635e-157c-4ae2-ad0c-568c144638d1/manager/0.log" Dec 03 15:19:25 crc kubenswrapper[4805]: I1203 15:19:25.708977 4805 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_nova-operator-controller-manager-697bc559fc-5rl85_05506111-ddb3-4ab9-93a0-ad2bedff67a4/kube-rbac-proxy/0.log" Dec 03 15:19:25 crc kubenswrapper[4805]: I1203 15:19:25.819881 4805 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_nova-operator-controller-manager-697bc559fc-5rl85_05506111-ddb3-4ab9-93a0-ad2bedff67a4/manager/0.log" Dec 03 15:19:25 crc kubenswrapper[4805]: I1203 15:19:25.956540 4805 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_octavia-operator-controller-manager-998648c74-8mbdb_0f040ce3-0a83-426b-8b9f-b63183d16b4a/kube-rbac-proxy/0.log" Dec 03 15:19:25 crc kubenswrapper[4805]: I1203 15:19:25.990378 4805 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_octavia-operator-controller-manager-998648c74-8mbdb_0f040ce3-0a83-426b-8b9f-b63183d16b4a/manager/0.log" Dec 03 15:19:26 crc kubenswrapper[4805]: I1203 15:19:26.074980 4805 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-baremetal-operator-controller-manager-64bc77cfd4x9mdd_b9d0830e-0f5d-4639-ab7d-92dcf07a29bc/kube-rbac-proxy/0.log" Dec 03 15:19:26 crc kubenswrapper[4805]: I1203 15:19:26.154415 4805 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-baremetal-operator-controller-manager-64bc77cfd4x9mdd_b9d0830e-0f5d-4639-ab7d-92dcf07a29bc/manager/0.log" Dec 03 15:19:26 crc kubenswrapper[4805]: I1203 15:19:26.550977 4805 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-index-p6vl8_f6e38dfc-ef6b-4813-8bd4-027bed8dfb81/registry-server/0.log" Dec 03 15:19:26 crc kubenswrapper[4805]: I1203 15:19:26.607682 4805 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-controller-operator-796df97b8b-rmv2c_9da93959-4eef-466e-aab4-1a36202b9970/operator/0.log" Dec 03 15:19:26 crc kubenswrapper[4805]: I1203 15:19:26.683676 4805 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ovn-operator-controller-manager-b6456fdb6-djzcm_42b37265-f5bd-4ac4-9b96-d758fbe1b31e/kube-rbac-proxy/0.log" Dec 03 15:19:26 crc kubenswrapper[4805]: I1203 15:19:26.840525 4805 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_placement-operator-controller-manager-78f8948974-b8l4l_f853eae6-829b-42f9-9198-3d7dcabd405f/kube-rbac-proxy/0.log" Dec 03 15:19:26 crc kubenswrapper[4805]: I1203 15:19:26.860904 4805 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ovn-operator-controller-manager-b6456fdb6-djzcm_42b37265-f5bd-4ac4-9b96-d758fbe1b31e/manager/0.log" Dec 03 15:19:26 crc kubenswrapper[4805]: I1203 15:19:26.956414 4805 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_placement-operator-controller-manager-78f8948974-b8l4l_f853eae6-829b-42f9-9198-3d7dcabd405f/manager/0.log" Dec 03 15:19:27 crc kubenswrapper[4805]: I1203 15:19:27.163504 4805 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_swift-operator-controller-manager-5f8c65bbfc-qrvpw_b51f4884-a980-4643-9deb-8668aead3876/kube-rbac-proxy/0.log" Dec 03 15:19:27 crc kubenswrapper[4805]: I1203 15:19:27.178524 4805 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_rabbitmq-cluster-operator-manager-668c99d594-5q6j2_0aa860af-81d3-44ba-9020-0ffef21cefb2/operator/0.log" Dec 03 15:19:27 crc kubenswrapper[4805]: I1203 15:19:27.456367 4805 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_telemetry-operator-controller-manager-cc9d8f87b-qfmmx_8cbfcd07-e136-4bfd-b3c6-6615c911d440/kube-rbac-proxy/0.log" Dec 03 15:19:27 crc kubenswrapper[4805]: I1203 15:19:27.499218 4805 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_swift-operator-controller-manager-5f8c65bbfc-qrvpw_b51f4884-a980-4643-9deb-8668aead3876/manager/0.log" Dec 03 15:19:27 crc kubenswrapper[4805]: I1203 15:19:27.648284 4805 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_telemetry-operator-controller-manager-cc9d8f87b-qfmmx_8cbfcd07-e136-4bfd-b3c6-6615c911d440/manager/0.log" Dec 03 15:19:27 crc kubenswrapper[4805]: I1203 15:19:27.791051 4805 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_test-operator-controller-manager-5854674fcc-d6pvv_ac4ba0f5-635d-4a72-a133-9b4aaf88f097/kube-rbac-proxy/0.log" Dec 03 15:19:27 crc kubenswrapper[4805]: I1203 15:19:27.813551 4805 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_test-operator-controller-manager-5854674fcc-d6pvv_ac4ba0f5-635d-4a72-a133-9b4aaf88f097/manager/0.log" Dec 03 15:19:27 crc kubenswrapper[4805]: I1203 15:19:27.960676 4805 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-controller-manager-cc68d789f-pkvr8_41c4618c-1f15-4ceb-825b-546e828e399a/manager/0.log" Dec 03 15:19:27 crc kubenswrapper[4805]: I1203 15:19:27.993277 4805 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_watcher-operator-controller-manager-769dc69bc-fl6mm_30776d04-200b-4c10-b48d-f51a926e469b/kube-rbac-proxy/0.log" Dec 03 15:19:28 crc kubenswrapper[4805]: I1203 15:19:28.029058 4805 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_watcher-operator-controller-manager-769dc69bc-fl6mm_30776d04-200b-4c10-b48d-f51a926e469b/manager/0.log" Dec 03 15:19:50 crc kubenswrapper[4805]: I1203 15:19:50.833352 4805 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_control-plane-machine-set-operator-78cbb6b69f-25khc_9bd2f61c-2b92-488f-b632-1e627b48a518/control-plane-machine-set-operator/0.log" Dec 03 15:19:50 crc kubenswrapper[4805]: I1203 15:19:50.904519 4805 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_machine-api-operator-5694c8668f-jvspg_dfd20bfc-5c11-4287-a4b3-df9e0ca4dcf1/kube-rbac-proxy/0.log" Dec 03 15:19:51 crc kubenswrapper[4805]: I1203 15:19:51.073440 4805 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_machine-api-operator-5694c8668f-jvspg_dfd20bfc-5c11-4287-a4b3-df9e0ca4dcf1/machine-api-operator/0.log" Dec 03 15:20:05 crc kubenswrapper[4805]: I1203 15:20:05.516660 4805 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-5b446d88c5-vj7mm_c7baa569-d061-4226-a134-2b54cf463f90/cert-manager-controller/0.log" Dec 03 15:20:05 crc kubenswrapper[4805]: I1203 15:20:05.664666 4805 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-cainjector-7f985d654d-ll5ml_70f90fc2-9a18-4737-8025-850782848e31/cert-manager-cainjector/0.log" Dec 03 15:20:05 crc kubenswrapper[4805]: I1203 15:20:05.709130 4805 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-webhook-5655c58dd6-mt9mr_55307bc8-2e9c-4ea9-b007-c40c05403707/cert-manager-webhook/0.log" Dec 03 15:20:17 crc kubenswrapper[4805]: I1203 15:20:17.883011 4805 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-console-plugin-7fbb5f6569-mbhzq_5d4ab6ea-87ca-46dc-a9ad-f315faf0a9e2/nmstate-console-plugin/0.log" Dec 03 15:20:18 crc kubenswrapper[4805]: I1203 15:20:18.044351 4805 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-handler-xqhj5_4f623c1b-94c0-43cd-ae9f-56597f4542b4/nmstate-handler/0.log" Dec 03 15:20:18 crc kubenswrapper[4805]: I1203 15:20:18.083062 4805 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-metrics-7f946cbc9-6r8jm_9d39f3b1-337a-4aeb-b1cf-9ca9290f7791/kube-rbac-proxy/0.log" Dec 03 15:20:18 crc kubenswrapper[4805]: I1203 15:20:18.136629 4805 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-metrics-7f946cbc9-6r8jm_9d39f3b1-337a-4aeb-b1cf-9ca9290f7791/nmstate-metrics/0.log" Dec 03 15:20:18 crc kubenswrapper[4805]: I1203 15:20:18.252383 4805 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-operator-5b5b58f5c8-tstgp_7f35b1a1-05c4-453a-9101-d43cf4fa3739/nmstate-operator/0.log" Dec 03 15:20:18 crc kubenswrapper[4805]: I1203 15:20:18.353031 4805 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-webhook-5f6d4c5ccb-2zszs_9654448a-2df5-4cfc-b6e5-401ad15bc52e/nmstate-webhook/0.log" Dec 03 15:20:33 crc kubenswrapper[4805]: I1203 15:20:33.102070 4805 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_controller-f8648f98b-n75gf_a3361a03-7368-4bb6-b209-485190285e82/kube-rbac-proxy/0.log" Dec 03 15:20:33 crc kubenswrapper[4805]: I1203 15:20:33.293854 4805 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_controller-f8648f98b-n75gf_a3361a03-7368-4bb6-b209-485190285e82/controller/0.log" Dec 03 15:20:33 crc kubenswrapper[4805]: I1203 15:20:33.353237 4805 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-jsrqx_2ba805a2-6c47-4649-b8dc-eb92fb614437/cp-frr-files/0.log" Dec 03 15:20:33 crc kubenswrapper[4805]: I1203 15:20:33.490632 4805 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-jsrqx_2ba805a2-6c47-4649-b8dc-eb92fb614437/cp-reloader/0.log" Dec 03 15:20:33 crc kubenswrapper[4805]: I1203 15:20:33.533824 4805 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-jsrqx_2ba805a2-6c47-4649-b8dc-eb92fb614437/cp-reloader/0.log" Dec 03 15:20:33 crc kubenswrapper[4805]: I1203 15:20:33.547265 4805 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-jsrqx_2ba805a2-6c47-4649-b8dc-eb92fb614437/cp-metrics/0.log" Dec 03 15:20:33 crc kubenswrapper[4805]: I1203 15:20:33.560652 4805 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-jsrqx_2ba805a2-6c47-4649-b8dc-eb92fb614437/cp-frr-files/0.log" Dec 03 15:20:33 crc kubenswrapper[4805]: I1203 15:20:33.722926 4805 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-jsrqx_2ba805a2-6c47-4649-b8dc-eb92fb614437/cp-reloader/0.log" Dec 03 15:20:33 crc kubenswrapper[4805]: I1203 15:20:33.738944 4805 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-jsrqx_2ba805a2-6c47-4649-b8dc-eb92fb614437/cp-frr-files/0.log" Dec 03 15:20:33 crc kubenswrapper[4805]: I1203 15:20:33.749740 4805 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-jsrqx_2ba805a2-6c47-4649-b8dc-eb92fb614437/cp-metrics/0.log" Dec 03 15:20:34 crc kubenswrapper[4805]: I1203 15:20:34.243669 4805 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-jsrqx_2ba805a2-6c47-4649-b8dc-eb92fb614437/cp-metrics/0.log" Dec 03 15:20:34 crc kubenswrapper[4805]: I1203 15:20:34.438197 4805 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-jsrqx_2ba805a2-6c47-4649-b8dc-eb92fb614437/cp-frr-files/0.log" Dec 03 15:20:34 crc kubenswrapper[4805]: I1203 15:20:34.452913 4805 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-jsrqx_2ba805a2-6c47-4649-b8dc-eb92fb614437/cp-reloader/0.log" Dec 03 15:20:34 crc kubenswrapper[4805]: I1203 15:20:34.463404 4805 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-jsrqx_2ba805a2-6c47-4649-b8dc-eb92fb614437/cp-metrics/0.log" Dec 03 15:20:34 crc kubenswrapper[4805]: I1203 15:20:34.492745 4805 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-jsrqx_2ba805a2-6c47-4649-b8dc-eb92fb614437/controller/0.log" Dec 03 15:20:34 crc kubenswrapper[4805]: I1203 15:20:34.662061 4805 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-jsrqx_2ba805a2-6c47-4649-b8dc-eb92fb614437/kube-rbac-proxy/0.log" Dec 03 15:20:34 crc kubenswrapper[4805]: I1203 15:20:34.682497 4805 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-jsrqx_2ba805a2-6c47-4649-b8dc-eb92fb614437/frr-metrics/0.log" Dec 03 15:20:34 crc kubenswrapper[4805]: I1203 15:20:34.734294 4805 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-jsrqx_2ba805a2-6c47-4649-b8dc-eb92fb614437/kube-rbac-proxy-frr/0.log" Dec 03 15:20:34 crc kubenswrapper[4805]: I1203 15:20:34.941026 4805 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-jsrqx_2ba805a2-6c47-4649-b8dc-eb92fb614437/reloader/0.log" Dec 03 15:20:34 crc kubenswrapper[4805]: I1203 15:20:34.951621 4805 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-webhook-server-7fcb986d4-lmm42_24cc3c3f-7f72-49c4-b531-093bf5851177/frr-k8s-webhook-server/0.log" Dec 03 15:20:35 crc kubenswrapper[4805]: I1203 15:20:35.262076 4805 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_metallb-operator-controller-manager-76979bd97f-kgnj5_33acb320-cfe5-4d01-9543-f56ec8c33633/manager/0.log" Dec 03 15:20:35 crc kubenswrapper[4805]: I1203 15:20:35.483063 4805 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_metallb-operator-webhook-server-5587f68f8b-4vcdf_f2e7792c-c1c1-41b0-8e90-fe082ba1c366/webhook-server/0.log" Dec 03 15:20:35 crc kubenswrapper[4805]: I1203 15:20:35.587829 4805 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_speaker-jpk6x_06e98f2f-39d1-4899-9a18-7cc27e006de9/kube-rbac-proxy/0.log" Dec 03 15:20:36 crc kubenswrapper[4805]: I1203 15:20:36.355605 4805 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_speaker-jpk6x_06e98f2f-39d1-4899-9a18-7cc27e006de9/speaker/0.log" Dec 03 15:20:36 crc kubenswrapper[4805]: I1203 15:20:36.365537 4805 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-jsrqx_2ba805a2-6c47-4649-b8dc-eb92fb614437/frr/0.log" Dec 03 15:20:48 crc kubenswrapper[4805]: I1203 15:20:48.184584 4805 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fsxzq6_abc1f29a-e8a4-425d-bbbd-17c106a39c5c/util/0.log" Dec 03 15:20:48 crc kubenswrapper[4805]: I1203 15:20:48.419451 4805 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fsxzq6_abc1f29a-e8a4-425d-bbbd-17c106a39c5c/pull/0.log" Dec 03 15:20:48 crc kubenswrapper[4805]: I1203 15:20:48.439232 4805 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fsxzq6_abc1f29a-e8a4-425d-bbbd-17c106a39c5c/util/0.log" Dec 03 15:20:48 crc kubenswrapper[4805]: I1203 15:20:48.455453 4805 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fsxzq6_abc1f29a-e8a4-425d-bbbd-17c106a39c5c/pull/0.log" Dec 03 15:20:48 crc kubenswrapper[4805]: I1203 15:20:48.596603 4805 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fsxzq6_abc1f29a-e8a4-425d-bbbd-17c106a39c5c/pull/0.log" Dec 03 15:20:48 crc kubenswrapper[4805]: I1203 15:20:48.598034 4805 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fsxzq6_abc1f29a-e8a4-425d-bbbd-17c106a39c5c/util/0.log" Dec 03 15:20:48 crc kubenswrapper[4805]: I1203 15:20:48.623826 4805 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fsxzq6_abc1f29a-e8a4-425d-bbbd-17c106a39c5c/extract/0.log" Dec 03 15:20:48 crc kubenswrapper[4805]: I1203 15:20:48.780700 4805 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210zddkr_95163734-0666-4f76-a68f-722bc3752b38/util/0.log" Dec 03 15:20:48 crc kubenswrapper[4805]: I1203 15:20:48.967203 4805 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210zddkr_95163734-0666-4f76-a68f-722bc3752b38/util/0.log" Dec 03 15:20:48 crc kubenswrapper[4805]: I1203 15:20:48.969122 4805 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210zddkr_95163734-0666-4f76-a68f-722bc3752b38/pull/0.log" Dec 03 15:20:49 crc kubenswrapper[4805]: I1203 15:20:49.056535 4805 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210zddkr_95163734-0666-4f76-a68f-722bc3752b38/pull/0.log" Dec 03 15:20:49 crc kubenswrapper[4805]: I1203 15:20:49.208540 4805 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210zddkr_95163734-0666-4f76-a68f-722bc3752b38/pull/0.log" Dec 03 15:20:49 crc kubenswrapper[4805]: I1203 15:20:49.217025 4805 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210zddkr_95163734-0666-4f76-a68f-722bc3752b38/util/0.log" Dec 03 15:20:49 crc kubenswrapper[4805]: I1203 15:20:49.238093 4805 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210zddkr_95163734-0666-4f76-a68f-722bc3752b38/extract/0.log" Dec 03 15:20:49 crc kubenswrapper[4805]: I1203 15:20:49.388603 4805 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83glh2z_449e7507-4d25-4399-b1bd-b174d65a5041/util/0.log" Dec 03 15:20:49 crc kubenswrapper[4805]: I1203 15:20:49.525080 4805 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83glh2z_449e7507-4d25-4399-b1bd-b174d65a5041/util/0.log" Dec 03 15:20:49 crc kubenswrapper[4805]: I1203 15:20:49.542403 4805 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83glh2z_449e7507-4d25-4399-b1bd-b174d65a5041/pull/0.log" Dec 03 15:20:49 crc kubenswrapper[4805]: I1203 15:20:49.546786 4805 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83glh2z_449e7507-4d25-4399-b1bd-b174d65a5041/pull/0.log" Dec 03 15:20:49 crc kubenswrapper[4805]: I1203 15:20:49.740960 4805 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83glh2z_449e7507-4d25-4399-b1bd-b174d65a5041/extract/0.log" Dec 03 15:20:49 crc kubenswrapper[4805]: I1203 15:20:49.757584 4805 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83glh2z_449e7507-4d25-4399-b1bd-b174d65a5041/pull/0.log" Dec 03 15:20:49 crc kubenswrapper[4805]: I1203 15:20:49.778452 4805 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83glh2z_449e7507-4d25-4399-b1bd-b174d65a5041/util/0.log" Dec 03 15:20:49 crc kubenswrapper[4805]: I1203 15:20:49.940621 4805 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-c4fdw_6fcdd680-c71f-4978-992a-65378f16e0a1/extract-utilities/0.log" Dec 03 15:20:50 crc kubenswrapper[4805]: I1203 15:20:50.101993 4805 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-c4fdw_6fcdd680-c71f-4978-992a-65378f16e0a1/extract-content/0.log" Dec 03 15:20:50 crc kubenswrapper[4805]: I1203 15:20:50.110656 4805 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-c4fdw_6fcdd680-c71f-4978-992a-65378f16e0a1/extract-utilities/0.log" Dec 03 15:20:50 crc kubenswrapper[4805]: I1203 15:20:50.131894 4805 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-c4fdw_6fcdd680-c71f-4978-992a-65378f16e0a1/extract-content/0.log" Dec 03 15:20:50 crc kubenswrapper[4805]: I1203 15:20:50.286268 4805 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-c4fdw_6fcdd680-c71f-4978-992a-65378f16e0a1/extract-content/0.log" Dec 03 15:20:50 crc kubenswrapper[4805]: I1203 15:20:50.294019 4805 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-c4fdw_6fcdd680-c71f-4978-992a-65378f16e0a1/extract-utilities/0.log" Dec 03 15:20:50 crc kubenswrapper[4805]: I1203 15:20:50.476324 4805 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-xfg6c_a93d60c0-f172-49ae-af40-0d5bbda4877f/extract-utilities/0.log" Dec 03 15:20:50 crc kubenswrapper[4805]: I1203 15:20:50.811233 4805 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-xfg6c_a93d60c0-f172-49ae-af40-0d5bbda4877f/extract-utilities/0.log" Dec 03 15:20:50 crc kubenswrapper[4805]: I1203 15:20:50.834925 4805 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-xfg6c_a93d60c0-f172-49ae-af40-0d5bbda4877f/extract-content/0.log" Dec 03 15:20:50 crc kubenswrapper[4805]: I1203 15:20:50.846425 4805 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-c4fdw_6fcdd680-c71f-4978-992a-65378f16e0a1/registry-server/0.log" Dec 03 15:20:50 crc kubenswrapper[4805]: I1203 15:20:50.848252 4805 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-xfg6c_a93d60c0-f172-49ae-af40-0d5bbda4877f/extract-content/0.log" Dec 03 15:20:51 crc kubenswrapper[4805]: I1203 15:20:51.059203 4805 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-xfg6c_a93d60c0-f172-49ae-af40-0d5bbda4877f/extract-utilities/0.log" Dec 03 15:20:51 crc kubenswrapper[4805]: I1203 15:20:51.062792 4805 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-xfg6c_a93d60c0-f172-49ae-af40-0d5bbda4877f/extract-content/0.log" Dec 03 15:20:51 crc kubenswrapper[4805]: I1203 15:20:51.329112 4805 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_marketplace-operator-79b997595-6qgcv_336b0e79-07ec-4899-8cba-a24a65458383/marketplace-operator/0.log" Dec 03 15:20:51 crc kubenswrapper[4805]: I1203 15:20:51.414215 4805 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-jftlh_0ab10c8b-6fa6-44a5-87af-211ef363ec7f/extract-utilities/0.log" Dec 03 15:20:52 crc kubenswrapper[4805]: I1203 15:20:52.005863 4805 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-xfg6c_a93d60c0-f172-49ae-af40-0d5bbda4877f/registry-server/0.log" Dec 03 15:20:52 crc kubenswrapper[4805]: I1203 15:20:52.053348 4805 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-jftlh_0ab10c8b-6fa6-44a5-87af-211ef363ec7f/extract-content/0.log" Dec 03 15:20:52 crc kubenswrapper[4805]: I1203 15:20:52.079133 4805 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-jftlh_0ab10c8b-6fa6-44a5-87af-211ef363ec7f/extract-utilities/0.log" Dec 03 15:20:52 crc kubenswrapper[4805]: I1203 15:20:52.079999 4805 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-jftlh_0ab10c8b-6fa6-44a5-87af-211ef363ec7f/extract-content/0.log" Dec 03 15:20:52 crc kubenswrapper[4805]: I1203 15:20:52.324574 4805 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-jftlh_0ab10c8b-6fa6-44a5-87af-211ef363ec7f/extract-utilities/0.log" Dec 03 15:20:52 crc kubenswrapper[4805]: I1203 15:20:52.346584 4805 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-jftlh_0ab10c8b-6fa6-44a5-87af-211ef363ec7f/extract-content/0.log" Dec 03 15:20:52 crc kubenswrapper[4805]: I1203 15:20:52.348689 4805 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-v8czz_33e30caf-a8b4-4472-9f81-20a67a0ef486/extract-utilities/0.log" Dec 03 15:20:52 crc kubenswrapper[4805]: I1203 15:20:52.492093 4805 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-jftlh_0ab10c8b-6fa6-44a5-87af-211ef363ec7f/registry-server/0.log" Dec 03 15:20:52 crc kubenswrapper[4805]: I1203 15:20:52.558824 4805 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-v8czz_33e30caf-a8b4-4472-9f81-20a67a0ef486/extract-content/0.log" Dec 03 15:20:52 crc kubenswrapper[4805]: I1203 15:20:52.562741 4805 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-v8czz_33e30caf-a8b4-4472-9f81-20a67a0ef486/extract-utilities/0.log" Dec 03 15:20:52 crc kubenswrapper[4805]: I1203 15:20:52.562953 4805 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-v8czz_33e30caf-a8b4-4472-9f81-20a67a0ef486/extract-content/0.log" Dec 03 15:20:52 crc kubenswrapper[4805]: I1203 15:20:52.776374 4805 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-v8czz_33e30caf-a8b4-4472-9f81-20a67a0ef486/extract-content/0.log" Dec 03 15:20:52 crc kubenswrapper[4805]: I1203 15:20:52.789427 4805 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-v8czz_33e30caf-a8b4-4472-9f81-20a67a0ef486/extract-utilities/0.log" Dec 03 15:20:53 crc kubenswrapper[4805]: I1203 15:20:53.372682 4805 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-v8czz_33e30caf-a8b4-4472-9f81-20a67a0ef486/registry-server/0.log" Dec 03 15:21:06 crc kubenswrapper[4805]: I1203 15:21:06.259090 4805 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-operators_obo-prometheus-operator-668cf9dfbb-rtvd2_063df7dd-86e9-4e2b-a354-9baa6e90ba05/prometheus-operator/0.log" Dec 03 15:21:06 crc kubenswrapper[4805]: I1203 15:21:06.437502 4805 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-operators_obo-prometheus-operator-admission-webhook-59d4564ff5-49hbr_94ea81e8-e4cc-4317-a369-b73314c79ab8/prometheus-operator-admission-webhook/0.log" Dec 03 15:21:06 crc kubenswrapper[4805]: I1203 15:21:06.461237 4805 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-operators_obo-prometheus-operator-admission-webhook-59d4564ff5-mh8fd_21da77ce-bcf7-4ad0-97f8-a0c9d85d13db/prometheus-operator-admission-webhook/0.log" Dec 03 15:21:06 crc kubenswrapper[4805]: I1203 15:21:06.611429 4805 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-operators_observability-operator-d8bb48f5d-m6b9n_76c82588-3cbd-4565-abf5-5a291f68160e/operator/0.log" Dec 03 15:21:06 crc kubenswrapper[4805]: I1203 15:21:06.713092 4805 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-operators_perses-operator-5446b9c989-22dsf_78288c6e-a51d-4870-8530-54415f661eb8/perses-operator/0.log" Dec 03 15:21:13 crc kubenswrapper[4805]: I1203 15:21:13.917247 4805 patch_prober.go:28] interesting pod/machine-config-daemon-gskh4 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 03 15:21:13 crc kubenswrapper[4805]: I1203 15:21:13.917879 4805 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-gskh4" podUID="6098937f-e3f6-45e8-a647-4994a79cd711" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 03 15:21:43 crc kubenswrapper[4805]: I1203 15:21:43.917804 4805 patch_prober.go:28] interesting pod/machine-config-daemon-gskh4 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 03 15:21:43 crc kubenswrapper[4805]: I1203 15:21:43.918955 4805 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-gskh4" podUID="6098937f-e3f6-45e8-a647-4994a79cd711" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 03 15:22:13 crc kubenswrapper[4805]: I1203 15:22:13.917277 4805 patch_prober.go:28] interesting pod/machine-config-daemon-gskh4 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 03 15:22:13 crc kubenswrapper[4805]: I1203 15:22:13.917912 4805 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-gskh4" podUID="6098937f-e3f6-45e8-a647-4994a79cd711" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 03 15:22:13 crc kubenswrapper[4805]: I1203 15:22:13.917964 4805 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-gskh4" Dec 03 15:22:13 crc kubenswrapper[4805]: I1203 15:22:13.918709 4805 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"c1a31f4f6e3e9fa851f0e438617fecf09bad8f6021091f02eae4b71f6b3e374c"} pod="openshift-machine-config-operator/machine-config-daemon-gskh4" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 03 15:22:13 crc kubenswrapper[4805]: I1203 15:22:13.918766 4805 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-gskh4" podUID="6098937f-e3f6-45e8-a647-4994a79cd711" containerName="machine-config-daemon" containerID="cri-o://c1a31f4f6e3e9fa851f0e438617fecf09bad8f6021091f02eae4b71f6b3e374c" gracePeriod=600 Dec 03 15:22:14 crc kubenswrapper[4805]: I1203 15:22:14.501575 4805 generic.go:334] "Generic (PLEG): container finished" podID="6098937f-e3f6-45e8-a647-4994a79cd711" containerID="c1a31f4f6e3e9fa851f0e438617fecf09bad8f6021091f02eae4b71f6b3e374c" exitCode=0 Dec 03 15:22:14 crc kubenswrapper[4805]: I1203 15:22:14.501639 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-gskh4" event={"ID":"6098937f-e3f6-45e8-a647-4994a79cd711","Type":"ContainerDied","Data":"c1a31f4f6e3e9fa851f0e438617fecf09bad8f6021091f02eae4b71f6b3e374c"} Dec 03 15:22:14 crc kubenswrapper[4805]: I1203 15:22:14.502225 4805 scope.go:117] "RemoveContainer" containerID="cbf73c6c9ba740c0a523c84ca23c0851292d6eba7a0fb47d38599acca964e9cc" Dec 03 15:22:15 crc kubenswrapper[4805]: I1203 15:22:15.513753 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-gskh4" event={"ID":"6098937f-e3f6-45e8-a647-4994a79cd711","Type":"ContainerStarted","Data":"7d1692e3c7d3a2b9dcbcaebd54978af4cba7ae377a17c65268d4048aa72d2c92"} Dec 03 15:22:46 crc kubenswrapper[4805]: I1203 15:22:46.428978 4805 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-nhfs4"] Dec 03 15:22:46 crc kubenswrapper[4805]: E1203 15:22:46.431236 4805 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e77ba0b2-883a-410f-ab6b-c9fb086a31d6" containerName="container-00" Dec 03 15:22:46 crc kubenswrapper[4805]: I1203 15:22:46.431346 4805 state_mem.go:107] "Deleted CPUSet assignment" podUID="e77ba0b2-883a-410f-ab6b-c9fb086a31d6" containerName="container-00" Dec 03 15:22:46 crc kubenswrapper[4805]: I1203 15:22:46.431882 4805 memory_manager.go:354] "RemoveStaleState removing state" podUID="e77ba0b2-883a-410f-ab6b-c9fb086a31d6" containerName="container-00" Dec 03 15:22:46 crc kubenswrapper[4805]: I1203 15:22:46.434114 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-nhfs4" Dec 03 15:22:46 crc kubenswrapper[4805]: I1203 15:22:46.450068 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-nhfs4"] Dec 03 15:22:46 crc kubenswrapper[4805]: I1203 15:22:46.475441 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6a801aa8-7e79-48d2-9838-0649431cd93c-catalog-content\") pod \"certified-operators-nhfs4\" (UID: \"6a801aa8-7e79-48d2-9838-0649431cd93c\") " pod="openshift-marketplace/certified-operators-nhfs4" Dec 03 15:22:46 crc kubenswrapper[4805]: I1203 15:22:46.475493 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gbbnb\" (UniqueName: \"kubernetes.io/projected/6a801aa8-7e79-48d2-9838-0649431cd93c-kube-api-access-gbbnb\") pod \"certified-operators-nhfs4\" (UID: \"6a801aa8-7e79-48d2-9838-0649431cd93c\") " pod="openshift-marketplace/certified-operators-nhfs4" Dec 03 15:22:46 crc kubenswrapper[4805]: I1203 15:22:46.475623 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6a801aa8-7e79-48d2-9838-0649431cd93c-utilities\") pod \"certified-operators-nhfs4\" (UID: \"6a801aa8-7e79-48d2-9838-0649431cd93c\") " pod="openshift-marketplace/certified-operators-nhfs4" Dec 03 15:22:46 crc kubenswrapper[4805]: I1203 15:22:46.577298 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6a801aa8-7e79-48d2-9838-0649431cd93c-utilities\") pod \"certified-operators-nhfs4\" (UID: \"6a801aa8-7e79-48d2-9838-0649431cd93c\") " pod="openshift-marketplace/certified-operators-nhfs4" Dec 03 15:22:46 crc kubenswrapper[4805]: I1203 15:22:46.577509 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6a801aa8-7e79-48d2-9838-0649431cd93c-catalog-content\") pod \"certified-operators-nhfs4\" (UID: \"6a801aa8-7e79-48d2-9838-0649431cd93c\") " pod="openshift-marketplace/certified-operators-nhfs4" Dec 03 15:22:46 crc kubenswrapper[4805]: I1203 15:22:46.577535 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gbbnb\" (UniqueName: \"kubernetes.io/projected/6a801aa8-7e79-48d2-9838-0649431cd93c-kube-api-access-gbbnb\") pod \"certified-operators-nhfs4\" (UID: \"6a801aa8-7e79-48d2-9838-0649431cd93c\") " pod="openshift-marketplace/certified-operators-nhfs4" Dec 03 15:22:46 crc kubenswrapper[4805]: I1203 15:22:46.578094 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6a801aa8-7e79-48d2-9838-0649431cd93c-utilities\") pod \"certified-operators-nhfs4\" (UID: \"6a801aa8-7e79-48d2-9838-0649431cd93c\") " pod="openshift-marketplace/certified-operators-nhfs4" Dec 03 15:22:46 crc kubenswrapper[4805]: I1203 15:22:46.578356 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6a801aa8-7e79-48d2-9838-0649431cd93c-catalog-content\") pod \"certified-operators-nhfs4\" (UID: \"6a801aa8-7e79-48d2-9838-0649431cd93c\") " pod="openshift-marketplace/certified-operators-nhfs4" Dec 03 15:22:46 crc kubenswrapper[4805]: I1203 15:22:46.609732 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gbbnb\" (UniqueName: \"kubernetes.io/projected/6a801aa8-7e79-48d2-9838-0649431cd93c-kube-api-access-gbbnb\") pod \"certified-operators-nhfs4\" (UID: \"6a801aa8-7e79-48d2-9838-0649431cd93c\") " pod="openshift-marketplace/certified-operators-nhfs4" Dec 03 15:22:46 crc kubenswrapper[4805]: I1203 15:22:46.767866 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-nhfs4" Dec 03 15:22:46 crc kubenswrapper[4805]: I1203 15:22:46.844690 4805 generic.go:334] "Generic (PLEG): container finished" podID="08ff2c92-4d49-4ac0-84c5-3403610205da" containerID="7cfbc3fe2001fe4c7e6cf09443ce6a6e860a40da1f2c016d071c05fbe7033424" exitCode=0 Dec 03 15:22:46 crc kubenswrapper[4805]: I1203 15:22:46.845006 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-n5xns/must-gather-5mnzj" event={"ID":"08ff2c92-4d49-4ac0-84c5-3403610205da","Type":"ContainerDied","Data":"7cfbc3fe2001fe4c7e6cf09443ce6a6e860a40da1f2c016d071c05fbe7033424"} Dec 03 15:22:46 crc kubenswrapper[4805]: I1203 15:22:46.845734 4805 scope.go:117] "RemoveContainer" containerID="7cfbc3fe2001fe4c7e6cf09443ce6a6e860a40da1f2c016d071c05fbe7033424" Dec 03 15:22:47 crc kubenswrapper[4805]: I1203 15:22:47.337053 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-nhfs4"] Dec 03 15:22:47 crc kubenswrapper[4805]: I1203 15:22:47.655403 4805 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-n5xns_must-gather-5mnzj_08ff2c92-4d49-4ac0-84c5-3403610205da/gather/0.log" Dec 03 15:22:47 crc kubenswrapper[4805]: I1203 15:22:47.859138 4805 generic.go:334] "Generic (PLEG): container finished" podID="6a801aa8-7e79-48d2-9838-0649431cd93c" containerID="926d9af1378cc6f1f4e7d25b6bc42db677e1e6440f5d7ca8c6124fbf25c4775a" exitCode=0 Dec 03 15:22:47 crc kubenswrapper[4805]: I1203 15:22:47.859214 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-nhfs4" event={"ID":"6a801aa8-7e79-48d2-9838-0649431cd93c","Type":"ContainerDied","Data":"926d9af1378cc6f1f4e7d25b6bc42db677e1e6440f5d7ca8c6124fbf25c4775a"} Dec 03 15:22:47 crc kubenswrapper[4805]: I1203 15:22:47.859303 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-nhfs4" event={"ID":"6a801aa8-7e79-48d2-9838-0649431cd93c","Type":"ContainerStarted","Data":"b3ade44c349cb1af5e8b49b37096759f58722f7b5d8efe8b34990dcb483943c9"} Dec 03 15:22:47 crc kubenswrapper[4805]: I1203 15:22:47.861981 4805 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 03 15:22:49 crc kubenswrapper[4805]: I1203 15:22:49.900477 4805 generic.go:334] "Generic (PLEG): container finished" podID="6a801aa8-7e79-48d2-9838-0649431cd93c" containerID="20cf102ce682571cac1a879dfe01299e5861dcf812346e7b8d966e235e4aedaf" exitCode=0 Dec 03 15:22:49 crc kubenswrapper[4805]: I1203 15:22:49.901053 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-nhfs4" event={"ID":"6a801aa8-7e79-48d2-9838-0649431cd93c","Type":"ContainerDied","Data":"20cf102ce682571cac1a879dfe01299e5861dcf812346e7b8d966e235e4aedaf"} Dec 03 15:22:50 crc kubenswrapper[4805]: I1203 15:22:50.915741 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-nhfs4" event={"ID":"6a801aa8-7e79-48d2-9838-0649431cd93c","Type":"ContainerStarted","Data":"11679b138fd85a1fe494ebf2daf1fe176659cc04a56d3928cddd525109948a07"} Dec 03 15:22:50 crc kubenswrapper[4805]: I1203 15:22:50.942910 4805 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-nhfs4" podStartSLOduration=2.446749948 podStartE2EDuration="4.94288875s" podCreationTimestamp="2025-12-03 15:22:46 +0000 UTC" firstStartedPulling="2025-12-03 15:22:47.861634097 +0000 UTC m=+4397.524551020" lastFinishedPulling="2025-12-03 15:22:50.357772899 +0000 UTC m=+4400.020689822" observedRunningTime="2025-12-03 15:22:50.937277819 +0000 UTC m=+4400.600194752" watchObservedRunningTime="2025-12-03 15:22:50.94288875 +0000 UTC m=+4400.605805673" Dec 03 15:22:55 crc kubenswrapper[4805]: I1203 15:22:55.436798 4805 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-n5xns/must-gather-5mnzj"] Dec 03 15:22:55 crc kubenswrapper[4805]: I1203 15:22:55.437629 4805 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-must-gather-n5xns/must-gather-5mnzj" podUID="08ff2c92-4d49-4ac0-84c5-3403610205da" containerName="copy" containerID="cri-o://313a0a66fa07afa0a361dc21ea0df9772ee5f6169321650079c0add6695204d7" gracePeriod=2 Dec 03 15:22:55 crc kubenswrapper[4805]: I1203 15:22:55.461659 4805 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-n5xns/must-gather-5mnzj"] Dec 03 15:22:55 crc kubenswrapper[4805]: I1203 15:22:55.965580 4805 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-n5xns_must-gather-5mnzj_08ff2c92-4d49-4ac0-84c5-3403610205da/copy/0.log" Dec 03 15:22:55 crc kubenswrapper[4805]: I1203 15:22:55.966525 4805 generic.go:334] "Generic (PLEG): container finished" podID="08ff2c92-4d49-4ac0-84c5-3403610205da" containerID="313a0a66fa07afa0a361dc21ea0df9772ee5f6169321650079c0add6695204d7" exitCode=143 Dec 03 15:22:55 crc kubenswrapper[4805]: I1203 15:22:55.966626 4805 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="2d2eb33a971d1f0548a5d3f2f9415ecc21721f07ad05ad9c1d45875758f5ed75" Dec 03 15:22:55 crc kubenswrapper[4805]: I1203 15:22:55.981878 4805 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-n5xns_must-gather-5mnzj_08ff2c92-4d49-4ac0-84c5-3403610205da/copy/0.log" Dec 03 15:22:55 crc kubenswrapper[4805]: I1203 15:22:55.982498 4805 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-n5xns/must-gather-5mnzj" Dec 03 15:22:56 crc kubenswrapper[4805]: I1203 15:22:56.075787 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vhp6q\" (UniqueName: \"kubernetes.io/projected/08ff2c92-4d49-4ac0-84c5-3403610205da-kube-api-access-vhp6q\") pod \"08ff2c92-4d49-4ac0-84c5-3403610205da\" (UID: \"08ff2c92-4d49-4ac0-84c5-3403610205da\") " Dec 03 15:22:56 crc kubenswrapper[4805]: I1203 15:22:56.076144 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/08ff2c92-4d49-4ac0-84c5-3403610205da-must-gather-output\") pod \"08ff2c92-4d49-4ac0-84c5-3403610205da\" (UID: \"08ff2c92-4d49-4ac0-84c5-3403610205da\") " Dec 03 15:22:56 crc kubenswrapper[4805]: I1203 15:22:56.082008 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/08ff2c92-4d49-4ac0-84c5-3403610205da-kube-api-access-vhp6q" (OuterVolumeSpecName: "kube-api-access-vhp6q") pod "08ff2c92-4d49-4ac0-84c5-3403610205da" (UID: "08ff2c92-4d49-4ac0-84c5-3403610205da"). InnerVolumeSpecName "kube-api-access-vhp6q". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 15:22:56 crc kubenswrapper[4805]: I1203 15:22:56.181399 4805 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vhp6q\" (UniqueName: \"kubernetes.io/projected/08ff2c92-4d49-4ac0-84c5-3403610205da-kube-api-access-vhp6q\") on node \"crc\" DevicePath \"\"" Dec 03 15:22:56 crc kubenswrapper[4805]: I1203 15:22:56.257735 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/08ff2c92-4d49-4ac0-84c5-3403610205da-must-gather-output" (OuterVolumeSpecName: "must-gather-output") pod "08ff2c92-4d49-4ac0-84c5-3403610205da" (UID: "08ff2c92-4d49-4ac0-84c5-3403610205da"). InnerVolumeSpecName "must-gather-output". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 15:22:56 crc kubenswrapper[4805]: I1203 15:22:56.283792 4805 reconciler_common.go:293] "Volume detached for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/08ff2c92-4d49-4ac0-84c5-3403610205da-must-gather-output\") on node \"crc\" DevicePath \"\"" Dec 03 15:22:56 crc kubenswrapper[4805]: I1203 15:22:56.705500 4805 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="08ff2c92-4d49-4ac0-84c5-3403610205da" path="/var/lib/kubelet/pods/08ff2c92-4d49-4ac0-84c5-3403610205da/volumes" Dec 03 15:22:56 crc kubenswrapper[4805]: I1203 15:22:56.768483 4805 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-nhfs4" Dec 03 15:22:56 crc kubenswrapper[4805]: I1203 15:22:56.768570 4805 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-nhfs4" Dec 03 15:22:56 crc kubenswrapper[4805]: I1203 15:22:56.975442 4805 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-n5xns/must-gather-5mnzj" Dec 03 15:22:57 crc kubenswrapper[4805]: I1203 15:22:57.441787 4805 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-nhfs4" Dec 03 15:22:57 crc kubenswrapper[4805]: I1203 15:22:57.493517 4805 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-nhfs4" Dec 03 15:22:57 crc kubenswrapper[4805]: I1203 15:22:57.681411 4805 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-nhfs4"] Dec 03 15:22:58 crc kubenswrapper[4805]: I1203 15:22:58.992796 4805 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-nhfs4" podUID="6a801aa8-7e79-48d2-9838-0649431cd93c" containerName="registry-server" containerID="cri-o://11679b138fd85a1fe494ebf2daf1fe176659cc04a56d3928cddd525109948a07" gracePeriod=2 Dec 03 15:22:59 crc kubenswrapper[4805]: I1203 15:22:59.980225 4805 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-nhfs4" Dec 03 15:23:00 crc kubenswrapper[4805]: I1203 15:23:00.012956 4805 generic.go:334] "Generic (PLEG): container finished" podID="6a801aa8-7e79-48d2-9838-0649431cd93c" containerID="11679b138fd85a1fe494ebf2daf1fe176659cc04a56d3928cddd525109948a07" exitCode=0 Dec 03 15:23:00 crc kubenswrapper[4805]: I1203 15:23:00.013014 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-nhfs4" event={"ID":"6a801aa8-7e79-48d2-9838-0649431cd93c","Type":"ContainerDied","Data":"11679b138fd85a1fe494ebf2daf1fe176659cc04a56d3928cddd525109948a07"} Dec 03 15:23:00 crc kubenswrapper[4805]: I1203 15:23:00.013048 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-nhfs4" event={"ID":"6a801aa8-7e79-48d2-9838-0649431cd93c","Type":"ContainerDied","Data":"b3ade44c349cb1af5e8b49b37096759f58722f7b5d8efe8b34990dcb483943c9"} Dec 03 15:23:00 crc kubenswrapper[4805]: I1203 15:23:00.013066 4805 scope.go:117] "RemoveContainer" containerID="11679b138fd85a1fe494ebf2daf1fe176659cc04a56d3928cddd525109948a07" Dec 03 15:23:00 crc kubenswrapper[4805]: I1203 15:23:00.013249 4805 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-nhfs4" Dec 03 15:23:00 crc kubenswrapper[4805]: I1203 15:23:00.038046 4805 scope.go:117] "RemoveContainer" containerID="20cf102ce682571cac1a879dfe01299e5861dcf812346e7b8d966e235e4aedaf" Dec 03 15:23:00 crc kubenswrapper[4805]: I1203 15:23:00.066648 4805 scope.go:117] "RemoveContainer" containerID="926d9af1378cc6f1f4e7d25b6bc42db677e1e6440f5d7ca8c6124fbf25c4775a" Dec 03 15:23:00 crc kubenswrapper[4805]: I1203 15:23:00.066713 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gbbnb\" (UniqueName: \"kubernetes.io/projected/6a801aa8-7e79-48d2-9838-0649431cd93c-kube-api-access-gbbnb\") pod \"6a801aa8-7e79-48d2-9838-0649431cd93c\" (UID: \"6a801aa8-7e79-48d2-9838-0649431cd93c\") " Dec 03 15:23:00 crc kubenswrapper[4805]: I1203 15:23:00.066988 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6a801aa8-7e79-48d2-9838-0649431cd93c-utilities\") pod \"6a801aa8-7e79-48d2-9838-0649431cd93c\" (UID: \"6a801aa8-7e79-48d2-9838-0649431cd93c\") " Dec 03 15:23:00 crc kubenswrapper[4805]: I1203 15:23:00.067044 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6a801aa8-7e79-48d2-9838-0649431cd93c-catalog-content\") pod \"6a801aa8-7e79-48d2-9838-0649431cd93c\" (UID: \"6a801aa8-7e79-48d2-9838-0649431cd93c\") " Dec 03 15:23:00 crc kubenswrapper[4805]: I1203 15:23:00.068401 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/6a801aa8-7e79-48d2-9838-0649431cd93c-utilities" (OuterVolumeSpecName: "utilities") pod "6a801aa8-7e79-48d2-9838-0649431cd93c" (UID: "6a801aa8-7e79-48d2-9838-0649431cd93c"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 15:23:00 crc kubenswrapper[4805]: I1203 15:23:00.090359 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6a801aa8-7e79-48d2-9838-0649431cd93c-kube-api-access-gbbnb" (OuterVolumeSpecName: "kube-api-access-gbbnb") pod "6a801aa8-7e79-48d2-9838-0649431cd93c" (UID: "6a801aa8-7e79-48d2-9838-0649431cd93c"). InnerVolumeSpecName "kube-api-access-gbbnb". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 15:23:00 crc kubenswrapper[4805]: I1203 15:23:00.114539 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/6a801aa8-7e79-48d2-9838-0649431cd93c-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "6a801aa8-7e79-48d2-9838-0649431cd93c" (UID: "6a801aa8-7e79-48d2-9838-0649431cd93c"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 15:23:00 crc kubenswrapper[4805]: I1203 15:23:00.169165 4805 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gbbnb\" (UniqueName: \"kubernetes.io/projected/6a801aa8-7e79-48d2-9838-0649431cd93c-kube-api-access-gbbnb\") on node \"crc\" DevicePath \"\"" Dec 03 15:23:00 crc kubenswrapper[4805]: I1203 15:23:00.169212 4805 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6a801aa8-7e79-48d2-9838-0649431cd93c-utilities\") on node \"crc\" DevicePath \"\"" Dec 03 15:23:00 crc kubenswrapper[4805]: I1203 15:23:00.169223 4805 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6a801aa8-7e79-48d2-9838-0649431cd93c-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 03 15:23:00 crc kubenswrapper[4805]: I1203 15:23:00.191272 4805 scope.go:117] "RemoveContainer" containerID="11679b138fd85a1fe494ebf2daf1fe176659cc04a56d3928cddd525109948a07" Dec 03 15:23:00 crc kubenswrapper[4805]: E1203 15:23:00.191780 4805 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"11679b138fd85a1fe494ebf2daf1fe176659cc04a56d3928cddd525109948a07\": container with ID starting with 11679b138fd85a1fe494ebf2daf1fe176659cc04a56d3928cddd525109948a07 not found: ID does not exist" containerID="11679b138fd85a1fe494ebf2daf1fe176659cc04a56d3928cddd525109948a07" Dec 03 15:23:00 crc kubenswrapper[4805]: I1203 15:23:00.191824 4805 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"11679b138fd85a1fe494ebf2daf1fe176659cc04a56d3928cddd525109948a07"} err="failed to get container status \"11679b138fd85a1fe494ebf2daf1fe176659cc04a56d3928cddd525109948a07\": rpc error: code = NotFound desc = could not find container \"11679b138fd85a1fe494ebf2daf1fe176659cc04a56d3928cddd525109948a07\": container with ID starting with 11679b138fd85a1fe494ebf2daf1fe176659cc04a56d3928cddd525109948a07 not found: ID does not exist" Dec 03 15:23:00 crc kubenswrapper[4805]: I1203 15:23:00.191867 4805 scope.go:117] "RemoveContainer" containerID="20cf102ce682571cac1a879dfe01299e5861dcf812346e7b8d966e235e4aedaf" Dec 03 15:23:00 crc kubenswrapper[4805]: E1203 15:23:00.192621 4805 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"20cf102ce682571cac1a879dfe01299e5861dcf812346e7b8d966e235e4aedaf\": container with ID starting with 20cf102ce682571cac1a879dfe01299e5861dcf812346e7b8d966e235e4aedaf not found: ID does not exist" containerID="20cf102ce682571cac1a879dfe01299e5861dcf812346e7b8d966e235e4aedaf" Dec 03 15:23:00 crc kubenswrapper[4805]: I1203 15:23:00.192654 4805 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"20cf102ce682571cac1a879dfe01299e5861dcf812346e7b8d966e235e4aedaf"} err="failed to get container status \"20cf102ce682571cac1a879dfe01299e5861dcf812346e7b8d966e235e4aedaf\": rpc error: code = NotFound desc = could not find container \"20cf102ce682571cac1a879dfe01299e5861dcf812346e7b8d966e235e4aedaf\": container with ID starting with 20cf102ce682571cac1a879dfe01299e5861dcf812346e7b8d966e235e4aedaf not found: ID does not exist" Dec 03 15:23:00 crc kubenswrapper[4805]: I1203 15:23:00.192677 4805 scope.go:117] "RemoveContainer" containerID="926d9af1378cc6f1f4e7d25b6bc42db677e1e6440f5d7ca8c6124fbf25c4775a" Dec 03 15:23:00 crc kubenswrapper[4805]: E1203 15:23:00.192924 4805 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"926d9af1378cc6f1f4e7d25b6bc42db677e1e6440f5d7ca8c6124fbf25c4775a\": container with ID starting with 926d9af1378cc6f1f4e7d25b6bc42db677e1e6440f5d7ca8c6124fbf25c4775a not found: ID does not exist" containerID="926d9af1378cc6f1f4e7d25b6bc42db677e1e6440f5d7ca8c6124fbf25c4775a" Dec 03 15:23:00 crc kubenswrapper[4805]: I1203 15:23:00.192954 4805 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"926d9af1378cc6f1f4e7d25b6bc42db677e1e6440f5d7ca8c6124fbf25c4775a"} err="failed to get container status \"926d9af1378cc6f1f4e7d25b6bc42db677e1e6440f5d7ca8c6124fbf25c4775a\": rpc error: code = NotFound desc = could not find container \"926d9af1378cc6f1f4e7d25b6bc42db677e1e6440f5d7ca8c6124fbf25c4775a\": container with ID starting with 926d9af1378cc6f1f4e7d25b6bc42db677e1e6440f5d7ca8c6124fbf25c4775a not found: ID does not exist" Dec 03 15:23:00 crc kubenswrapper[4805]: I1203 15:23:00.350622 4805 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-nhfs4"] Dec 03 15:23:00 crc kubenswrapper[4805]: I1203 15:23:00.362975 4805 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-nhfs4"] Dec 03 15:23:00 crc kubenswrapper[4805]: I1203 15:23:00.709404 4805 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6a801aa8-7e79-48d2-9838-0649431cd93c" path="/var/lib/kubelet/pods/6a801aa8-7e79-48d2-9838-0649431cd93c/volumes" Dec 03 15:23:13 crc kubenswrapper[4805]: I1203 15:23:13.899104 4805 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack-operators/openstack-operator-controller-operator-796df97b8b-rmv2c" podUID="9da93959-4eef-466e-aab4-1a36202b9970" containerName="operator" probeResult="failure" output="Get \"http://10.217.0.57:8081/readyz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 03 15:23:21 crc kubenswrapper[4805]: I1203 15:23:21.322062 4805 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-f8mbc"] Dec 03 15:23:21 crc kubenswrapper[4805]: E1203 15:23:21.323153 4805 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="08ff2c92-4d49-4ac0-84c5-3403610205da" containerName="copy" Dec 03 15:23:21 crc kubenswrapper[4805]: I1203 15:23:21.323169 4805 state_mem.go:107] "Deleted CPUSet assignment" podUID="08ff2c92-4d49-4ac0-84c5-3403610205da" containerName="copy" Dec 03 15:23:21 crc kubenswrapper[4805]: E1203 15:23:21.323194 4805 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6a801aa8-7e79-48d2-9838-0649431cd93c" containerName="registry-server" Dec 03 15:23:21 crc kubenswrapper[4805]: I1203 15:23:21.323201 4805 state_mem.go:107] "Deleted CPUSet assignment" podUID="6a801aa8-7e79-48d2-9838-0649431cd93c" containerName="registry-server" Dec 03 15:23:21 crc kubenswrapper[4805]: E1203 15:23:21.323221 4805 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6a801aa8-7e79-48d2-9838-0649431cd93c" containerName="extract-content" Dec 03 15:23:21 crc kubenswrapper[4805]: I1203 15:23:21.323228 4805 state_mem.go:107] "Deleted CPUSet assignment" podUID="6a801aa8-7e79-48d2-9838-0649431cd93c" containerName="extract-content" Dec 03 15:23:21 crc kubenswrapper[4805]: E1203 15:23:21.323239 4805 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="08ff2c92-4d49-4ac0-84c5-3403610205da" containerName="gather" Dec 03 15:23:21 crc kubenswrapper[4805]: I1203 15:23:21.323245 4805 state_mem.go:107] "Deleted CPUSet assignment" podUID="08ff2c92-4d49-4ac0-84c5-3403610205da" containerName="gather" Dec 03 15:23:21 crc kubenswrapper[4805]: E1203 15:23:21.323262 4805 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6a801aa8-7e79-48d2-9838-0649431cd93c" containerName="extract-utilities" Dec 03 15:23:21 crc kubenswrapper[4805]: I1203 15:23:21.323269 4805 state_mem.go:107] "Deleted CPUSet assignment" podUID="6a801aa8-7e79-48d2-9838-0649431cd93c" containerName="extract-utilities" Dec 03 15:23:21 crc kubenswrapper[4805]: I1203 15:23:21.323471 4805 memory_manager.go:354] "RemoveStaleState removing state" podUID="08ff2c92-4d49-4ac0-84c5-3403610205da" containerName="copy" Dec 03 15:23:21 crc kubenswrapper[4805]: I1203 15:23:21.323487 4805 memory_manager.go:354] "RemoveStaleState removing state" podUID="6a801aa8-7e79-48d2-9838-0649431cd93c" containerName="registry-server" Dec 03 15:23:21 crc kubenswrapper[4805]: I1203 15:23:21.323498 4805 memory_manager.go:354] "RemoveStaleState removing state" podUID="08ff2c92-4d49-4ac0-84c5-3403610205da" containerName="gather" Dec 03 15:23:21 crc kubenswrapper[4805]: I1203 15:23:21.325242 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-f8mbc" Dec 03 15:23:21 crc kubenswrapper[4805]: I1203 15:23:21.335590 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-f8mbc"] Dec 03 15:23:21 crc kubenswrapper[4805]: I1203 15:23:21.504349 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5pxgh\" (UniqueName: \"kubernetes.io/projected/38b80aeb-575b-4337-837d-3520bb8690a2-kube-api-access-5pxgh\") pod \"redhat-operators-f8mbc\" (UID: \"38b80aeb-575b-4337-837d-3520bb8690a2\") " pod="openshift-marketplace/redhat-operators-f8mbc" Dec 03 15:23:21 crc kubenswrapper[4805]: I1203 15:23:21.504757 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/38b80aeb-575b-4337-837d-3520bb8690a2-utilities\") pod \"redhat-operators-f8mbc\" (UID: \"38b80aeb-575b-4337-837d-3520bb8690a2\") " pod="openshift-marketplace/redhat-operators-f8mbc" Dec 03 15:23:21 crc kubenswrapper[4805]: I1203 15:23:21.504805 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/38b80aeb-575b-4337-837d-3520bb8690a2-catalog-content\") pod \"redhat-operators-f8mbc\" (UID: \"38b80aeb-575b-4337-837d-3520bb8690a2\") " pod="openshift-marketplace/redhat-operators-f8mbc" Dec 03 15:23:21 crc kubenswrapper[4805]: I1203 15:23:21.606543 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/38b80aeb-575b-4337-837d-3520bb8690a2-utilities\") pod \"redhat-operators-f8mbc\" (UID: \"38b80aeb-575b-4337-837d-3520bb8690a2\") " pod="openshift-marketplace/redhat-operators-f8mbc" Dec 03 15:23:21 crc kubenswrapper[4805]: I1203 15:23:21.606687 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/38b80aeb-575b-4337-837d-3520bb8690a2-catalog-content\") pod \"redhat-operators-f8mbc\" (UID: \"38b80aeb-575b-4337-837d-3520bb8690a2\") " pod="openshift-marketplace/redhat-operators-f8mbc" Dec 03 15:23:21 crc kubenswrapper[4805]: I1203 15:23:21.606894 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5pxgh\" (UniqueName: \"kubernetes.io/projected/38b80aeb-575b-4337-837d-3520bb8690a2-kube-api-access-5pxgh\") pod \"redhat-operators-f8mbc\" (UID: \"38b80aeb-575b-4337-837d-3520bb8690a2\") " pod="openshift-marketplace/redhat-operators-f8mbc" Dec 03 15:23:21 crc kubenswrapper[4805]: I1203 15:23:21.607078 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/38b80aeb-575b-4337-837d-3520bb8690a2-utilities\") pod \"redhat-operators-f8mbc\" (UID: \"38b80aeb-575b-4337-837d-3520bb8690a2\") " pod="openshift-marketplace/redhat-operators-f8mbc" Dec 03 15:23:21 crc kubenswrapper[4805]: I1203 15:23:21.607266 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/38b80aeb-575b-4337-837d-3520bb8690a2-catalog-content\") pod \"redhat-operators-f8mbc\" (UID: \"38b80aeb-575b-4337-837d-3520bb8690a2\") " pod="openshift-marketplace/redhat-operators-f8mbc" Dec 03 15:23:21 crc kubenswrapper[4805]: I1203 15:23:21.630696 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5pxgh\" (UniqueName: \"kubernetes.io/projected/38b80aeb-575b-4337-837d-3520bb8690a2-kube-api-access-5pxgh\") pod \"redhat-operators-f8mbc\" (UID: \"38b80aeb-575b-4337-837d-3520bb8690a2\") " pod="openshift-marketplace/redhat-operators-f8mbc" Dec 03 15:23:21 crc kubenswrapper[4805]: I1203 15:23:21.658734 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-f8mbc" Dec 03 15:23:22 crc kubenswrapper[4805]: I1203 15:23:22.147476 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-f8mbc"] Dec 03 15:23:22 crc kubenswrapper[4805]: I1203 15:23:22.238198 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-f8mbc" event={"ID":"38b80aeb-575b-4337-837d-3520bb8690a2","Type":"ContainerStarted","Data":"f5e911e91c49a602fee56485da4d327cc2e609ae1521e3b054094635a42b39ba"} Dec 03 15:23:23 crc kubenswrapper[4805]: I1203 15:23:23.248986 4805 generic.go:334] "Generic (PLEG): container finished" podID="38b80aeb-575b-4337-837d-3520bb8690a2" containerID="083cfa768e839b50963c4be1f2f04fc7723282eda487335e3959d9edc28beb48" exitCode=0 Dec 03 15:23:23 crc kubenswrapper[4805]: I1203 15:23:23.249059 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-f8mbc" event={"ID":"38b80aeb-575b-4337-837d-3520bb8690a2","Type":"ContainerDied","Data":"083cfa768e839b50963c4be1f2f04fc7723282eda487335e3959d9edc28beb48"} Dec 03 15:23:25 crc kubenswrapper[4805]: I1203 15:23:25.274870 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-f8mbc" event={"ID":"38b80aeb-575b-4337-837d-3520bb8690a2","Type":"ContainerStarted","Data":"f91ddc215a11c0d2908240bdf52268a463969a260e217b2afd1c8ac4858e2de6"} Dec 03 15:23:27 crc kubenswrapper[4805]: I1203 15:23:27.297188 4805 generic.go:334] "Generic (PLEG): container finished" podID="38b80aeb-575b-4337-837d-3520bb8690a2" containerID="f91ddc215a11c0d2908240bdf52268a463969a260e217b2afd1c8ac4858e2de6" exitCode=0 Dec 03 15:23:27 crc kubenswrapper[4805]: I1203 15:23:27.297253 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-f8mbc" event={"ID":"38b80aeb-575b-4337-837d-3520bb8690a2","Type":"ContainerDied","Data":"f91ddc215a11c0d2908240bdf52268a463969a260e217b2afd1c8ac4858e2de6"} Dec 03 15:23:28 crc kubenswrapper[4805]: I1203 15:23:28.309644 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-f8mbc" event={"ID":"38b80aeb-575b-4337-837d-3520bb8690a2","Type":"ContainerStarted","Data":"732a19de6353ca13c12b856ee6fd929b489f7c5bf3666df158221d1d7abb9527"} Dec 03 15:23:28 crc kubenswrapper[4805]: I1203 15:23:28.334702 4805 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-f8mbc" podStartSLOduration=2.887727898 podStartE2EDuration="7.334676833s" podCreationTimestamp="2025-12-03 15:23:21 +0000 UTC" firstStartedPulling="2025-12-03 15:23:23.25192001 +0000 UTC m=+4432.914836933" lastFinishedPulling="2025-12-03 15:23:27.698868935 +0000 UTC m=+4437.361785868" observedRunningTime="2025-12-03 15:23:28.331509678 +0000 UTC m=+4437.994426601" watchObservedRunningTime="2025-12-03 15:23:28.334676833 +0000 UTC m=+4437.997593756" Dec 03 15:23:31 crc kubenswrapper[4805]: I1203 15:23:31.659459 4805 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-f8mbc" Dec 03 15:23:31 crc kubenswrapper[4805]: I1203 15:23:31.660172 4805 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-f8mbc" Dec 03 15:23:32 crc kubenswrapper[4805]: I1203 15:23:32.706517 4805 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-f8mbc" podUID="38b80aeb-575b-4337-837d-3520bb8690a2" containerName="registry-server" probeResult="failure" output=< Dec 03 15:23:32 crc kubenswrapper[4805]: timeout: failed to connect service ":50051" within 1s Dec 03 15:23:32 crc kubenswrapper[4805]: > Dec 03 15:23:41 crc kubenswrapper[4805]: I1203 15:23:41.709280 4805 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-f8mbc" Dec 03 15:23:41 crc kubenswrapper[4805]: I1203 15:23:41.758765 4805 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-f8mbc" Dec 03 15:23:41 crc kubenswrapper[4805]: I1203 15:23:41.956907 4805 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-f8mbc"] Dec 03 15:23:43 crc kubenswrapper[4805]: I1203 15:23:43.450720 4805 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-f8mbc" podUID="38b80aeb-575b-4337-837d-3520bb8690a2" containerName="registry-server" containerID="cri-o://732a19de6353ca13c12b856ee6fd929b489f7c5bf3666df158221d1d7abb9527" gracePeriod=2 Dec 03 15:23:43 crc kubenswrapper[4805]: I1203 15:23:43.909263 4805 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-f8mbc" Dec 03 15:23:43 crc kubenswrapper[4805]: I1203 15:23:43.950598 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/38b80aeb-575b-4337-837d-3520bb8690a2-catalog-content\") pod \"38b80aeb-575b-4337-837d-3520bb8690a2\" (UID: \"38b80aeb-575b-4337-837d-3520bb8690a2\") " Dec 03 15:23:43 crc kubenswrapper[4805]: I1203 15:23:43.950729 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/38b80aeb-575b-4337-837d-3520bb8690a2-utilities\") pod \"38b80aeb-575b-4337-837d-3520bb8690a2\" (UID: \"38b80aeb-575b-4337-837d-3520bb8690a2\") " Dec 03 15:23:43 crc kubenswrapper[4805]: I1203 15:23:43.950753 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5pxgh\" (UniqueName: \"kubernetes.io/projected/38b80aeb-575b-4337-837d-3520bb8690a2-kube-api-access-5pxgh\") pod \"38b80aeb-575b-4337-837d-3520bb8690a2\" (UID: \"38b80aeb-575b-4337-837d-3520bb8690a2\") " Dec 03 15:23:43 crc kubenswrapper[4805]: I1203 15:23:43.956366 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/38b80aeb-575b-4337-837d-3520bb8690a2-utilities" (OuterVolumeSpecName: "utilities") pod "38b80aeb-575b-4337-837d-3520bb8690a2" (UID: "38b80aeb-575b-4337-837d-3520bb8690a2"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 15:23:43 crc kubenswrapper[4805]: I1203 15:23:43.975337 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/38b80aeb-575b-4337-837d-3520bb8690a2-kube-api-access-5pxgh" (OuterVolumeSpecName: "kube-api-access-5pxgh") pod "38b80aeb-575b-4337-837d-3520bb8690a2" (UID: "38b80aeb-575b-4337-837d-3520bb8690a2"). InnerVolumeSpecName "kube-api-access-5pxgh". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 15:23:44 crc kubenswrapper[4805]: I1203 15:23:44.053133 4805 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/38b80aeb-575b-4337-837d-3520bb8690a2-utilities\") on node \"crc\" DevicePath \"\"" Dec 03 15:23:44 crc kubenswrapper[4805]: I1203 15:23:44.053176 4805 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5pxgh\" (UniqueName: \"kubernetes.io/projected/38b80aeb-575b-4337-837d-3520bb8690a2-kube-api-access-5pxgh\") on node \"crc\" DevicePath \"\"" Dec 03 15:23:44 crc kubenswrapper[4805]: I1203 15:23:44.114379 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/38b80aeb-575b-4337-837d-3520bb8690a2-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "38b80aeb-575b-4337-837d-3520bb8690a2" (UID: "38b80aeb-575b-4337-837d-3520bb8690a2"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 15:23:44 crc kubenswrapper[4805]: I1203 15:23:44.155704 4805 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/38b80aeb-575b-4337-837d-3520bb8690a2-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 03 15:23:44 crc kubenswrapper[4805]: I1203 15:23:44.464766 4805 generic.go:334] "Generic (PLEG): container finished" podID="38b80aeb-575b-4337-837d-3520bb8690a2" containerID="732a19de6353ca13c12b856ee6fd929b489f7c5bf3666df158221d1d7abb9527" exitCode=0 Dec 03 15:23:44 crc kubenswrapper[4805]: I1203 15:23:44.464820 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-f8mbc" event={"ID":"38b80aeb-575b-4337-837d-3520bb8690a2","Type":"ContainerDied","Data":"732a19de6353ca13c12b856ee6fd929b489f7c5bf3666df158221d1d7abb9527"} Dec 03 15:23:44 crc kubenswrapper[4805]: I1203 15:23:44.464887 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-f8mbc" event={"ID":"38b80aeb-575b-4337-837d-3520bb8690a2","Type":"ContainerDied","Data":"f5e911e91c49a602fee56485da4d327cc2e609ae1521e3b054094635a42b39ba"} Dec 03 15:23:44 crc kubenswrapper[4805]: I1203 15:23:44.464942 4805 scope.go:117] "RemoveContainer" containerID="732a19de6353ca13c12b856ee6fd929b489f7c5bf3666df158221d1d7abb9527" Dec 03 15:23:44 crc kubenswrapper[4805]: I1203 15:23:44.464996 4805 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-f8mbc" Dec 03 15:23:44 crc kubenswrapper[4805]: I1203 15:23:44.496546 4805 scope.go:117] "RemoveContainer" containerID="f91ddc215a11c0d2908240bdf52268a463969a260e217b2afd1c8ac4858e2de6" Dec 03 15:23:44 crc kubenswrapper[4805]: I1203 15:23:44.529631 4805 scope.go:117] "RemoveContainer" containerID="083cfa768e839b50963c4be1f2f04fc7723282eda487335e3959d9edc28beb48" Dec 03 15:23:44 crc kubenswrapper[4805]: I1203 15:23:44.543269 4805 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-f8mbc"] Dec 03 15:23:44 crc kubenswrapper[4805]: I1203 15:23:44.552180 4805 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-f8mbc"] Dec 03 15:23:44 crc kubenswrapper[4805]: I1203 15:23:44.583490 4805 scope.go:117] "RemoveContainer" containerID="732a19de6353ca13c12b856ee6fd929b489f7c5bf3666df158221d1d7abb9527" Dec 03 15:23:44 crc kubenswrapper[4805]: E1203 15:23:44.583937 4805 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"732a19de6353ca13c12b856ee6fd929b489f7c5bf3666df158221d1d7abb9527\": container with ID starting with 732a19de6353ca13c12b856ee6fd929b489f7c5bf3666df158221d1d7abb9527 not found: ID does not exist" containerID="732a19de6353ca13c12b856ee6fd929b489f7c5bf3666df158221d1d7abb9527" Dec 03 15:23:44 crc kubenswrapper[4805]: I1203 15:23:44.583992 4805 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"732a19de6353ca13c12b856ee6fd929b489f7c5bf3666df158221d1d7abb9527"} err="failed to get container status \"732a19de6353ca13c12b856ee6fd929b489f7c5bf3666df158221d1d7abb9527\": rpc error: code = NotFound desc = could not find container \"732a19de6353ca13c12b856ee6fd929b489f7c5bf3666df158221d1d7abb9527\": container with ID starting with 732a19de6353ca13c12b856ee6fd929b489f7c5bf3666df158221d1d7abb9527 not found: ID does not exist" Dec 03 15:23:44 crc kubenswrapper[4805]: I1203 15:23:44.584028 4805 scope.go:117] "RemoveContainer" containerID="f91ddc215a11c0d2908240bdf52268a463969a260e217b2afd1c8ac4858e2de6" Dec 03 15:23:44 crc kubenswrapper[4805]: E1203 15:23:44.584366 4805 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f91ddc215a11c0d2908240bdf52268a463969a260e217b2afd1c8ac4858e2de6\": container with ID starting with f91ddc215a11c0d2908240bdf52268a463969a260e217b2afd1c8ac4858e2de6 not found: ID does not exist" containerID="f91ddc215a11c0d2908240bdf52268a463969a260e217b2afd1c8ac4858e2de6" Dec 03 15:23:44 crc kubenswrapper[4805]: I1203 15:23:44.584395 4805 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f91ddc215a11c0d2908240bdf52268a463969a260e217b2afd1c8ac4858e2de6"} err="failed to get container status \"f91ddc215a11c0d2908240bdf52268a463969a260e217b2afd1c8ac4858e2de6\": rpc error: code = NotFound desc = could not find container \"f91ddc215a11c0d2908240bdf52268a463969a260e217b2afd1c8ac4858e2de6\": container with ID starting with f91ddc215a11c0d2908240bdf52268a463969a260e217b2afd1c8ac4858e2de6 not found: ID does not exist" Dec 03 15:23:44 crc kubenswrapper[4805]: I1203 15:23:44.584412 4805 scope.go:117] "RemoveContainer" containerID="083cfa768e839b50963c4be1f2f04fc7723282eda487335e3959d9edc28beb48" Dec 03 15:23:44 crc kubenswrapper[4805]: E1203 15:23:44.584712 4805 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"083cfa768e839b50963c4be1f2f04fc7723282eda487335e3959d9edc28beb48\": container with ID starting with 083cfa768e839b50963c4be1f2f04fc7723282eda487335e3959d9edc28beb48 not found: ID does not exist" containerID="083cfa768e839b50963c4be1f2f04fc7723282eda487335e3959d9edc28beb48" Dec 03 15:23:44 crc kubenswrapper[4805]: I1203 15:23:44.584784 4805 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"083cfa768e839b50963c4be1f2f04fc7723282eda487335e3959d9edc28beb48"} err="failed to get container status \"083cfa768e839b50963c4be1f2f04fc7723282eda487335e3959d9edc28beb48\": rpc error: code = NotFound desc = could not find container \"083cfa768e839b50963c4be1f2f04fc7723282eda487335e3959d9edc28beb48\": container with ID starting with 083cfa768e839b50963c4be1f2f04fc7723282eda487335e3959d9edc28beb48 not found: ID does not exist" Dec 03 15:23:44 crc kubenswrapper[4805]: I1203 15:23:44.706730 4805 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="38b80aeb-575b-4337-837d-3520bb8690a2" path="/var/lib/kubelet/pods/38b80aeb-575b-4337-837d-3520bb8690a2/volumes" Dec 03 15:23:47 crc kubenswrapper[4805]: I1203 15:23:47.984356 4805 scope.go:117] "RemoveContainer" containerID="7cfbc3fe2001fe4c7e6cf09443ce6a6e860a40da1f2c016d071c05fbe7033424" Dec 03 15:23:48 crc kubenswrapper[4805]: I1203 15:23:48.072183 4805 scope.go:117] "RemoveContainer" containerID="313a0a66fa07afa0a361dc21ea0df9772ee5f6169321650079c0add6695204d7" Dec 03 15:24:43 crc kubenswrapper[4805]: I1203 15:24:43.917143 4805 patch_prober.go:28] interesting pod/machine-config-daemon-gskh4 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 03 15:24:43 crc kubenswrapper[4805]: I1203 15:24:43.917782 4805 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-gskh4" podUID="6098937f-e3f6-45e8-a647-4994a79cd711" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 03 15:25:03 crc kubenswrapper[4805]: I1203 15:25:03.995233 4805 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-tffmk"] Dec 03 15:25:03 crc kubenswrapper[4805]: E1203 15:25:03.996130 4805 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="38b80aeb-575b-4337-837d-3520bb8690a2" containerName="extract-utilities" Dec 03 15:25:03 crc kubenswrapper[4805]: I1203 15:25:03.996144 4805 state_mem.go:107] "Deleted CPUSet assignment" podUID="38b80aeb-575b-4337-837d-3520bb8690a2" containerName="extract-utilities" Dec 03 15:25:03 crc kubenswrapper[4805]: E1203 15:25:03.996160 4805 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="38b80aeb-575b-4337-837d-3520bb8690a2" containerName="extract-content" Dec 03 15:25:03 crc kubenswrapper[4805]: I1203 15:25:03.996165 4805 state_mem.go:107] "Deleted CPUSet assignment" podUID="38b80aeb-575b-4337-837d-3520bb8690a2" containerName="extract-content" Dec 03 15:25:03 crc kubenswrapper[4805]: E1203 15:25:03.996190 4805 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="38b80aeb-575b-4337-837d-3520bb8690a2" containerName="registry-server" Dec 03 15:25:03 crc kubenswrapper[4805]: I1203 15:25:03.996196 4805 state_mem.go:107] "Deleted CPUSet assignment" podUID="38b80aeb-575b-4337-837d-3520bb8690a2" containerName="registry-server" Dec 03 15:25:03 crc kubenswrapper[4805]: I1203 15:25:03.996408 4805 memory_manager.go:354] "RemoveStaleState removing state" podUID="38b80aeb-575b-4337-837d-3520bb8690a2" containerName="registry-server" Dec 03 15:25:03 crc kubenswrapper[4805]: I1203 15:25:03.997870 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-tffmk" Dec 03 15:25:04 crc kubenswrapper[4805]: I1203 15:25:04.004206 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-tffmk"] Dec 03 15:25:04 crc kubenswrapper[4805]: I1203 15:25:04.133781 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c0647465-ae5b-4aa6-9a27-e6e3a6b8219f-utilities\") pod \"community-operators-tffmk\" (UID: \"c0647465-ae5b-4aa6-9a27-e6e3a6b8219f\") " pod="openshift-marketplace/community-operators-tffmk" Dec 03 15:25:04 crc kubenswrapper[4805]: I1203 15:25:04.133873 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c0647465-ae5b-4aa6-9a27-e6e3a6b8219f-catalog-content\") pod \"community-operators-tffmk\" (UID: \"c0647465-ae5b-4aa6-9a27-e6e3a6b8219f\") " pod="openshift-marketplace/community-operators-tffmk" Dec 03 15:25:04 crc kubenswrapper[4805]: I1203 15:25:04.133978 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bnlnm\" (UniqueName: \"kubernetes.io/projected/c0647465-ae5b-4aa6-9a27-e6e3a6b8219f-kube-api-access-bnlnm\") pod \"community-operators-tffmk\" (UID: \"c0647465-ae5b-4aa6-9a27-e6e3a6b8219f\") " pod="openshift-marketplace/community-operators-tffmk" Dec 03 15:25:04 crc kubenswrapper[4805]: I1203 15:25:04.236243 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c0647465-ae5b-4aa6-9a27-e6e3a6b8219f-utilities\") pod \"community-operators-tffmk\" (UID: \"c0647465-ae5b-4aa6-9a27-e6e3a6b8219f\") " pod="openshift-marketplace/community-operators-tffmk" Dec 03 15:25:04 crc kubenswrapper[4805]: I1203 15:25:04.236334 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c0647465-ae5b-4aa6-9a27-e6e3a6b8219f-catalog-content\") pod \"community-operators-tffmk\" (UID: \"c0647465-ae5b-4aa6-9a27-e6e3a6b8219f\") " pod="openshift-marketplace/community-operators-tffmk" Dec 03 15:25:04 crc kubenswrapper[4805]: I1203 15:25:04.236453 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bnlnm\" (UniqueName: \"kubernetes.io/projected/c0647465-ae5b-4aa6-9a27-e6e3a6b8219f-kube-api-access-bnlnm\") pod \"community-operators-tffmk\" (UID: \"c0647465-ae5b-4aa6-9a27-e6e3a6b8219f\") " pod="openshift-marketplace/community-operators-tffmk" Dec 03 15:25:04 crc kubenswrapper[4805]: I1203 15:25:04.237209 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c0647465-ae5b-4aa6-9a27-e6e3a6b8219f-utilities\") pod \"community-operators-tffmk\" (UID: \"c0647465-ae5b-4aa6-9a27-e6e3a6b8219f\") " pod="openshift-marketplace/community-operators-tffmk" Dec 03 15:25:04 crc kubenswrapper[4805]: I1203 15:25:04.237785 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c0647465-ae5b-4aa6-9a27-e6e3a6b8219f-catalog-content\") pod \"community-operators-tffmk\" (UID: \"c0647465-ae5b-4aa6-9a27-e6e3a6b8219f\") " pod="openshift-marketplace/community-operators-tffmk" Dec 03 15:25:04 crc kubenswrapper[4805]: I1203 15:25:04.268103 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bnlnm\" (UniqueName: \"kubernetes.io/projected/c0647465-ae5b-4aa6-9a27-e6e3a6b8219f-kube-api-access-bnlnm\") pod \"community-operators-tffmk\" (UID: \"c0647465-ae5b-4aa6-9a27-e6e3a6b8219f\") " pod="openshift-marketplace/community-operators-tffmk" Dec 03 15:25:04 crc kubenswrapper[4805]: I1203 15:25:04.329404 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-tffmk" Dec 03 15:25:04 crc kubenswrapper[4805]: I1203 15:25:04.886202 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-tffmk"] Dec 03 15:25:05 crc kubenswrapper[4805]: I1203 15:25:05.305093 4805 generic.go:334] "Generic (PLEG): container finished" podID="c0647465-ae5b-4aa6-9a27-e6e3a6b8219f" containerID="9332cec20127a4ffa3e629cfe8d78f9a1058f4877c7111a41ac899e7ff7ff11a" exitCode=0 Dec 03 15:25:05 crc kubenswrapper[4805]: I1203 15:25:05.305208 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-tffmk" event={"ID":"c0647465-ae5b-4aa6-9a27-e6e3a6b8219f","Type":"ContainerDied","Data":"9332cec20127a4ffa3e629cfe8d78f9a1058f4877c7111a41ac899e7ff7ff11a"} Dec 03 15:25:05 crc kubenswrapper[4805]: I1203 15:25:05.305493 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-tffmk" event={"ID":"c0647465-ae5b-4aa6-9a27-e6e3a6b8219f","Type":"ContainerStarted","Data":"db70b0b4fa1eeeef97a4a93dd3ce3f1f13c5b1be57cf63a1c1b1698ffac4e4c5"} Dec 03 15:25:07 crc kubenswrapper[4805]: I1203 15:25:07.330707 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-tffmk" event={"ID":"c0647465-ae5b-4aa6-9a27-e6e3a6b8219f","Type":"ContainerStarted","Data":"e81a02dc64bab0ca9023bb91fda11cc8c6cb6051b7e69775fae1769c02869bd2"} Dec 03 15:25:08 crc kubenswrapper[4805]: I1203 15:25:08.351128 4805 generic.go:334] "Generic (PLEG): container finished" podID="c0647465-ae5b-4aa6-9a27-e6e3a6b8219f" containerID="e81a02dc64bab0ca9023bb91fda11cc8c6cb6051b7e69775fae1769c02869bd2" exitCode=0 Dec 03 15:25:08 crc kubenswrapper[4805]: I1203 15:25:08.351510 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-tffmk" event={"ID":"c0647465-ae5b-4aa6-9a27-e6e3a6b8219f","Type":"ContainerDied","Data":"e81a02dc64bab0ca9023bb91fda11cc8c6cb6051b7e69775fae1769c02869bd2"} Dec 03 15:25:09 crc kubenswrapper[4805]: I1203 15:25:09.361711 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-tffmk" event={"ID":"c0647465-ae5b-4aa6-9a27-e6e3a6b8219f","Type":"ContainerStarted","Data":"6648ce9b8ea65002088b9873dfeea38d0455ceea7ec2cb31fab85763d034ebb7"} Dec 03 15:25:09 crc kubenswrapper[4805]: I1203 15:25:09.400631 4805 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-tffmk" podStartSLOduration=2.950855234 podStartE2EDuration="6.40060544s" podCreationTimestamp="2025-12-03 15:25:03 +0000 UTC" firstStartedPulling="2025-12-03 15:25:05.307098681 +0000 UTC m=+4534.970015604" lastFinishedPulling="2025-12-03 15:25:08.756848887 +0000 UTC m=+4538.419765810" observedRunningTime="2025-12-03 15:25:09.393020665 +0000 UTC m=+4539.055937588" watchObservedRunningTime="2025-12-03 15:25:09.40060544 +0000 UTC m=+4539.063522363" Dec 03 15:25:13 crc kubenswrapper[4805]: I1203 15:25:13.918160 4805 patch_prober.go:28] interesting pod/machine-config-daemon-gskh4 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 03 15:25:13 crc kubenswrapper[4805]: I1203 15:25:13.918734 4805 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-gskh4" podUID="6098937f-e3f6-45e8-a647-4994a79cd711" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 03 15:25:14 crc kubenswrapper[4805]: I1203 15:25:14.331120 4805 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-tffmk" Dec 03 15:25:14 crc kubenswrapper[4805]: I1203 15:25:14.331175 4805 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-tffmk" Dec 03 15:25:14 crc kubenswrapper[4805]: I1203 15:25:14.384685 4805 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-tffmk" Dec 03 15:25:14 crc kubenswrapper[4805]: I1203 15:25:14.465957 4805 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-tffmk" Dec 03 15:25:14 crc kubenswrapper[4805]: I1203 15:25:14.626569 4805 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-tffmk"] Dec 03 15:25:16 crc kubenswrapper[4805]: I1203 15:25:16.432116 4805 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-tffmk" podUID="c0647465-ae5b-4aa6-9a27-e6e3a6b8219f" containerName="registry-server" containerID="cri-o://6648ce9b8ea65002088b9873dfeea38d0455ceea7ec2cb31fab85763d034ebb7" gracePeriod=2 Dec 03 15:25:16 crc kubenswrapper[4805]: I1203 15:25:16.961243 4805 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-tffmk" Dec 03 15:25:16 crc kubenswrapper[4805]: I1203 15:25:16.995201 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bnlnm\" (UniqueName: \"kubernetes.io/projected/c0647465-ae5b-4aa6-9a27-e6e3a6b8219f-kube-api-access-bnlnm\") pod \"c0647465-ae5b-4aa6-9a27-e6e3a6b8219f\" (UID: \"c0647465-ae5b-4aa6-9a27-e6e3a6b8219f\") " Dec 03 15:25:16 crc kubenswrapper[4805]: I1203 15:25:16.995283 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c0647465-ae5b-4aa6-9a27-e6e3a6b8219f-catalog-content\") pod \"c0647465-ae5b-4aa6-9a27-e6e3a6b8219f\" (UID: \"c0647465-ae5b-4aa6-9a27-e6e3a6b8219f\") " Dec 03 15:25:16 crc kubenswrapper[4805]: I1203 15:25:16.995418 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c0647465-ae5b-4aa6-9a27-e6e3a6b8219f-utilities\") pod \"c0647465-ae5b-4aa6-9a27-e6e3a6b8219f\" (UID: \"c0647465-ae5b-4aa6-9a27-e6e3a6b8219f\") " Dec 03 15:25:16 crc kubenswrapper[4805]: I1203 15:25:16.996352 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c0647465-ae5b-4aa6-9a27-e6e3a6b8219f-utilities" (OuterVolumeSpecName: "utilities") pod "c0647465-ae5b-4aa6-9a27-e6e3a6b8219f" (UID: "c0647465-ae5b-4aa6-9a27-e6e3a6b8219f"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 15:25:17 crc kubenswrapper[4805]: I1203 15:25:17.002978 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c0647465-ae5b-4aa6-9a27-e6e3a6b8219f-kube-api-access-bnlnm" (OuterVolumeSpecName: "kube-api-access-bnlnm") pod "c0647465-ae5b-4aa6-9a27-e6e3a6b8219f" (UID: "c0647465-ae5b-4aa6-9a27-e6e3a6b8219f"). InnerVolumeSpecName "kube-api-access-bnlnm". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 15:25:17 crc kubenswrapper[4805]: I1203 15:25:17.065980 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c0647465-ae5b-4aa6-9a27-e6e3a6b8219f-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "c0647465-ae5b-4aa6-9a27-e6e3a6b8219f" (UID: "c0647465-ae5b-4aa6-9a27-e6e3a6b8219f"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 15:25:17 crc kubenswrapper[4805]: I1203 15:25:17.097981 4805 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bnlnm\" (UniqueName: \"kubernetes.io/projected/c0647465-ae5b-4aa6-9a27-e6e3a6b8219f-kube-api-access-bnlnm\") on node \"crc\" DevicePath \"\"" Dec 03 15:25:17 crc kubenswrapper[4805]: I1203 15:25:17.098022 4805 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c0647465-ae5b-4aa6-9a27-e6e3a6b8219f-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 03 15:25:17 crc kubenswrapper[4805]: I1203 15:25:17.098039 4805 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c0647465-ae5b-4aa6-9a27-e6e3a6b8219f-utilities\") on node \"crc\" DevicePath \"\"" Dec 03 15:25:17 crc kubenswrapper[4805]: I1203 15:25:17.444028 4805 generic.go:334] "Generic (PLEG): container finished" podID="c0647465-ae5b-4aa6-9a27-e6e3a6b8219f" containerID="6648ce9b8ea65002088b9873dfeea38d0455ceea7ec2cb31fab85763d034ebb7" exitCode=0 Dec 03 15:25:17 crc kubenswrapper[4805]: I1203 15:25:17.444095 4805 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-tffmk" Dec 03 15:25:17 crc kubenswrapper[4805]: I1203 15:25:17.444108 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-tffmk" event={"ID":"c0647465-ae5b-4aa6-9a27-e6e3a6b8219f","Type":"ContainerDied","Data":"6648ce9b8ea65002088b9873dfeea38d0455ceea7ec2cb31fab85763d034ebb7"} Dec 03 15:25:17 crc kubenswrapper[4805]: I1203 15:25:17.444155 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-tffmk" event={"ID":"c0647465-ae5b-4aa6-9a27-e6e3a6b8219f","Type":"ContainerDied","Data":"db70b0b4fa1eeeef97a4a93dd3ce3f1f13c5b1be57cf63a1c1b1698ffac4e4c5"} Dec 03 15:25:17 crc kubenswrapper[4805]: I1203 15:25:17.444197 4805 scope.go:117] "RemoveContainer" containerID="6648ce9b8ea65002088b9873dfeea38d0455ceea7ec2cb31fab85763d034ebb7" Dec 03 15:25:17 crc kubenswrapper[4805]: I1203 15:25:17.470375 4805 scope.go:117] "RemoveContainer" containerID="e81a02dc64bab0ca9023bb91fda11cc8c6cb6051b7e69775fae1769c02869bd2" Dec 03 15:25:17 crc kubenswrapper[4805]: I1203 15:25:17.485344 4805 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-tffmk"] Dec 03 15:25:17 crc kubenswrapper[4805]: I1203 15:25:17.494568 4805 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-tffmk"] Dec 03 15:25:17 crc kubenswrapper[4805]: I1203 15:25:17.494701 4805 scope.go:117] "RemoveContainer" containerID="9332cec20127a4ffa3e629cfe8d78f9a1058f4877c7111a41ac899e7ff7ff11a" Dec 03 15:25:17 crc kubenswrapper[4805]: I1203 15:25:17.546432 4805 scope.go:117] "RemoveContainer" containerID="6648ce9b8ea65002088b9873dfeea38d0455ceea7ec2cb31fab85763d034ebb7" Dec 03 15:25:17 crc kubenswrapper[4805]: E1203 15:25:17.547509 4805 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6648ce9b8ea65002088b9873dfeea38d0455ceea7ec2cb31fab85763d034ebb7\": container with ID starting with 6648ce9b8ea65002088b9873dfeea38d0455ceea7ec2cb31fab85763d034ebb7 not found: ID does not exist" containerID="6648ce9b8ea65002088b9873dfeea38d0455ceea7ec2cb31fab85763d034ebb7" Dec 03 15:25:17 crc kubenswrapper[4805]: I1203 15:25:17.547549 4805 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6648ce9b8ea65002088b9873dfeea38d0455ceea7ec2cb31fab85763d034ebb7"} err="failed to get container status \"6648ce9b8ea65002088b9873dfeea38d0455ceea7ec2cb31fab85763d034ebb7\": rpc error: code = NotFound desc = could not find container \"6648ce9b8ea65002088b9873dfeea38d0455ceea7ec2cb31fab85763d034ebb7\": container with ID starting with 6648ce9b8ea65002088b9873dfeea38d0455ceea7ec2cb31fab85763d034ebb7 not found: ID does not exist" Dec 03 15:25:17 crc kubenswrapper[4805]: I1203 15:25:17.547578 4805 scope.go:117] "RemoveContainer" containerID="e81a02dc64bab0ca9023bb91fda11cc8c6cb6051b7e69775fae1769c02869bd2" Dec 03 15:25:17 crc kubenswrapper[4805]: E1203 15:25:17.547970 4805 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e81a02dc64bab0ca9023bb91fda11cc8c6cb6051b7e69775fae1769c02869bd2\": container with ID starting with e81a02dc64bab0ca9023bb91fda11cc8c6cb6051b7e69775fae1769c02869bd2 not found: ID does not exist" containerID="e81a02dc64bab0ca9023bb91fda11cc8c6cb6051b7e69775fae1769c02869bd2" Dec 03 15:25:17 crc kubenswrapper[4805]: I1203 15:25:17.548014 4805 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e81a02dc64bab0ca9023bb91fda11cc8c6cb6051b7e69775fae1769c02869bd2"} err="failed to get container status \"e81a02dc64bab0ca9023bb91fda11cc8c6cb6051b7e69775fae1769c02869bd2\": rpc error: code = NotFound desc = could not find container \"e81a02dc64bab0ca9023bb91fda11cc8c6cb6051b7e69775fae1769c02869bd2\": container with ID starting with e81a02dc64bab0ca9023bb91fda11cc8c6cb6051b7e69775fae1769c02869bd2 not found: ID does not exist" Dec 03 15:25:17 crc kubenswrapper[4805]: I1203 15:25:17.548043 4805 scope.go:117] "RemoveContainer" containerID="9332cec20127a4ffa3e629cfe8d78f9a1058f4877c7111a41ac899e7ff7ff11a" Dec 03 15:25:17 crc kubenswrapper[4805]: E1203 15:25:17.548385 4805 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9332cec20127a4ffa3e629cfe8d78f9a1058f4877c7111a41ac899e7ff7ff11a\": container with ID starting with 9332cec20127a4ffa3e629cfe8d78f9a1058f4877c7111a41ac899e7ff7ff11a not found: ID does not exist" containerID="9332cec20127a4ffa3e629cfe8d78f9a1058f4877c7111a41ac899e7ff7ff11a" Dec 03 15:25:17 crc kubenswrapper[4805]: I1203 15:25:17.548436 4805 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9332cec20127a4ffa3e629cfe8d78f9a1058f4877c7111a41ac899e7ff7ff11a"} err="failed to get container status \"9332cec20127a4ffa3e629cfe8d78f9a1058f4877c7111a41ac899e7ff7ff11a\": rpc error: code = NotFound desc = could not find container \"9332cec20127a4ffa3e629cfe8d78f9a1058f4877c7111a41ac899e7ff7ff11a\": container with ID starting with 9332cec20127a4ffa3e629cfe8d78f9a1058f4877c7111a41ac899e7ff7ff11a not found: ID does not exist" Dec 03 15:25:18 crc kubenswrapper[4805]: I1203 15:25:18.708448 4805 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c0647465-ae5b-4aa6-9a27-e6e3a6b8219f" path="/var/lib/kubelet/pods/c0647465-ae5b-4aa6-9a27-e6e3a6b8219f/volumes" Dec 03 15:25:43 crc kubenswrapper[4805]: I1203 15:25:43.917484 4805 patch_prober.go:28] interesting pod/machine-config-daemon-gskh4 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 03 15:25:43 crc kubenswrapper[4805]: I1203 15:25:43.918201 4805 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-gskh4" podUID="6098937f-e3f6-45e8-a647-4994a79cd711" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 03 15:25:43 crc kubenswrapper[4805]: I1203 15:25:43.918284 4805 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-gskh4" Dec 03 15:25:43 crc kubenswrapper[4805]: I1203 15:25:43.919057 4805 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"7d1692e3c7d3a2b9dcbcaebd54978af4cba7ae377a17c65268d4048aa72d2c92"} pod="openshift-machine-config-operator/machine-config-daemon-gskh4" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 03 15:25:43 crc kubenswrapper[4805]: I1203 15:25:43.919112 4805 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-gskh4" podUID="6098937f-e3f6-45e8-a647-4994a79cd711" containerName="machine-config-daemon" containerID="cri-o://7d1692e3c7d3a2b9dcbcaebd54978af4cba7ae377a17c65268d4048aa72d2c92" gracePeriod=600 Dec 03 15:25:44 crc kubenswrapper[4805]: E1203 15:25:44.114362 4805 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-gskh4_openshift-machine-config-operator(6098937f-e3f6-45e8-a647-4994a79cd711)\"" pod="openshift-machine-config-operator/machine-config-daemon-gskh4" podUID="6098937f-e3f6-45e8-a647-4994a79cd711" Dec 03 15:25:44 crc kubenswrapper[4805]: I1203 15:25:44.711536 4805 generic.go:334] "Generic (PLEG): container finished" podID="6098937f-e3f6-45e8-a647-4994a79cd711" containerID="7d1692e3c7d3a2b9dcbcaebd54978af4cba7ae377a17c65268d4048aa72d2c92" exitCode=0 Dec 03 15:25:44 crc kubenswrapper[4805]: I1203 15:25:44.711590 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-gskh4" event={"ID":"6098937f-e3f6-45e8-a647-4994a79cd711","Type":"ContainerDied","Data":"7d1692e3c7d3a2b9dcbcaebd54978af4cba7ae377a17c65268d4048aa72d2c92"} Dec 03 15:25:44 crc kubenswrapper[4805]: I1203 15:25:44.711656 4805 scope.go:117] "RemoveContainer" containerID="c1a31f4f6e3e9fa851f0e438617fecf09bad8f6021091f02eae4b71f6b3e374c" Dec 03 15:25:44 crc kubenswrapper[4805]: I1203 15:25:44.712439 4805 scope.go:117] "RemoveContainer" containerID="7d1692e3c7d3a2b9dcbcaebd54978af4cba7ae377a17c65268d4048aa72d2c92" Dec 03 15:25:44 crc kubenswrapper[4805]: E1203 15:25:44.712729 4805 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-gskh4_openshift-machine-config-operator(6098937f-e3f6-45e8-a647-4994a79cd711)\"" pod="openshift-machine-config-operator/machine-config-daemon-gskh4" podUID="6098937f-e3f6-45e8-a647-4994a79cd711" Dec 03 15:25:54 crc kubenswrapper[4805]: I1203 15:25:54.796047 4805 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-7hg4c/must-gather-2mm9s"] Dec 03 15:25:54 crc kubenswrapper[4805]: E1203 15:25:54.797033 4805 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c0647465-ae5b-4aa6-9a27-e6e3a6b8219f" containerName="extract-utilities" Dec 03 15:25:54 crc kubenswrapper[4805]: I1203 15:25:54.797047 4805 state_mem.go:107] "Deleted CPUSet assignment" podUID="c0647465-ae5b-4aa6-9a27-e6e3a6b8219f" containerName="extract-utilities" Dec 03 15:25:54 crc kubenswrapper[4805]: E1203 15:25:54.797073 4805 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c0647465-ae5b-4aa6-9a27-e6e3a6b8219f" containerName="extract-content" Dec 03 15:25:54 crc kubenswrapper[4805]: I1203 15:25:54.797079 4805 state_mem.go:107] "Deleted CPUSet assignment" podUID="c0647465-ae5b-4aa6-9a27-e6e3a6b8219f" containerName="extract-content" Dec 03 15:25:54 crc kubenswrapper[4805]: E1203 15:25:54.797109 4805 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c0647465-ae5b-4aa6-9a27-e6e3a6b8219f" containerName="registry-server" Dec 03 15:25:54 crc kubenswrapper[4805]: I1203 15:25:54.797115 4805 state_mem.go:107] "Deleted CPUSet assignment" podUID="c0647465-ae5b-4aa6-9a27-e6e3a6b8219f" containerName="registry-server" Dec 03 15:25:54 crc kubenswrapper[4805]: I1203 15:25:54.797372 4805 memory_manager.go:354] "RemoveStaleState removing state" podUID="c0647465-ae5b-4aa6-9a27-e6e3a6b8219f" containerName="registry-server" Dec 03 15:25:54 crc kubenswrapper[4805]: I1203 15:25:54.798459 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-7hg4c/must-gather-2mm9s" Dec 03 15:25:54 crc kubenswrapper[4805]: I1203 15:25:54.802511 4805 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-must-gather-7hg4c"/"kube-root-ca.crt" Dec 03 15:25:54 crc kubenswrapper[4805]: I1203 15:25:54.802759 4805 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-must-gather-7hg4c"/"openshift-service-ca.crt" Dec 03 15:25:54 crc kubenswrapper[4805]: I1203 15:25:54.823157 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-must-gather-7hg4c/must-gather-2mm9s"] Dec 03 15:25:54 crc kubenswrapper[4805]: I1203 15:25:54.906282 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/6093636d-db50-435e-b563-2ea4aef4e8c9-must-gather-output\") pod \"must-gather-2mm9s\" (UID: \"6093636d-db50-435e-b563-2ea4aef4e8c9\") " pod="openshift-must-gather-7hg4c/must-gather-2mm9s" Dec 03 15:25:54 crc kubenswrapper[4805]: I1203 15:25:54.906339 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lm479\" (UniqueName: \"kubernetes.io/projected/6093636d-db50-435e-b563-2ea4aef4e8c9-kube-api-access-lm479\") pod \"must-gather-2mm9s\" (UID: \"6093636d-db50-435e-b563-2ea4aef4e8c9\") " pod="openshift-must-gather-7hg4c/must-gather-2mm9s" Dec 03 15:25:55 crc kubenswrapper[4805]: I1203 15:25:55.008430 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/6093636d-db50-435e-b563-2ea4aef4e8c9-must-gather-output\") pod \"must-gather-2mm9s\" (UID: \"6093636d-db50-435e-b563-2ea4aef4e8c9\") " pod="openshift-must-gather-7hg4c/must-gather-2mm9s" Dec 03 15:25:55 crc kubenswrapper[4805]: I1203 15:25:55.008489 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lm479\" (UniqueName: \"kubernetes.io/projected/6093636d-db50-435e-b563-2ea4aef4e8c9-kube-api-access-lm479\") pod \"must-gather-2mm9s\" (UID: \"6093636d-db50-435e-b563-2ea4aef4e8c9\") " pod="openshift-must-gather-7hg4c/must-gather-2mm9s" Dec 03 15:25:55 crc kubenswrapper[4805]: I1203 15:25:55.009300 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/6093636d-db50-435e-b563-2ea4aef4e8c9-must-gather-output\") pod \"must-gather-2mm9s\" (UID: \"6093636d-db50-435e-b563-2ea4aef4e8c9\") " pod="openshift-must-gather-7hg4c/must-gather-2mm9s" Dec 03 15:25:55 crc kubenswrapper[4805]: I1203 15:25:55.036953 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lm479\" (UniqueName: \"kubernetes.io/projected/6093636d-db50-435e-b563-2ea4aef4e8c9-kube-api-access-lm479\") pod \"must-gather-2mm9s\" (UID: \"6093636d-db50-435e-b563-2ea4aef4e8c9\") " pod="openshift-must-gather-7hg4c/must-gather-2mm9s" Dec 03 15:25:55 crc kubenswrapper[4805]: I1203 15:25:55.119194 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-7hg4c/must-gather-2mm9s" Dec 03 15:25:55 crc kubenswrapper[4805]: I1203 15:25:55.724116 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-must-gather-7hg4c/must-gather-2mm9s"] Dec 03 15:25:55 crc kubenswrapper[4805]: I1203 15:25:55.819528 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-7hg4c/must-gather-2mm9s" event={"ID":"6093636d-db50-435e-b563-2ea4aef4e8c9","Type":"ContainerStarted","Data":"f89cdcf4b84ef5a2a4ed311d4d099fe913c60ba40d30334efe526de0ee7fcc25"} Dec 03 15:25:56 crc kubenswrapper[4805]: I1203 15:25:56.831351 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-7hg4c/must-gather-2mm9s" event={"ID":"6093636d-db50-435e-b563-2ea4aef4e8c9","Type":"ContainerStarted","Data":"9dcab8d81985fdec27ab73016a77ccac5cd1d0b236c583da6fc032e6b7b9df87"} Dec 03 15:25:56 crc kubenswrapper[4805]: I1203 15:25:56.832604 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-7hg4c/must-gather-2mm9s" event={"ID":"6093636d-db50-435e-b563-2ea4aef4e8c9","Type":"ContainerStarted","Data":"cd4b89422b4de3d0665d78ffea319d0d1c3bd669d7c50fb2f68c2ea5780b99c5"} Dec 03 15:25:56 crc kubenswrapper[4805]: I1203 15:25:56.849272 4805 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-must-gather-7hg4c/must-gather-2mm9s" podStartSLOduration=2.84925225 podStartE2EDuration="2.84925225s" podCreationTimestamp="2025-12-03 15:25:54 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 15:25:56.846867286 +0000 UTC m=+4586.509784219" watchObservedRunningTime="2025-12-03 15:25:56.84925225 +0000 UTC m=+4586.512169173" Dec 03 15:25:58 crc kubenswrapper[4805]: I1203 15:25:58.695058 4805 scope.go:117] "RemoveContainer" containerID="7d1692e3c7d3a2b9dcbcaebd54978af4cba7ae377a17c65268d4048aa72d2c92" Dec 03 15:25:58 crc kubenswrapper[4805]: E1203 15:25:58.695804 4805 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-gskh4_openshift-machine-config-operator(6098937f-e3f6-45e8-a647-4994a79cd711)\"" pod="openshift-machine-config-operator/machine-config-daemon-gskh4" podUID="6098937f-e3f6-45e8-a647-4994a79cd711" Dec 03 15:25:59 crc kubenswrapper[4805]: I1203 15:25:59.832401 4805 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-7hg4c/crc-debug-sm9sv"] Dec 03 15:25:59 crc kubenswrapper[4805]: I1203 15:25:59.834506 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-7hg4c/crc-debug-sm9sv" Dec 03 15:25:59 crc kubenswrapper[4805]: I1203 15:25:59.837202 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-must-gather-7hg4c"/"default-dockercfg-fn2js" Dec 03 15:26:00 crc kubenswrapper[4805]: I1203 15:26:00.009630 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fxsll\" (UniqueName: \"kubernetes.io/projected/ab193626-9048-4d2e-bfbd-39f4a9672aa8-kube-api-access-fxsll\") pod \"crc-debug-sm9sv\" (UID: \"ab193626-9048-4d2e-bfbd-39f4a9672aa8\") " pod="openshift-must-gather-7hg4c/crc-debug-sm9sv" Dec 03 15:26:00 crc kubenswrapper[4805]: I1203 15:26:00.009688 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/ab193626-9048-4d2e-bfbd-39f4a9672aa8-host\") pod \"crc-debug-sm9sv\" (UID: \"ab193626-9048-4d2e-bfbd-39f4a9672aa8\") " pod="openshift-must-gather-7hg4c/crc-debug-sm9sv" Dec 03 15:26:00 crc kubenswrapper[4805]: I1203 15:26:00.111545 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fxsll\" (UniqueName: \"kubernetes.io/projected/ab193626-9048-4d2e-bfbd-39f4a9672aa8-kube-api-access-fxsll\") pod \"crc-debug-sm9sv\" (UID: \"ab193626-9048-4d2e-bfbd-39f4a9672aa8\") " pod="openshift-must-gather-7hg4c/crc-debug-sm9sv" Dec 03 15:26:00 crc kubenswrapper[4805]: I1203 15:26:00.111883 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/ab193626-9048-4d2e-bfbd-39f4a9672aa8-host\") pod \"crc-debug-sm9sv\" (UID: \"ab193626-9048-4d2e-bfbd-39f4a9672aa8\") " pod="openshift-must-gather-7hg4c/crc-debug-sm9sv" Dec 03 15:26:00 crc kubenswrapper[4805]: I1203 15:26:00.111970 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/ab193626-9048-4d2e-bfbd-39f4a9672aa8-host\") pod \"crc-debug-sm9sv\" (UID: \"ab193626-9048-4d2e-bfbd-39f4a9672aa8\") " pod="openshift-must-gather-7hg4c/crc-debug-sm9sv" Dec 03 15:26:00 crc kubenswrapper[4805]: I1203 15:26:00.146572 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fxsll\" (UniqueName: \"kubernetes.io/projected/ab193626-9048-4d2e-bfbd-39f4a9672aa8-kube-api-access-fxsll\") pod \"crc-debug-sm9sv\" (UID: \"ab193626-9048-4d2e-bfbd-39f4a9672aa8\") " pod="openshift-must-gather-7hg4c/crc-debug-sm9sv" Dec 03 15:26:00 crc kubenswrapper[4805]: I1203 15:26:00.163609 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-7hg4c/crc-debug-sm9sv" Dec 03 15:26:00 crc kubenswrapper[4805]: W1203 15:26:00.216029 4805 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podab193626_9048_4d2e_bfbd_39f4a9672aa8.slice/crio-2e4f57d783f318d9950e24ba0d6926f7e84ead0e76398098ae47df78612aca84 WatchSource:0}: Error finding container 2e4f57d783f318d9950e24ba0d6926f7e84ead0e76398098ae47df78612aca84: Status 404 returned error can't find the container with id 2e4f57d783f318d9950e24ba0d6926f7e84ead0e76398098ae47df78612aca84 Dec 03 15:26:00 crc kubenswrapper[4805]: I1203 15:26:00.873256 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-7hg4c/crc-debug-sm9sv" event={"ID":"ab193626-9048-4d2e-bfbd-39f4a9672aa8","Type":"ContainerStarted","Data":"a83cfa2c0582b7024a22927ba43f38d9e3119dc6901ca48927056e1a9158e8d9"} Dec 03 15:26:00 crc kubenswrapper[4805]: I1203 15:26:00.874932 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-7hg4c/crc-debug-sm9sv" event={"ID":"ab193626-9048-4d2e-bfbd-39f4a9672aa8","Type":"ContainerStarted","Data":"2e4f57d783f318d9950e24ba0d6926f7e84ead0e76398098ae47df78612aca84"} Dec 03 15:26:00 crc kubenswrapper[4805]: I1203 15:26:00.905668 4805 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-must-gather-7hg4c/crc-debug-sm9sv" podStartSLOduration=1.905646008 podStartE2EDuration="1.905646008s" podCreationTimestamp="2025-12-03 15:25:59 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 15:26:00.894157698 +0000 UTC m=+4590.557074621" watchObservedRunningTime="2025-12-03 15:26:00.905646008 +0000 UTC m=+4590.568562931" Dec 03 15:26:10 crc kubenswrapper[4805]: I1203 15:26:10.701696 4805 scope.go:117] "RemoveContainer" containerID="7d1692e3c7d3a2b9dcbcaebd54978af4cba7ae377a17c65268d4048aa72d2c92" Dec 03 15:26:10 crc kubenswrapper[4805]: E1203 15:26:10.703023 4805 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-gskh4_openshift-machine-config-operator(6098937f-e3f6-45e8-a647-4994a79cd711)\"" pod="openshift-machine-config-operator/machine-config-daemon-gskh4" podUID="6098937f-e3f6-45e8-a647-4994a79cd711" Dec 03 15:26:12 crc kubenswrapper[4805]: I1203 15:26:12.018607 4805 generic.go:334] "Generic (PLEG): container finished" podID="ab193626-9048-4d2e-bfbd-39f4a9672aa8" containerID="a83cfa2c0582b7024a22927ba43f38d9e3119dc6901ca48927056e1a9158e8d9" exitCode=0 Dec 03 15:26:12 crc kubenswrapper[4805]: I1203 15:26:12.018723 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-7hg4c/crc-debug-sm9sv" event={"ID":"ab193626-9048-4d2e-bfbd-39f4a9672aa8","Type":"ContainerDied","Data":"a83cfa2c0582b7024a22927ba43f38d9e3119dc6901ca48927056e1a9158e8d9"} Dec 03 15:26:13 crc kubenswrapper[4805]: I1203 15:26:13.848471 4805 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-7hg4c/crc-debug-sm9sv" Dec 03 15:26:13 crc kubenswrapper[4805]: I1203 15:26:13.884135 4805 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-7hg4c/crc-debug-sm9sv"] Dec 03 15:26:13 crc kubenswrapper[4805]: I1203 15:26:13.897934 4805 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-7hg4c/crc-debug-sm9sv"] Dec 03 15:26:13 crc kubenswrapper[4805]: I1203 15:26:13.902537 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/ab193626-9048-4d2e-bfbd-39f4a9672aa8-host\") pod \"ab193626-9048-4d2e-bfbd-39f4a9672aa8\" (UID: \"ab193626-9048-4d2e-bfbd-39f4a9672aa8\") " Dec 03 15:26:13 crc kubenswrapper[4805]: I1203 15:26:13.902648 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fxsll\" (UniqueName: \"kubernetes.io/projected/ab193626-9048-4d2e-bfbd-39f4a9672aa8-kube-api-access-fxsll\") pod \"ab193626-9048-4d2e-bfbd-39f4a9672aa8\" (UID: \"ab193626-9048-4d2e-bfbd-39f4a9672aa8\") " Dec 03 15:26:13 crc kubenswrapper[4805]: I1203 15:26:13.904297 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/ab193626-9048-4d2e-bfbd-39f4a9672aa8-host" (OuterVolumeSpecName: "host") pod "ab193626-9048-4d2e-bfbd-39f4a9672aa8" (UID: "ab193626-9048-4d2e-bfbd-39f4a9672aa8"). InnerVolumeSpecName "host". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 03 15:26:13 crc kubenswrapper[4805]: I1203 15:26:13.924749 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ab193626-9048-4d2e-bfbd-39f4a9672aa8-kube-api-access-fxsll" (OuterVolumeSpecName: "kube-api-access-fxsll") pod "ab193626-9048-4d2e-bfbd-39f4a9672aa8" (UID: "ab193626-9048-4d2e-bfbd-39f4a9672aa8"). InnerVolumeSpecName "kube-api-access-fxsll". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 15:26:14 crc kubenswrapper[4805]: I1203 15:26:14.005829 4805 reconciler_common.go:293] "Volume detached for volume \"host\" (UniqueName: \"kubernetes.io/host-path/ab193626-9048-4d2e-bfbd-39f4a9672aa8-host\") on node \"crc\" DevicePath \"\"" Dec 03 15:26:14 crc kubenswrapper[4805]: I1203 15:26:14.005889 4805 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fxsll\" (UniqueName: \"kubernetes.io/projected/ab193626-9048-4d2e-bfbd-39f4a9672aa8-kube-api-access-fxsll\") on node \"crc\" DevicePath \"\"" Dec 03 15:26:14 crc kubenswrapper[4805]: I1203 15:26:14.041611 4805 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="2e4f57d783f318d9950e24ba0d6926f7e84ead0e76398098ae47df78612aca84" Dec 03 15:26:14 crc kubenswrapper[4805]: I1203 15:26:14.041694 4805 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-7hg4c/crc-debug-sm9sv" Dec 03 15:26:14 crc kubenswrapper[4805]: I1203 15:26:14.708908 4805 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ab193626-9048-4d2e-bfbd-39f4a9672aa8" path="/var/lib/kubelet/pods/ab193626-9048-4d2e-bfbd-39f4a9672aa8/volumes" Dec 03 15:26:15 crc kubenswrapper[4805]: I1203 15:26:15.092562 4805 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-7hg4c/crc-debug-76hqb"] Dec 03 15:26:15 crc kubenswrapper[4805]: E1203 15:26:15.093428 4805 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ab193626-9048-4d2e-bfbd-39f4a9672aa8" containerName="container-00" Dec 03 15:26:15 crc kubenswrapper[4805]: I1203 15:26:15.093447 4805 state_mem.go:107] "Deleted CPUSet assignment" podUID="ab193626-9048-4d2e-bfbd-39f4a9672aa8" containerName="container-00" Dec 03 15:26:15 crc kubenswrapper[4805]: I1203 15:26:15.093652 4805 memory_manager.go:354] "RemoveStaleState removing state" podUID="ab193626-9048-4d2e-bfbd-39f4a9672aa8" containerName="container-00" Dec 03 15:26:15 crc kubenswrapper[4805]: I1203 15:26:15.094335 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-7hg4c/crc-debug-76hqb" Dec 03 15:26:15 crc kubenswrapper[4805]: I1203 15:26:15.098113 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-must-gather-7hg4c"/"default-dockercfg-fn2js" Dec 03 15:26:15 crc kubenswrapper[4805]: I1203 15:26:15.231082 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/1a1dd071-ae0f-45bd-a7a4-afa5d281c904-host\") pod \"crc-debug-76hqb\" (UID: \"1a1dd071-ae0f-45bd-a7a4-afa5d281c904\") " pod="openshift-must-gather-7hg4c/crc-debug-76hqb" Dec 03 15:26:15 crc kubenswrapper[4805]: I1203 15:26:15.231450 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vjhbx\" (UniqueName: \"kubernetes.io/projected/1a1dd071-ae0f-45bd-a7a4-afa5d281c904-kube-api-access-vjhbx\") pod \"crc-debug-76hqb\" (UID: \"1a1dd071-ae0f-45bd-a7a4-afa5d281c904\") " pod="openshift-must-gather-7hg4c/crc-debug-76hqb" Dec 03 15:26:15 crc kubenswrapper[4805]: I1203 15:26:15.333054 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/1a1dd071-ae0f-45bd-a7a4-afa5d281c904-host\") pod \"crc-debug-76hqb\" (UID: \"1a1dd071-ae0f-45bd-a7a4-afa5d281c904\") " pod="openshift-must-gather-7hg4c/crc-debug-76hqb" Dec 03 15:26:15 crc kubenswrapper[4805]: I1203 15:26:15.333125 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vjhbx\" (UniqueName: \"kubernetes.io/projected/1a1dd071-ae0f-45bd-a7a4-afa5d281c904-kube-api-access-vjhbx\") pod \"crc-debug-76hqb\" (UID: \"1a1dd071-ae0f-45bd-a7a4-afa5d281c904\") " pod="openshift-must-gather-7hg4c/crc-debug-76hqb" Dec 03 15:26:15 crc kubenswrapper[4805]: I1203 15:26:15.333348 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/1a1dd071-ae0f-45bd-a7a4-afa5d281c904-host\") pod \"crc-debug-76hqb\" (UID: \"1a1dd071-ae0f-45bd-a7a4-afa5d281c904\") " pod="openshift-must-gather-7hg4c/crc-debug-76hqb" Dec 03 15:26:15 crc kubenswrapper[4805]: I1203 15:26:15.394098 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vjhbx\" (UniqueName: \"kubernetes.io/projected/1a1dd071-ae0f-45bd-a7a4-afa5d281c904-kube-api-access-vjhbx\") pod \"crc-debug-76hqb\" (UID: \"1a1dd071-ae0f-45bd-a7a4-afa5d281c904\") " pod="openshift-must-gather-7hg4c/crc-debug-76hqb" Dec 03 15:26:15 crc kubenswrapper[4805]: I1203 15:26:15.412752 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-7hg4c/crc-debug-76hqb" Dec 03 15:26:15 crc kubenswrapper[4805]: W1203 15:26:15.457007 4805 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod1a1dd071_ae0f_45bd_a7a4_afa5d281c904.slice/crio-d17d460eaf0968a35db1bf59cf1d51c2e1e05c03f7686383283d5b9d3eeabff0 WatchSource:0}: Error finding container d17d460eaf0968a35db1bf59cf1d51c2e1e05c03f7686383283d5b9d3eeabff0: Status 404 returned error can't find the container with id d17d460eaf0968a35db1bf59cf1d51c2e1e05c03f7686383283d5b9d3eeabff0 Dec 03 15:26:16 crc kubenswrapper[4805]: I1203 15:26:16.060203 4805 generic.go:334] "Generic (PLEG): container finished" podID="1a1dd071-ae0f-45bd-a7a4-afa5d281c904" containerID="4d5e47f3aa1cfa8c117b7dd4cbef461c49dd3ef478c60ee5a797698ca669c1cd" exitCode=1 Dec 03 15:26:16 crc kubenswrapper[4805]: I1203 15:26:16.060297 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-7hg4c/crc-debug-76hqb" event={"ID":"1a1dd071-ae0f-45bd-a7a4-afa5d281c904","Type":"ContainerDied","Data":"4d5e47f3aa1cfa8c117b7dd4cbef461c49dd3ef478c60ee5a797698ca669c1cd"} Dec 03 15:26:16 crc kubenswrapper[4805]: I1203 15:26:16.060760 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-7hg4c/crc-debug-76hqb" event={"ID":"1a1dd071-ae0f-45bd-a7a4-afa5d281c904","Type":"ContainerStarted","Data":"d17d460eaf0968a35db1bf59cf1d51c2e1e05c03f7686383283d5b9d3eeabff0"} Dec 03 15:26:16 crc kubenswrapper[4805]: I1203 15:26:16.104571 4805 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-7hg4c/crc-debug-76hqb"] Dec 03 15:26:16 crc kubenswrapper[4805]: I1203 15:26:16.117784 4805 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-7hg4c/crc-debug-76hqb"] Dec 03 15:26:17 crc kubenswrapper[4805]: I1203 15:26:17.181057 4805 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-7hg4c/crc-debug-76hqb" Dec 03 15:26:17 crc kubenswrapper[4805]: I1203 15:26:17.281264 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vjhbx\" (UniqueName: \"kubernetes.io/projected/1a1dd071-ae0f-45bd-a7a4-afa5d281c904-kube-api-access-vjhbx\") pod \"1a1dd071-ae0f-45bd-a7a4-afa5d281c904\" (UID: \"1a1dd071-ae0f-45bd-a7a4-afa5d281c904\") " Dec 03 15:26:17 crc kubenswrapper[4805]: I1203 15:26:17.282413 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/1a1dd071-ae0f-45bd-a7a4-afa5d281c904-host\") pod \"1a1dd071-ae0f-45bd-a7a4-afa5d281c904\" (UID: \"1a1dd071-ae0f-45bd-a7a4-afa5d281c904\") " Dec 03 15:26:17 crc kubenswrapper[4805]: I1203 15:26:17.282659 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/1a1dd071-ae0f-45bd-a7a4-afa5d281c904-host" (OuterVolumeSpecName: "host") pod "1a1dd071-ae0f-45bd-a7a4-afa5d281c904" (UID: "1a1dd071-ae0f-45bd-a7a4-afa5d281c904"). InnerVolumeSpecName "host". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 03 15:26:17 crc kubenswrapper[4805]: I1203 15:26:17.283317 4805 reconciler_common.go:293] "Volume detached for volume \"host\" (UniqueName: \"kubernetes.io/host-path/1a1dd071-ae0f-45bd-a7a4-afa5d281c904-host\") on node \"crc\" DevicePath \"\"" Dec 03 15:26:17 crc kubenswrapper[4805]: I1203 15:26:17.288122 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1a1dd071-ae0f-45bd-a7a4-afa5d281c904-kube-api-access-vjhbx" (OuterVolumeSpecName: "kube-api-access-vjhbx") pod "1a1dd071-ae0f-45bd-a7a4-afa5d281c904" (UID: "1a1dd071-ae0f-45bd-a7a4-afa5d281c904"). InnerVolumeSpecName "kube-api-access-vjhbx". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 15:26:17 crc kubenswrapper[4805]: I1203 15:26:17.385433 4805 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vjhbx\" (UniqueName: \"kubernetes.io/projected/1a1dd071-ae0f-45bd-a7a4-afa5d281c904-kube-api-access-vjhbx\") on node \"crc\" DevicePath \"\"" Dec 03 15:26:18 crc kubenswrapper[4805]: I1203 15:26:18.081411 4805 scope.go:117] "RemoveContainer" containerID="4d5e47f3aa1cfa8c117b7dd4cbef461c49dd3ef478c60ee5a797698ca669c1cd" Dec 03 15:26:18 crc kubenswrapper[4805]: I1203 15:26:18.081429 4805 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-7hg4c/crc-debug-76hqb" Dec 03 15:26:18 crc kubenswrapper[4805]: I1203 15:26:18.707104 4805 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1a1dd071-ae0f-45bd-a7a4-afa5d281c904" path="/var/lib/kubelet/pods/1a1dd071-ae0f-45bd-a7a4-afa5d281c904/volumes" Dec 03 15:26:22 crc kubenswrapper[4805]: I1203 15:26:22.694919 4805 scope.go:117] "RemoveContainer" containerID="7d1692e3c7d3a2b9dcbcaebd54978af4cba7ae377a17c65268d4048aa72d2c92" Dec 03 15:26:22 crc kubenswrapper[4805]: E1203 15:26:22.695918 4805 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-gskh4_openshift-machine-config-operator(6098937f-e3f6-45e8-a647-4994a79cd711)\"" pod="openshift-machine-config-operator/machine-config-daemon-gskh4" podUID="6098937f-e3f6-45e8-a647-4994a79cd711" Dec 03 15:26:37 crc kubenswrapper[4805]: I1203 15:26:37.695069 4805 scope.go:117] "RemoveContainer" containerID="7d1692e3c7d3a2b9dcbcaebd54978af4cba7ae377a17c65268d4048aa72d2c92" Dec 03 15:26:37 crc kubenswrapper[4805]: E1203 15:26:37.695943 4805 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-gskh4_openshift-machine-config-operator(6098937f-e3f6-45e8-a647-4994a79cd711)\"" pod="openshift-machine-config-operator/machine-config-daemon-gskh4" podUID="6098937f-e3f6-45e8-a647-4994a79cd711" Dec 03 15:26:52 crc kubenswrapper[4805]: I1203 15:26:52.695274 4805 scope.go:117] "RemoveContainer" containerID="7d1692e3c7d3a2b9dcbcaebd54978af4cba7ae377a17c65268d4048aa72d2c92" Dec 03 15:26:52 crc kubenswrapper[4805]: E1203 15:26:52.697671 4805 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-gskh4_openshift-machine-config-operator(6098937f-e3f6-45e8-a647-4994a79cd711)\"" pod="openshift-machine-config-operator/machine-config-daemon-gskh4" podUID="6098937f-e3f6-45e8-a647-4994a79cd711" Dec 03 15:27:03 crc kubenswrapper[4805]: I1203 15:27:03.695009 4805 scope.go:117] "RemoveContainer" containerID="7d1692e3c7d3a2b9dcbcaebd54978af4cba7ae377a17c65268d4048aa72d2c92" Dec 03 15:27:03 crc kubenswrapper[4805]: E1203 15:27:03.695830 4805 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-gskh4_openshift-machine-config-operator(6098937f-e3f6-45e8-a647-4994a79cd711)\"" pod="openshift-machine-config-operator/machine-config-daemon-gskh4" podUID="6098937f-e3f6-45e8-a647-4994a79cd711" Dec 03 15:27:15 crc kubenswrapper[4805]: I1203 15:27:15.695015 4805 scope.go:117] "RemoveContainer" containerID="7d1692e3c7d3a2b9dcbcaebd54978af4cba7ae377a17c65268d4048aa72d2c92" Dec 03 15:27:15 crc kubenswrapper[4805]: E1203 15:27:15.695730 4805 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-gskh4_openshift-machine-config-operator(6098937f-e3f6-45e8-a647-4994a79cd711)\"" pod="openshift-machine-config-operator/machine-config-daemon-gskh4" podUID="6098937f-e3f6-45e8-a647-4994a79cd711" Dec 03 15:27:25 crc kubenswrapper[4805]: I1203 15:27:25.831028 4805 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_alertmanager-metric-storage-0_a0412bf9-ae66-4d99-9bb3-ea2fe4637678/init-config-reloader/0.log" Dec 03 15:27:26 crc kubenswrapper[4805]: I1203 15:27:26.061410 4805 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_alertmanager-metric-storage-0_a0412bf9-ae66-4d99-9bb3-ea2fe4637678/init-config-reloader/0.log" Dec 03 15:27:26 crc kubenswrapper[4805]: I1203 15:27:26.107319 4805 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_alertmanager-metric-storage-0_a0412bf9-ae66-4d99-9bb3-ea2fe4637678/alertmanager/0.log" Dec 03 15:27:26 crc kubenswrapper[4805]: I1203 15:27:26.113276 4805 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_alertmanager-metric-storage-0_a0412bf9-ae66-4d99-9bb3-ea2fe4637678/config-reloader/0.log" Dec 03 15:27:26 crc kubenswrapper[4805]: I1203 15:27:26.248138 4805 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_aodh-0_4297360e-d110-459b-b8b9-75f4a201921e/aodh-api/0.log" Dec 03 15:27:26 crc kubenswrapper[4805]: I1203 15:27:26.274358 4805 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_aodh-0_4297360e-d110-459b-b8b9-75f4a201921e/aodh-evaluator/0.log" Dec 03 15:27:26 crc kubenswrapper[4805]: I1203 15:27:26.327008 4805 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_aodh-0_4297360e-d110-459b-b8b9-75f4a201921e/aodh-listener/0.log" Dec 03 15:27:26 crc kubenswrapper[4805]: I1203 15:27:26.412646 4805 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_aodh-0_4297360e-d110-459b-b8b9-75f4a201921e/aodh-notifier/0.log" Dec 03 15:27:26 crc kubenswrapper[4805]: I1203 15:27:26.463323 4805 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-api-58794647f4-rtjlq_5a55a37b-c678-481e-bf36-a8961f5b26ed/barbican-api/0.log" Dec 03 15:27:26 crc kubenswrapper[4805]: I1203 15:27:26.530419 4805 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-api-58794647f4-rtjlq_5a55a37b-c678-481e-bf36-a8961f5b26ed/barbican-api-log/0.log" Dec 03 15:27:26 crc kubenswrapper[4805]: I1203 15:27:26.630676 4805 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-keystone-listener-8484c694-6frjp_18349f43-5c89-4c11-8ff2-fee30d6cceb7/barbican-keystone-listener/0.log" Dec 03 15:27:26 crc kubenswrapper[4805]: I1203 15:27:26.688510 4805 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-keystone-listener-8484c694-6frjp_18349f43-5c89-4c11-8ff2-fee30d6cceb7/barbican-keystone-listener-log/0.log" Dec 03 15:27:26 crc kubenswrapper[4805]: I1203 15:27:26.821193 4805 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-worker-8559568497-vwxwh_93ad1063-a321-44a3-bd91-c00c7ba1ef96/barbican-worker/0.log" Dec 03 15:27:26 crc kubenswrapper[4805]: I1203 15:27:26.845049 4805 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-worker-8559568497-vwxwh_93ad1063-a321-44a3-bd91-c00c7ba1ef96/barbican-worker-log/0.log" Dec 03 15:27:26 crc kubenswrapper[4805]: I1203 15:27:26.980906 4805 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_bootstrap-edpm-deployment-openstack-edpm-ipam-brwqg_7dde2b60-d63b-4989-b76d-ec802af8f951/bootstrap-edpm-deployment-openstack-edpm-ipam/0.log" Dec 03 15:27:27 crc kubenswrapper[4805]: I1203 15:27:27.150961 4805 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_df32824d-ca3b-4ee5-88a2-1e4591fafc04/ceilometer-central-agent/0.log" Dec 03 15:27:27 crc kubenswrapper[4805]: I1203 15:27:27.177890 4805 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_df32824d-ca3b-4ee5-88a2-1e4591fafc04/ceilometer-notification-agent/0.log" Dec 03 15:27:27 crc kubenswrapper[4805]: I1203 15:27:27.200433 4805 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_df32824d-ca3b-4ee5-88a2-1e4591fafc04/proxy-httpd/0.log" Dec 03 15:27:27 crc kubenswrapper[4805]: I1203 15:27:27.695244 4805 scope.go:117] "RemoveContainer" containerID="7d1692e3c7d3a2b9dcbcaebd54978af4cba7ae377a17c65268d4048aa72d2c92" Dec 03 15:27:27 crc kubenswrapper[4805]: E1203 15:27:27.695552 4805 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-gskh4_openshift-machine-config-operator(6098937f-e3f6-45e8-a647-4994a79cd711)\"" pod="openshift-machine-config-operator/machine-config-daemon-gskh4" podUID="6098937f-e3f6-45e8-a647-4994a79cd711" Dec 03 15:27:28 crc kubenswrapper[4805]: I1203 15:27:28.079134 4805 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_df32824d-ca3b-4ee5-88a2-1e4591fafc04/sg-core/0.log" Dec 03 15:27:28 crc kubenswrapper[4805]: I1203 15:27:28.102872 4805 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-api-0_7b5b9f5a-5af6-41d7-be26-3f4c091098b7/cinder-api-log/0.log" Dec 03 15:27:28 crc kubenswrapper[4805]: I1203 15:27:28.184933 4805 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-api-0_7b5b9f5a-5af6-41d7-be26-3f4c091098b7/cinder-api/0.log" Dec 03 15:27:28 crc kubenswrapper[4805]: I1203 15:27:28.347658 4805 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-scheduler-0_6d34a555-7003-4796-a15a-85a71c489bc9/probe/0.log" Dec 03 15:27:28 crc kubenswrapper[4805]: I1203 15:27:28.430546 4805 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-scheduler-0_6d34a555-7003-4796-a15a-85a71c489bc9/cinder-scheduler/0.log" Dec 03 15:27:28 crc kubenswrapper[4805]: I1203 15:27:28.529610 4805 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_configure-network-edpm-deployment-openstack-edpm-ipam-6wncz_546a9824-86ca-46ed-b6fa-3384de230115/configure-network-edpm-deployment-openstack-edpm-ipam/0.log" Dec 03 15:27:28 crc kubenswrapper[4805]: I1203 15:27:28.669342 4805 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_configure-os-edpm-deployment-openstack-edpm-ipam-5bpz7_2aa97cdd-3fbb-4e80-b82a-fb688c8c92d0/configure-os-edpm-deployment-openstack-edpm-ipam/0.log" Dec 03 15:27:28 crc kubenswrapper[4805]: I1203 15:27:28.739215 4805 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_dnsmasq-dns-bb85b8995-trmpv_7d390cdb-18d5-49b3-8b24-1c2203442042/init/0.log" Dec 03 15:27:28 crc kubenswrapper[4805]: I1203 15:27:28.958220 4805 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_dnsmasq-dns-bb85b8995-trmpv_7d390cdb-18d5-49b3-8b24-1c2203442042/init/0.log" Dec 03 15:27:28 crc kubenswrapper[4805]: I1203 15:27:28.971643 4805 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_dnsmasq-dns-bb85b8995-trmpv_7d390cdb-18d5-49b3-8b24-1c2203442042/dnsmasq-dns/0.log" Dec 03 15:27:28 crc kubenswrapper[4805]: I1203 15:27:28.986904 4805 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_download-cache-edpm-deployment-openstack-edpm-ipam-nvbww_a6f87a69-c7bb-44d8-9f65-82b1b4f8885a/download-cache-edpm-deployment-openstack-edpm-ipam/0.log" Dec 03 15:27:29 crc kubenswrapper[4805]: I1203 15:27:29.707029 4805 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-external-api-0_6a4ca0f4-cff3-4c57-ab1b-3a01d7d6576c/glance-httpd/0.log" Dec 03 15:27:29 crc kubenswrapper[4805]: I1203 15:27:29.720415 4805 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-internal-api-0_cbb9e604-033b-4586-b0fd-7300a5e30897/glance-httpd/0.log" Dec 03 15:27:29 crc kubenswrapper[4805]: I1203 15:27:29.725830 4805 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-internal-api-0_cbb9e604-033b-4586-b0fd-7300a5e30897/glance-log/0.log" Dec 03 15:27:29 crc kubenswrapper[4805]: I1203 15:27:29.755460 4805 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-external-api-0_6a4ca0f4-cff3-4c57-ab1b-3a01d7d6576c/glance-log/0.log" Dec 03 15:27:30 crc kubenswrapper[4805]: I1203 15:27:30.547971 4805 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_heat-api-84746cbddc-kc7hm_078bbd61-70d1-4e2c-ab8c-eeba4127f662/heat-api/0.log" Dec 03 15:27:30 crc kubenswrapper[4805]: I1203 15:27:30.607370 4805 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_heat-engine-76dcfc959c-swfs4_854cde2d-c692-4ab4-9172-02ad6a2b2f6a/heat-engine/0.log" Dec 03 15:27:30 crc kubenswrapper[4805]: I1203 15:27:30.684186 4805 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_install-certs-edpm-deployment-openstack-edpm-ipam-tw4dn_9cd62272-1077-4e13-8e99-c158fbb8ba1a/install-certs-edpm-deployment-openstack-edpm-ipam/0.log" Dec 03 15:27:30 crc kubenswrapper[4805]: I1203 15:27:30.839359 4805 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_heat-cfnapi-68589fbcdd-snszz_cde129c0-8757-4996-8974-c553a5a48f17/heat-cfnapi/0.log" Dec 03 15:27:30 crc kubenswrapper[4805]: I1203 15:27:30.874100 4805 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_install-os-edpm-deployment-openstack-edpm-ipam-v79c2_b96a6fd3-b6a0-4580-9504-9ac4237ad37e/install-os-edpm-deployment-openstack-edpm-ipam/0.log" Dec 03 15:27:31 crc kubenswrapper[4805]: I1203 15:27:31.094018 4805 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_keystone-585fb5cffc-n26gq_3e723db7-3ea3-4360-b5b8-49f1e0524f8b/keystone-api/0.log" Dec 03 15:27:31 crc kubenswrapper[4805]: I1203 15:27:31.103066 4805 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_keystone-cron-29412901-wgktp_ae49c111-f5d4-4884-857c-d03296d16312/keystone-cron/0.log" Dec 03 15:27:31 crc kubenswrapper[4805]: I1203 15:27:31.241216 4805 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_kube-state-metrics-0_91f65152-9858-4a6f-ba45-9e0556490322/kube-state-metrics/0.log" Dec 03 15:27:31 crc kubenswrapper[4805]: I1203 15:27:31.366137 4805 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_libvirt-edpm-deployment-openstack-edpm-ipam-lmlms_1776968b-6613-4b93-bb53-a91c3b6c3d2c/libvirt-edpm-deployment-openstack-edpm-ipam/0.log" Dec 03 15:27:31 crc kubenswrapper[4805]: I1203 15:27:31.663753 4805 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-9764b468f-7sr7n_31537e49-7f76-4c8e-8c14-3fe55f9615b9/neutron-httpd/0.log" Dec 03 15:27:31 crc kubenswrapper[4805]: I1203 15:27:31.695310 4805 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-9764b468f-7sr7n_31537e49-7f76-4c8e-8c14-3fe55f9615b9/neutron-api/0.log" Dec 03 15:27:31 crc kubenswrapper[4805]: I1203 15:27:31.833098 4805 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-metadata-edpm-deployment-openstack-edpm-ipam-s4fzz_779c86b3-bd17-4893-b976-e668607675f2/neutron-metadata-edpm-deployment-openstack-edpm-ipam/0.log" Dec 03 15:27:32 crc kubenswrapper[4805]: I1203 15:27:32.122332 4805 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-api-0_65386594-9cac-40ef-b3a8-2f3ee225bfb6/nova-api-log/0.log" Dec 03 15:27:32 crc kubenswrapper[4805]: I1203 15:27:32.444316 4805 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell0-conductor-0_bc143d9b-fc34-4d12-a500-67de79d3c71a/nova-cell0-conductor-conductor/0.log" Dec 03 15:27:32 crc kubenswrapper[4805]: I1203 15:27:32.598325 4805 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell1-conductor-0_b3635de5-a6a4-465c-8793-bce58d357da6/nova-cell1-conductor-conductor/0.log" Dec 03 15:27:32 crc kubenswrapper[4805]: I1203 15:27:32.658852 4805 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-api-0_65386594-9cac-40ef-b3a8-2f3ee225bfb6/nova-api-api/0.log" Dec 03 15:27:32 crc kubenswrapper[4805]: I1203 15:27:32.810607 4805 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell1-novncproxy-0_d386b865-1a97-468b-ba0e-7733abb94034/nova-cell1-novncproxy-novncproxy/0.log" Dec 03 15:27:32 crc kubenswrapper[4805]: I1203 15:27:32.911591 4805 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-edpm-deployment-openstack-edpm-ipam-75h75_7e18e1db-7529-412b-bf01-17cb78e689b9/nova-edpm-deployment-openstack-edpm-ipam/0.log" Dec 03 15:27:33 crc kubenswrapper[4805]: I1203 15:27:33.221656 4805 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-metadata-0_9acb704f-0e77-458e-b31e-7b32cf8619e5/nova-metadata-log/0.log" Dec 03 15:27:33 crc kubenswrapper[4805]: I1203 15:27:33.550756 4805 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-scheduler-0_e30e8472-484d-47a9-bb8f-839d41ec04be/nova-scheduler-scheduler/0.log" Dec 03 15:27:33 crc kubenswrapper[4805]: I1203 15:27:33.746861 4805 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-cell1-galera-0_c13cbc51-432a-419a-9469-37fd31a82dd9/mysql-bootstrap/0.log" Dec 03 15:27:33 crc kubenswrapper[4805]: I1203 15:27:33.885432 4805 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-cell1-galera-0_c13cbc51-432a-419a-9469-37fd31a82dd9/galera/0.log" Dec 03 15:27:33 crc kubenswrapper[4805]: I1203 15:27:33.904031 4805 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-cell1-galera-0_c13cbc51-432a-419a-9469-37fd31a82dd9/mysql-bootstrap/0.log" Dec 03 15:27:34 crc kubenswrapper[4805]: I1203 15:27:34.147251 4805 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-galera-0_5b4bb902-dba6-4ec6-97bf-532d996e918d/mysql-bootstrap/0.log" Dec 03 15:27:34 crc kubenswrapper[4805]: I1203 15:27:34.378292 4805 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-galera-0_5b4bb902-dba6-4ec6-97bf-532d996e918d/mysql-bootstrap/0.log" Dec 03 15:27:34 crc kubenswrapper[4805]: I1203 15:27:34.509566 4805 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-galera-0_5b4bb902-dba6-4ec6-97bf-532d996e918d/galera/0.log" Dec 03 15:27:34 crc kubenswrapper[4805]: I1203 15:27:34.671074 4805 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstackclient_49d35952-8526-46a9-80f5-b0fe3a359abd/openstackclient/0.log" Dec 03 15:27:34 crc kubenswrapper[4805]: I1203 15:27:34.787641 4805 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-metrics-7kmnp_76002c62-88cb-4f74-816a-3b2be4f8ba4c/openstack-network-exporter/0.log" Dec 03 15:27:35 crc kubenswrapper[4805]: I1203 15:27:35.060999 4805 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-dqwfk_4335461c-d070-4c88-9d40-980109c57b4c/ovsdb-server-init/0.log" Dec 03 15:27:35 crc kubenswrapper[4805]: I1203 15:27:35.099736 4805 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-metadata-0_9acb704f-0e77-458e-b31e-7b32cf8619e5/nova-metadata-metadata/0.log" Dec 03 15:27:35 crc kubenswrapper[4805]: I1203 15:27:35.213930 4805 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-dqwfk_4335461c-d070-4c88-9d40-980109c57b4c/ovsdb-server-init/0.log" Dec 03 15:27:35 crc kubenswrapper[4805]: I1203 15:27:35.217589 4805 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-dqwfk_4335461c-d070-4c88-9d40-980109c57b4c/ovs-vswitchd/0.log" Dec 03 15:27:35 crc kubenswrapper[4805]: I1203 15:27:35.408315 4805 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-dqwfk_4335461c-d070-4c88-9d40-980109c57b4c/ovsdb-server/0.log" Dec 03 15:27:35 crc kubenswrapper[4805]: I1203 15:27:35.536105 4805 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-zhnwt_139abefe-12d4-4d8f-bdc6-43e71c818a65/ovn-controller/0.log" Dec 03 15:27:35 crc kubenswrapper[4805]: I1203 15:27:35.740321 4805 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-edpm-deployment-openstack-edpm-ipam-hbx5q_af960cbb-259d-409e-9140-2d66af349200/ovn-edpm-deployment-openstack-edpm-ipam/0.log" Dec 03 15:27:35 crc kubenswrapper[4805]: I1203 15:27:35.807983 4805 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-northd-0_8e6ca40c-f7f2-445e-8e22-71f8d4d4693b/openstack-network-exporter/0.log" Dec 03 15:27:35 crc kubenswrapper[4805]: I1203 15:27:35.836368 4805 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-northd-0_8e6ca40c-f7f2-445e-8e22-71f8d4d4693b/ovn-northd/0.log" Dec 03 15:27:36 crc kubenswrapper[4805]: I1203 15:27:36.025682 4805 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-nb-0_00ac9536-1677-4a3a-be82-7c8d8df91656/openstack-network-exporter/0.log" Dec 03 15:27:36 crc kubenswrapper[4805]: I1203 15:27:36.063737 4805 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-nb-0_00ac9536-1677-4a3a-be82-7c8d8df91656/ovsdbserver-nb/0.log" Dec 03 15:27:36 crc kubenswrapper[4805]: I1203 15:27:36.332199 4805 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-sb-0_17f337d8-be50-4c6c-8364-208b394be6d3/openstack-network-exporter/0.log" Dec 03 15:27:36 crc kubenswrapper[4805]: I1203 15:27:36.428379 4805 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-sb-0_17f337d8-be50-4c6c-8364-208b394be6d3/ovsdbserver-sb/0.log" Dec 03 15:27:36 crc kubenswrapper[4805]: I1203 15:27:36.472571 4805 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_placement-98cb4d7c6-cbjgt_d099a9d3-b941-499a-829f-7590e1fe7b0a/placement-api/0.log" Dec 03 15:27:36 crc kubenswrapper[4805]: I1203 15:27:36.651617 4805 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_prometheus-metric-storage-0_2cf69241-d762-41ec-9512-84643c4d9dc1/init-config-reloader/0.log" Dec 03 15:27:36 crc kubenswrapper[4805]: I1203 15:27:36.721056 4805 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_placement-98cb4d7c6-cbjgt_d099a9d3-b941-499a-829f-7590e1fe7b0a/placement-log/0.log" Dec 03 15:27:36 crc kubenswrapper[4805]: I1203 15:27:36.901649 4805 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_prometheus-metric-storage-0_2cf69241-d762-41ec-9512-84643c4d9dc1/init-config-reloader/0.log" Dec 03 15:27:36 crc kubenswrapper[4805]: I1203 15:27:36.919575 4805 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_prometheus-metric-storage-0_2cf69241-d762-41ec-9512-84643c4d9dc1/config-reloader/0.log" Dec 03 15:27:36 crc kubenswrapper[4805]: I1203 15:27:36.957656 4805 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_prometheus-metric-storage-0_2cf69241-d762-41ec-9512-84643c4d9dc1/prometheus/0.log" Dec 03 15:27:37 crc kubenswrapper[4805]: I1203 15:27:37.013432 4805 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_prometheus-metric-storage-0_2cf69241-d762-41ec-9512-84643c4d9dc1/thanos-sidecar/0.log" Dec 03 15:27:37 crc kubenswrapper[4805]: I1203 15:27:37.199505 4805 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-cell1-server-0_fff4a752-b61c-4c51-bc0b-428c9affbb00/setup-container/0.log" Dec 03 15:27:37 crc kubenswrapper[4805]: I1203 15:27:37.671623 4805 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-cell1-server-0_fff4a752-b61c-4c51-bc0b-428c9affbb00/rabbitmq/0.log" Dec 03 15:27:37 crc kubenswrapper[4805]: I1203 15:27:37.779007 4805 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-server-0_0664c16b-c578-4685-b1ba-98c5d5f3fe07/setup-container/0.log" Dec 03 15:27:37 crc kubenswrapper[4805]: I1203 15:27:37.824321 4805 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-cell1-server-0_fff4a752-b61c-4c51-bc0b-428c9affbb00/setup-container/0.log" Dec 03 15:27:38 crc kubenswrapper[4805]: I1203 15:27:38.119283 4805 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-server-0_0664c16b-c578-4685-b1ba-98c5d5f3fe07/setup-container/0.log" Dec 03 15:27:38 crc kubenswrapper[4805]: I1203 15:27:38.129861 4805 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_reboot-os-edpm-deployment-openstack-edpm-ipam-k5scz_ba862dc9-b0f4-4ba3-888a-5eae1601dfdb/reboot-os-edpm-deployment-openstack-edpm-ipam/0.log" Dec 03 15:27:38 crc kubenswrapper[4805]: I1203 15:27:38.143326 4805 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-server-0_0664c16b-c578-4685-b1ba-98c5d5f3fe07/rabbitmq/0.log" Dec 03 15:27:38 crc kubenswrapper[4805]: I1203 15:27:38.404334 4805 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_redhat-edpm-deployment-openstack-edpm-ipam-ltzhs_30551b1d-dce7-4c0f-85c9-300f32bf0fc5/redhat-edpm-deployment-openstack-edpm-ipam/0.log" Dec 03 15:27:38 crc kubenswrapper[4805]: I1203 15:27:38.511815 4805 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_repo-setup-edpm-deployment-openstack-edpm-ipam-lghw6_a63daecb-effd-412b-9aeb-9400cd81d763/repo-setup-edpm-deployment-openstack-edpm-ipam/0.log" Dec 03 15:27:38 crc kubenswrapper[4805]: I1203 15:27:38.687192 4805 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_run-os-edpm-deployment-openstack-edpm-ipam-lp9h9_3183d9eb-f2b5-4ac6-8cb1-917cff8a8ded/run-os-edpm-deployment-openstack-edpm-ipam/0.log" Dec 03 15:27:38 crc kubenswrapper[4805]: I1203 15:27:38.815522 4805 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ssh-known-hosts-edpm-deployment-hzqxj_5221f418-44bd-4db1-96c8-788e0df137c7/ssh-known-hosts-edpm-deployment/0.log" Dec 03 15:27:39 crc kubenswrapper[4805]: I1203 15:27:39.086869 4805 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-proxy-78bdd85b87-5tf8n_0fdb9aee-e509-4e04-88b7-61fc216cfc34/proxy-server/0.log" Dec 03 15:27:39 crc kubenswrapper[4805]: I1203 15:27:39.213279 4805 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-proxy-78bdd85b87-5tf8n_0fdb9aee-e509-4e04-88b7-61fc216cfc34/proxy-httpd/0.log" Dec 03 15:27:39 crc kubenswrapper[4805]: I1203 15:27:39.232080 4805 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-ring-rebalance-h55gd_acfd02d5-fb0a-4427-95e0-4c9f956ea137/swift-ring-rebalance/0.log" Dec 03 15:27:39 crc kubenswrapper[4805]: I1203 15:27:39.405495 4805 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_4fbc8d7f-d47f-4d91-915b-504820aac08a/account-auditor/0.log" Dec 03 15:27:39 crc kubenswrapper[4805]: I1203 15:27:39.426823 4805 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_4fbc8d7f-d47f-4d91-915b-504820aac08a/account-reaper/0.log" Dec 03 15:27:39 crc kubenswrapper[4805]: I1203 15:27:39.505307 4805 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_4fbc8d7f-d47f-4d91-915b-504820aac08a/account-replicator/0.log" Dec 03 15:27:39 crc kubenswrapper[4805]: I1203 15:27:39.620487 4805 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_4fbc8d7f-d47f-4d91-915b-504820aac08a/container-auditor/0.log" Dec 03 15:27:39 crc kubenswrapper[4805]: I1203 15:27:39.699585 4805 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_4fbc8d7f-d47f-4d91-915b-504820aac08a/account-server/0.log" Dec 03 15:27:39 crc kubenswrapper[4805]: I1203 15:27:39.719754 4805 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_4fbc8d7f-d47f-4d91-915b-504820aac08a/container-replicator/0.log" Dec 03 15:27:39 crc kubenswrapper[4805]: I1203 15:27:39.757454 4805 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_4fbc8d7f-d47f-4d91-915b-504820aac08a/container-server/0.log" Dec 03 15:27:39 crc kubenswrapper[4805]: I1203 15:27:39.837614 4805 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_4fbc8d7f-d47f-4d91-915b-504820aac08a/container-updater/0.log" Dec 03 15:27:39 crc kubenswrapper[4805]: I1203 15:27:39.964896 4805 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_4fbc8d7f-d47f-4d91-915b-504820aac08a/object-expirer/0.log" Dec 03 15:27:40 crc kubenswrapper[4805]: I1203 15:27:40.040761 4805 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_4fbc8d7f-d47f-4d91-915b-504820aac08a/object-auditor/0.log" Dec 03 15:27:40 crc kubenswrapper[4805]: I1203 15:27:40.078155 4805 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_4fbc8d7f-d47f-4d91-915b-504820aac08a/object-replicator/0.log" Dec 03 15:27:40 crc kubenswrapper[4805]: I1203 15:27:40.079148 4805 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_4fbc8d7f-d47f-4d91-915b-504820aac08a/object-server/0.log" Dec 03 15:27:40 crc kubenswrapper[4805]: I1203 15:27:40.170356 4805 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_4fbc8d7f-d47f-4d91-915b-504820aac08a/object-updater/0.log" Dec 03 15:27:40 crc kubenswrapper[4805]: I1203 15:27:40.329282 4805 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_4fbc8d7f-d47f-4d91-915b-504820aac08a/rsync/0.log" Dec 03 15:27:40 crc kubenswrapper[4805]: I1203 15:27:40.332814 4805 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_4fbc8d7f-d47f-4d91-915b-504820aac08a/swift-recon-cron/0.log" Dec 03 15:27:40 crc kubenswrapper[4805]: I1203 15:27:40.463350 4805 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_telemetry-edpm-deployment-openstack-edpm-ipam-ctxx4_4c03480a-7dce-48dd-9e5d-1aea5c540a72/telemetry-edpm-deployment-openstack-edpm-ipam/0.log" Dec 03 15:27:40 crc kubenswrapper[4805]: I1203 15:27:40.581250 4805 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_validate-network-edpm-deployment-openstack-edpm-ipam-7px67_dd64758f-9fe5-4f00-8f60-4bba03a0f98a/validate-network-edpm-deployment-openstack-edpm-ipam/0.log" Dec 03 15:27:42 crc kubenswrapper[4805]: I1203 15:27:42.695155 4805 scope.go:117] "RemoveContainer" containerID="7d1692e3c7d3a2b9dcbcaebd54978af4cba7ae377a17c65268d4048aa72d2c92" Dec 03 15:27:42 crc kubenswrapper[4805]: E1203 15:27:42.701193 4805 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-gskh4_openshift-machine-config-operator(6098937f-e3f6-45e8-a647-4994a79cd711)\"" pod="openshift-machine-config-operator/machine-config-daemon-gskh4" podUID="6098937f-e3f6-45e8-a647-4994a79cd711" Dec 03 15:27:51 crc kubenswrapper[4805]: I1203 15:27:51.568478 4805 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_memcached-0_3c84e84f-43fa-4163-81b2-489ec6a0d834/memcached/0.log" Dec 03 15:27:54 crc kubenswrapper[4805]: I1203 15:27:54.695145 4805 scope.go:117] "RemoveContainer" containerID="7d1692e3c7d3a2b9dcbcaebd54978af4cba7ae377a17c65268d4048aa72d2c92" Dec 03 15:27:54 crc kubenswrapper[4805]: E1203 15:27:54.697145 4805 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-gskh4_openshift-machine-config-operator(6098937f-e3f6-45e8-a647-4994a79cd711)\"" pod="openshift-machine-config-operator/machine-config-daemon-gskh4" podUID="6098937f-e3f6-45e8-a647-4994a79cd711" Dec 03 15:28:06 crc kubenswrapper[4805]: I1203 15:28:06.695006 4805 scope.go:117] "RemoveContainer" containerID="7d1692e3c7d3a2b9dcbcaebd54978af4cba7ae377a17c65268d4048aa72d2c92" Dec 03 15:28:06 crc kubenswrapper[4805]: E1203 15:28:06.696096 4805 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-gskh4_openshift-machine-config-operator(6098937f-e3f6-45e8-a647-4994a79cd711)\"" pod="openshift-machine-config-operator/machine-config-daemon-gskh4" podUID="6098937f-e3f6-45e8-a647-4994a79cd711" Dec 03 15:28:10 crc kubenswrapper[4805]: I1203 15:28:10.826161 4805 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_01b9b0248b373c1370a002b5a413a9a23eeeadf6c66c17ba9e1f19741esbt54_7811f073-78a5-49a2-8b6e-336505b43756/util/0.log" Dec 03 15:28:10 crc kubenswrapper[4805]: I1203 15:28:10.908018 4805 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_01b9b0248b373c1370a002b5a413a9a23eeeadf6c66c17ba9e1f19741esbt54_7811f073-78a5-49a2-8b6e-336505b43756/util/0.log" Dec 03 15:28:11 crc kubenswrapper[4805]: I1203 15:28:11.064286 4805 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_01b9b0248b373c1370a002b5a413a9a23eeeadf6c66c17ba9e1f19741esbt54_7811f073-78a5-49a2-8b6e-336505b43756/pull/0.log" Dec 03 15:28:11 crc kubenswrapper[4805]: I1203 15:28:11.068903 4805 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_01b9b0248b373c1370a002b5a413a9a23eeeadf6c66c17ba9e1f19741esbt54_7811f073-78a5-49a2-8b6e-336505b43756/pull/0.log" Dec 03 15:28:11 crc kubenswrapper[4805]: I1203 15:28:11.243421 4805 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_01b9b0248b373c1370a002b5a413a9a23eeeadf6c66c17ba9e1f19741esbt54_7811f073-78a5-49a2-8b6e-336505b43756/pull/0.log" Dec 03 15:28:11 crc kubenswrapper[4805]: I1203 15:28:11.265742 4805 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_01b9b0248b373c1370a002b5a413a9a23eeeadf6c66c17ba9e1f19741esbt54_7811f073-78a5-49a2-8b6e-336505b43756/util/0.log" Dec 03 15:28:11 crc kubenswrapper[4805]: I1203 15:28:11.275633 4805 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_01b9b0248b373c1370a002b5a413a9a23eeeadf6c66c17ba9e1f19741esbt54_7811f073-78a5-49a2-8b6e-336505b43756/extract/0.log" Dec 03 15:28:11 crc kubenswrapper[4805]: I1203 15:28:11.445078 4805 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_barbican-operator-controller-manager-7d9dfd778-jc4dw_cbf3813c-fb86-49d2-8fac-25487ff08601/kube-rbac-proxy/0.log" Dec 03 15:28:11 crc kubenswrapper[4805]: I1203 15:28:11.530696 4805 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_barbican-operator-controller-manager-7d9dfd778-jc4dw_cbf3813c-fb86-49d2-8fac-25487ff08601/manager/0.log" Dec 03 15:28:11 crc kubenswrapper[4805]: I1203 15:28:11.566152 4805 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_cinder-operator-controller-manager-859b6ccc6-2nzgx_f9b6b508-1cf6-40aa-829e-b9cf264921e5/kube-rbac-proxy/0.log" Dec 03 15:28:11 crc kubenswrapper[4805]: I1203 15:28:11.702885 4805 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_cinder-operator-controller-manager-859b6ccc6-2nzgx_f9b6b508-1cf6-40aa-829e-b9cf264921e5/manager/0.log" Dec 03 15:28:11 crc kubenswrapper[4805]: I1203 15:28:11.743606 4805 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_designate-operator-controller-manager-78b4bc895b-w56sr_326305da-3798-498c-9b47-0aeb71bd2205/kube-rbac-proxy/0.log" Dec 03 15:28:11 crc kubenswrapper[4805]: I1203 15:28:11.767409 4805 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_designate-operator-controller-manager-78b4bc895b-w56sr_326305da-3798-498c-9b47-0aeb71bd2205/manager/0.log" Dec 03 15:28:11 crc kubenswrapper[4805]: I1203 15:28:11.951663 4805 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_glance-operator-controller-manager-77987cd8cd-pt9fw_5615ce63-1ec1-4a77-8e44-49e6d114eec6/kube-rbac-proxy/0.log" Dec 03 15:28:12 crc kubenswrapper[4805]: I1203 15:28:12.058613 4805 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_glance-operator-controller-manager-77987cd8cd-pt9fw_5615ce63-1ec1-4a77-8e44-49e6d114eec6/manager/0.log" Dec 03 15:28:12 crc kubenswrapper[4805]: I1203 15:28:12.155527 4805 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_heat-operator-controller-manager-5f64f6f8bb-86sbc_a748cfe4-2e54-4a7b-a353-e1dfea6c14b3/kube-rbac-proxy/0.log" Dec 03 15:28:12 crc kubenswrapper[4805]: I1203 15:28:12.272457 4805 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_horizon-operator-controller-manager-68c6d99b8f-kxbr8_b29e4120-0cbc-4de5-a4b6-97b97761f49a/kube-rbac-proxy/0.log" Dec 03 15:28:12 crc kubenswrapper[4805]: I1203 15:28:12.273162 4805 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_heat-operator-controller-manager-5f64f6f8bb-86sbc_a748cfe4-2e54-4a7b-a353-e1dfea6c14b3/manager/0.log" Dec 03 15:28:12 crc kubenswrapper[4805]: I1203 15:28:12.369681 4805 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_horizon-operator-controller-manager-68c6d99b8f-kxbr8_b29e4120-0cbc-4de5-a4b6-97b97761f49a/manager/0.log" Dec 03 15:28:12 crc kubenswrapper[4805]: I1203 15:28:12.494951 4805 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_infra-operator-controller-manager-57548d458d-77t9s_f21790d8-af47-4be7-8758-f038bda76908/kube-rbac-proxy/0.log" Dec 03 15:28:12 crc kubenswrapper[4805]: I1203 15:28:12.686321 4805 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_infra-operator-controller-manager-57548d458d-77t9s_f21790d8-af47-4be7-8758-f038bda76908/manager/0.log" Dec 03 15:28:13 crc kubenswrapper[4805]: I1203 15:28:13.298510 4805 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ironic-operator-controller-manager-6c548fd776-mz4pw_1ad4fa11-2f11-4a06-9944-bba8c7445e84/kube-rbac-proxy/0.log" Dec 03 15:28:13 crc kubenswrapper[4805]: I1203 15:28:13.309147 4805 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_keystone-operator-controller-manager-7765d96ddf-qzzz9_ffa72db0-588a-4a78-a1cd-8710f5ce4f41/kube-rbac-proxy/0.log" Dec 03 15:28:13 crc kubenswrapper[4805]: I1203 15:28:13.339301 4805 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ironic-operator-controller-manager-6c548fd776-mz4pw_1ad4fa11-2f11-4a06-9944-bba8c7445e84/manager/0.log" Dec 03 15:28:13 crc kubenswrapper[4805]: I1203 15:28:13.548884 4805 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_manila-operator-controller-manager-7c79b5df47-d2thc_967f26e7-3027-4e23-9f02-d111d714a396/kube-rbac-proxy/0.log" Dec 03 15:28:13 crc kubenswrapper[4805]: I1203 15:28:13.562680 4805 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_keystone-operator-controller-manager-7765d96ddf-qzzz9_ffa72db0-588a-4a78-a1cd-8710f5ce4f41/manager/0.log" Dec 03 15:28:13 crc kubenswrapper[4805]: I1203 15:28:13.597868 4805 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_manila-operator-controller-manager-7c79b5df47-d2thc_967f26e7-3027-4e23-9f02-d111d714a396/manager/0.log" Dec 03 15:28:13 crc kubenswrapper[4805]: I1203 15:28:13.758850 4805 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_mariadb-operator-controller-manager-56bbcc9d85-n6m2d_2d72009f-b9ef-4c9d-b224-5b133a3cd93a/kube-rbac-proxy/0.log" Dec 03 15:28:13 crc kubenswrapper[4805]: I1203 15:28:13.809681 4805 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_mariadb-operator-controller-manager-56bbcc9d85-n6m2d_2d72009f-b9ef-4c9d-b224-5b133a3cd93a/manager/0.log" Dec 03 15:28:13 crc kubenswrapper[4805]: I1203 15:28:13.983529 4805 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_neutron-operator-controller-manager-5fdfd5b6b5-4mlhk_58aa635e-157c-4ae2-ad0c-568c144638d1/kube-rbac-proxy/0.log" Dec 03 15:28:14 crc kubenswrapper[4805]: I1203 15:28:14.002747 4805 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_neutron-operator-controller-manager-5fdfd5b6b5-4mlhk_58aa635e-157c-4ae2-ad0c-568c144638d1/manager/0.log" Dec 03 15:28:14 crc kubenswrapper[4805]: I1203 15:28:14.091502 4805 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_nova-operator-controller-manager-697bc559fc-5rl85_05506111-ddb3-4ab9-93a0-ad2bedff67a4/kube-rbac-proxy/0.log" Dec 03 15:28:14 crc kubenswrapper[4805]: I1203 15:28:14.248497 4805 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_nova-operator-controller-manager-697bc559fc-5rl85_05506111-ddb3-4ab9-93a0-ad2bedff67a4/manager/0.log" Dec 03 15:28:14 crc kubenswrapper[4805]: I1203 15:28:14.282499 4805 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_octavia-operator-controller-manager-998648c74-8mbdb_0f040ce3-0a83-426b-8b9f-b63183d16b4a/kube-rbac-proxy/0.log" Dec 03 15:28:14 crc kubenswrapper[4805]: I1203 15:28:14.332306 4805 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_octavia-operator-controller-manager-998648c74-8mbdb_0f040ce3-0a83-426b-8b9f-b63183d16b4a/manager/0.log" Dec 03 15:28:14 crc kubenswrapper[4805]: I1203 15:28:14.436917 4805 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-baremetal-operator-controller-manager-64bc77cfd4x9mdd_b9d0830e-0f5d-4639-ab7d-92dcf07a29bc/kube-rbac-proxy/0.log" Dec 03 15:28:15 crc kubenswrapper[4805]: I1203 15:28:15.125744 4805 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-baremetal-operator-controller-manager-64bc77cfd4x9mdd_b9d0830e-0f5d-4639-ab7d-92dcf07a29bc/manager/0.log" Dec 03 15:28:15 crc kubenswrapper[4805]: I1203 15:28:15.359118 4805 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-controller-operator-796df97b8b-rmv2c_9da93959-4eef-466e-aab4-1a36202b9970/operator/0.log" Dec 03 15:28:15 crc kubenswrapper[4805]: I1203 15:28:15.433490 4805 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-index-p6vl8_f6e38dfc-ef6b-4813-8bd4-027bed8dfb81/registry-server/0.log" Dec 03 15:28:15 crc kubenswrapper[4805]: I1203 15:28:15.655030 4805 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ovn-operator-controller-manager-b6456fdb6-djzcm_42b37265-f5bd-4ac4-9b96-d758fbe1b31e/kube-rbac-proxy/0.log" Dec 03 15:28:15 crc kubenswrapper[4805]: I1203 15:28:15.687957 4805 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ovn-operator-controller-manager-b6456fdb6-djzcm_42b37265-f5bd-4ac4-9b96-d758fbe1b31e/manager/0.log" Dec 03 15:28:15 crc kubenswrapper[4805]: I1203 15:28:15.775776 4805 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_placement-operator-controller-manager-78f8948974-b8l4l_f853eae6-829b-42f9-9198-3d7dcabd405f/kube-rbac-proxy/0.log" Dec 03 15:28:15 crc kubenswrapper[4805]: I1203 15:28:15.933707 4805 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_placement-operator-controller-manager-78f8948974-b8l4l_f853eae6-829b-42f9-9198-3d7dcabd405f/manager/0.log" Dec 03 15:28:15 crc kubenswrapper[4805]: I1203 15:28:15.959136 4805 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_rabbitmq-cluster-operator-manager-668c99d594-5q6j2_0aa860af-81d3-44ba-9020-0ffef21cefb2/operator/0.log" Dec 03 15:28:16 crc kubenswrapper[4805]: I1203 15:28:16.124998 4805 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_swift-operator-controller-manager-5f8c65bbfc-qrvpw_b51f4884-a980-4643-9deb-8668aead3876/kube-rbac-proxy/0.log" Dec 03 15:28:16 crc kubenswrapper[4805]: I1203 15:28:16.266137 4805 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_swift-operator-controller-manager-5f8c65bbfc-qrvpw_b51f4884-a980-4643-9deb-8668aead3876/manager/0.log" Dec 03 15:28:16 crc kubenswrapper[4805]: I1203 15:28:16.295115 4805 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_telemetry-operator-controller-manager-cc9d8f87b-qfmmx_8cbfcd07-e136-4bfd-b3c6-6615c911d440/kube-rbac-proxy/0.log" Dec 03 15:28:16 crc kubenswrapper[4805]: I1203 15:28:16.528067 4805 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_test-operator-controller-manager-5854674fcc-d6pvv_ac4ba0f5-635d-4a72-a133-9b4aaf88f097/manager/0.log" Dec 03 15:28:16 crc kubenswrapper[4805]: I1203 15:28:16.583129 4805 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_test-operator-controller-manager-5854674fcc-d6pvv_ac4ba0f5-635d-4a72-a133-9b4aaf88f097/kube-rbac-proxy/0.log" Dec 03 15:28:16 crc kubenswrapper[4805]: I1203 15:28:16.670058 4805 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_telemetry-operator-controller-manager-cc9d8f87b-qfmmx_8cbfcd07-e136-4bfd-b3c6-6615c911d440/manager/0.log" Dec 03 15:28:16 crc kubenswrapper[4805]: I1203 15:28:16.739363 4805 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_watcher-operator-controller-manager-769dc69bc-fl6mm_30776d04-200b-4c10-b48d-f51a926e469b/kube-rbac-proxy/0.log" Dec 03 15:28:16 crc kubenswrapper[4805]: I1203 15:28:16.824982 4805 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_watcher-operator-controller-manager-769dc69bc-fl6mm_30776d04-200b-4c10-b48d-f51a926e469b/manager/0.log" Dec 03 15:28:16 crc kubenswrapper[4805]: I1203 15:28:16.862042 4805 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-controller-manager-cc68d789f-pkvr8_41c4618c-1f15-4ceb-825b-546e828e399a/manager/0.log" Dec 03 15:28:20 crc kubenswrapper[4805]: I1203 15:28:20.703775 4805 scope.go:117] "RemoveContainer" containerID="7d1692e3c7d3a2b9dcbcaebd54978af4cba7ae377a17c65268d4048aa72d2c92" Dec 03 15:28:20 crc kubenswrapper[4805]: E1203 15:28:20.704947 4805 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-gskh4_openshift-machine-config-operator(6098937f-e3f6-45e8-a647-4994a79cd711)\"" pod="openshift-machine-config-operator/machine-config-daemon-gskh4" podUID="6098937f-e3f6-45e8-a647-4994a79cd711" Dec 03 15:28:33 crc kubenswrapper[4805]: I1203 15:28:33.694563 4805 scope.go:117] "RemoveContainer" containerID="7d1692e3c7d3a2b9dcbcaebd54978af4cba7ae377a17c65268d4048aa72d2c92" Dec 03 15:28:33 crc kubenswrapper[4805]: E1203 15:28:33.695404 4805 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-gskh4_openshift-machine-config-operator(6098937f-e3f6-45e8-a647-4994a79cd711)\"" pod="openshift-machine-config-operator/machine-config-daemon-gskh4" podUID="6098937f-e3f6-45e8-a647-4994a79cd711" Dec 03 15:28:38 crc kubenswrapper[4805]: I1203 15:28:38.284703 4805 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_control-plane-machine-set-operator-78cbb6b69f-25khc_9bd2f61c-2b92-488f-b632-1e627b48a518/control-plane-machine-set-operator/0.log" Dec 03 15:28:38 crc kubenswrapper[4805]: I1203 15:28:38.492978 4805 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_machine-api-operator-5694c8668f-jvspg_dfd20bfc-5c11-4287-a4b3-df9e0ca4dcf1/kube-rbac-proxy/0.log" Dec 03 15:28:38 crc kubenswrapper[4805]: I1203 15:28:38.557147 4805 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_machine-api-operator-5694c8668f-jvspg_dfd20bfc-5c11-4287-a4b3-df9e0ca4dcf1/machine-api-operator/0.log" Dec 03 15:28:47 crc kubenswrapper[4805]: I1203 15:28:47.695280 4805 scope.go:117] "RemoveContainer" containerID="7d1692e3c7d3a2b9dcbcaebd54978af4cba7ae377a17c65268d4048aa72d2c92" Dec 03 15:28:47 crc kubenswrapper[4805]: E1203 15:28:47.696256 4805 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-gskh4_openshift-machine-config-operator(6098937f-e3f6-45e8-a647-4994a79cd711)\"" pod="openshift-machine-config-operator/machine-config-daemon-gskh4" podUID="6098937f-e3f6-45e8-a647-4994a79cd711" Dec 03 15:28:54 crc kubenswrapper[4805]: I1203 15:28:54.141947 4805 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-5b446d88c5-vj7mm_c7baa569-d061-4226-a134-2b54cf463f90/cert-manager-controller/0.log" Dec 03 15:28:54 crc kubenswrapper[4805]: I1203 15:28:54.382735 4805 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-cainjector-7f985d654d-ll5ml_70f90fc2-9a18-4737-8025-850782848e31/cert-manager-cainjector/0.log" Dec 03 15:28:54 crc kubenswrapper[4805]: I1203 15:28:54.431146 4805 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-webhook-5655c58dd6-mt9mr_55307bc8-2e9c-4ea9-b007-c40c05403707/cert-manager-webhook/0.log" Dec 03 15:28:58 crc kubenswrapper[4805]: I1203 15:28:58.695622 4805 scope.go:117] "RemoveContainer" containerID="7d1692e3c7d3a2b9dcbcaebd54978af4cba7ae377a17c65268d4048aa72d2c92" Dec 03 15:28:58 crc kubenswrapper[4805]: E1203 15:28:58.696611 4805 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-gskh4_openshift-machine-config-operator(6098937f-e3f6-45e8-a647-4994a79cd711)\"" pod="openshift-machine-config-operator/machine-config-daemon-gskh4" podUID="6098937f-e3f6-45e8-a647-4994a79cd711" Dec 03 15:29:10 crc kubenswrapper[4805]: I1203 15:29:10.226583 4805 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-console-plugin-7fbb5f6569-mbhzq_5d4ab6ea-87ca-46dc-a9ad-f315faf0a9e2/nmstate-console-plugin/0.log" Dec 03 15:29:10 crc kubenswrapper[4805]: I1203 15:29:10.313815 4805 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-handler-xqhj5_4f623c1b-94c0-43cd-ae9f-56597f4542b4/nmstate-handler/0.log" Dec 03 15:29:10 crc kubenswrapper[4805]: I1203 15:29:10.457092 4805 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-metrics-7f946cbc9-6r8jm_9d39f3b1-337a-4aeb-b1cf-9ca9290f7791/kube-rbac-proxy/0.log" Dec 03 15:29:10 crc kubenswrapper[4805]: I1203 15:29:10.498047 4805 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-metrics-7f946cbc9-6r8jm_9d39f3b1-337a-4aeb-b1cf-9ca9290f7791/nmstate-metrics/0.log" Dec 03 15:29:10 crc kubenswrapper[4805]: I1203 15:29:10.672918 4805 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-operator-5b5b58f5c8-tstgp_7f35b1a1-05c4-453a-9101-d43cf4fa3739/nmstate-operator/0.log" Dec 03 15:29:10 crc kubenswrapper[4805]: I1203 15:29:10.760495 4805 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-webhook-5f6d4c5ccb-2zszs_9654448a-2df5-4cfc-b6e5-401ad15bc52e/nmstate-webhook/0.log" Dec 03 15:29:11 crc kubenswrapper[4805]: I1203 15:29:11.695071 4805 scope.go:117] "RemoveContainer" containerID="7d1692e3c7d3a2b9dcbcaebd54978af4cba7ae377a17c65268d4048aa72d2c92" Dec 03 15:29:11 crc kubenswrapper[4805]: E1203 15:29:11.695457 4805 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-gskh4_openshift-machine-config-operator(6098937f-e3f6-45e8-a647-4994a79cd711)\"" pod="openshift-machine-config-operator/machine-config-daemon-gskh4" podUID="6098937f-e3f6-45e8-a647-4994a79cd711" Dec 03 15:29:25 crc kubenswrapper[4805]: I1203 15:29:25.936394 4805 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_controller-f8648f98b-n75gf_a3361a03-7368-4bb6-b209-485190285e82/kube-rbac-proxy/0.log" Dec 03 15:29:26 crc kubenswrapper[4805]: I1203 15:29:26.048493 4805 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_controller-f8648f98b-n75gf_a3361a03-7368-4bb6-b209-485190285e82/controller/0.log" Dec 03 15:29:26 crc kubenswrapper[4805]: I1203 15:29:26.137038 4805 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-jsrqx_2ba805a2-6c47-4649-b8dc-eb92fb614437/cp-frr-files/0.log" Dec 03 15:29:26 crc kubenswrapper[4805]: I1203 15:29:26.293022 4805 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-jsrqx_2ba805a2-6c47-4649-b8dc-eb92fb614437/cp-reloader/0.log" Dec 03 15:29:26 crc kubenswrapper[4805]: I1203 15:29:26.338665 4805 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-jsrqx_2ba805a2-6c47-4649-b8dc-eb92fb614437/cp-frr-files/0.log" Dec 03 15:29:26 crc kubenswrapper[4805]: I1203 15:29:26.350231 4805 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-jsrqx_2ba805a2-6c47-4649-b8dc-eb92fb614437/cp-metrics/0.log" Dec 03 15:29:26 crc kubenswrapper[4805]: I1203 15:29:26.398716 4805 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-jsrqx_2ba805a2-6c47-4649-b8dc-eb92fb614437/cp-reloader/0.log" Dec 03 15:29:26 crc kubenswrapper[4805]: I1203 15:29:26.574000 4805 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-jsrqx_2ba805a2-6c47-4649-b8dc-eb92fb614437/cp-reloader/0.log" Dec 03 15:29:26 crc kubenswrapper[4805]: I1203 15:29:26.631978 4805 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-jsrqx_2ba805a2-6c47-4649-b8dc-eb92fb614437/cp-metrics/0.log" Dec 03 15:29:26 crc kubenswrapper[4805]: I1203 15:29:26.645331 4805 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-jsrqx_2ba805a2-6c47-4649-b8dc-eb92fb614437/cp-metrics/0.log" Dec 03 15:29:26 crc kubenswrapper[4805]: I1203 15:29:26.667032 4805 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-jsrqx_2ba805a2-6c47-4649-b8dc-eb92fb614437/cp-frr-files/0.log" Dec 03 15:29:26 crc kubenswrapper[4805]: I1203 15:29:26.695214 4805 scope.go:117] "RemoveContainer" containerID="7d1692e3c7d3a2b9dcbcaebd54978af4cba7ae377a17c65268d4048aa72d2c92" Dec 03 15:29:26 crc kubenswrapper[4805]: E1203 15:29:26.695544 4805 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-gskh4_openshift-machine-config-operator(6098937f-e3f6-45e8-a647-4994a79cd711)\"" pod="openshift-machine-config-operator/machine-config-daemon-gskh4" podUID="6098937f-e3f6-45e8-a647-4994a79cd711" Dec 03 15:29:26 crc kubenswrapper[4805]: I1203 15:29:26.824135 4805 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-jsrqx_2ba805a2-6c47-4649-b8dc-eb92fb614437/cp-frr-files/0.log" Dec 03 15:29:26 crc kubenswrapper[4805]: I1203 15:29:26.825228 4805 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-jsrqx_2ba805a2-6c47-4649-b8dc-eb92fb614437/cp-reloader/0.log" Dec 03 15:29:26 crc kubenswrapper[4805]: I1203 15:29:26.869929 4805 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-jsrqx_2ba805a2-6c47-4649-b8dc-eb92fb614437/cp-metrics/0.log" Dec 03 15:29:26 crc kubenswrapper[4805]: I1203 15:29:26.878893 4805 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-jsrqx_2ba805a2-6c47-4649-b8dc-eb92fb614437/controller/0.log" Dec 03 15:29:27 crc kubenswrapper[4805]: I1203 15:29:27.020941 4805 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-jsrqx_2ba805a2-6c47-4649-b8dc-eb92fb614437/frr-metrics/0.log" Dec 03 15:29:27 crc kubenswrapper[4805]: I1203 15:29:27.042199 4805 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-jsrqx_2ba805a2-6c47-4649-b8dc-eb92fb614437/kube-rbac-proxy/0.log" Dec 03 15:29:27 crc kubenswrapper[4805]: I1203 15:29:27.103799 4805 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-jsrqx_2ba805a2-6c47-4649-b8dc-eb92fb614437/kube-rbac-proxy-frr/0.log" Dec 03 15:29:27 crc kubenswrapper[4805]: I1203 15:29:27.255665 4805 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-2qfdn"] Dec 03 15:29:27 crc kubenswrapper[4805]: E1203 15:29:27.256131 4805 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1a1dd071-ae0f-45bd-a7a4-afa5d281c904" containerName="container-00" Dec 03 15:29:27 crc kubenswrapper[4805]: I1203 15:29:27.256148 4805 state_mem.go:107] "Deleted CPUSet assignment" podUID="1a1dd071-ae0f-45bd-a7a4-afa5d281c904" containerName="container-00" Dec 03 15:29:27 crc kubenswrapper[4805]: I1203 15:29:27.256333 4805 memory_manager.go:354] "RemoveStaleState removing state" podUID="1a1dd071-ae0f-45bd-a7a4-afa5d281c904" containerName="container-00" Dec 03 15:29:27 crc kubenswrapper[4805]: I1203 15:29:27.257802 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-2qfdn" Dec 03 15:29:27 crc kubenswrapper[4805]: I1203 15:29:27.262705 4805 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-jsrqx_2ba805a2-6c47-4649-b8dc-eb92fb614437/reloader/0.log" Dec 03 15:29:27 crc kubenswrapper[4805]: I1203 15:29:27.289064 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-2qfdn"] Dec 03 15:29:27 crc kubenswrapper[4805]: I1203 15:29:27.302016 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5l5fg\" (UniqueName: \"kubernetes.io/projected/5b943fe1-4f5c-44b0-8d46-f10eabb00c6b-kube-api-access-5l5fg\") pod \"redhat-marketplace-2qfdn\" (UID: \"5b943fe1-4f5c-44b0-8d46-f10eabb00c6b\") " pod="openshift-marketplace/redhat-marketplace-2qfdn" Dec 03 15:29:27 crc kubenswrapper[4805]: I1203 15:29:27.302172 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5b943fe1-4f5c-44b0-8d46-f10eabb00c6b-utilities\") pod \"redhat-marketplace-2qfdn\" (UID: \"5b943fe1-4f5c-44b0-8d46-f10eabb00c6b\") " pod="openshift-marketplace/redhat-marketplace-2qfdn" Dec 03 15:29:27 crc kubenswrapper[4805]: I1203 15:29:27.302922 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5b943fe1-4f5c-44b0-8d46-f10eabb00c6b-catalog-content\") pod \"redhat-marketplace-2qfdn\" (UID: \"5b943fe1-4f5c-44b0-8d46-f10eabb00c6b\") " pod="openshift-marketplace/redhat-marketplace-2qfdn" Dec 03 15:29:27 crc kubenswrapper[4805]: I1203 15:29:27.387711 4805 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-webhook-server-7fcb986d4-lmm42_24cc3c3f-7f72-49c4-b531-093bf5851177/frr-k8s-webhook-server/0.log" Dec 03 15:29:27 crc kubenswrapper[4805]: I1203 15:29:27.404433 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5b943fe1-4f5c-44b0-8d46-f10eabb00c6b-utilities\") pod \"redhat-marketplace-2qfdn\" (UID: \"5b943fe1-4f5c-44b0-8d46-f10eabb00c6b\") " pod="openshift-marketplace/redhat-marketplace-2qfdn" Dec 03 15:29:27 crc kubenswrapper[4805]: I1203 15:29:27.404505 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5b943fe1-4f5c-44b0-8d46-f10eabb00c6b-catalog-content\") pod \"redhat-marketplace-2qfdn\" (UID: \"5b943fe1-4f5c-44b0-8d46-f10eabb00c6b\") " pod="openshift-marketplace/redhat-marketplace-2qfdn" Dec 03 15:29:27 crc kubenswrapper[4805]: I1203 15:29:27.404661 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5l5fg\" (UniqueName: \"kubernetes.io/projected/5b943fe1-4f5c-44b0-8d46-f10eabb00c6b-kube-api-access-5l5fg\") pod \"redhat-marketplace-2qfdn\" (UID: \"5b943fe1-4f5c-44b0-8d46-f10eabb00c6b\") " pod="openshift-marketplace/redhat-marketplace-2qfdn" Dec 03 15:29:27 crc kubenswrapper[4805]: I1203 15:29:27.405067 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5b943fe1-4f5c-44b0-8d46-f10eabb00c6b-utilities\") pod \"redhat-marketplace-2qfdn\" (UID: \"5b943fe1-4f5c-44b0-8d46-f10eabb00c6b\") " pod="openshift-marketplace/redhat-marketplace-2qfdn" Dec 03 15:29:27 crc kubenswrapper[4805]: I1203 15:29:27.405109 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5b943fe1-4f5c-44b0-8d46-f10eabb00c6b-catalog-content\") pod \"redhat-marketplace-2qfdn\" (UID: \"5b943fe1-4f5c-44b0-8d46-f10eabb00c6b\") " pod="openshift-marketplace/redhat-marketplace-2qfdn" Dec 03 15:29:27 crc kubenswrapper[4805]: I1203 15:29:27.449664 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5l5fg\" (UniqueName: \"kubernetes.io/projected/5b943fe1-4f5c-44b0-8d46-f10eabb00c6b-kube-api-access-5l5fg\") pod \"redhat-marketplace-2qfdn\" (UID: \"5b943fe1-4f5c-44b0-8d46-f10eabb00c6b\") " pod="openshift-marketplace/redhat-marketplace-2qfdn" Dec 03 15:29:27 crc kubenswrapper[4805]: I1203 15:29:27.588994 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-2qfdn" Dec 03 15:29:27 crc kubenswrapper[4805]: I1203 15:29:27.690711 4805 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_metallb-operator-controller-manager-76979bd97f-kgnj5_33acb320-cfe5-4d01-9543-f56ec8c33633/manager/0.log" Dec 03 15:29:27 crc kubenswrapper[4805]: I1203 15:29:27.971232 4805 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_metallb-operator-webhook-server-5587f68f8b-4vcdf_f2e7792c-c1c1-41b0-8e90-fe082ba1c366/webhook-server/0.log" Dec 03 15:29:28 crc kubenswrapper[4805]: I1203 15:29:28.145776 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-2qfdn"] Dec 03 15:29:28 crc kubenswrapper[4805]: I1203 15:29:28.203014 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-2qfdn" event={"ID":"5b943fe1-4f5c-44b0-8d46-f10eabb00c6b","Type":"ContainerStarted","Data":"469248e9142a9ac5b391e1b6e18ddd60db310a8f6461cdd63b493f17fd8e9618"} Dec 03 15:29:28 crc kubenswrapper[4805]: I1203 15:29:28.218058 4805 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_speaker-jpk6x_06e98f2f-39d1-4899-9a18-7cc27e006de9/kube-rbac-proxy/0.log" Dec 03 15:29:28 crc kubenswrapper[4805]: I1203 15:29:28.739648 4805 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_speaker-jpk6x_06e98f2f-39d1-4899-9a18-7cc27e006de9/speaker/0.log" Dec 03 15:29:28 crc kubenswrapper[4805]: I1203 15:29:28.972708 4805 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-jsrqx_2ba805a2-6c47-4649-b8dc-eb92fb614437/frr/0.log" Dec 03 15:29:29 crc kubenswrapper[4805]: I1203 15:29:29.214400 4805 generic.go:334] "Generic (PLEG): container finished" podID="5b943fe1-4f5c-44b0-8d46-f10eabb00c6b" containerID="c1564dbd24b9e6e4cbfc0773945e4db307f64f564a10e74de69ee31efdc200d9" exitCode=0 Dec 03 15:29:29 crc kubenswrapper[4805]: I1203 15:29:29.214439 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-2qfdn" event={"ID":"5b943fe1-4f5c-44b0-8d46-f10eabb00c6b","Type":"ContainerDied","Data":"c1564dbd24b9e6e4cbfc0773945e4db307f64f564a10e74de69ee31efdc200d9"} Dec 03 15:29:29 crc kubenswrapper[4805]: I1203 15:29:29.216286 4805 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 03 15:29:30 crc kubenswrapper[4805]: I1203 15:29:30.243340 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-2qfdn" event={"ID":"5b943fe1-4f5c-44b0-8d46-f10eabb00c6b","Type":"ContainerStarted","Data":"64987d72a7e3ba553970ebd65b3868b3372c2b5f14c061cb1086e686b0ea3c6a"} Dec 03 15:29:31 crc kubenswrapper[4805]: I1203 15:29:31.257138 4805 generic.go:334] "Generic (PLEG): container finished" podID="5b943fe1-4f5c-44b0-8d46-f10eabb00c6b" containerID="64987d72a7e3ba553970ebd65b3868b3372c2b5f14c061cb1086e686b0ea3c6a" exitCode=0 Dec 03 15:29:31 crc kubenswrapper[4805]: I1203 15:29:31.257198 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-2qfdn" event={"ID":"5b943fe1-4f5c-44b0-8d46-f10eabb00c6b","Type":"ContainerDied","Data":"64987d72a7e3ba553970ebd65b3868b3372c2b5f14c061cb1086e686b0ea3c6a"} Dec 03 15:29:32 crc kubenswrapper[4805]: I1203 15:29:32.269150 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-2qfdn" event={"ID":"5b943fe1-4f5c-44b0-8d46-f10eabb00c6b","Type":"ContainerStarted","Data":"c2779a61a9e3026c56d2a4c193ead5867c3effcb015ebf3ba3979e2861eac0fb"} Dec 03 15:29:32 crc kubenswrapper[4805]: I1203 15:29:32.292183 4805 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-2qfdn" podStartSLOduration=2.882360224 podStartE2EDuration="5.292157909s" podCreationTimestamp="2025-12-03 15:29:27 +0000 UTC" firstStartedPulling="2025-12-03 15:29:29.216088065 +0000 UTC m=+4798.879004988" lastFinishedPulling="2025-12-03 15:29:31.62588575 +0000 UTC m=+4801.288802673" observedRunningTime="2025-12-03 15:29:32.286143717 +0000 UTC m=+4801.949060640" watchObservedRunningTime="2025-12-03 15:29:32.292157909 +0000 UTC m=+4801.955074832" Dec 03 15:29:37 crc kubenswrapper[4805]: I1203 15:29:37.590152 4805 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-2qfdn" Dec 03 15:29:37 crc kubenswrapper[4805]: I1203 15:29:37.590535 4805 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-2qfdn" Dec 03 15:29:37 crc kubenswrapper[4805]: I1203 15:29:37.640602 4805 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-2qfdn" Dec 03 15:29:38 crc kubenswrapper[4805]: I1203 15:29:38.384630 4805 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-2qfdn" Dec 03 15:29:38 crc kubenswrapper[4805]: I1203 15:29:38.439155 4805 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-2qfdn"] Dec 03 15:29:38 crc kubenswrapper[4805]: I1203 15:29:38.696077 4805 scope.go:117] "RemoveContainer" containerID="7d1692e3c7d3a2b9dcbcaebd54978af4cba7ae377a17c65268d4048aa72d2c92" Dec 03 15:29:38 crc kubenswrapper[4805]: E1203 15:29:38.698231 4805 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-gskh4_openshift-machine-config-operator(6098937f-e3f6-45e8-a647-4994a79cd711)\"" pod="openshift-machine-config-operator/machine-config-daemon-gskh4" podUID="6098937f-e3f6-45e8-a647-4994a79cd711" Dec 03 15:29:40 crc kubenswrapper[4805]: I1203 15:29:40.345576 4805 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-2qfdn" podUID="5b943fe1-4f5c-44b0-8d46-f10eabb00c6b" containerName="registry-server" containerID="cri-o://c2779a61a9e3026c56d2a4c193ead5867c3effcb015ebf3ba3979e2861eac0fb" gracePeriod=2 Dec 03 15:29:40 crc kubenswrapper[4805]: I1203 15:29:40.912264 4805 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-2qfdn" Dec 03 15:29:40 crc kubenswrapper[4805]: I1203 15:29:40.981970 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5b943fe1-4f5c-44b0-8d46-f10eabb00c6b-utilities\") pod \"5b943fe1-4f5c-44b0-8d46-f10eabb00c6b\" (UID: \"5b943fe1-4f5c-44b0-8d46-f10eabb00c6b\") " Dec 03 15:29:40 crc kubenswrapper[4805]: I1203 15:29:40.982245 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5l5fg\" (UniqueName: \"kubernetes.io/projected/5b943fe1-4f5c-44b0-8d46-f10eabb00c6b-kube-api-access-5l5fg\") pod \"5b943fe1-4f5c-44b0-8d46-f10eabb00c6b\" (UID: \"5b943fe1-4f5c-44b0-8d46-f10eabb00c6b\") " Dec 03 15:29:40 crc kubenswrapper[4805]: I1203 15:29:40.982320 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5b943fe1-4f5c-44b0-8d46-f10eabb00c6b-catalog-content\") pod \"5b943fe1-4f5c-44b0-8d46-f10eabb00c6b\" (UID: \"5b943fe1-4f5c-44b0-8d46-f10eabb00c6b\") " Dec 03 15:29:40 crc kubenswrapper[4805]: I1203 15:29:40.983518 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5b943fe1-4f5c-44b0-8d46-f10eabb00c6b-utilities" (OuterVolumeSpecName: "utilities") pod "5b943fe1-4f5c-44b0-8d46-f10eabb00c6b" (UID: "5b943fe1-4f5c-44b0-8d46-f10eabb00c6b"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 15:29:40 crc kubenswrapper[4805]: I1203 15:29:40.987681 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5b943fe1-4f5c-44b0-8d46-f10eabb00c6b-kube-api-access-5l5fg" (OuterVolumeSpecName: "kube-api-access-5l5fg") pod "5b943fe1-4f5c-44b0-8d46-f10eabb00c6b" (UID: "5b943fe1-4f5c-44b0-8d46-f10eabb00c6b"). InnerVolumeSpecName "kube-api-access-5l5fg". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 15:29:41 crc kubenswrapper[4805]: I1203 15:29:41.009731 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5b943fe1-4f5c-44b0-8d46-f10eabb00c6b-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "5b943fe1-4f5c-44b0-8d46-f10eabb00c6b" (UID: "5b943fe1-4f5c-44b0-8d46-f10eabb00c6b"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 15:29:41 crc kubenswrapper[4805]: I1203 15:29:41.084865 4805 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5l5fg\" (UniqueName: \"kubernetes.io/projected/5b943fe1-4f5c-44b0-8d46-f10eabb00c6b-kube-api-access-5l5fg\") on node \"crc\" DevicePath \"\"" Dec 03 15:29:41 crc kubenswrapper[4805]: I1203 15:29:41.084914 4805 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5b943fe1-4f5c-44b0-8d46-f10eabb00c6b-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 03 15:29:41 crc kubenswrapper[4805]: I1203 15:29:41.084927 4805 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5b943fe1-4f5c-44b0-8d46-f10eabb00c6b-utilities\") on node \"crc\" DevicePath \"\"" Dec 03 15:29:41 crc kubenswrapper[4805]: I1203 15:29:41.177043 4805 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fsxzq6_abc1f29a-e8a4-425d-bbbd-17c106a39c5c/util/0.log" Dec 03 15:29:41 crc kubenswrapper[4805]: I1203 15:29:41.358195 4805 generic.go:334] "Generic (PLEG): container finished" podID="5b943fe1-4f5c-44b0-8d46-f10eabb00c6b" containerID="c2779a61a9e3026c56d2a4c193ead5867c3effcb015ebf3ba3979e2861eac0fb" exitCode=0 Dec 03 15:29:41 crc kubenswrapper[4805]: I1203 15:29:41.358251 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-2qfdn" event={"ID":"5b943fe1-4f5c-44b0-8d46-f10eabb00c6b","Type":"ContainerDied","Data":"c2779a61a9e3026c56d2a4c193ead5867c3effcb015ebf3ba3979e2861eac0fb"} Dec 03 15:29:41 crc kubenswrapper[4805]: I1203 15:29:41.358265 4805 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-2qfdn" Dec 03 15:29:41 crc kubenswrapper[4805]: I1203 15:29:41.358319 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-2qfdn" event={"ID":"5b943fe1-4f5c-44b0-8d46-f10eabb00c6b","Type":"ContainerDied","Data":"469248e9142a9ac5b391e1b6e18ddd60db310a8f6461cdd63b493f17fd8e9618"} Dec 03 15:29:41 crc kubenswrapper[4805]: I1203 15:29:41.358350 4805 scope.go:117] "RemoveContainer" containerID="c2779a61a9e3026c56d2a4c193ead5867c3effcb015ebf3ba3979e2861eac0fb" Dec 03 15:29:41 crc kubenswrapper[4805]: I1203 15:29:41.398268 4805 scope.go:117] "RemoveContainer" containerID="64987d72a7e3ba553970ebd65b3868b3372c2b5f14c061cb1086e686b0ea3c6a" Dec 03 15:29:41 crc kubenswrapper[4805]: I1203 15:29:41.409140 4805 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-2qfdn"] Dec 03 15:29:41 crc kubenswrapper[4805]: I1203 15:29:41.423304 4805 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-2qfdn"] Dec 03 15:29:41 crc kubenswrapper[4805]: I1203 15:29:41.431052 4805 scope.go:117] "RemoveContainer" containerID="c1564dbd24b9e6e4cbfc0773945e4db307f64f564a10e74de69ee31efdc200d9" Dec 03 15:29:41 crc kubenswrapper[4805]: I1203 15:29:41.470533 4805 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fsxzq6_abc1f29a-e8a4-425d-bbbd-17c106a39c5c/util/0.log" Dec 03 15:29:41 crc kubenswrapper[4805]: I1203 15:29:41.491057 4805 scope.go:117] "RemoveContainer" containerID="c2779a61a9e3026c56d2a4c193ead5867c3effcb015ebf3ba3979e2861eac0fb" Dec 03 15:29:41 crc kubenswrapper[4805]: E1203 15:29:41.493475 4805 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c2779a61a9e3026c56d2a4c193ead5867c3effcb015ebf3ba3979e2861eac0fb\": container with ID starting with c2779a61a9e3026c56d2a4c193ead5867c3effcb015ebf3ba3979e2861eac0fb not found: ID does not exist" containerID="c2779a61a9e3026c56d2a4c193ead5867c3effcb015ebf3ba3979e2861eac0fb" Dec 03 15:29:41 crc kubenswrapper[4805]: I1203 15:29:41.493520 4805 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c2779a61a9e3026c56d2a4c193ead5867c3effcb015ebf3ba3979e2861eac0fb"} err="failed to get container status \"c2779a61a9e3026c56d2a4c193ead5867c3effcb015ebf3ba3979e2861eac0fb\": rpc error: code = NotFound desc = could not find container \"c2779a61a9e3026c56d2a4c193ead5867c3effcb015ebf3ba3979e2861eac0fb\": container with ID starting with c2779a61a9e3026c56d2a4c193ead5867c3effcb015ebf3ba3979e2861eac0fb not found: ID does not exist" Dec 03 15:29:41 crc kubenswrapper[4805]: I1203 15:29:41.493554 4805 scope.go:117] "RemoveContainer" containerID="64987d72a7e3ba553970ebd65b3868b3372c2b5f14c061cb1086e686b0ea3c6a" Dec 03 15:29:41 crc kubenswrapper[4805]: E1203 15:29:41.494117 4805 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"64987d72a7e3ba553970ebd65b3868b3372c2b5f14c061cb1086e686b0ea3c6a\": container with ID starting with 64987d72a7e3ba553970ebd65b3868b3372c2b5f14c061cb1086e686b0ea3c6a not found: ID does not exist" containerID="64987d72a7e3ba553970ebd65b3868b3372c2b5f14c061cb1086e686b0ea3c6a" Dec 03 15:29:41 crc kubenswrapper[4805]: I1203 15:29:41.494197 4805 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"64987d72a7e3ba553970ebd65b3868b3372c2b5f14c061cb1086e686b0ea3c6a"} err="failed to get container status \"64987d72a7e3ba553970ebd65b3868b3372c2b5f14c061cb1086e686b0ea3c6a\": rpc error: code = NotFound desc = could not find container \"64987d72a7e3ba553970ebd65b3868b3372c2b5f14c061cb1086e686b0ea3c6a\": container with ID starting with 64987d72a7e3ba553970ebd65b3868b3372c2b5f14c061cb1086e686b0ea3c6a not found: ID does not exist" Dec 03 15:29:41 crc kubenswrapper[4805]: I1203 15:29:41.494227 4805 scope.go:117] "RemoveContainer" containerID="c1564dbd24b9e6e4cbfc0773945e4db307f64f564a10e74de69ee31efdc200d9" Dec 03 15:29:41 crc kubenswrapper[4805]: I1203 15:29:41.494532 4805 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fsxzq6_abc1f29a-e8a4-425d-bbbd-17c106a39c5c/pull/0.log" Dec 03 15:29:41 crc kubenswrapper[4805]: I1203 15:29:41.494515 4805 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fsxzq6_abc1f29a-e8a4-425d-bbbd-17c106a39c5c/pull/0.log" Dec 03 15:29:41 crc kubenswrapper[4805]: E1203 15:29:41.496280 4805 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c1564dbd24b9e6e4cbfc0773945e4db307f64f564a10e74de69ee31efdc200d9\": container with ID starting with c1564dbd24b9e6e4cbfc0773945e4db307f64f564a10e74de69ee31efdc200d9 not found: ID does not exist" containerID="c1564dbd24b9e6e4cbfc0773945e4db307f64f564a10e74de69ee31efdc200d9" Dec 03 15:29:41 crc kubenswrapper[4805]: I1203 15:29:41.496321 4805 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c1564dbd24b9e6e4cbfc0773945e4db307f64f564a10e74de69ee31efdc200d9"} err="failed to get container status \"c1564dbd24b9e6e4cbfc0773945e4db307f64f564a10e74de69ee31efdc200d9\": rpc error: code = NotFound desc = could not find container \"c1564dbd24b9e6e4cbfc0773945e4db307f64f564a10e74de69ee31efdc200d9\": container with ID starting with c1564dbd24b9e6e4cbfc0773945e4db307f64f564a10e74de69ee31efdc200d9 not found: ID does not exist" Dec 03 15:29:41 crc kubenswrapper[4805]: I1203 15:29:41.721302 4805 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fsxzq6_abc1f29a-e8a4-425d-bbbd-17c106a39c5c/extract/0.log" Dec 03 15:29:41 crc kubenswrapper[4805]: I1203 15:29:41.764785 4805 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fsxzq6_abc1f29a-e8a4-425d-bbbd-17c106a39c5c/pull/0.log" Dec 03 15:29:41 crc kubenswrapper[4805]: I1203 15:29:41.767362 4805 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fsxzq6_abc1f29a-e8a4-425d-bbbd-17c106a39c5c/util/0.log" Dec 03 15:29:41 crc kubenswrapper[4805]: I1203 15:29:41.928181 4805 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210zddkr_95163734-0666-4f76-a68f-722bc3752b38/util/0.log" Dec 03 15:29:42 crc kubenswrapper[4805]: I1203 15:29:42.097954 4805 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210zddkr_95163734-0666-4f76-a68f-722bc3752b38/util/0.log" Dec 03 15:29:42 crc kubenswrapper[4805]: I1203 15:29:42.106133 4805 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210zddkr_95163734-0666-4f76-a68f-722bc3752b38/pull/0.log" Dec 03 15:29:42 crc kubenswrapper[4805]: I1203 15:29:42.135084 4805 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210zddkr_95163734-0666-4f76-a68f-722bc3752b38/pull/0.log" Dec 03 15:29:42 crc kubenswrapper[4805]: I1203 15:29:42.285521 4805 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210zddkr_95163734-0666-4f76-a68f-722bc3752b38/extract/0.log" Dec 03 15:29:42 crc kubenswrapper[4805]: I1203 15:29:42.285575 4805 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210zddkr_95163734-0666-4f76-a68f-722bc3752b38/pull/0.log" Dec 03 15:29:42 crc kubenswrapper[4805]: I1203 15:29:42.287175 4805 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210zddkr_95163734-0666-4f76-a68f-722bc3752b38/util/0.log" Dec 03 15:29:42 crc kubenswrapper[4805]: I1203 15:29:42.456726 4805 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83glh2z_449e7507-4d25-4399-b1bd-b174d65a5041/util/0.log" Dec 03 15:29:42 crc kubenswrapper[4805]: I1203 15:29:42.673986 4805 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83glh2z_449e7507-4d25-4399-b1bd-b174d65a5041/pull/0.log" Dec 03 15:29:42 crc kubenswrapper[4805]: I1203 15:29:42.686462 4805 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83glh2z_449e7507-4d25-4399-b1bd-b174d65a5041/pull/0.log" Dec 03 15:29:42 crc kubenswrapper[4805]: I1203 15:29:42.692476 4805 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83glh2z_449e7507-4d25-4399-b1bd-b174d65a5041/util/0.log" Dec 03 15:29:42 crc kubenswrapper[4805]: I1203 15:29:42.706701 4805 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5b943fe1-4f5c-44b0-8d46-f10eabb00c6b" path="/var/lib/kubelet/pods/5b943fe1-4f5c-44b0-8d46-f10eabb00c6b/volumes" Dec 03 15:29:42 crc kubenswrapper[4805]: I1203 15:29:42.915830 4805 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83glh2z_449e7507-4d25-4399-b1bd-b174d65a5041/util/0.log" Dec 03 15:29:42 crc kubenswrapper[4805]: I1203 15:29:42.918635 4805 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83glh2z_449e7507-4d25-4399-b1bd-b174d65a5041/pull/0.log" Dec 03 15:29:42 crc kubenswrapper[4805]: I1203 15:29:42.927312 4805 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83glh2z_449e7507-4d25-4399-b1bd-b174d65a5041/extract/0.log" Dec 03 15:29:43 crc kubenswrapper[4805]: I1203 15:29:43.087155 4805 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-c4fdw_6fcdd680-c71f-4978-992a-65378f16e0a1/extract-utilities/0.log" Dec 03 15:29:43 crc kubenswrapper[4805]: I1203 15:29:43.307804 4805 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-c4fdw_6fcdd680-c71f-4978-992a-65378f16e0a1/extract-content/0.log" Dec 03 15:29:43 crc kubenswrapper[4805]: I1203 15:29:43.314929 4805 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-c4fdw_6fcdd680-c71f-4978-992a-65378f16e0a1/extract-utilities/0.log" Dec 03 15:29:43 crc kubenswrapper[4805]: I1203 15:29:43.352552 4805 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-c4fdw_6fcdd680-c71f-4978-992a-65378f16e0a1/extract-content/0.log" Dec 03 15:29:43 crc kubenswrapper[4805]: I1203 15:29:43.504429 4805 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-c4fdw_6fcdd680-c71f-4978-992a-65378f16e0a1/extract-content/0.log" Dec 03 15:29:43 crc kubenswrapper[4805]: I1203 15:29:43.586216 4805 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-c4fdw_6fcdd680-c71f-4978-992a-65378f16e0a1/extract-utilities/0.log" Dec 03 15:29:43 crc kubenswrapper[4805]: I1203 15:29:43.746008 4805 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-xfg6c_a93d60c0-f172-49ae-af40-0d5bbda4877f/extract-utilities/0.log" Dec 03 15:29:44 crc kubenswrapper[4805]: I1203 15:29:44.006436 4805 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-xfg6c_a93d60c0-f172-49ae-af40-0d5bbda4877f/extract-utilities/0.log" Dec 03 15:29:44 crc kubenswrapper[4805]: I1203 15:29:44.050058 4805 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-xfg6c_a93d60c0-f172-49ae-af40-0d5bbda4877f/extract-content/0.log" Dec 03 15:29:44 crc kubenswrapper[4805]: I1203 15:29:44.054171 4805 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-xfg6c_a93d60c0-f172-49ae-af40-0d5bbda4877f/extract-content/0.log" Dec 03 15:29:44 crc kubenswrapper[4805]: I1203 15:29:44.146099 4805 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-c4fdw_6fcdd680-c71f-4978-992a-65378f16e0a1/registry-server/0.log" Dec 03 15:29:44 crc kubenswrapper[4805]: I1203 15:29:44.464781 4805 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-xfg6c_a93d60c0-f172-49ae-af40-0d5bbda4877f/extract-utilities/0.log" Dec 03 15:29:44 crc kubenswrapper[4805]: I1203 15:29:44.536709 4805 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-xfg6c_a93d60c0-f172-49ae-af40-0d5bbda4877f/extract-content/0.log" Dec 03 15:29:45 crc kubenswrapper[4805]: I1203 15:29:45.036718 4805 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_marketplace-operator-79b997595-6qgcv_336b0e79-07ec-4899-8cba-a24a65458383/marketplace-operator/0.log" Dec 03 15:29:45 crc kubenswrapper[4805]: I1203 15:29:45.283678 4805 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-xfg6c_a93d60c0-f172-49ae-af40-0d5bbda4877f/registry-server/0.log" Dec 03 15:29:45 crc kubenswrapper[4805]: I1203 15:29:45.785169 4805 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-jftlh_0ab10c8b-6fa6-44a5-87af-211ef363ec7f/extract-utilities/0.log" Dec 03 15:29:46 crc kubenswrapper[4805]: I1203 15:29:46.021001 4805 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-jftlh_0ab10c8b-6fa6-44a5-87af-211ef363ec7f/extract-content/0.log" Dec 03 15:29:46 crc kubenswrapper[4805]: I1203 15:29:46.056390 4805 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-jftlh_0ab10c8b-6fa6-44a5-87af-211ef363ec7f/extract-utilities/0.log" Dec 03 15:29:46 crc kubenswrapper[4805]: I1203 15:29:46.071893 4805 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-jftlh_0ab10c8b-6fa6-44a5-87af-211ef363ec7f/extract-content/0.log" Dec 03 15:29:46 crc kubenswrapper[4805]: I1203 15:29:46.304276 4805 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-jftlh_0ab10c8b-6fa6-44a5-87af-211ef363ec7f/extract-utilities/0.log" Dec 03 15:29:46 crc kubenswrapper[4805]: I1203 15:29:46.337308 4805 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-jftlh_0ab10c8b-6fa6-44a5-87af-211ef363ec7f/extract-content/0.log" Dec 03 15:29:46 crc kubenswrapper[4805]: I1203 15:29:46.353499 4805 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-v8czz_33e30caf-a8b4-4472-9f81-20a67a0ef486/extract-utilities/0.log" Dec 03 15:29:46 crc kubenswrapper[4805]: I1203 15:29:46.479363 4805 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-jftlh_0ab10c8b-6fa6-44a5-87af-211ef363ec7f/registry-server/0.log" Dec 03 15:29:46 crc kubenswrapper[4805]: I1203 15:29:46.516099 4805 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-v8czz_33e30caf-a8b4-4472-9f81-20a67a0ef486/extract-utilities/0.log" Dec 03 15:29:46 crc kubenswrapper[4805]: I1203 15:29:46.589906 4805 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-v8czz_33e30caf-a8b4-4472-9f81-20a67a0ef486/extract-content/0.log" Dec 03 15:29:46 crc kubenswrapper[4805]: I1203 15:29:46.645788 4805 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-v8czz_33e30caf-a8b4-4472-9f81-20a67a0ef486/extract-content/0.log" Dec 03 15:29:47 crc kubenswrapper[4805]: I1203 15:29:47.182504 4805 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-v8czz_33e30caf-a8b4-4472-9f81-20a67a0ef486/extract-content/0.log" Dec 03 15:29:47 crc kubenswrapper[4805]: I1203 15:29:47.274829 4805 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-v8czz_33e30caf-a8b4-4472-9f81-20a67a0ef486/extract-utilities/0.log" Dec 03 15:29:47 crc kubenswrapper[4805]: I1203 15:29:47.906991 4805 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-v8czz_33e30caf-a8b4-4472-9f81-20a67a0ef486/registry-server/0.log" Dec 03 15:29:50 crc kubenswrapper[4805]: I1203 15:29:50.702791 4805 scope.go:117] "RemoveContainer" containerID="7d1692e3c7d3a2b9dcbcaebd54978af4cba7ae377a17c65268d4048aa72d2c92" Dec 03 15:29:50 crc kubenswrapper[4805]: E1203 15:29:50.703387 4805 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-gskh4_openshift-machine-config-operator(6098937f-e3f6-45e8-a647-4994a79cd711)\"" pod="openshift-machine-config-operator/machine-config-daemon-gskh4" podUID="6098937f-e3f6-45e8-a647-4994a79cd711" Dec 03 15:30:00 crc kubenswrapper[4805]: I1203 15:30:00.153564 4805 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29412930-4pn7m"] Dec 03 15:30:00 crc kubenswrapper[4805]: E1203 15:30:00.154557 4805 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5b943fe1-4f5c-44b0-8d46-f10eabb00c6b" containerName="extract-content" Dec 03 15:30:00 crc kubenswrapper[4805]: I1203 15:30:00.154572 4805 state_mem.go:107] "Deleted CPUSet assignment" podUID="5b943fe1-4f5c-44b0-8d46-f10eabb00c6b" containerName="extract-content" Dec 03 15:30:00 crc kubenswrapper[4805]: E1203 15:30:00.154588 4805 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5b943fe1-4f5c-44b0-8d46-f10eabb00c6b" containerName="registry-server" Dec 03 15:30:00 crc kubenswrapper[4805]: I1203 15:30:00.154594 4805 state_mem.go:107] "Deleted CPUSet assignment" podUID="5b943fe1-4f5c-44b0-8d46-f10eabb00c6b" containerName="registry-server" Dec 03 15:30:00 crc kubenswrapper[4805]: E1203 15:30:00.154607 4805 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5b943fe1-4f5c-44b0-8d46-f10eabb00c6b" containerName="extract-utilities" Dec 03 15:30:00 crc kubenswrapper[4805]: I1203 15:30:00.154614 4805 state_mem.go:107] "Deleted CPUSet assignment" podUID="5b943fe1-4f5c-44b0-8d46-f10eabb00c6b" containerName="extract-utilities" Dec 03 15:30:00 crc kubenswrapper[4805]: I1203 15:30:00.154800 4805 memory_manager.go:354] "RemoveStaleState removing state" podUID="5b943fe1-4f5c-44b0-8d46-f10eabb00c6b" containerName="registry-server" Dec 03 15:30:00 crc kubenswrapper[4805]: I1203 15:30:00.155959 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29412930-4pn7m" Dec 03 15:30:00 crc kubenswrapper[4805]: I1203 15:30:00.158036 4805 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Dec 03 15:30:00 crc kubenswrapper[4805]: I1203 15:30:00.159756 4805 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Dec 03 15:30:00 crc kubenswrapper[4805]: I1203 15:30:00.168147 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29412930-4pn7m"] Dec 03 15:30:00 crc kubenswrapper[4805]: I1203 15:30:00.288223 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/0bbf15d5-33ba-4931-90a5-8b4fb7d2aba6-secret-volume\") pod \"collect-profiles-29412930-4pn7m\" (UID: \"0bbf15d5-33ba-4931-90a5-8b4fb7d2aba6\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29412930-4pn7m" Dec 03 15:30:00 crc kubenswrapper[4805]: I1203 15:30:00.288279 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-42mrq\" (UniqueName: \"kubernetes.io/projected/0bbf15d5-33ba-4931-90a5-8b4fb7d2aba6-kube-api-access-42mrq\") pod \"collect-profiles-29412930-4pn7m\" (UID: \"0bbf15d5-33ba-4931-90a5-8b4fb7d2aba6\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29412930-4pn7m" Dec 03 15:30:00 crc kubenswrapper[4805]: I1203 15:30:00.288502 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/0bbf15d5-33ba-4931-90a5-8b4fb7d2aba6-config-volume\") pod \"collect-profiles-29412930-4pn7m\" (UID: \"0bbf15d5-33ba-4931-90a5-8b4fb7d2aba6\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29412930-4pn7m" Dec 03 15:30:00 crc kubenswrapper[4805]: I1203 15:30:00.390223 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/0bbf15d5-33ba-4931-90a5-8b4fb7d2aba6-config-volume\") pod \"collect-profiles-29412930-4pn7m\" (UID: \"0bbf15d5-33ba-4931-90a5-8b4fb7d2aba6\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29412930-4pn7m" Dec 03 15:30:00 crc kubenswrapper[4805]: I1203 15:30:00.390322 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/0bbf15d5-33ba-4931-90a5-8b4fb7d2aba6-secret-volume\") pod \"collect-profiles-29412930-4pn7m\" (UID: \"0bbf15d5-33ba-4931-90a5-8b4fb7d2aba6\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29412930-4pn7m" Dec 03 15:30:00 crc kubenswrapper[4805]: I1203 15:30:00.390355 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-42mrq\" (UniqueName: \"kubernetes.io/projected/0bbf15d5-33ba-4931-90a5-8b4fb7d2aba6-kube-api-access-42mrq\") pod \"collect-profiles-29412930-4pn7m\" (UID: \"0bbf15d5-33ba-4931-90a5-8b4fb7d2aba6\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29412930-4pn7m" Dec 03 15:30:00 crc kubenswrapper[4805]: I1203 15:30:00.391325 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/0bbf15d5-33ba-4931-90a5-8b4fb7d2aba6-config-volume\") pod \"collect-profiles-29412930-4pn7m\" (UID: \"0bbf15d5-33ba-4931-90a5-8b4fb7d2aba6\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29412930-4pn7m" Dec 03 15:30:00 crc kubenswrapper[4805]: I1203 15:30:00.397933 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/0bbf15d5-33ba-4931-90a5-8b4fb7d2aba6-secret-volume\") pod \"collect-profiles-29412930-4pn7m\" (UID: \"0bbf15d5-33ba-4931-90a5-8b4fb7d2aba6\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29412930-4pn7m" Dec 03 15:30:00 crc kubenswrapper[4805]: I1203 15:30:00.413609 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-42mrq\" (UniqueName: \"kubernetes.io/projected/0bbf15d5-33ba-4931-90a5-8b4fb7d2aba6-kube-api-access-42mrq\") pod \"collect-profiles-29412930-4pn7m\" (UID: \"0bbf15d5-33ba-4931-90a5-8b4fb7d2aba6\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29412930-4pn7m" Dec 03 15:30:00 crc kubenswrapper[4805]: I1203 15:30:00.492028 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29412930-4pn7m" Dec 03 15:30:01 crc kubenswrapper[4805]: I1203 15:30:01.039597 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29412930-4pn7m"] Dec 03 15:30:01 crc kubenswrapper[4805]: I1203 15:30:01.605527 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29412930-4pn7m" event={"ID":"0bbf15d5-33ba-4931-90a5-8b4fb7d2aba6","Type":"ContainerStarted","Data":"4be39edebe07370e4016a957eb068c915d74f2a7b9b07f33d5a5be9274a640c6"} Dec 03 15:30:01 crc kubenswrapper[4805]: I1203 15:30:01.605579 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29412930-4pn7m" event={"ID":"0bbf15d5-33ba-4931-90a5-8b4fb7d2aba6","Type":"ContainerStarted","Data":"544ee4d9b632fc6e63c2d812981685ea822e1f994065663f2472a00125214740"} Dec 03 15:30:01 crc kubenswrapper[4805]: I1203 15:30:01.694618 4805 scope.go:117] "RemoveContainer" containerID="7d1692e3c7d3a2b9dcbcaebd54978af4cba7ae377a17c65268d4048aa72d2c92" Dec 03 15:30:01 crc kubenswrapper[4805]: E1203 15:30:01.694992 4805 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-gskh4_openshift-machine-config-operator(6098937f-e3f6-45e8-a647-4994a79cd711)\"" pod="openshift-machine-config-operator/machine-config-daemon-gskh4" podUID="6098937f-e3f6-45e8-a647-4994a79cd711" Dec 03 15:30:02 crc kubenswrapper[4805]: I1203 15:30:02.615982 4805 generic.go:334] "Generic (PLEG): container finished" podID="0bbf15d5-33ba-4931-90a5-8b4fb7d2aba6" containerID="4be39edebe07370e4016a957eb068c915d74f2a7b9b07f33d5a5be9274a640c6" exitCode=0 Dec 03 15:30:02 crc kubenswrapper[4805]: I1203 15:30:02.616088 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29412930-4pn7m" event={"ID":"0bbf15d5-33ba-4931-90a5-8b4fb7d2aba6","Type":"ContainerDied","Data":"4be39edebe07370e4016a957eb068c915d74f2a7b9b07f33d5a5be9274a640c6"} Dec 03 15:30:02 crc kubenswrapper[4805]: I1203 15:30:02.718031 4805 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-operators_obo-prometheus-operator-668cf9dfbb-rtvd2_063df7dd-86e9-4e2b-a354-9baa6e90ba05/prometheus-operator/0.log" Dec 03 15:30:02 crc kubenswrapper[4805]: I1203 15:30:02.972643 4805 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-operators_obo-prometheus-operator-admission-webhook-59d4564ff5-49hbr_94ea81e8-e4cc-4317-a369-b73314c79ab8/prometheus-operator-admission-webhook/0.log" Dec 03 15:30:03 crc kubenswrapper[4805]: I1203 15:30:03.028196 4805 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29412930-4pn7m" Dec 03 15:30:03 crc kubenswrapper[4805]: I1203 15:30:03.049384 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/0bbf15d5-33ba-4931-90a5-8b4fb7d2aba6-config-volume\") pod \"0bbf15d5-33ba-4931-90a5-8b4fb7d2aba6\" (UID: \"0bbf15d5-33ba-4931-90a5-8b4fb7d2aba6\") " Dec 03 15:30:03 crc kubenswrapper[4805]: I1203 15:30:03.049508 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/0bbf15d5-33ba-4931-90a5-8b4fb7d2aba6-secret-volume\") pod \"0bbf15d5-33ba-4931-90a5-8b4fb7d2aba6\" (UID: \"0bbf15d5-33ba-4931-90a5-8b4fb7d2aba6\") " Dec 03 15:30:03 crc kubenswrapper[4805]: I1203 15:30:03.049548 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-42mrq\" (UniqueName: \"kubernetes.io/projected/0bbf15d5-33ba-4931-90a5-8b4fb7d2aba6-kube-api-access-42mrq\") pod \"0bbf15d5-33ba-4931-90a5-8b4fb7d2aba6\" (UID: \"0bbf15d5-33ba-4931-90a5-8b4fb7d2aba6\") " Dec 03 15:30:03 crc kubenswrapper[4805]: I1203 15:30:03.050259 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0bbf15d5-33ba-4931-90a5-8b4fb7d2aba6-config-volume" (OuterVolumeSpecName: "config-volume") pod "0bbf15d5-33ba-4931-90a5-8b4fb7d2aba6" (UID: "0bbf15d5-33ba-4931-90a5-8b4fb7d2aba6"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 15:30:03 crc kubenswrapper[4805]: I1203 15:30:03.051411 4805 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/0bbf15d5-33ba-4931-90a5-8b4fb7d2aba6-config-volume\") on node \"crc\" DevicePath \"\"" Dec 03 15:30:03 crc kubenswrapper[4805]: I1203 15:30:03.059044 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0bbf15d5-33ba-4931-90a5-8b4fb7d2aba6-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "0bbf15d5-33ba-4931-90a5-8b4fb7d2aba6" (UID: "0bbf15d5-33ba-4931-90a5-8b4fb7d2aba6"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 15:30:03 crc kubenswrapper[4805]: I1203 15:30:03.060322 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0bbf15d5-33ba-4931-90a5-8b4fb7d2aba6-kube-api-access-42mrq" (OuterVolumeSpecName: "kube-api-access-42mrq") pod "0bbf15d5-33ba-4931-90a5-8b4fb7d2aba6" (UID: "0bbf15d5-33ba-4931-90a5-8b4fb7d2aba6"). InnerVolumeSpecName "kube-api-access-42mrq". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 15:30:03 crc kubenswrapper[4805]: I1203 15:30:03.083338 4805 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-operators_obo-prometheus-operator-admission-webhook-59d4564ff5-mh8fd_21da77ce-bcf7-4ad0-97f8-a0c9d85d13db/prometheus-operator-admission-webhook/0.log" Dec 03 15:30:03 crc kubenswrapper[4805]: I1203 15:30:03.153045 4805 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/0bbf15d5-33ba-4931-90a5-8b4fb7d2aba6-secret-volume\") on node \"crc\" DevicePath \"\"" Dec 03 15:30:03 crc kubenswrapper[4805]: I1203 15:30:03.153086 4805 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-42mrq\" (UniqueName: \"kubernetes.io/projected/0bbf15d5-33ba-4931-90a5-8b4fb7d2aba6-kube-api-access-42mrq\") on node \"crc\" DevicePath \"\"" Dec 03 15:30:03 crc kubenswrapper[4805]: I1203 15:30:03.329785 4805 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-operators_perses-operator-5446b9c989-22dsf_78288c6e-a51d-4870-8530-54415f661eb8/perses-operator/0.log" Dec 03 15:30:03 crc kubenswrapper[4805]: I1203 15:30:03.367717 4805 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-operators_observability-operator-d8bb48f5d-m6b9n_76c82588-3cbd-4565-abf5-5a291f68160e/operator/0.log" Dec 03 15:30:03 crc kubenswrapper[4805]: I1203 15:30:03.625884 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29412930-4pn7m" event={"ID":"0bbf15d5-33ba-4931-90a5-8b4fb7d2aba6","Type":"ContainerDied","Data":"544ee4d9b632fc6e63c2d812981685ea822e1f994065663f2472a00125214740"} Dec 03 15:30:03 crc kubenswrapper[4805]: I1203 15:30:03.625921 4805 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29412930-4pn7m" Dec 03 15:30:03 crc kubenswrapper[4805]: I1203 15:30:03.625932 4805 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="544ee4d9b632fc6e63c2d812981685ea822e1f994065663f2472a00125214740" Dec 03 15:30:04 crc kubenswrapper[4805]: I1203 15:30:04.102832 4805 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29412885-nz8rt"] Dec 03 15:30:04 crc kubenswrapper[4805]: I1203 15:30:04.116716 4805 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29412885-nz8rt"] Dec 03 15:30:04 crc kubenswrapper[4805]: I1203 15:30:04.707349 4805 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9b9690b7-2a97-4043-a398-071a05fc9aea" path="/var/lib/kubelet/pods/9b9690b7-2a97-4043-a398-071a05fc9aea/volumes" Dec 03 15:30:16 crc kubenswrapper[4805]: I1203 15:30:16.696521 4805 scope.go:117] "RemoveContainer" containerID="7d1692e3c7d3a2b9dcbcaebd54978af4cba7ae377a17c65268d4048aa72d2c92" Dec 03 15:30:16 crc kubenswrapper[4805]: E1203 15:30:16.697380 4805 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-gskh4_openshift-machine-config-operator(6098937f-e3f6-45e8-a647-4994a79cd711)\"" pod="openshift-machine-config-operator/machine-config-daemon-gskh4" podUID="6098937f-e3f6-45e8-a647-4994a79cd711" Dec 03 15:30:29 crc kubenswrapper[4805]: I1203 15:30:29.694202 4805 scope.go:117] "RemoveContainer" containerID="7d1692e3c7d3a2b9dcbcaebd54978af4cba7ae377a17c65268d4048aa72d2c92" Dec 03 15:30:29 crc kubenswrapper[4805]: E1203 15:30:29.694976 4805 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-gskh4_openshift-machine-config-operator(6098937f-e3f6-45e8-a647-4994a79cd711)\"" pod="openshift-machine-config-operator/machine-config-daemon-gskh4" podUID="6098937f-e3f6-45e8-a647-4994a79cd711" Dec 03 15:30:42 crc kubenswrapper[4805]: I1203 15:30:42.698724 4805 scope.go:117] "RemoveContainer" containerID="7d1692e3c7d3a2b9dcbcaebd54978af4cba7ae377a17c65268d4048aa72d2c92" Dec 03 15:30:42 crc kubenswrapper[4805]: E1203 15:30:42.700361 4805 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-gskh4_openshift-machine-config-operator(6098937f-e3f6-45e8-a647-4994a79cd711)\"" pod="openshift-machine-config-operator/machine-config-daemon-gskh4" podUID="6098937f-e3f6-45e8-a647-4994a79cd711" Dec 03 15:30:48 crc kubenswrapper[4805]: I1203 15:30:48.305310 4805 scope.go:117] "RemoveContainer" containerID="52bf529a439b949affea292e2c996d7d0ec64acd9b851e33f7bd9cfde8608c0c" Dec 03 15:30:56 crc kubenswrapper[4805]: I1203 15:30:56.694464 4805 scope.go:117] "RemoveContainer" containerID="7d1692e3c7d3a2b9dcbcaebd54978af4cba7ae377a17c65268d4048aa72d2c92" Dec 03 15:30:57 crc kubenswrapper[4805]: I1203 15:30:57.076783 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-gskh4" event={"ID":"6098937f-e3f6-45e8-a647-4994a79cd711","Type":"ContainerStarted","Data":"8a6f7e8ec0c2240c0eb0a039e259a0f769944a70b22cf0a08378ac2f9c4cfeb8"} Dec 03 15:31:39 crc kubenswrapper[4805]: I1203 15:31:39.510950 4805 generic.go:334] "Generic (PLEG): container finished" podID="6093636d-db50-435e-b563-2ea4aef4e8c9" containerID="cd4b89422b4de3d0665d78ffea319d0d1c3bd669d7c50fb2f68c2ea5780b99c5" exitCode=0 Dec 03 15:31:39 crc kubenswrapper[4805]: I1203 15:31:39.511049 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-7hg4c/must-gather-2mm9s" event={"ID":"6093636d-db50-435e-b563-2ea4aef4e8c9","Type":"ContainerDied","Data":"cd4b89422b4de3d0665d78ffea319d0d1c3bd669d7c50fb2f68c2ea5780b99c5"} Dec 03 15:31:39 crc kubenswrapper[4805]: I1203 15:31:39.512696 4805 scope.go:117] "RemoveContainer" containerID="cd4b89422b4de3d0665d78ffea319d0d1c3bd669d7c50fb2f68c2ea5780b99c5" Dec 03 15:31:39 crc kubenswrapper[4805]: I1203 15:31:39.897040 4805 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-7hg4c_must-gather-2mm9s_6093636d-db50-435e-b563-2ea4aef4e8c9/gather/0.log" Dec 03 15:31:50 crc kubenswrapper[4805]: I1203 15:31:50.827306 4805 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-7hg4c/must-gather-2mm9s"] Dec 03 15:31:50 crc kubenswrapper[4805]: I1203 15:31:50.828450 4805 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-must-gather-7hg4c/must-gather-2mm9s" podUID="6093636d-db50-435e-b563-2ea4aef4e8c9" containerName="copy" containerID="cri-o://9dcab8d81985fdec27ab73016a77ccac5cd1d0b236c583da6fc032e6b7b9df87" gracePeriod=2 Dec 03 15:31:50 crc kubenswrapper[4805]: I1203 15:31:50.848591 4805 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-7hg4c/must-gather-2mm9s"] Dec 03 15:31:51 crc kubenswrapper[4805]: I1203 15:31:51.277570 4805 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-7hg4c_must-gather-2mm9s_6093636d-db50-435e-b563-2ea4aef4e8c9/copy/0.log" Dec 03 15:31:51 crc kubenswrapper[4805]: I1203 15:31:51.278448 4805 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-7hg4c/must-gather-2mm9s" Dec 03 15:31:51 crc kubenswrapper[4805]: I1203 15:31:51.367874 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lm479\" (UniqueName: \"kubernetes.io/projected/6093636d-db50-435e-b563-2ea4aef4e8c9-kube-api-access-lm479\") pod \"6093636d-db50-435e-b563-2ea4aef4e8c9\" (UID: \"6093636d-db50-435e-b563-2ea4aef4e8c9\") " Dec 03 15:31:51 crc kubenswrapper[4805]: I1203 15:31:51.367984 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/6093636d-db50-435e-b563-2ea4aef4e8c9-must-gather-output\") pod \"6093636d-db50-435e-b563-2ea4aef4e8c9\" (UID: \"6093636d-db50-435e-b563-2ea4aef4e8c9\") " Dec 03 15:31:51 crc kubenswrapper[4805]: I1203 15:31:51.373556 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6093636d-db50-435e-b563-2ea4aef4e8c9-kube-api-access-lm479" (OuterVolumeSpecName: "kube-api-access-lm479") pod "6093636d-db50-435e-b563-2ea4aef4e8c9" (UID: "6093636d-db50-435e-b563-2ea4aef4e8c9"). InnerVolumeSpecName "kube-api-access-lm479". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 15:31:51 crc kubenswrapper[4805]: I1203 15:31:51.470672 4805 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lm479\" (UniqueName: \"kubernetes.io/projected/6093636d-db50-435e-b563-2ea4aef4e8c9-kube-api-access-lm479\") on node \"crc\" DevicePath \"\"" Dec 03 15:31:51 crc kubenswrapper[4805]: I1203 15:31:51.521824 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/6093636d-db50-435e-b563-2ea4aef4e8c9-must-gather-output" (OuterVolumeSpecName: "must-gather-output") pod "6093636d-db50-435e-b563-2ea4aef4e8c9" (UID: "6093636d-db50-435e-b563-2ea4aef4e8c9"). InnerVolumeSpecName "must-gather-output". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 15:31:51 crc kubenswrapper[4805]: I1203 15:31:51.572968 4805 reconciler_common.go:293] "Volume detached for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/6093636d-db50-435e-b563-2ea4aef4e8c9-must-gather-output\") on node \"crc\" DevicePath \"\"" Dec 03 15:31:51 crc kubenswrapper[4805]: I1203 15:31:51.628757 4805 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-7hg4c_must-gather-2mm9s_6093636d-db50-435e-b563-2ea4aef4e8c9/copy/0.log" Dec 03 15:31:51 crc kubenswrapper[4805]: I1203 15:31:51.629432 4805 generic.go:334] "Generic (PLEG): container finished" podID="6093636d-db50-435e-b563-2ea4aef4e8c9" containerID="9dcab8d81985fdec27ab73016a77ccac5cd1d0b236c583da6fc032e6b7b9df87" exitCode=143 Dec 03 15:31:51 crc kubenswrapper[4805]: I1203 15:31:51.629475 4805 scope.go:117] "RemoveContainer" containerID="9dcab8d81985fdec27ab73016a77ccac5cd1d0b236c583da6fc032e6b7b9df87" Dec 03 15:31:51 crc kubenswrapper[4805]: I1203 15:31:51.629533 4805 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-7hg4c/must-gather-2mm9s" Dec 03 15:31:51 crc kubenswrapper[4805]: I1203 15:31:51.651779 4805 scope.go:117] "RemoveContainer" containerID="cd4b89422b4de3d0665d78ffea319d0d1c3bd669d7c50fb2f68c2ea5780b99c5" Dec 03 15:31:51 crc kubenswrapper[4805]: I1203 15:31:51.717159 4805 scope.go:117] "RemoveContainer" containerID="9dcab8d81985fdec27ab73016a77ccac5cd1d0b236c583da6fc032e6b7b9df87" Dec 03 15:31:51 crc kubenswrapper[4805]: E1203 15:31:51.717538 4805 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9dcab8d81985fdec27ab73016a77ccac5cd1d0b236c583da6fc032e6b7b9df87\": container with ID starting with 9dcab8d81985fdec27ab73016a77ccac5cd1d0b236c583da6fc032e6b7b9df87 not found: ID does not exist" containerID="9dcab8d81985fdec27ab73016a77ccac5cd1d0b236c583da6fc032e6b7b9df87" Dec 03 15:31:51 crc kubenswrapper[4805]: I1203 15:31:51.717637 4805 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9dcab8d81985fdec27ab73016a77ccac5cd1d0b236c583da6fc032e6b7b9df87"} err="failed to get container status \"9dcab8d81985fdec27ab73016a77ccac5cd1d0b236c583da6fc032e6b7b9df87\": rpc error: code = NotFound desc = could not find container \"9dcab8d81985fdec27ab73016a77ccac5cd1d0b236c583da6fc032e6b7b9df87\": container with ID starting with 9dcab8d81985fdec27ab73016a77ccac5cd1d0b236c583da6fc032e6b7b9df87 not found: ID does not exist" Dec 03 15:31:51 crc kubenswrapper[4805]: I1203 15:31:51.717728 4805 scope.go:117] "RemoveContainer" containerID="cd4b89422b4de3d0665d78ffea319d0d1c3bd669d7c50fb2f68c2ea5780b99c5" Dec 03 15:31:51 crc kubenswrapper[4805]: E1203 15:31:51.718027 4805 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"cd4b89422b4de3d0665d78ffea319d0d1c3bd669d7c50fb2f68c2ea5780b99c5\": container with ID starting with cd4b89422b4de3d0665d78ffea319d0d1c3bd669d7c50fb2f68c2ea5780b99c5 not found: ID does not exist" containerID="cd4b89422b4de3d0665d78ffea319d0d1c3bd669d7c50fb2f68c2ea5780b99c5" Dec 03 15:31:51 crc kubenswrapper[4805]: I1203 15:31:51.718128 4805 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"cd4b89422b4de3d0665d78ffea319d0d1c3bd669d7c50fb2f68c2ea5780b99c5"} err="failed to get container status \"cd4b89422b4de3d0665d78ffea319d0d1c3bd669d7c50fb2f68c2ea5780b99c5\": rpc error: code = NotFound desc = could not find container \"cd4b89422b4de3d0665d78ffea319d0d1c3bd669d7c50fb2f68c2ea5780b99c5\": container with ID starting with cd4b89422b4de3d0665d78ffea319d0d1c3bd669d7c50fb2f68c2ea5780b99c5 not found: ID does not exist" Dec 03 15:31:52 crc kubenswrapper[4805]: I1203 15:31:52.706991 4805 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6093636d-db50-435e-b563-2ea4aef4e8c9" path="/var/lib/kubelet/pods/6093636d-db50-435e-b563-2ea4aef4e8c9/volumes" Dec 03 15:32:48 crc kubenswrapper[4805]: I1203 15:32:48.376309 4805 scope.go:117] "RemoveContainer" containerID="a83cfa2c0582b7024a22927ba43f38d9e3119dc6901ca48927056e1a9158e8d9" Dec 03 15:33:13 crc kubenswrapper[4805]: I1203 15:33:13.917951 4805 patch_prober.go:28] interesting pod/machine-config-daemon-gskh4 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 03 15:33:13 crc kubenswrapper[4805]: I1203 15:33:13.918789 4805 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-gskh4" podUID="6098937f-e3f6-45e8-a647-4994a79cd711" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 03 15:33:43 crc kubenswrapper[4805]: I1203 15:33:43.918229 4805 patch_prober.go:28] interesting pod/machine-config-daemon-gskh4 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 03 15:33:43 crc kubenswrapper[4805]: I1203 15:33:43.918695 4805 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-gskh4" podUID="6098937f-e3f6-45e8-a647-4994a79cd711" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 03 15:33:47 crc kubenswrapper[4805]: I1203 15:33:47.169148 4805 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-7dkn9"] Dec 03 15:33:47 crc kubenswrapper[4805]: E1203 15:33:47.170257 4805 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6093636d-db50-435e-b563-2ea4aef4e8c9" containerName="gather" Dec 03 15:33:47 crc kubenswrapper[4805]: I1203 15:33:47.170273 4805 state_mem.go:107] "Deleted CPUSet assignment" podUID="6093636d-db50-435e-b563-2ea4aef4e8c9" containerName="gather" Dec 03 15:33:47 crc kubenswrapper[4805]: E1203 15:33:47.170307 4805 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6093636d-db50-435e-b563-2ea4aef4e8c9" containerName="copy" Dec 03 15:33:47 crc kubenswrapper[4805]: I1203 15:33:47.170313 4805 state_mem.go:107] "Deleted CPUSet assignment" podUID="6093636d-db50-435e-b563-2ea4aef4e8c9" containerName="copy" Dec 03 15:33:47 crc kubenswrapper[4805]: E1203 15:33:47.170337 4805 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0bbf15d5-33ba-4931-90a5-8b4fb7d2aba6" containerName="collect-profiles" Dec 03 15:33:47 crc kubenswrapper[4805]: I1203 15:33:47.170344 4805 state_mem.go:107] "Deleted CPUSet assignment" podUID="0bbf15d5-33ba-4931-90a5-8b4fb7d2aba6" containerName="collect-profiles" Dec 03 15:33:47 crc kubenswrapper[4805]: I1203 15:33:47.170536 4805 memory_manager.go:354] "RemoveStaleState removing state" podUID="6093636d-db50-435e-b563-2ea4aef4e8c9" containerName="gather" Dec 03 15:33:47 crc kubenswrapper[4805]: I1203 15:33:47.170555 4805 memory_manager.go:354] "RemoveStaleState removing state" podUID="6093636d-db50-435e-b563-2ea4aef4e8c9" containerName="copy" Dec 03 15:33:47 crc kubenswrapper[4805]: I1203 15:33:47.170561 4805 memory_manager.go:354] "RemoveStaleState removing state" podUID="0bbf15d5-33ba-4931-90a5-8b4fb7d2aba6" containerName="collect-profiles" Dec 03 15:33:47 crc kubenswrapper[4805]: I1203 15:33:47.172053 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-7dkn9" Dec 03 15:33:47 crc kubenswrapper[4805]: I1203 15:33:47.185923 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-7dkn9"] Dec 03 15:33:47 crc kubenswrapper[4805]: I1203 15:33:47.326738 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0eb7ff56-b089-4ba1-83b4-a7d100a19540-utilities\") pod \"redhat-operators-7dkn9\" (UID: \"0eb7ff56-b089-4ba1-83b4-a7d100a19540\") " pod="openshift-marketplace/redhat-operators-7dkn9" Dec 03 15:33:47 crc kubenswrapper[4805]: I1203 15:33:47.326795 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bhszl\" (UniqueName: \"kubernetes.io/projected/0eb7ff56-b089-4ba1-83b4-a7d100a19540-kube-api-access-bhszl\") pod \"redhat-operators-7dkn9\" (UID: \"0eb7ff56-b089-4ba1-83b4-a7d100a19540\") " pod="openshift-marketplace/redhat-operators-7dkn9" Dec 03 15:33:47 crc kubenswrapper[4805]: I1203 15:33:47.326831 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0eb7ff56-b089-4ba1-83b4-a7d100a19540-catalog-content\") pod \"redhat-operators-7dkn9\" (UID: \"0eb7ff56-b089-4ba1-83b4-a7d100a19540\") " pod="openshift-marketplace/redhat-operators-7dkn9" Dec 03 15:33:47 crc kubenswrapper[4805]: I1203 15:33:47.429739 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0eb7ff56-b089-4ba1-83b4-a7d100a19540-utilities\") pod \"redhat-operators-7dkn9\" (UID: \"0eb7ff56-b089-4ba1-83b4-a7d100a19540\") " pod="openshift-marketplace/redhat-operators-7dkn9" Dec 03 15:33:47 crc kubenswrapper[4805]: I1203 15:33:47.429801 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bhszl\" (UniqueName: \"kubernetes.io/projected/0eb7ff56-b089-4ba1-83b4-a7d100a19540-kube-api-access-bhszl\") pod \"redhat-operators-7dkn9\" (UID: \"0eb7ff56-b089-4ba1-83b4-a7d100a19540\") " pod="openshift-marketplace/redhat-operators-7dkn9" Dec 03 15:33:47 crc kubenswrapper[4805]: I1203 15:33:47.429851 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0eb7ff56-b089-4ba1-83b4-a7d100a19540-catalog-content\") pod \"redhat-operators-7dkn9\" (UID: \"0eb7ff56-b089-4ba1-83b4-a7d100a19540\") " pod="openshift-marketplace/redhat-operators-7dkn9" Dec 03 15:33:47 crc kubenswrapper[4805]: I1203 15:33:47.430481 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0eb7ff56-b089-4ba1-83b4-a7d100a19540-catalog-content\") pod \"redhat-operators-7dkn9\" (UID: \"0eb7ff56-b089-4ba1-83b4-a7d100a19540\") " pod="openshift-marketplace/redhat-operators-7dkn9" Dec 03 15:33:47 crc kubenswrapper[4805]: I1203 15:33:47.430763 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0eb7ff56-b089-4ba1-83b4-a7d100a19540-utilities\") pod \"redhat-operators-7dkn9\" (UID: \"0eb7ff56-b089-4ba1-83b4-a7d100a19540\") " pod="openshift-marketplace/redhat-operators-7dkn9" Dec 03 15:33:47 crc kubenswrapper[4805]: I1203 15:33:47.590663 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bhszl\" (UniqueName: \"kubernetes.io/projected/0eb7ff56-b089-4ba1-83b4-a7d100a19540-kube-api-access-bhszl\") pod \"redhat-operators-7dkn9\" (UID: \"0eb7ff56-b089-4ba1-83b4-a7d100a19540\") " pod="openshift-marketplace/redhat-operators-7dkn9" Dec 03 15:33:47 crc kubenswrapper[4805]: I1203 15:33:47.819136 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-7dkn9" Dec 03 15:33:48 crc kubenswrapper[4805]: I1203 15:33:48.336991 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-7dkn9"] Dec 03 15:33:48 crc kubenswrapper[4805]: I1203 15:33:48.734313 4805 generic.go:334] "Generic (PLEG): container finished" podID="0eb7ff56-b089-4ba1-83b4-a7d100a19540" containerID="1c654520e747619ce09eedf8632bec35afbac7a325819e06a1f50361da214962" exitCode=0 Dec 03 15:33:48 crc kubenswrapper[4805]: I1203 15:33:48.734351 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-7dkn9" event={"ID":"0eb7ff56-b089-4ba1-83b4-a7d100a19540","Type":"ContainerDied","Data":"1c654520e747619ce09eedf8632bec35afbac7a325819e06a1f50361da214962"} Dec 03 15:33:48 crc kubenswrapper[4805]: I1203 15:33:48.734387 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-7dkn9" event={"ID":"0eb7ff56-b089-4ba1-83b4-a7d100a19540","Type":"ContainerStarted","Data":"aa6c470176d3c3895911a156fb41a046325f3727267ff7367e80841e0783008a"} Dec 03 15:33:50 crc kubenswrapper[4805]: I1203 15:33:50.756667 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-7dkn9" event={"ID":"0eb7ff56-b089-4ba1-83b4-a7d100a19540","Type":"ContainerStarted","Data":"618f9b03e7dc31bbc1d8550439f02cdb7634f7e865a4a09c1eeef47bdf55ef25"} Dec 03 15:33:51 crc kubenswrapper[4805]: I1203 15:33:51.769507 4805 generic.go:334] "Generic (PLEG): container finished" podID="0eb7ff56-b089-4ba1-83b4-a7d100a19540" containerID="618f9b03e7dc31bbc1d8550439f02cdb7634f7e865a4a09c1eeef47bdf55ef25" exitCode=0 Dec 03 15:33:51 crc kubenswrapper[4805]: I1203 15:33:51.769633 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-7dkn9" event={"ID":"0eb7ff56-b089-4ba1-83b4-a7d100a19540","Type":"ContainerDied","Data":"618f9b03e7dc31bbc1d8550439f02cdb7634f7e865a4a09c1eeef47bdf55ef25"} Dec 03 15:33:52 crc kubenswrapper[4805]: I1203 15:33:52.786158 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-7dkn9" event={"ID":"0eb7ff56-b089-4ba1-83b4-a7d100a19540","Type":"ContainerStarted","Data":"87f64222147e2db669f3f110530071b2d07f767794a9d1beb4b5f74f36c729eb"} Dec 03 15:33:52 crc kubenswrapper[4805]: I1203 15:33:52.816904 4805 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-7dkn9" podStartSLOduration=2.353160729 podStartE2EDuration="5.816882159s" podCreationTimestamp="2025-12-03 15:33:47 +0000 UTC" firstStartedPulling="2025-12-03 15:33:48.736280704 +0000 UTC m=+5058.399197637" lastFinishedPulling="2025-12-03 15:33:52.200002144 +0000 UTC m=+5061.862919067" observedRunningTime="2025-12-03 15:33:52.808803951 +0000 UTC m=+5062.471720874" watchObservedRunningTime="2025-12-03 15:33:52.816882159 +0000 UTC m=+5062.479799082" Dec 03 15:33:54 crc kubenswrapper[4805]: I1203 15:33:54.141208 4805 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-wp57t"] Dec 03 15:33:54 crc kubenswrapper[4805]: I1203 15:33:54.144330 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-wp57t" Dec 03 15:33:54 crc kubenswrapper[4805]: I1203 15:33:54.155981 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-wp57t"] Dec 03 15:33:54 crc kubenswrapper[4805]: I1203 15:33:54.176705 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d9dcff62-3c2e-4101-8a45-94c6b826b31a-utilities\") pod \"certified-operators-wp57t\" (UID: \"d9dcff62-3c2e-4101-8a45-94c6b826b31a\") " pod="openshift-marketplace/certified-operators-wp57t" Dec 03 15:33:54 crc kubenswrapper[4805]: I1203 15:33:54.176781 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d9dcff62-3c2e-4101-8a45-94c6b826b31a-catalog-content\") pod \"certified-operators-wp57t\" (UID: \"d9dcff62-3c2e-4101-8a45-94c6b826b31a\") " pod="openshift-marketplace/certified-operators-wp57t" Dec 03 15:33:54 crc kubenswrapper[4805]: I1203 15:33:54.176902 4805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-28vng\" (UniqueName: \"kubernetes.io/projected/d9dcff62-3c2e-4101-8a45-94c6b826b31a-kube-api-access-28vng\") pod \"certified-operators-wp57t\" (UID: \"d9dcff62-3c2e-4101-8a45-94c6b826b31a\") " pod="openshift-marketplace/certified-operators-wp57t" Dec 03 15:33:54 crc kubenswrapper[4805]: I1203 15:33:54.279381 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d9dcff62-3c2e-4101-8a45-94c6b826b31a-catalog-content\") pod \"certified-operators-wp57t\" (UID: \"d9dcff62-3c2e-4101-8a45-94c6b826b31a\") " pod="openshift-marketplace/certified-operators-wp57t" Dec 03 15:33:54 crc kubenswrapper[4805]: I1203 15:33:54.279491 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-28vng\" (UniqueName: \"kubernetes.io/projected/d9dcff62-3c2e-4101-8a45-94c6b826b31a-kube-api-access-28vng\") pod \"certified-operators-wp57t\" (UID: \"d9dcff62-3c2e-4101-8a45-94c6b826b31a\") " pod="openshift-marketplace/certified-operators-wp57t" Dec 03 15:33:54 crc kubenswrapper[4805]: I1203 15:33:54.279651 4805 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d9dcff62-3c2e-4101-8a45-94c6b826b31a-utilities\") pod \"certified-operators-wp57t\" (UID: \"d9dcff62-3c2e-4101-8a45-94c6b826b31a\") " pod="openshift-marketplace/certified-operators-wp57t" Dec 03 15:33:54 crc kubenswrapper[4805]: I1203 15:33:54.280176 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d9dcff62-3c2e-4101-8a45-94c6b826b31a-catalog-content\") pod \"certified-operators-wp57t\" (UID: \"d9dcff62-3c2e-4101-8a45-94c6b826b31a\") " pod="openshift-marketplace/certified-operators-wp57t" Dec 03 15:33:54 crc kubenswrapper[4805]: I1203 15:33:54.280194 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d9dcff62-3c2e-4101-8a45-94c6b826b31a-utilities\") pod \"certified-operators-wp57t\" (UID: \"d9dcff62-3c2e-4101-8a45-94c6b826b31a\") " pod="openshift-marketplace/certified-operators-wp57t" Dec 03 15:33:54 crc kubenswrapper[4805]: I1203 15:33:54.308272 4805 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-28vng\" (UniqueName: \"kubernetes.io/projected/d9dcff62-3c2e-4101-8a45-94c6b826b31a-kube-api-access-28vng\") pod \"certified-operators-wp57t\" (UID: \"d9dcff62-3c2e-4101-8a45-94c6b826b31a\") " pod="openshift-marketplace/certified-operators-wp57t" Dec 03 15:33:54 crc kubenswrapper[4805]: I1203 15:33:54.472910 4805 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-wp57t" Dec 03 15:33:55 crc kubenswrapper[4805]: I1203 15:33:55.079334 4805 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-wp57t"] Dec 03 15:33:55 crc kubenswrapper[4805]: W1203 15:33:55.099373 4805 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd9dcff62_3c2e_4101_8a45_94c6b826b31a.slice/crio-e116dba7637a91a8bc7d01eda1292b7e8da753ca59d8499b7dd3cfd7ae3aa45e WatchSource:0}: Error finding container e116dba7637a91a8bc7d01eda1292b7e8da753ca59d8499b7dd3cfd7ae3aa45e: Status 404 returned error can't find the container with id e116dba7637a91a8bc7d01eda1292b7e8da753ca59d8499b7dd3cfd7ae3aa45e Dec 03 15:33:55 crc kubenswrapper[4805]: I1203 15:33:55.817145 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-wp57t" event={"ID":"d9dcff62-3c2e-4101-8a45-94c6b826b31a","Type":"ContainerStarted","Data":"e116dba7637a91a8bc7d01eda1292b7e8da753ca59d8499b7dd3cfd7ae3aa45e"} Dec 03 15:33:56 crc kubenswrapper[4805]: I1203 15:33:56.836754 4805 generic.go:334] "Generic (PLEG): container finished" podID="d9dcff62-3c2e-4101-8a45-94c6b826b31a" containerID="8c57c57e2d02c90c88abcb2a4080ee2909d162a27f685da1970bac4c77b3c293" exitCode=0 Dec 03 15:33:56 crc kubenswrapper[4805]: I1203 15:33:56.837309 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-wp57t" event={"ID":"d9dcff62-3c2e-4101-8a45-94c6b826b31a","Type":"ContainerDied","Data":"8c57c57e2d02c90c88abcb2a4080ee2909d162a27f685da1970bac4c77b3c293"} Dec 03 15:33:57 crc kubenswrapper[4805]: I1203 15:33:57.819390 4805 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-7dkn9" Dec 03 15:33:57 crc kubenswrapper[4805]: I1203 15:33:57.820016 4805 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-7dkn9" Dec 03 15:33:57 crc kubenswrapper[4805]: I1203 15:33:57.851897 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-wp57t" event={"ID":"d9dcff62-3c2e-4101-8a45-94c6b826b31a","Type":"ContainerStarted","Data":"aa71fd37da619f4e8608ed27f65bfa361f802ba7a004629ca5c58979404f768b"} Dec 03 15:33:57 crc kubenswrapper[4805]: I1203 15:33:57.872757 4805 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-7dkn9" Dec 03 15:33:57 crc kubenswrapper[4805]: I1203 15:33:57.927779 4805 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-7dkn9" Dec 03 15:33:58 crc kubenswrapper[4805]: I1203 15:33:58.864453 4805 generic.go:334] "Generic (PLEG): container finished" podID="d9dcff62-3c2e-4101-8a45-94c6b826b31a" containerID="aa71fd37da619f4e8608ed27f65bfa361f802ba7a004629ca5c58979404f768b" exitCode=0 Dec 03 15:33:58 crc kubenswrapper[4805]: I1203 15:33:58.864531 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-wp57t" event={"ID":"d9dcff62-3c2e-4101-8a45-94c6b826b31a","Type":"ContainerDied","Data":"aa71fd37da619f4e8608ed27f65bfa361f802ba7a004629ca5c58979404f768b"} Dec 03 15:33:59 crc kubenswrapper[4805]: I1203 15:33:59.876339 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-wp57t" event={"ID":"d9dcff62-3c2e-4101-8a45-94c6b826b31a","Type":"ContainerStarted","Data":"ea1ec205461b421a355389ec20f1de8bfdd4cb91fab212dffc44ed2be46db349"} Dec 03 15:33:59 crc kubenswrapper[4805]: I1203 15:33:59.900722 4805 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-wp57t" podStartSLOduration=3.474312344 podStartE2EDuration="5.900700664s" podCreationTimestamp="2025-12-03 15:33:54 +0000 UTC" firstStartedPulling="2025-12-03 15:33:56.841391437 +0000 UTC m=+5066.504308360" lastFinishedPulling="2025-12-03 15:33:59.267779757 +0000 UTC m=+5068.930696680" observedRunningTime="2025-12-03 15:33:59.897144148 +0000 UTC m=+5069.560061091" watchObservedRunningTime="2025-12-03 15:33:59.900700664 +0000 UTC m=+5069.563617587" Dec 03 15:34:00 crc kubenswrapper[4805]: I1203 15:34:00.132301 4805 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-7dkn9"] Dec 03 15:34:00 crc kubenswrapper[4805]: I1203 15:34:00.132640 4805 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-7dkn9" podUID="0eb7ff56-b089-4ba1-83b4-a7d100a19540" containerName="registry-server" containerID="cri-o://87f64222147e2db669f3f110530071b2d07f767794a9d1beb4b5f74f36c729eb" gracePeriod=2 Dec 03 15:34:02 crc kubenswrapper[4805]: I1203 15:34:02.907044 4805 generic.go:334] "Generic (PLEG): container finished" podID="0eb7ff56-b089-4ba1-83b4-a7d100a19540" containerID="87f64222147e2db669f3f110530071b2d07f767794a9d1beb4b5f74f36c729eb" exitCode=0 Dec 03 15:34:02 crc kubenswrapper[4805]: I1203 15:34:02.907117 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-7dkn9" event={"ID":"0eb7ff56-b089-4ba1-83b4-a7d100a19540","Type":"ContainerDied","Data":"87f64222147e2db669f3f110530071b2d07f767794a9d1beb4b5f74f36c729eb"} Dec 03 15:34:03 crc kubenswrapper[4805]: I1203 15:34:03.285490 4805 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-7dkn9" Dec 03 15:34:03 crc kubenswrapper[4805]: I1203 15:34:03.469250 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0eb7ff56-b089-4ba1-83b4-a7d100a19540-catalog-content\") pod \"0eb7ff56-b089-4ba1-83b4-a7d100a19540\" (UID: \"0eb7ff56-b089-4ba1-83b4-a7d100a19540\") " Dec 03 15:34:03 crc kubenswrapper[4805]: I1203 15:34:03.469762 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bhszl\" (UniqueName: \"kubernetes.io/projected/0eb7ff56-b089-4ba1-83b4-a7d100a19540-kube-api-access-bhszl\") pod \"0eb7ff56-b089-4ba1-83b4-a7d100a19540\" (UID: \"0eb7ff56-b089-4ba1-83b4-a7d100a19540\") " Dec 03 15:34:03 crc kubenswrapper[4805]: I1203 15:34:03.469970 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0eb7ff56-b089-4ba1-83b4-a7d100a19540-utilities\") pod \"0eb7ff56-b089-4ba1-83b4-a7d100a19540\" (UID: \"0eb7ff56-b089-4ba1-83b4-a7d100a19540\") " Dec 03 15:34:03 crc kubenswrapper[4805]: I1203 15:34:03.470709 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0eb7ff56-b089-4ba1-83b4-a7d100a19540-utilities" (OuterVolumeSpecName: "utilities") pod "0eb7ff56-b089-4ba1-83b4-a7d100a19540" (UID: "0eb7ff56-b089-4ba1-83b4-a7d100a19540"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 15:34:03 crc kubenswrapper[4805]: I1203 15:34:03.481527 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0eb7ff56-b089-4ba1-83b4-a7d100a19540-kube-api-access-bhszl" (OuterVolumeSpecName: "kube-api-access-bhszl") pod "0eb7ff56-b089-4ba1-83b4-a7d100a19540" (UID: "0eb7ff56-b089-4ba1-83b4-a7d100a19540"). InnerVolumeSpecName "kube-api-access-bhszl". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 15:34:03 crc kubenswrapper[4805]: I1203 15:34:03.572658 4805 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0eb7ff56-b089-4ba1-83b4-a7d100a19540-utilities\") on node \"crc\" DevicePath \"\"" Dec 03 15:34:03 crc kubenswrapper[4805]: I1203 15:34:03.572695 4805 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bhszl\" (UniqueName: \"kubernetes.io/projected/0eb7ff56-b089-4ba1-83b4-a7d100a19540-kube-api-access-bhszl\") on node \"crc\" DevicePath \"\"" Dec 03 15:34:03 crc kubenswrapper[4805]: I1203 15:34:03.596279 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0eb7ff56-b089-4ba1-83b4-a7d100a19540-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "0eb7ff56-b089-4ba1-83b4-a7d100a19540" (UID: "0eb7ff56-b089-4ba1-83b4-a7d100a19540"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 15:34:03 crc kubenswrapper[4805]: I1203 15:34:03.674441 4805 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0eb7ff56-b089-4ba1-83b4-a7d100a19540-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 03 15:34:03 crc kubenswrapper[4805]: I1203 15:34:03.951268 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-7dkn9" event={"ID":"0eb7ff56-b089-4ba1-83b4-a7d100a19540","Type":"ContainerDied","Data":"aa6c470176d3c3895911a156fb41a046325f3727267ff7367e80841e0783008a"} Dec 03 15:34:03 crc kubenswrapper[4805]: I1203 15:34:03.951327 4805 scope.go:117] "RemoveContainer" containerID="87f64222147e2db669f3f110530071b2d07f767794a9d1beb4b5f74f36c729eb" Dec 03 15:34:03 crc kubenswrapper[4805]: I1203 15:34:03.951387 4805 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-7dkn9" Dec 03 15:34:03 crc kubenswrapper[4805]: I1203 15:34:03.978428 4805 scope.go:117] "RemoveContainer" containerID="618f9b03e7dc31bbc1d8550439f02cdb7634f7e865a4a09c1eeef47bdf55ef25" Dec 03 15:34:04 crc kubenswrapper[4805]: I1203 15:34:04.005636 4805 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-7dkn9"] Dec 03 15:34:04 crc kubenswrapper[4805]: I1203 15:34:04.015465 4805 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-7dkn9"] Dec 03 15:34:04 crc kubenswrapper[4805]: I1203 15:34:04.018710 4805 scope.go:117] "RemoveContainer" containerID="1c654520e747619ce09eedf8632bec35afbac7a325819e06a1f50361da214962" Dec 03 15:34:04 crc kubenswrapper[4805]: I1203 15:34:04.473335 4805 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-wp57t" Dec 03 15:34:04 crc kubenswrapper[4805]: I1203 15:34:04.473668 4805 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-wp57t" Dec 03 15:34:04 crc kubenswrapper[4805]: I1203 15:34:04.527591 4805 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-wp57t" Dec 03 15:34:04 crc kubenswrapper[4805]: I1203 15:34:04.706974 4805 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0eb7ff56-b089-4ba1-83b4-a7d100a19540" path="/var/lib/kubelet/pods/0eb7ff56-b089-4ba1-83b4-a7d100a19540/volumes" Dec 03 15:34:05 crc kubenswrapper[4805]: I1203 15:34:05.017051 4805 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-wp57t" Dec 03 15:34:06 crc kubenswrapper[4805]: I1203 15:34:06.931264 4805 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-wp57t"] Dec 03 15:34:06 crc kubenswrapper[4805]: I1203 15:34:06.979589 4805 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-wp57t" podUID="d9dcff62-3c2e-4101-8a45-94c6b826b31a" containerName="registry-server" containerID="cri-o://ea1ec205461b421a355389ec20f1de8bfdd4cb91fab212dffc44ed2be46db349" gracePeriod=2 Dec 03 15:34:07 crc kubenswrapper[4805]: I1203 15:34:07.474947 4805 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-wp57t" Dec 03 15:34:07 crc kubenswrapper[4805]: I1203 15:34:07.661984 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-28vng\" (UniqueName: \"kubernetes.io/projected/d9dcff62-3c2e-4101-8a45-94c6b826b31a-kube-api-access-28vng\") pod \"d9dcff62-3c2e-4101-8a45-94c6b826b31a\" (UID: \"d9dcff62-3c2e-4101-8a45-94c6b826b31a\") " Dec 03 15:34:07 crc kubenswrapper[4805]: I1203 15:34:07.662439 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d9dcff62-3c2e-4101-8a45-94c6b826b31a-catalog-content\") pod \"d9dcff62-3c2e-4101-8a45-94c6b826b31a\" (UID: \"d9dcff62-3c2e-4101-8a45-94c6b826b31a\") " Dec 03 15:34:07 crc kubenswrapper[4805]: I1203 15:34:07.662667 4805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d9dcff62-3c2e-4101-8a45-94c6b826b31a-utilities\") pod \"d9dcff62-3c2e-4101-8a45-94c6b826b31a\" (UID: \"d9dcff62-3c2e-4101-8a45-94c6b826b31a\") " Dec 03 15:34:07 crc kubenswrapper[4805]: I1203 15:34:07.664256 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d9dcff62-3c2e-4101-8a45-94c6b826b31a-utilities" (OuterVolumeSpecName: "utilities") pod "d9dcff62-3c2e-4101-8a45-94c6b826b31a" (UID: "d9dcff62-3c2e-4101-8a45-94c6b826b31a"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 15:34:07 crc kubenswrapper[4805]: I1203 15:34:07.675997 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d9dcff62-3c2e-4101-8a45-94c6b826b31a-kube-api-access-28vng" (OuterVolumeSpecName: "kube-api-access-28vng") pod "d9dcff62-3c2e-4101-8a45-94c6b826b31a" (UID: "d9dcff62-3c2e-4101-8a45-94c6b826b31a"). InnerVolumeSpecName "kube-api-access-28vng". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 15:34:07 crc kubenswrapper[4805]: I1203 15:34:07.715515 4805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d9dcff62-3c2e-4101-8a45-94c6b826b31a-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "d9dcff62-3c2e-4101-8a45-94c6b826b31a" (UID: "d9dcff62-3c2e-4101-8a45-94c6b826b31a"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 15:34:07 crc kubenswrapper[4805]: I1203 15:34:07.765314 4805 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d9dcff62-3c2e-4101-8a45-94c6b826b31a-utilities\") on node \"crc\" DevicePath \"\"" Dec 03 15:34:07 crc kubenswrapper[4805]: I1203 15:34:07.765360 4805 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-28vng\" (UniqueName: \"kubernetes.io/projected/d9dcff62-3c2e-4101-8a45-94c6b826b31a-kube-api-access-28vng\") on node \"crc\" DevicePath \"\"" Dec 03 15:34:07 crc kubenswrapper[4805]: I1203 15:34:07.765377 4805 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d9dcff62-3c2e-4101-8a45-94c6b826b31a-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 03 15:34:07 crc kubenswrapper[4805]: I1203 15:34:07.990416 4805 generic.go:334] "Generic (PLEG): container finished" podID="d9dcff62-3c2e-4101-8a45-94c6b826b31a" containerID="ea1ec205461b421a355389ec20f1de8bfdd4cb91fab212dffc44ed2be46db349" exitCode=0 Dec 03 15:34:07 crc kubenswrapper[4805]: I1203 15:34:07.990463 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-wp57t" event={"ID":"d9dcff62-3c2e-4101-8a45-94c6b826b31a","Type":"ContainerDied","Data":"ea1ec205461b421a355389ec20f1de8bfdd4cb91fab212dffc44ed2be46db349"} Dec 03 15:34:07 crc kubenswrapper[4805]: I1203 15:34:07.990489 4805 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-wp57t" Dec 03 15:34:07 crc kubenswrapper[4805]: I1203 15:34:07.990506 4805 scope.go:117] "RemoveContainer" containerID="ea1ec205461b421a355389ec20f1de8bfdd4cb91fab212dffc44ed2be46db349" Dec 03 15:34:07 crc kubenswrapper[4805]: I1203 15:34:07.990493 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-wp57t" event={"ID":"d9dcff62-3c2e-4101-8a45-94c6b826b31a","Type":"ContainerDied","Data":"e116dba7637a91a8bc7d01eda1292b7e8da753ca59d8499b7dd3cfd7ae3aa45e"} Dec 03 15:34:08 crc kubenswrapper[4805]: I1203 15:34:08.016308 4805 scope.go:117] "RemoveContainer" containerID="aa71fd37da619f4e8608ed27f65bfa361f802ba7a004629ca5c58979404f768b" Dec 03 15:34:08 crc kubenswrapper[4805]: I1203 15:34:08.040708 4805 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-wp57t"] Dec 03 15:34:08 crc kubenswrapper[4805]: I1203 15:34:08.051307 4805 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-wp57t"] Dec 03 15:34:08 crc kubenswrapper[4805]: I1203 15:34:08.407023 4805 scope.go:117] "RemoveContainer" containerID="8c57c57e2d02c90c88abcb2a4080ee2909d162a27f685da1970bac4c77b3c293" Dec 03 15:34:08 crc kubenswrapper[4805]: I1203 15:34:08.581140 4805 scope.go:117] "RemoveContainer" containerID="ea1ec205461b421a355389ec20f1de8bfdd4cb91fab212dffc44ed2be46db349" Dec 03 15:34:08 crc kubenswrapper[4805]: E1203 15:34:08.581673 4805 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ea1ec205461b421a355389ec20f1de8bfdd4cb91fab212dffc44ed2be46db349\": container with ID starting with ea1ec205461b421a355389ec20f1de8bfdd4cb91fab212dffc44ed2be46db349 not found: ID does not exist" containerID="ea1ec205461b421a355389ec20f1de8bfdd4cb91fab212dffc44ed2be46db349" Dec 03 15:34:08 crc kubenswrapper[4805]: I1203 15:34:08.581704 4805 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ea1ec205461b421a355389ec20f1de8bfdd4cb91fab212dffc44ed2be46db349"} err="failed to get container status \"ea1ec205461b421a355389ec20f1de8bfdd4cb91fab212dffc44ed2be46db349\": rpc error: code = NotFound desc = could not find container \"ea1ec205461b421a355389ec20f1de8bfdd4cb91fab212dffc44ed2be46db349\": container with ID starting with ea1ec205461b421a355389ec20f1de8bfdd4cb91fab212dffc44ed2be46db349 not found: ID does not exist" Dec 03 15:34:08 crc kubenswrapper[4805]: I1203 15:34:08.581726 4805 scope.go:117] "RemoveContainer" containerID="aa71fd37da619f4e8608ed27f65bfa361f802ba7a004629ca5c58979404f768b" Dec 03 15:34:08 crc kubenswrapper[4805]: E1203 15:34:08.581954 4805 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"aa71fd37da619f4e8608ed27f65bfa361f802ba7a004629ca5c58979404f768b\": container with ID starting with aa71fd37da619f4e8608ed27f65bfa361f802ba7a004629ca5c58979404f768b not found: ID does not exist" containerID="aa71fd37da619f4e8608ed27f65bfa361f802ba7a004629ca5c58979404f768b" Dec 03 15:34:08 crc kubenswrapper[4805]: I1203 15:34:08.582000 4805 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"aa71fd37da619f4e8608ed27f65bfa361f802ba7a004629ca5c58979404f768b"} err="failed to get container status \"aa71fd37da619f4e8608ed27f65bfa361f802ba7a004629ca5c58979404f768b\": rpc error: code = NotFound desc = could not find container \"aa71fd37da619f4e8608ed27f65bfa361f802ba7a004629ca5c58979404f768b\": container with ID starting with aa71fd37da619f4e8608ed27f65bfa361f802ba7a004629ca5c58979404f768b not found: ID does not exist" Dec 03 15:34:08 crc kubenswrapper[4805]: I1203 15:34:08.582017 4805 scope.go:117] "RemoveContainer" containerID="8c57c57e2d02c90c88abcb2a4080ee2909d162a27f685da1970bac4c77b3c293" Dec 03 15:34:08 crc kubenswrapper[4805]: E1203 15:34:08.582217 4805 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8c57c57e2d02c90c88abcb2a4080ee2909d162a27f685da1970bac4c77b3c293\": container with ID starting with 8c57c57e2d02c90c88abcb2a4080ee2909d162a27f685da1970bac4c77b3c293 not found: ID does not exist" containerID="8c57c57e2d02c90c88abcb2a4080ee2909d162a27f685da1970bac4c77b3c293" Dec 03 15:34:08 crc kubenswrapper[4805]: I1203 15:34:08.582237 4805 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8c57c57e2d02c90c88abcb2a4080ee2909d162a27f685da1970bac4c77b3c293"} err="failed to get container status \"8c57c57e2d02c90c88abcb2a4080ee2909d162a27f685da1970bac4c77b3c293\": rpc error: code = NotFound desc = could not find container \"8c57c57e2d02c90c88abcb2a4080ee2909d162a27f685da1970bac4c77b3c293\": container with ID starting with 8c57c57e2d02c90c88abcb2a4080ee2909d162a27f685da1970bac4c77b3c293 not found: ID does not exist" Dec 03 15:34:08 crc kubenswrapper[4805]: I1203 15:34:08.706760 4805 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d9dcff62-3c2e-4101-8a45-94c6b826b31a" path="/var/lib/kubelet/pods/d9dcff62-3c2e-4101-8a45-94c6b826b31a/volumes" Dec 03 15:34:13 crc kubenswrapper[4805]: I1203 15:34:13.917019 4805 patch_prober.go:28] interesting pod/machine-config-daemon-gskh4 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 03 15:34:13 crc kubenswrapper[4805]: I1203 15:34:13.917532 4805 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-gskh4" podUID="6098937f-e3f6-45e8-a647-4994a79cd711" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 03 15:34:13 crc kubenswrapper[4805]: I1203 15:34:13.917573 4805 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-gskh4" Dec 03 15:34:13 crc kubenswrapper[4805]: I1203 15:34:13.918305 4805 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"8a6f7e8ec0c2240c0eb0a039e259a0f769944a70b22cf0a08378ac2f9c4cfeb8"} pod="openshift-machine-config-operator/machine-config-daemon-gskh4" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 03 15:34:13 crc kubenswrapper[4805]: I1203 15:34:13.918364 4805 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-gskh4" podUID="6098937f-e3f6-45e8-a647-4994a79cd711" containerName="machine-config-daemon" containerID="cri-o://8a6f7e8ec0c2240c0eb0a039e259a0f769944a70b22cf0a08378ac2f9c4cfeb8" gracePeriod=600 Dec 03 15:34:14 crc kubenswrapper[4805]: I1203 15:34:14.095618 4805 generic.go:334] "Generic (PLEG): container finished" podID="6098937f-e3f6-45e8-a647-4994a79cd711" containerID="8a6f7e8ec0c2240c0eb0a039e259a0f769944a70b22cf0a08378ac2f9c4cfeb8" exitCode=0 Dec 03 15:34:14 crc kubenswrapper[4805]: I1203 15:34:14.095670 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-gskh4" event={"ID":"6098937f-e3f6-45e8-a647-4994a79cd711","Type":"ContainerDied","Data":"8a6f7e8ec0c2240c0eb0a039e259a0f769944a70b22cf0a08378ac2f9c4cfeb8"} Dec 03 15:34:14 crc kubenswrapper[4805]: I1203 15:34:14.095709 4805 scope.go:117] "RemoveContainer" containerID="7d1692e3c7d3a2b9dcbcaebd54978af4cba7ae377a17c65268d4048aa72d2c92" Dec 03 15:34:15 crc kubenswrapper[4805]: I1203 15:34:15.108674 4805 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-gskh4" event={"ID":"6098937f-e3f6-45e8-a647-4994a79cd711","Type":"ContainerStarted","Data":"8cd2de2aa9cabd703ee0826942f65426d0e3ae064edae39c61b357127b42de64"} var/home/core/zuul-output/logs/crc-cloud-workdir-crc-all-logs.tar.gz0000644000175000000000000000005515114054373024450 0ustar coreroot‹íÁ  ÷Om7 €7šÞ'(var/home/core/zuul-output/logs/crc-cloud/0000755000175000000000000000000015114054374017366 5ustar corerootvar/home/core/zuul-output/artifacts/0000755000175000017500000000000015114041702016500 5ustar corecorevar/home/core/zuul-output/docs/0000755000175000017500000000000015114041702015450 5ustar corecore